update timestamps usage
Browse files
README.md
CHANGED
@@ -319,7 +319,7 @@ predicted_text = output[0].text
|
|
319 |
|
320 |
```
|
321 |
|
322 |
-
canary-180m-flash can also
|
323 |
```python
|
324 |
output = canary_model.transcribe(
|
325 |
['filepath.wav'],
|
@@ -331,6 +331,7 @@ word_level_timestamps = output[0].timestamp['word']
|
|
331 |
segment_level_timestamps = output[0].timestamp['segment']
|
332 |
|
333 |
```
|
|
|
334 |
|
335 |
To use canary-180m-flash for transcribing other supported languages or perform Speech-to-Text translation or provide word-level timestamps, specify the input as jsonl manifest file, where each line in the file is a dictionary containing the following fields:
|
336 |
|
|
|
319 |
|
320 |
```
|
321 |
|
322 |
+
canary-180m-flash can also predict word-level and segment-level timestamps
|
323 |
```python
|
324 |
output = canary_model.transcribe(
|
325 |
['filepath.wav'],
|
|
|
331 |
segment_level_timestamps = output[0].timestamp['segment']
|
332 |
|
333 |
```
|
334 |
+
To predict timestamps for audio files longer than 10 seconds, we recommend using the longform inference script (explained in the next section) with `chunk_len_in_secs=10.0`.
|
335 |
|
336 |
To use canary-180m-flash for transcribing other supported languages or perform Speech-to-Text translation or provide word-level timestamps, specify the input as jsonl manifest file, where each line in the file is a dictionary containing the following fields:
|
337 |
|