Update README.md
Browse files
README.md
CHANGED
@@ -258,7 +258,7 @@ The Whisper model is intrinsically designed to work on audio samples of up to 30
|
|
258 |
algorithm, it can be used to transcribe audio samples of up to arbitrary length. This is possible through Transformers
|
259 |
[`pipeline`](https://huggingface.co/docs/transformers/main_classes/pipelines#transformers.AutomaticSpeechRecognitionPipeline)
|
260 |
method. Chunking is enabled by setting `chunk_length_s=30` when instantiating the pipeline. With chunking enabled, the pipeline
|
261 |
-
can be run with batched inference. It can also be extended to predict
|
262 |
|
263 |
```python
|
264 |
>>> import torch
|
|
|
258 |
algorithm, it can be used to transcribe audio samples of up to arbitrary length. This is possible through Transformers
|
259 |
[`pipeline`](https://huggingface.co/docs/transformers/main_classes/pipelines#transformers.AutomaticSpeechRecognitionPipeline)
|
260 |
method. Chunking is enabled by setting `chunk_length_s=30` when instantiating the pipeline. With chunking enabled, the pipeline
|
261 |
+
can be run with batched inference. It can also be extended to predict word level timestamps by passing `return_timestamps="word"`:
|
262 |
|
263 |
```python
|
264 |
>>> import torch
|