Balalaika
Collection
11 items
•
Updated
•
5
Solves the problem of extremely slow operation of the model on some devices and adds the ability to run inference directly from the GPU code.
!pip install git+https://github.com/NikiPshg/T-one-cuda-onnx.git
from tone import StreamingCTCPipeline, read_audio, read_example_audio
audio = read_example_audio() # or read_audio("your_audio.flac")
# device_id device_id if the graphics card is not found, the CPU is used
pipeline = StreamingCTCPipeline.from_hugging_face(device_id=0)
print(pipeline.forward_offline(audio)) # offline recognition using onnx cuda