Joshua's picture

Joshua

Xenova

AI & ML interests

None yet

Recent Activity

Organizations

Hugging Face's profile picture Google's profile picture Gradio's profile picture Hugging Face Internal Testing Organization's profile picture onnx's profile picture Huggingface.js's profile picture Huggingface Projects's profile picture Whisper Distillation's profile picture Hugging Face OSS Metrics's profile picture Core ML Projects's profile picture Blog-explorers's profile picture Useful Sensors Inc.'s profile picture Terrain Diffusion's profile picture Hugging Face Smol Models Research's profile picture kotol's profile picture Static Templates's profile picture gg-hf's profile picture Nerfies's profile picture WebML Community's profile picture Llava Hugging Face's profile picture Hugging Face - Visual Blocks's profile picture Social Post Explorers's profile picture hsramall's profile picture gg-tt's profile picture ONNX Community's profile picture OuteAI's profile picture Hugging Face Discord Community's profile picture LLHF's profile picture SLLHF's profile picture Hugging Quants's profile picture nltpt's profile picture Hugging Face Party @ PyTorch Conference's profile picture rmbg's profile picture ExecuTorch Community's profile picture Hello Games's profile picture G2P (Grapheme To Phoneme) Exploration's profile picture ONNX Internal Testing Organization's profile picture gg-hf-g's profile picture

Xenova's activity

New activity in Xenova/modnet about 8 hours ago

Testing device-level configs

#5 opened about 8 hours ago by
Xenova
New activity in Xenova/webgpu-jina-clip 1 day ago

Open source?

1
#4 opened 1 day ago by
seanyboi
replied to their post 6 days ago
view reply

The model itself has a maximum context length, so you can't feed everything through the model at once, unfortunately. To solve this, I implemented streaming in v1.2.0, which you can use as follows:

import { KokoroTTS } from "kokoro-js";

const model_id = "onnx-community/Kokoro-82M-v1.0-ONNX";
const tts = await KokoroTTS.from_pretrained(model_id, {
  dtype: "fp32", // Options: "fp32", "fp16", "q8", "q4", "q4f16"
  // device: "webgpu", // Options: "wasm", "webgpu" (web) or "cpu" (node).
});

const text = "Kokoro is an open-weight TTS model with 82 million parameters. Despite its lightweight architecture, it delivers comparable quality to larger models while being significantly faster and more cost-efficient. With Apache-licensed weights, Kokoro can be deployed anywhere from production environments to personal projects. It can even run 100% locally in your browser, powered by Transformers.js!";
const stream = tts.stream(text);
let i = 0;
for await (const { text, phonemes, audio } of stream) {
  console.log({ text, phonemes });
  audio.save(`audio-${i++}.wav`);
}
New activity in webml-community/phi-3.5-webgpu 8 days ago

Update Transformers.js

#2 opened 8 days ago by
Xenova