Pinokio

Store

Related tags
Multi-Voice Text-to-Speech for Stories and Audiobooks. Supports Kokoro and Chatterbox TTS engines with GPU acceleration.
ComfyuiFeatured
The most powerful and modular diffusion model GUI, api and backend with a graph/nodes interface. https://github.com/comfyanonymous/ComfyUI
Check-ins27 check-ins
Platforms
GPUNVIDIAAMDApple
A multi-voice AI audiobook generator built on Qwen3-TTS — annotate scripts with an LLM, assign unique voices to each character, per-line style instructions for delivery, clone voices from reference audio, design new voices from text descriptions, train custom voices with LoRA fine-tuning, and export to MP3 or Audacity multi-track projects
Check-ins2 check-ins
VoiceboxFeatured
Local-first voice synthesis studio powered by Qwen3-TTS.
Check-ins8 check-ins
Platforms
GPUNVIDIAAMDApple
Kokoro, KittenTTS, Higgs audio, Chatterbox/Multi, Fish-Speech, F5 & index-tts & indextts2, VoxCPM and VibeVoice in one app
Check-ins13 check-ins
Wan2GPFeatured
Super Optimized Gradio UI for AI video creation for GPU poor machines (6GB+ VRAM). Supports Wan 2.1/2.2, Qwen, Hunyuan Video, LTX Video and Flux. https://github.com/deepbeepmeep/Wan2GP
Check-ins52 check-ins
GPUNVIDIAAMDApple
High-quality text-to-speech with Beautiful Web UI & API, optimized for Apple Silicon using MLX. Features include Custom Voice (preset speakers), Voice Design (natural language), and Voice Cloning. With enhanced features for saving custom voices and long-form / endless TTS streaming.
Check-ins20 check-ins
Platforms
GPUNVIDIAAMDApple
Qwen3-TTSFeatured
Qwen3-TTS is an open-source series of TTS models developed by the Qwen team
Check-ins14 check-ins
Platforms
GPUNVIDIAAMDApple
e2-f5-ttsFeatured
F5-TTS: A Fairytaler that Fakes Fluent and Faithful Speech with Flow Matching https://huggingface.co/spaces/mrfakename/E2-F5-TTS
Kokoro, KittenTTS, Higgs audio, Chatterbox/Multi, Fish-Speech, F5 & index-tts & indextts2, VoxCPM and VibeVoice in one app
Check-ins6 check-ins
GPUNVIDIAAMDApple
Qwen3-TTS is an open-source series of TTS models developed by the Qwen team at Alibaba Cloud, supporting stable, expressive, and streaming speech generation, free-form voice design, and vivid voice cloning.
Check-insNo check-ins yet
Platforms
GPUNVIDIAAMDApple
Qwen3-TTS is an open-source series of TTS models developed by the Qwen team
Check-ins7 check-ins
GPUNVIDIAAMDApple
Qwen3-TTS is an open-source series of TTS models developed by the Qwen team at Alibaba Cloud, supporting stable, expressive, and streaming speech generation, free-form voice design, and vivid voice cloning.
Check-insNo check-ins yet
Platforms
GPUNVIDIAAMDApple
A production-ready FastAPI server for Qwen3-TTS models, supporting CustomVoice (preset speakers), VoiceDesign (natural language voice design), and Base models (voice cloning).
Check-insNo check-ins yet
Platforms
GPUNVIDIAAMDApple
Realtime streaming TTS demo using microsoft/VibeVoice-Realtime-0.5B
Check-ins3 check-ins
[NVIDIA ONLY] AllTalk-TTS is a unified UI for E5-TTS, XTTS, Vite TTS, Piper TTS, Parler TTS and RVC, based on CoquiTTS, including a finetune mode.
OpenAudioFeatured
Multilingual Text-to-Speech with Voice Cloning (Supports: English, Japanese, Korean, Chinese, French, German, Arabic, and Spanish) https://github.com/fishaudio/fish-speech
This application converts written text into spoken words. Users input text and can optionally provide reference audio to specify the speaker's voice. The result is a generated audio file that reads...
Check-insNo check-ins yet
Platforms
GPUNVIDIAAMDApple
🎙️ Qwen3-TTS-DubFlow: An open-source, human-in-the-loop AI dubbing workbench for novels, games, podcasts, and more. Features a "Design-then-Clone" workflow powered by Qwen3-TTS to achieve consistent identity and context-aware emotional performance.
Check-insNo check-ins yet
Platforms
GPUNVIDIAAMDApple
Gradio-based web interface for the LuxTTS voice cloning and text-to-speech model, enabling users to generate customized speech from text using uploaded or recorded audio references with adjustable parameters like speed, guidance scale, and inference steps.