Voxta Documentation
  • Voxta Server
  • Voxta Cloud
  • Virt-A-Mate
  • Voxy
Documentation What is Voxta? Getting Started Installing Install Server Update Server Interface Menu Chat Profile Characters Services Add Service Wizard Benchmarks Chat History Diagnostics Test Services Prompt Formatting Prompt Templates Playgrounds Terminal Creator Studio Scenarios Memory Books Actions Events Contexts Templates Flags Scripting App Triggers App Triggersv2 Packages Messages Asset Formats Services Services Overview ✨ Voxta Cloud ✨ F5-TTS ✨ Coqui ✨ Azure Speech Service ✨ Deepgram ✨ ExLlamaV2 ✨ llama.cpp ✨ WhisperLive ✨ ElevenLabs ✨ KoboldAI ✨ NovelAI ✨ Text To Speech HTTP API ✨ OpenAI ✨ OpenAI Compatible ✨ OpenRouter ✨ Text Generation Web UI 🔘 Silero 🔘 Vosk 🚧 Azure Bing Search 🚧 Text Generation Inference 🚧 Windows Speech ✨ ChromaDB ✨ Florence-2 Vision ✨ Kokoro TTS NAudio Voxta Utilities Articles Large Language Models RunPod XTTS Server LLM Parameters Troubleshooting & Solutions Glossary FAQ Voxta Cloud Credits FAQ Voxta server FAQ Voxta VAM plugin FAQ About Patreon Terms AI Safety

Text Generation Inference

About Text Generation Inference

A Rust, Python and gRPC server for text generation inference. Used in production at HuggingFace to power Hugging Chat, the Inference API and Inference Endpoint.

https://github.com/huggingface/text-generation-inference

 Last updated: 0001-01-01 Prev Next
On this page
  • About Text Generation Inference
2023-2025 Voxta
Made with  from Docura