Hugging Face
By integrating Fluents with Hugging Face, businesses can leverage advanced AI orchestration, ensuring smooth campaign operation and robust compliance with ease.
Run Custom Hugging Face Models as Your Fluents Conversation Engine
Hugging Face is the world's largest open-source AI model repository, hosting tens of thousands of LLMs including fine-tuned variants, domain-specific models, and the latest open-weight releases. For Fluents deployments that need a custom or specialized model — rather than a general-purpose frontier LLM — Hugging Face is where those models live.
Fluents can be configured to call Hugging Face Inference Endpoints as its conversation engine, enabling organizations to run their own fine-tuned models trained on their specific domain data — insurance claims language, medical terminology, legal intake patterns — for higher accuracy on their particular call types.
Use a fine-tuned model trained on your domain — insurance claims, medical intake, legal qualification — as the conversation engine for higher task-specific accuracy
Open-weight models from Hugging Face give organizations full control over model weights, enabling private deployment with zero third-party data processing
Access the latest open-source releases (Llama, Mixtral, Falcon, Phi) through Hugging Face Inference Endpoints as soon as they're published

When General Models Aren't Enough
General-purpose frontier models like Gemini are excellent at most voice AI tasks. But some organizations have highly specialized requirements: an insurance carrier whose agents must navigate extremely specific policy language, a healthcare network whose agents handle complex clinical terminology, or a legal firm whose agents need to qualify leads across a nuanced fact pattern. Fine-tuned models trained on your actual call transcripts and domain data can outperform general models on these specialized tasks.
Insurance: Models Trained on Claims Language
A carrier that has thousands of recorded FNOL calls can fine-tune a Llama or Mistral model on that data — training it to recognize the specific phrases, edge cases, and exceptions that appear in their claims calls. Hosted on Hugging Face and connected to Fluents as the conversation engine, that fine-tuned model handles claims intake with higher accuracy than any general model trained on web data.
Healthcare: Clinical Terminology Fine-Tuning
Medical terminology is dense and precise. A clinical AI model fine-tuned on medical literature and patient communication transcripts handles the nuances of healthcare conversations — recognizing symptom descriptions, medication names, and procedure references — with greater accuracy than general LLMs.
Full Data Control With Private Endpoints
Hugging Face Inference Endpoints support private deployment — your model runs on dedicated infrastructure, no data is shared with other users, and the model weights are under your control. For organizations with strict data governance requirements, this is the maximum-control path for LLM deployment in Fluents.
Calls That Just Work
No per-minute taxes. No brittle workflows. Just enterprise-grade reliability with API-level flexibility.
Request a New Integration
We’re constantly expanding our library. If your stack isn’t covered yet, request it here — we’ll support niche tools and co-build connectors.
Other Integrations
Dive deeper with setup guides, API references, and partner tutorials to unlock the full potential of Fluents integrations.
Fluents + Keragon
Automate Patient Communication with Fluents Voice AI The Fluents connector for Keragon bridges the gap between your healthcare data and action. By integrating Fluents' powerful Voice AI directly into your Keragon workflows, you can automatically trigger outbound phone calls to patients or staff based on real-time events.
Fluents + MailerLite empowers real-time voice integration into your email campaigns, enhancing orchestration and maintaining compliance across channels.
Fluents + BotPenguin empower real campaigns with seamless integration, compliance assurance, and enhanced communication orchestration.
“Fluents made it incredibly fast to get our AI agent live. It replaced an answering service that cost 5x more - and performed better. Trusted partner, excellent quality, zero hassle.”

.avif)
FAQs
Questions about Hugging Face models in Fluents.
No. You can use any publicly available model hosted on Hugging Face as your conversation engine — including the latest Llama, Mixtral, or Phi releases. Fine-tuning is an option for teams that want to optimize for their specific domain, but it's not required to use Hugging Face-hosted models in Fluents.
Gemini is Fluents' default and performs best for general-purpose voice AI. Hugging Face becomes the better choice when you have a fine-tuned model that outperforms Gemini on your specific call type, or when you need full model weight control for compliance reasons. Contact the team to discuss whether fine-tuning makes sense for your use case.
Custom model endpoint configuration is an enterprise feature. Contact the Fluents team to discuss your requirements.