TEN Agent is a conversational voice AI agent powered by TEN, integrating Deepseek, Gemini, OpenAI, RTC, and hardware like ESP32. It enables realtime AI capabilities like seeing, hearing, and speaking, and is fully compatible with platforms like Dify and Coze.
TEN Agent is a conversational voice AI agent powered by TEN, integrating DeepSeek, Gemini, OpenAI, RTC, and hardware like ESP32. It enables realtime AI capabilities like seeing, hearing, and speaking, and is fully compatible with platforms like Dify and Coze.
Community Channel | Purpose |
---|---|
Join our Discord community to connect with developers | |
Follow TEN Framework on X for updates and announcements | |
Join our WeChat group for Chinese community discussions |
[!IMPORTANT]
Star Our Repository ⭐️
Get instant notifications for new releases and updates. Your support helps us grow and improve TEN Agent!
TEN Agent now integrates with Llama 4, Meta’s latest large language model. With no setup or waiting required, you can simply start a real-time conversation with TEN Agent.
TEN Agent now integrates seamlessly with MCP servers, expanding its LLM capabilities. To get started:
This integration allows you to leverage MCP’s diverse servers offerings while maintaining TEN Agent’s powerful conversational abilities.
Build engaging AI avatars with TEN Agent using Trulience’s diverse collection of free avatar options. To get it up and running, you only need 2 steps:
TEN is a very versatile framework. That said, TEN Agent is compatible with DeepSeek R1, try experiencing realtime conversations with DeepSeek R1!
TEN Agent is now running on the Espressif ESP32-S3 Korvo V3 development board, an excellent way to integrate realtime communication with LLM on hardware.
Try Google Gemini Multimodal Live API with realtime vision and realtime screenshare detection capabilities, it is a ready-to-use extension, along with powerful tools like Weather Check and Web Search integrated perfectly into TEN Agent.
Describe a topic and ask TEN Agent to tell you a story while also generating images of the story to provide a more immersive experience for kids.
TEN offers a great support to make the realtime interactive experience even better on other LLM platform as well, check out docs for more.
TEN seamlessly integrates with Coze platform to enhance real-time interactive experiences. Check out our documentation to learn how to leverage these powerful integrations.
Category | Requirements |
---|---|
Keys | • Agora App ID and App Certificate (free minutes every month) • OpenAI API key (any LLM that is compatible with OpenAI) • Deepgram ASR (free credits available with signup) • Elevenlabs TTS (free credits available with signup) |
Installation | • Docker / Docker Compose • Node.js(LTS) v18 |
Minimum System Requirements | • CPU >= 2 Core • RAM >= 4 GB |
[!NOTE]
macOS: Docker setting on Apple Silicon
Uncheck “Use Rosetta for x86/amd64 emulation” in Docker settings, it may result in slower build times on ARM, but performance will be normal when deployed to x64 servers.
.env
file from .env.example
cp ./.env.example ./.env
.env
AGORA_APP_ID=
AGORA_APP_CERTIFICATE=
docker compose up -d
docker exec -it ten_agent_dev bash
graph
( ~5min - ~8min)check the /examples
folder for more examples
# use the default agent
task use
# or use the demo agent
task use AGENT=agents/examples/demo
task run
Now, we have successfully set up the playground. This is just the beginning of TEN Agent. There are many different ways to explore and utilize TEN Agent. To learn more, please refer to the documentation.
GitHub offers free Codespace for each repository, you can run the playground in Codespace without using Docker.Also, the speed of Codespace is much faster than localhost.
Check out this guide for more details.
Playground and Demo server different purposes, in a nut shell, think it as Playground is for you to customize you agent, and Demo is for you to deploy your agent.
Check out this guide for more details.
Once you have customized your agent (either by using the playground or editing property.json
directly), you can deploy it by creating a release Docker image for your service.
Read the Deployment Guide for detailed information about deployment.
coming soon…
1️⃣ TEN Agent App: Core application that manages extensions and data flow based on graph configuration
2️⃣ Dev Server: port:49480
- local server for development purposes.
3️⃣ Web Server: port:8080
- Golang server handling HTTP requests and agent process management
4️⃣ Front-end UI:
port:3000
Playground - To customize and test your agent configurations.port:3002
Demo - To deploy your agent without module picker.Project | Preview |
---|---|
🏚️ TEN Framework TEN, a AI agent framework to create various AI agents which supports real-time conversation. |
![]() |
🎙️ TEN Agent TEN Agent is a conversational voice AI agent powered by TEN, integrating Deepseek, Gemini, OpenAI, RTC, and hardware like ESP32. It enables realtime AI capabilities like seeing, hearing, and speaking, and is fully compatible with platforms like Dify and Coze. |
![]() |
🎨 TMAN Designer alpha TMAN Designer is low/no code option to make a cool voice agent. With it’s easy-to-use workflow UI, you can build things easily. It comes with runtime, dark/light themes, integrated editors and integrated terminals. |
![]() |
📒 TEN Portal The official site of TEN framework, it has documentation, blog and showcases. |
![]() |
We welcome all forms of open-source collaboration! Whether you’re fixing bugs, adding features, improving documentation, or sharing ideas - your contributions help advance personalized AI tools. Check out our GitHub Issues and Projects to find ways to contribute and show your skills. Together, we can build something amazing!
[!TIP]
Welcome all kinds of contributions 🙏
Join us in building TEN better! Every contribution makes a difference, from code to documentation. Share your TEN Agent projects on social media with to inspire others!
Connect with TEN maintainer @elliotchen100 on 𝕏 or @cyfyifanchen on GitHub for project updates, discussions and collaboration opportunities.
Contributions are welcome! Please read the contribution guidelines first.
This project is Apache 2.0 licensed.