EVA OS
Intelligent Multimodal Core
Our Advantages
EVA OS is AutoArk’s core multimodal system, combining speech, video, and vision with ASR/TTS and system orchestration to enable real-time, natural interaction. We welcome developers to join as Maintainers.
Multimodal Interaction
Voice, vision, and text work seamlessly together for natural real-time interaction.
Low-Latency Speech & Visual
ASR, translation, and TTS deliver fast, stable, and interruption-free communication.
Custom Voices & Emotion
Create unique voice styles with fine-grained emotional expression.
Zero-Code Multimodal Agents
Build powerful multimodal agents effortlessly with simple drag-and-drop tools.
EVA OS
Full Hardware Development Support
Complete open-source firmware for all
supported boards.
ESP32、RK 3562/3562/3568...
EVA OS
Full Hardware Development Solutions
Hardware Product Use Cases

Aqi Companion Robot
Powered by EVA OS with full-stack multimodal capabilities, delivering sub-350ms real-time voice and vision interaction.

Kidodo Early Education Robot
Designed for children aged 0–10, combining emotional companionship and science-based learning with AI-driven interactive stories.

Confidential Smartphone Assistant Project
A next-generation AI assistant built on EVA’s multimodal model, enabling natural dialogue, real-time perception, and intelligent decision-making. Currently in confidential development.

Confidential AR Glasses Project
Lightweight AR intelligent eyewear with visual understanding, spatial sensing, and multimodal interaction. Designed for seamless real-world integration. Technical details remain confidential.

Confidential In-Vehicle Assistant Project
An intelligent in-car assistant combining voice and visual capabilities for real-time perception and multi-scene assistance. Developed in partnership with automotive teams.
An open-source, real-time multimodal agent engine that lets your devices hear, see, and act.