Mini-Omni: Language Models Can Hear, Talk While Thinking in Streaming
ð¤ Hugging Face | ð Github | ð Technical report
Mini-Omni is an open-source multimodel large language model that can **hear, talk while thinking**. Featuring real-time end-to-end speech input and **streaming audio output** conversational capabilities.## Features â **Real-time speech-to-speech** conversational capabilities. No extra ASR or TTS models required. â **Talking while thinking**, with the ability to generate text and audio at the same time. â **Streaming audio outupt** capabilities. â With "Audio-to-Text" and "Audio-to-Audio" **batch inference** to further boost the performance. **NOTE**: please refer to https://github.com/gpt-omni/mini-omni for more details.