mirror of
https://github.com/gpt-omni/mini-omni
synced 2024-11-16 05:03:47 +00:00
92 lines
2.4 KiB
Markdown
92 lines
2.4 KiB
Markdown
|
|
# Mini-Omni
|
|
|
|
<p align="center">
|
|
Mini-Omni: Language Models Can Hear, Talk While Thinking in Streaming
|
|
<p>
|
|
|
|
<p align="center">
|
|
🤗 <a href="">Hugging Face</a> | 📖 <a href="https://github.com/gpt-omni/mini-omni">Github</a>
|
|
| 📑 <a href="">Technical report (coming soon)</a>
|
|
<p>
|
|
|
|
Mini-Omni is an open-source multimodel large language model that can **hear, talk while thinking**. Featuring real-time end-to-end speech input and **streaming audio output** conversational capabilities.
|
|
|
|
<p align="center">
|
|
<img src="data/figures/frameworkv3.jpg" width="100%"/>
|
|
<p>
|
|
|
|
|
|
## Features
|
|
|
|
✅ **Real-time speech-to-speech** conversational capabilities. No extra ASR or TTS models required.
|
|
|
|
✅ **Talking while thinking**, with the ability to generate text and audio at the same time.
|
|
|
|
✅ **Streaming audio outupt** capabilities.
|
|
|
|
✅ With "Audio-to-Text" and "Audio-to-Audio" **batch inference** to further boost the performance.
|
|
|
|
## Demo
|
|
|
|
![](./data/demo_streamlit.mov)
|
|
|
|
## Install
|
|
|
|
Create a new conda environment and install the required packages:
|
|
|
|
```sh
|
|
conda create -n omni python=3.10
|
|
conda activate omni
|
|
|
|
git clone https://github.com/gpt-omni/mini-omni.git
|
|
cd mini-omni
|
|
pip install -r requirements.txt
|
|
```
|
|
|
|
## Quick start
|
|
|
|
**Interactive demo**
|
|
|
|
- start server
|
|
```sh
|
|
conda activate omni
|
|
cd mini-omni
|
|
python3 server.py --ip '0.0.0.0' --port 60808
|
|
```
|
|
|
|
- run streamlit demo
|
|
|
|
NOTE: you need to run streamlit locally with PyAudio installed.
|
|
|
|
```sh
|
|
pip install PyAudio==0.2.14
|
|
API_URL=http://0.0.0.0:60808/chat streamlit run webui/omni_streamlit.py
|
|
```
|
|
|
|
- run gradio demo
|
|
```sh
|
|
API_URL=http://0.0.0.0:60808/chat python3 webui/omni_gradio.py
|
|
```
|
|
|
|
example:
|
|
|
|
![](./data/demo_gradio.mov)
|
|
|
|
**Local test**
|
|
|
|
```sh
|
|
conda activate omni
|
|
cd mini-omni
|
|
# test run the preset audio samples and questions
|
|
python inference.py
|
|
```
|
|
|
|
## Acknowledgements
|
|
|
|
- [Qwen2](https://github.com/QwenLM/Qwen2/) as the LLM backbone.
|
|
- [litGPT](https://github.com/Lightning-AI/litgpt/) for training and inference.
|
|
- [whisper](https://github.com/openai/whisper/) for audio encoding.
|
|
- [snac](https://github.com/hubertsiuzdak/snac/) for audio decoding.
|
|
- [CosyVoice](https://github.com/FunAudioLLM/CosyVoice) for generating synthetic speech.
|
|
- [OpenOrca](https://huggingface.co/datasets/Open-Orca/OpenOrca) and [MOSS](https://github.com/OpenMOSS/MOSS/tree/main) for alignment. |