mirror of
https://github.com/gpt-omni/mini-omni
synced 2024-11-25 05:21:39 +00:00
Merge branch 'main' of github.com:gpt-omni/mini-omni
This commit is contained in:
commit
2b08f18735
@ -24,7 +24,7 @@ Mini-Omni is an open-source multimodal large language model that can **hear, tal
|
|||||||
|
|
||||||
✅ **Talking while thinking**, with the ability to generate text and audio at the same time.
|
✅ **Talking while thinking**, with the ability to generate text and audio at the same time.
|
||||||
|
|
||||||
✅ **Streaming audio outupt** capabilities.
|
✅ **Streaming audio output** capabilities.
|
||||||
|
|
||||||
✅ With "Audio-to-Text" and "Audio-to-Audio" **batch inference** to further boost the performance.
|
✅ With "Audio-to-Text" and "Audio-to-Audio" **batch inference** to further boost the performance.
|
||||||
|
|
||||||
|
@ -399,7 +399,7 @@ class OmniInference:
|
|||||||
model = self.model
|
model = self.model
|
||||||
|
|
||||||
with self.fabric.init_tensor():
|
with self.fabric.init_tensor():
|
||||||
model.set_kv_cache(batch_size=2)
|
model.set_kv_cache(batch_size=2,device=self.device)
|
||||||
|
|
||||||
mel, leng = load_audio(audio_path)
|
mel, leng = load_audio(audio_path)
|
||||||
audio_feature, input_ids = get_input_ids_whisper_ATBatch(mel, leng, self.whispermodel, self.device)
|
audio_feature, input_ids = get_input_ids_whisper_ATBatch(mel, leng, self.whispermodel, self.device)
|
||||||
|
@ -46,9 +46,9 @@ def create_app():
|
|||||||
return server.server
|
return server.server
|
||||||
|
|
||||||
|
|
||||||
def serve(ip='0.0.0.0', port=60808):
|
def serve(ip='0.0.0.0', port=60808, device='cuda:0'):
|
||||||
|
|
||||||
OmniChatServer(ip, port=port, run_app=True)
|
OmniChatServer(ip, port=port,run_app=True, device=device)
|
||||||
|
|
||||||
|
|
||||||
if __name__ == "__main__":
|
if __name__ == "__main__":
|
||||||
|
Loading…
Reference in New Issue
Block a user