mirror of
https://github.com/gpt-omni/mini-omni
synced 2024-11-25 05:21:39 +00:00
Merge branch 'main' of github.com:gpt-omni/mini-omni
This commit is contained in:
commit
2b08f18735
@ -24,7 +24,7 @@ Mini-Omni is an open-source multimodal large language model that can **hear, tal
|
||||
|
||||
✅ **Talking while thinking**, with the ability to generate text and audio at the same time.
|
||||
|
||||
✅ **Streaming audio outupt** capabilities.
|
||||
✅ **Streaming audio output** capabilities.
|
||||
|
||||
✅ With "Audio-to-Text" and "Audio-to-Audio" **batch inference** to further boost the performance.
|
||||
|
||||
|
@ -399,7 +399,7 @@ class OmniInference:
|
||||
model = self.model
|
||||
|
||||
with self.fabric.init_tensor():
|
||||
model.set_kv_cache(batch_size=2)
|
||||
model.set_kv_cache(batch_size=2,device=self.device)
|
||||
|
||||
mel, leng = load_audio(audio_path)
|
||||
audio_feature, input_ids = get_input_ids_whisper_ATBatch(mel, leng, self.whispermodel, self.device)
|
||||
|
Loading…
Reference in New Issue
Block a user