mac support

#7
by kunci115 - opened

forked from mini omni, it works end to end with mac m1, m2, m3 chip, but whisper seems a problem while transcribing the voice, if anyone can take a look what's wrong
https://github.com/kunci115/mini-omni-mac-support

Could you please more specifically describe "seems a problem". What are you observing and what are you expecting to happen?

transcribing is not expected as the input speak, its too far, seems a problem means there is something wrong with the way I'm using whisper in that code

https://github.com/kunci115/mini-omni-mac-support/blob/main/inference.py#L359
the model is trained with whisper small, but you load tiny. That might be the problem?

Change it back to small, but still the same result

https://github.com/openai/whisper/blob/279133e3107392276dc509148da1f41bfb532c7e/whisper/model.py#L256
please refer to the whisper model, we use embed_audio for encode the audio feature(use only the encoder part of the whisper), and I think it's not the same as your output.logits which is the output of decoder?

Sign up or log in to comment