Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Large-v3 hallucination problems #4

Open
MaximeDde opened this issue Mar 13, 2024 · 5 comments
Open

Large-v3 hallucination problems #4

MaximeDde opened this issue Mar 13, 2024 · 5 comments

Comments

@MaximeDde
Copy link

Hello Thomas, and thanks again for this great cog !

I've been finding more and more problems with the large-v3 model though - much more hallucinations than with large-v2, and a global precision really, really lower than with its predecessor. (I've seen it documented in several places, I can find the links again for you but https://deepgram.com/learn/whisper-v3-results this article went the more in-depth I believe).

Do you think there could be a way to have this cog with the large-v2 model instead ? If that doesn't mean too much work on your side !!

@thomasmol
Copy link
Owner

Hi there, thanks!
I know it's a common problem unfortunately. I just pushed an update so the model uses a new version of faster-whisper (1.0.1) which should mitigate the problems with v3 over v2. Let me know if it fixes it for you!

@MaximeDde
Copy link
Author

I'll give it a try and give feedback here, thank you for the update !

@thomasmol
Copy link
Owner

had to revert back to an older version of faster-whisper unfortunately, seems to be an issue with inference speed of the diarization. probably because of some conflicting python/cuda packages. will update when this is fixed

@MaximeDde
Copy link
Author

Yes, I saw that... I reverted back to the version just before :( Hopefully they can solve this soon ! Could you please keep me posted ?

@thomasmol
Copy link
Owner

yes will do!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants