You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I know flash attention is implemented in transformers.
So maybe what should be done is to be able to use either openai-whisper or transformers backend in whisper-timestamped...
But that sounds like a big project.
I added (HuggingFace) transformers backend in whisper-timestamped.
Unfortunately, this does not unlock the use of flash attention, because when enabling flash attention in transformers, there is no more possibility of accessing attention weights (and this is require to guess timestamps).
Also, I did not notice any speedup by using HF transformers instead of openai-whisper implementation.
No description provided.
The text was updated successfully, but these errors were encountered: