Skip to content

It does not work normally on the RTX 5070 TI. #1260

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Open
c3xingchen opened this issue Mar 10, 2025 · 12 comments
Open

It does not work normally on the RTX 5070 TI. #1260

c3xingchen opened this issue Mar 10, 2025 · 12 comments

Comments

@c3xingchen
Copy link

The RTX 5070 TI encountered a RuntimeError: cuBLAS failed with status CUBLAS_STATUS_NOT_SUPPORTED. When the --compute_type float32 option was added, it could run, but long videos did not work properly. In videos longer than 2 hours, the program only worked for the first 33 minutes and then abruptly stopped.

@Purfview
Copy link
Contributor

Check what is your cuBLAS version or what CUDA Toolkit version you have installed?

@c3xingchen
Copy link
Author

Check what is your cuBLAS version or what CUDA Toolkit version you have installed?

I am not a specialist in this area of technology. I have not independently installed cBLAS and CUDA Toolkit myself. I use PotPlayer to generate subtitles using this model on my laptop with an RTX 3060, which works fine. However, when I try the same process on my RTX 5070 Ti, it encounters issues.

@Purfview
Copy link
Contributor

RTX 5070

Don't know why compute_type auto [int8] doesn't work with these GPUs, use --compute_type float16

I use PotPlayer to generate subtitles using this model

Then you are in the wrong repo, go there: https://github.com/Purfview/whisper-standalone-win

@ictsmc
Copy link

ictsmc commented Mar 12, 2025

I also can't use it with my 5070ti. Basically, all 50 series cards are unusable.

@Purfview
Copy link
Contributor

I also can't use it with my 5070ti. Basically, all 50 series cards are unusable.

Use proper settings.

@teddybear082
Copy link

RTX 5070

Don't know why compute_type auto [int8] doesn't work with these GPUs, use --compute_type float16

I use PotPlayer to generate subtitles using this model

Then you are in the wrong repo, go there: https://github.com/Purfview/whisper-standalone-win

Thanks for this, just helped me as well, can confirm did not work on auto or int8 but did work on float16.

@Purfview
Copy link
Contributor

@teddybear082 @ictsmc, do you have this issue when using Python and this repo?

@teddybear082
Copy link

@teddybear082 @ictsmc, do you have this issue when using Python and this repo?

I’m using faster whisper python library via WingmanAI by ShipBit: https://github.com/ShipBit/wingman-ai. They use pyinstaller to turn the python into an exe I believe and faster-whisper is one of the dependencies.

@Purfview
Copy link
Contributor

Purfview commented Mar 29, 2025

So not the Python directly. Kinda strange that in my repo I have lots of reports about this, but I don't see any reports in Python repos.

Btw, similar reports about pyannote and 50xxx GPUs, but none in pyannote repo too.

@teddybear082
Copy link

So not the Python directly. Kinda strange that in my repo I have lots of reports about this, but I don't see any reports in Python repos.

Btw, similar reports about pyannote and 50xxx GPUs, but none in pyannote repo too.

What do you mean not the python directly? Isn't this the repo for the faster-whisper pipy python project? Wingman depends on faster-whisper=1.1.1 python library I believe. I may just be confusing what you mean.

@Purfview
Copy link
Contributor

Purfview commented Mar 29, 2025

I meant using Python directly, not the exe compiled with pyinstaller.
And it strange that all reports about 50xx comes only from "exe" repos.

Maybe because the original default for compute_type is not "auto", I don't remember now, it could be "default".

@Purfview
Copy link
Contributor

Here is link to the issue at CTranslate2: OpenNMT/CTranslate2#1865

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

4 participants