Comments (16)
- I dunno, you tell me if GPU processing works on Windows 7.
- You can download cuBLAS nad cuDNN libs from there: https://github.com/Purfview/whisper-standalone-win/releases/tag/libs
Place libs in the same folder where Faster-Whisper executable is.
from whisper-standalone-win.
Sorry for reopening this. Not sure if I need a new issue. Can anyone help solve this?
Faster-Whisper r125 running on: CPU
"D:\whisper-fast\__main__.py", line 445, in <module>
"D:\whisper-fast\__main__.py", line 355, in cli
"faster_whisper\transcribe.py", line 123, in __init__
RuntimeError: mkl_malloc: failed to allocate memory
[1164] Failed to execute script '__main__' due to unhandled exception!
The files from cuBLAS.and.cuDNN.7z are where Whisper is. In NVIDIA Control Panel, only 3D Settings are available, where High-performance NVIDIA processor is applied to all. Among the settings is CUDA GPUs - All. From NVIDIA System Information:
CUDA Cores: 48
Core clock: 475 MHz
Shader clock: 950 MHz
Memory data rate: 1334 MHz
Memory interface: 64-bit
Memory bandwidth: 10.67 GB/s
Total available graphics memory: 2533 MB
Dedicated video memory: 1024 MB DDR3
System video memory: 0 MB
Shared system memory: 1509 MB
NVCUDA.DLL v7.5.15
from whisper-standalone-win.
Dedicated video memory: 1024 MB DDR3
This looks very low.
What is you GPU, CPU and RAM?
Try model=tiny
.
from whisper-standalone-win.
CPU processing worked with the Tiny model on 4/1GB RAM/VRAM. But GPU processing still won't work even on 8/2GB in Windows 7. I wonder if the DLLs in cuBLAS.and.cuDNN.7z are not for 7.
2) Off this topic a bit, I need time-coding more than transcription. In the SRT file, all the start times follow the end times (these are correct) immediately (e.g. End time 1 = Start time 2), resulting in impossibly long show times without interruptions. Is this due to the model used?
from whisper-standalone-win.
- What model is your GPU.
- What version of Standalone Faster-Whisper you are using?
- Post a command line parameters you are using.
from whisper-standalone-win.
My GPU is GF610M, 1GB. I can get details from a GPU reporting utility. I'm studying [https://docs.nvidia.com/deploy/pdf/CUDA_Compatibility.pdf](compatibility issues). My omission. A search for 'Whisper CUDA' found that you need to add '--device cuda' to enable it. Now I get Faster-Whisper r125 running on: CUDA, but errors too. The same as with 'RuntimeError: mkl_malloc: failed to allocate memory' when using the model medium with CPU processing:
"D:\whisper-fast\__main__.py", line 445, in <module>
"D:\whisper-fast\__main__.py", line 355, in cli
"faster_whisper\transcribe.py", line 123, in __init__
RuntimeError: CUDA failed with error initialization error
[3032] Failed to execute script '__main__' due to unhandled exception!
Errors when using --help:
File "D:\whisper-fast\__main__.py", line 445, in <module>
File "D:\whisper-fast\__main__.py", line 277, in cli
File "argparse.py", line 1768, in parse_args
File "argparse.py", line 1800, in parse_known_args
File "argparse.py", line 2006, in _parse_known_args
File "argparse.py", line 1946, in consume_optional
File "argparse.py", line 1874, in take_action
File "argparse.py", line 1044, in __call__
File "argparse.py", line 2494, in print_help
File "argparse.py", line 2500, in _print_message
File "encodings\cp1251.py", line 19, in encode
UnicodeEncodeError: 'charmap' codec can't encode character '\xbf' in position 88
21: character maps to <undefined>
from whisper-standalone-win.
Correction: Compatibility issues
from whisper-standalone-win.
You need newer GPU.
For better timestamps get latest r128 version.
from whisper-standalone-win.
I get similar errors with Faster-Whisper r134+++ when trying to run on CUDA. Is GTX960M, 2GB VRAM, still not enough?
from whisper-standalone-win.
Run what? What errors?
Maybe problem is with mobile GPU or its drivers, or whatever.
from whisper-standalone-win.
Errors (as above) with -h (help not really necessary, just testing) and this
Whisper-Faster_r134 %Audio% --language en --output_format srt --task transcribe --model tiny --device cuda
Is GTX960M, 2GB VRAM, not enough for CUDA?
from whisper-standalone-win.
To solve problem with --help
you may need to change language to English in Console Panel>Region and Language.
2GB is enough for tiny model.
In laptop you maybe need to use the second GPU device, try: --device cuda:1
.
from whisper-standalone-win.
'cuda:1' didn't help. Thank you! I'll try to find the cause and share it if there is one.
I suspected the reason for errors with -h was the locale because of 'cp1251.py … UnicodeEncodeError'.
from whisper-standalone-win.
The title of this issue should be 'How to make it run on CUDA'.
I have so far failed to do it on three architectures of Nvidia GPUs (oldest to newest): Fermi, Maxwell, Hopper (probably Hopper as I didn't check the model, but it's built in late 2022; RuntimeError: CUDA failed with error CUDA driver version is insufficient for CUDA runtime version.) The subject matter is complex to the uninitiated. I refer anyone to this table in hopes of finding suitable drivers or anything else sooner.
from whisper-standalone-win.
The title of this issue should be 'How to make it run on CUDA'.
It should be as the original issue. You shouldn't post the different issues here.
from whisper-standalone-win.
I meant a new title for this issue.
from whisper-standalone-win.
Related Issues (20)
- Whisper Timestamped HOT 11
- Google Soundstorm HOT 1
- Virus Reported In Latest Version HOT 5
- Hallucination loop HOT 24
- Faster-Whisper inside Docker fails HOT 1
- Seems like it's using all cpu cores? HOT 1
- RuntimeError: Library cublas64_11.dll is not found or cannot be loaded HOT 1
- Could not load library libcudnn_ops_infer.so.8 HOT 11
- missing segments HOT 30
- dot HOT 4
- Support for WhisperX standalone HOT 6
- Support for Linux ARM64 Architecture HOT 1
- Support for other language models HOT 1
- please release Whisper-Faster_r172.1_linux.zip HOT 4
- Large V2 versus Large V3 HOT 4
- RecursionError HOT 2
- missing lines HOT 6
- Help needed to do tests HOT 58
- [Feature-Request] Output pure text with 1 sentence per line HOT 1
- [Question] max_line_width and max_line_count HOT 3
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from whisper-standalone-win.