Comments (15)
with model large v3
from stt.
what output at cmd ?
set cuda_cmd_type =int8 retry
from stt.
CMD OUTPUT -- NOTHING EXCEPT 1983 I MENTIONED LAST ISSUE.
I WILL TRY int8 LATER. THANKS A LOT.
from stt.
the problem was happened again, I will change to cpu.
from stt.
cuda_cmd_type = int8_float16
if source code deploy, line 107
segments,info = modelobj.transcribe(wav_file, beam_size=1,best_of=1,temperature=0, vad_filter=True, vad_parameters=dict(min_silence_duration_ms=500),language=language)
change to
segments,info = modelobj.transcribe(wav_file, beam_size=5,best_of=5, vad_filter=True, vad_parameters=dict(min_silence_duration_ms=500),language=language)
from stt.
cpu test passed with normal result srt files forabout 5500s, will try cuda_cmd_type = int8_float16 later.
from stt.
one mp4 file with 100MB size, about 68 mins, cuda with int8_float16 , the result display nothing after recognize.
another mp4 file with 100MB size, about 62 mins, cuda with int8_float16, the result displayed.
from stt.
cannot show
update to 0.91 and open set.ini and try adjusting several parameters at the bottom, all with comments. You can try to adjust them according to the maximum and minimum GPU consumption
from stt.
wow its hard to ajust.... I dont know how to ajust, comments not so clear.
from stt.
web_address=127.0.0.1:9977
lang=en
devtype=cuda
cuda_com_type=float32
beam_size=5
best_of=5
vad=true
temperature=1
condition_on_previous_text=true
This is the best effect, but it also consumes the most GPU
web_address=127.0.0.1:9977
lang=en
devtype=cuda
cuda_com_type=int8
beam_size=1
best_of=1
vad=false
temperature=0
condition_on_previous_text=false
This is the most GPU efficient configuration and the effect is relatively poor
from stt.
The speech recognition to subtitle function in this project is the same as the current project's speech recognition function, both are fast wheelers. Perhaps you can download this and give it a try.
https://github.com/jianchang512/pyvideotrans
from stt.
web_address=127.0.0.1:9977
lang=en
devtype=cuda
cuda_com_type=float32
beam_size=5
best_of=5
vad=true
temperature=1
condition_on_previous_text=true
consume all my GPU that is nice
but this performance velocity or what si difference at final?
from stt.
{'web_address': '127.0.0.1:9977', 'lang': 'en', 'devtype': 'cuda', 'cuda_com_type': 'float32', 'beam_size': 5, 'best_of': 5, 'vad': True, 'temperature': 1, 'condition_on_previous_text': True}
The browser is open. If it does not open automatically, please open the URL manually http://127.0.0.1:9977
res.status_code=200
d={'version': 'v0.0.91', 'version_num': 91}
2024-01-29 21:28:53.8140166 [W:onnxruntime:Default, onnxruntime_pybind_state.cc:1983 onnxruntime::python::CreateInferencePybindStateModule] Init provider bridge failed.
CUDA failed with error out of memory
what option i have for prevent our of memory?
from stt.
pyvideotrans perfcet tools
from stt.
{'web_address': '127.0.0.1:9977', 'lang': 'en', 'devtype': 'cuda', 'cuda_com_type': 'float32', 'beam_size': 5, 'best_of': 5, 'vad': True, 'temperature': 1, 'condition_on_previous_text': True}
The browser is open. If it does not open automatically, please open the URL manually http://127.0.0.1:9977 res.status_code=200 d={'version': 'v0.0.91', 'version_num': 91} 2024-01-29 21:28:53.8140166 [W:onnxruntime:Default, onnxruntime_pybind_state.cc:1983 onnxruntime::python::CreateInferencePybindStateModule] Init provider bridge failed. CUDA failed with error out of memory what option i have for prevent our of memory?
web_address=127.0.0.1:9977
lang=en
devtype=cuda
cuda_com_type=int8
beam_size=1
best_of=1
vad=false
temperature=0
condition_on_previous_text=false
This is the most GPU efficient configuration and the effect is relatively poor
from stt.
Related Issues (20)
- 处理结果为null
- Error opening output file HOT 1
- 不支持m4a音频文件
- 文件不完整 HOT 1
- model HOT 1
- 数字识别不太好,有什么解决办法吗? HOT 1
- API报错None Model file does not exist HOT 1
- 请问是否有ROCm(类似CUDA的AMD GPU深度学习框架)支持 HOT 1
- 无法通过内网访问页面和接口 HOT 1
- 配置怎么算的,比如长视频用large的模型,多长时间的视频要用的大概什么样的配置【内存、GPU】 HOT 1
- 进度条卡着0%不动 HOT 1
- Language
- What is the reason for the following error message when the project is identified?
- large-v3 HOT 1
- RuntimeError: parallel_for failed: cudaErrorNoKernelImageForDevice: no kernel image is available for execution on the device
- large-v3是不是只支持英文转换
- 各位大佬,请问为什么使用了large-v3,却依然出现识别幻觉?
- SRT subtitle incorrect timestamp
- GPL3.0!不怎么友好
- 此项目可以自己训练语音改进码 HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from stt.