Comments (3)
Could it be the program running out of memory?
If you are using different shapes for each image, there could be an increase in memory consumption per image.
We are caching the priors for each image shape (see here).
You could try to disable the caching, however, this will increase the runtime quite a bit.
from dsfd-pytorch-inference.
Thank you for you quick answer.
Could it be the program running out of memory?
Yes maybe it could be possible but why RetinaNetRestNet50 is not running out of memory on the same images ?
Both detectors do not save in cache the images ?
You could try to disable the caching, however, this will increase the runtime quite a bit.
Ok thank you for the advice it I want to disable the caching I juste have to comment the line you sent ?
And you are saying that it will be slower, but do you know approximately in what proportions ?
from dsfd-pytorch-inference.
Yes maybe it could be possible but why RetinaNetRestNet50 is not running out of memory on the same images?
That's true. We're caching the priors in RetinaNet somewhat different, therefore, it might not run into the same issue.
Ok thank you for the advice it I want to disable the caching I juste have to comment the line you sent ?
Yep!
And you are saying that it will be slower, but do you know approximately in what proportions ?
This is highly dependent on the dataset you have. If all images has a different shape, then it should be equally fast.
Another point is that it might be that you're running out of memory on your GPU. You could limit the max resolution to (for example) 1080.
from dsfd-pytorch-inference.
Related Issues (20)
- detector.detect_with_landmarks error HOT 2
- How to control device HOT 1
- Memory keeps increasing HOT 4
- Problem with importing face_detection.build_detector HOT 3
- Validation accuracy HOT 1
- TRT inference error HOT 3
- Broken DSFD checkpoint URL HOT 3
- Receiving HTTP 308 error while creating detector HOT 1
- Run on CPU HOT 2
- Returning the face landmarks? HOT 2
- Model consumes GPU-0 even if I specify other gpus HOT 2
- Inference Speed Results HOT 2
- How can I use my own model that I trained for recognizing people's age HOT 1
- Zeros in confidence score HOT 1
- Save and load Pytorch models from another directory HOT 1
- Install with Pip not working anymore HOT 1
- RuntimeWarning: overflow encountered in scalar multiply HOT 1
- 1. Thank you! 2. little documentation improvement
- Detector gets slower in a for loop HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from dsfd-pytorch-inference.