Comments (13)
Thanks for sharing the information. I'll try to implement some improvements soon.
from tensorrt_demos.
Thanks for letting me know. Could you share more information about your test case? I will look into it.
- Are you testing this on a Jetson Nano or TX2 or else? Which TensorRT version is used?
- How large (what resolution: 1280x720 or else) is the input image or video?
- Roughly how many (ground-truth) faces are there in the test image?
- How much slower is my code comparing to jetson_nano_demo?
from tensorrt_demos.
Im testing on Jetson Nano
I find out i did not change the min face size LOL,when i change it FPS is nearly same in mtcnn with tensorrt or not
from tensorrt_demos.
I do have a few tricks which could speed up MTCNN. I haven't found time to implement them into the code.
I'd appreciate you sharing with me about your test case: image resolution and how many faces, etc. It could help me to make better design decisions later on.
Thanks.
from tensorrt_demos.
I tested with resolution 1280x720 and 0,1,2 faces cases,i will test more and share result
I'm waiting for your implement.
from tensorrt_demos.
@tranmanhdat I have optimized the TrtPNet code. Comparing to my Sep. 30 code, the current version runs over 30% faster overall. For example, FPS number improved from 5.15 to 6.94 when I test with the same Avengers picture on Jetson Nano.
Please check out my latest code, recompile 'create_engines.cpp' and rebuild the TensorRT engine files. Then run your test again.
I plan to write a blog post about how I implement the optimization soon.
from tensorrt_demos.
thank for sharing,i wil test on my case soon.
could you explain why the current version run 30% over faster,do you change some arguments?
from tensorrt_demos.
Please read my new blog post, Optimizing TensorRT MTCNN, for details.
from tensorrt_demos.
thank for your optimization!
I tested on my computer for your last update and i realized that without tensorRT it's run faster,the result FPS you can see in 2 files i attach.
i config with the same arguments,i used a large video but i was select a ROI ( 750x1000) for testing.
i used scale_factor=709,min_face_size=90 and set the max_batch=3.
FPS_withoutTensorRT.txt
FPS_withTensorRT.txt
from tensorrt_demos.
Thanks for sharing the test results with me. I don't plan to do further optimization on my code for now. But allow me to make a couple of comments:
-
My code was optimized for 1280x720 image inputs. For 750x1000 images, my code would need to do rescaling (additional computation) on every image. In addition, there would be padded zero pixels on the rght-hand side (for PNet) which are wasted. To get the best performance with my implementation, you could modify input blob dimension in det1_relu.prototxt, as well as the corresponding code in
utils/mtcnn.py
. (Refer to my "Optimizing TensorRT MTCNN" blog post for details.) -
In the TensorFlow (non-TensorRT) case, it seems that FPS number fluctuated (34~69) significantly during your test. Do you know why?
from tensorrt_demos.
I changed image to 1280x720 but with tensorRT still lower without tensorRT,
i can't define why FPS increase/decrease so fastly and unstable,i realize that when image have faces FPS will decrease ( base on number of faces in image) and it will increase fast when image dont have any face
from tensorrt_demos.
Thanks again for sharing the result.
from tensorrt_demos.
your implemence show a better performance when have many faces in iamge ( about 3,4 and above) ! it's greate!
from tensorrt_demos.
Related Issues (20)
- Calib.table not created Deepstream
- Error when install install_pycuda.sh file HOT 2
- Any plan about SwinIR? HOT 1
- Batch Processed Image Inference HOT 2
- Failure to Build ONNX from Custom Yolo HOT 2
- Error when running on separate thread HOT 2
- wrong many bounding box and wrong predict untrained class id HOT 2
- error happening when sh ./install_pycuda.sh HOT 3
- Integration With AGX Orin HOT 4
- yolov3 onnx format with low opset, suitable for tensorrt 6.5 HOT 1
- Trouble with linker on yolo plugin x86 HOT 2
- Xavier onnx to TensorRT error HOT 1
- How to prevent tensorrt from fusing the final layers when trying to convert tiny yolov3 to int8? HOT 2
- No output when running my own yolov4
- Failed to build the TensorRT engine --int8
- How do I free up memory when I want to switch models
- Cudnn initialization error: I already installed Cudnn but Why?
- Is there any way to avoid using the reshape operation after inference?
- DRIVE AGX Orin
- Facing Issue with yolo to onnx conversion with pre-trained yolov3-tiny model.
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from tensorrt_demos.