Comments (7)
I HAVE FOUND A BUG IN MY CODE, IT WAS A FALSE ALARM
Just a quick update, for the sake of future usage - I have found a tiny (but major, obviously) bug on my python NMS, that was replacing the original torch implementation of NMS from the original YoloX repository
After fixing the bug - I was able to assert that the overall YoloX .pth -> .onnx -> .pb (TF1.X format) is almost perfect, meaning identical up to a tiny fraction of an error between the 3 models (the error is negligible for most scenarios, and is almost zero when testing on multiple scenarios of model weights and benchmarks)
Thanks for the support
from onnx2tf.
pip show onnx2tf
Name: onnx2tf
Version: 1.18.14
onnx2tf -i yolox_s.onnx -cotof
With the understanding that you are looking at the same conversion log as I am, the output is an exact match. It must be a problem with your logic or TensorFlow v1.
dup: #507
Frankly, I don't recommend the Protocol Buffer, and I don't think Google is nearly willing to do it either.
from onnx2tf.
Thanks for your quick reply
I wan to point out that I am using TF 2.13, so maybe the "TF<2.10.0" tag is irrelevant
Also - I do not think that this is a duplicate of #507 due to the specific code that both of us (that was me on #507 :) ) added and is not available anywhere on this repository. It is a very complex set of lines that without them - one cannot make a good usage of the .pb model.
As for your tip - Protocol Buffer is being used in TF by default, so basically you are suggesting to not use .pb models at all?
I was not aware that PBs were not a good practice, could you elaborate and maybe point out to a better solution for TF?
from onnx2tf.
I wan to point out that I am using TF 2.13, so maybe the "TF<2.10.0" tag is irrelevant
The .pb of TF v1.x and the saved_model.pb of TF v2.x have distinctly different specifications. For example, in v1.x .pb, differences in the parameters available for Resize and bugs in TensorFlow's internal implementation are still present. As a splice, Keras is forcibly merged and forcibly redirected to V1 logic via compat.v1.
In other words, we are aware that no maintenance (including bugs in the specification) has been done on the parts of the system where the V1-based internal logic is called.
I'm not going to go back and read TensorFlow code that is two to three years old logic and try to solve the problem.
https://github.com/tensorflow/tensorflow/tree/v1.15.5
from onnx2tf.
Oh, I get what you are saying...
Thanks for the explanation.
Maybe I will try and see if I could replace my TF V1.x server to something more stable
And once again - thanks for the wonderful work that you are doing here.
from onnx2tf.
If there is no activity within the next two days, this issue will be closed automatically.
from onnx2tf.
Excellent. Thank you for sharing.
from onnx2tf.
Related Issues (20)
- Conversion of YoloV5, conversion of StridedSlice from 5d to 4d HOT 7
- Failed to convert Real-ESRGAN x4-general (NCHW) HOT 1
- Error when onnx2tf: Floating point exception (core dumped) HOT 10
- Full int8 quantized model doesn't work when "flatc" is installed on the system HOT 8
- ValueError: Exception encountered when calling layer "tf.math.multiply_46" (type TFOpLambda).
- 0 trainable parameters HOT 1
- Facing onnx2tf converstion issue with resnet model mentioned in README HOT 1
- pre_explicit_broadcast should not expend scalar tensor HOT 13
- "GEMM: Dimension mismatch" during value validation phase HOT 4
- Exclude selected operations from INT8 quantized model HOT 18
- ValueError in MaxPool Layer during ONNX to TensorFlow Conversion with 3D Input HOT 5
- The tflite detection result converted from the yolo NAS model is inaccurate. HOT 5
- Converting quantized onnx with Q/DQ representation to full int8 TFLite model HOT 2
- Unable to replace flatten HOT 4
- How to convert onnx(NCHW) to NHWC and write back to onnx file Because my HW takes onnx file as input in NHWC format HOT 3
- How to convert onnx(NCHW) to NHWC(without adding transpose) and write back to onnx file HOT 1
- How to convert onnx(NCHW) to NHWC(without adding transpose) and write back to onnx file HOT 3
- Please do reply for the last time either this issue or @ How to convert onnx(NCHW) to NHWC(without adding transpose) and write back to onnx file #586 HOT 1
- ONNX Split operation converts to a tf StridedSlice operation with outputs in the wrong order. HOT 2
- Depth_Anything to ONNX model's conversion failed HOT 1
Recommend Projects
-
React
A declarative, efficient, and flexible JavaScript library for building user interfaces.
-
Vue.js
🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.
-
Typescript
TypeScript is a superset of JavaScript that compiles to clean JavaScript output.
-
TensorFlow
An Open Source Machine Learning Framework for Everyone
-
Django
The Web framework for perfectionists with deadlines.
-
Laravel
A PHP framework for web artisans
-
D3
Bring data to life with SVG, Canvas and HTML. 📊📈🎉
-
Recommend Topics
-
javascript
JavaScript (JS) is a lightweight interpreted programming language with first-class functions.
-
web
Some thing interesting about web. New door for the world.
-
server
A server is a program made to process requests and deliver data to clients.
-
Machine learning
Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.
-
Visualization
Some thing interesting about visualization, use data art
-
Game
Some thing interesting about game, make everyone happy.
Recommend Org
-
Facebook
We are working to build community through open source technology. NB: members must have two-factor auth.
-
Microsoft
Open source projects and samples from Microsoft.
-
Google
Google ❤️ Open Source for everyone.
-
Alibaba
Alibaba Open Source for everyone
-
D3
Data-Driven Documents codes.
-
Tencent
China tencent open source team.
from onnx2tf.