GithubHelp home page GithubHelp logo

Comments (6)

github-actions avatar github-actions commented on August 19, 2024

👋 Hello @hank1101444, thank you for your interest in Ultralytics YOLOv8 🚀! We recommend a visit to the Docs for new users where you can find many Python and CLI usage examples and where many of the most common questions may already be answered.

If this is a 🐛 Bug Report, please provide a minimum reproducible example to help us debug it.

If this is a custom training ❓ Question, please provide as much information as possible, including dataset image examples and training logs, and verify you are following our Tips for Best Training Results.

Join the vibrant Ultralytics Discord 🎧 community for real-time conversations and collaborations. This platform offers a perfect space to inquire, showcase your work, and connect with fellow Ultralytics users.

Install

Pip install the ultralytics package including all requirements in a Python>=3.8 environment with PyTorch>=1.8.

pip install ultralytics

Environments

YOLOv8 may be run in any of the following up-to-date verified environments (with all dependencies including CUDA/CUDNN, Python and PyTorch preinstalled):

Status

Ultralytics CI

If this badge is green, all Ultralytics CI tests are currently passing. CI tests verify correct operation of all YOLOv8 Modes and Tasks on macOS, Windows, and Ubuntu every 24 hours and on every commit.

from ultralytics.

glenn-jocher avatar glenn-jocher commented on August 19, 2024

Hello!

You're correct in your understanding of the iou and conf parameters relating to the non-maximum suppression (NMS) process, rather than directly affecting the confusion matrix or mAP calculations. These parameters help in filtering out overlapping bounding boxes based on their confidence scores and IoU values during detection.

For modifying the parameters like conf and iou_thres in the confusion matrix calculation, you can indeed adjust these values in the __init__ method of the ConfusionMatrix class within ultralytics/utils/metrics.py. This adjustment will allow you to tailor the confusion matrix calculations to better fit your specific requirements.

Regarding the calculation of mAP using the largest mean F1 score, YOLOv8 typically computes mAP across a range of IoU thresholds (from 0.5 to 0.95) to provide a more comprehensive evaluation metric that balances both precision and recall across these thresholds.

If you have any more questions or need further clarification, feel free to ask. Happy coding! 😊

from ultralytics.

zwcity avatar zwcity commented on August 19, 2024

@glenn-jocher
Hello, why did v8 choose the default 0.45 as the IOU threshold for the confusion matrix instead of 0.5?

from ultralytics.

zwcity avatar zwcity commented on August 19, 2024

@glenn-jocher
Hello, I also noticed that there are two approaches used when calculating the predicted results to match the true labels. One is when calculating the map,

match_predictions():
                        ....
                        matches = matches[iou[matches[:, 0], matches[:, 1]].argsort()[::-1]]
                        matches = matches[np.unique(matches[:, 1], return_index=True)[1]]
                        # matches = matches[matches[:, 2].argsort()[::-1]]
                        matches = matches[np.unique(matches[:, 0], return_index=True)[1]]
                        .....

and the other is when calculating the confusion matrix,

process_batch():
                .....
                matches = matches[matches[:, 2].argsort()[::-1]]
                matches = matches[np.unique(matches[:, 1], return_index=True)[1]]
                matches = matches[matches[:, 2].argsort()[::-1]]
                matches = matches[np.unique(matches[:, 0], return_index=True)[1]]
                .......

Why is there such a difference? It seems more reasonable to calculate the confusion matrix in this way, considering the order of calculating the maximum IOU twice. When using the first calculation method, for a GT with multiple pre target boxes that can be matched, it will default to matching the first target that meets the requirements, rather than the target box corresponding to the maximum iou.

from ultralytics.

glenn-jocher avatar glenn-jocher commented on August 19, 2024

@zwcity hello!

The difference in approaches for matching predictions to true labels between mAP calculation and confusion matrix generation primarily stems from their distinct objectives and the metrics they aim to optimize.

For mAP calculation, the goal is to evaluate the model's precision and recall across different confidence thresholds, which often involves sorting by IoU and then by confidence to prioritize higher quality predictions. This method ensures that the best possible matches are considered first, enhancing the precision metric crucial for mAP.

On the other hand, the confusion matrix computation focuses more on the overall accuracy and class-wise performance, where sorting by confidence twice helps in accurately assigning predictions to true labels, reflecting a more direct measure of prediction correctness.

Each method is tailored to provide the most relevant insights for its respective metric, which is why you see a variation in the sorting and matching process. This nuanced approach helps in fine-tuning the performance evaluation according to the specific needs of each metric.

If you have more questions or need further clarification, feel free to ask. Happy to help! 😊

from ultralytics.

github-actions avatar github-actions commented on August 19, 2024

👋 Hello there! We wanted to give you a friendly reminder that this issue has not had any recent activity and may be closed soon, but don't worry - you can always reopen it if needed. If you still have any questions or concerns, please feel free to let us know how we can help.

For additional resources and information, please see the links below:

Feel free to inform us of any other issues you discover or feature requests that come to mind in the future. Pull Requests (PRs) are also always welcomed!

Thank you for your contributions to YOLO 🚀 and Vision AI ⭐

from ultralytics.

Related Issues (20)

Recommend Projects

  • React photo React

    A declarative, efficient, and flexible JavaScript library for building user interfaces.

  • Vue.js photo Vue.js

    🖖 Vue.js is a progressive, incrementally-adoptable JavaScript framework for building UI on the web.

  • Typescript photo Typescript

    TypeScript is a superset of JavaScript that compiles to clean JavaScript output.

  • TensorFlow photo TensorFlow

    An Open Source Machine Learning Framework for Everyone

  • Django photo Django

    The Web framework for perfectionists with deadlines.

  • D3 photo D3

    Bring data to life with SVG, Canvas and HTML. 📊📈🎉

Recommend Topics

  • javascript

    JavaScript (JS) is a lightweight interpreted programming language with first-class functions.

  • web

    Some thing interesting about web. New door for the world.

  • server

    A server is a program made to process requests and deliver data to clients.

  • Machine learning

    Machine learning is a way of modeling and interpreting data that allows a piece of software to respond intelligently.

  • Game

    Some thing interesting about game, make everyone happy.

Recommend Org

  • Facebook photo Facebook

    We are working to build community through open source technology. NB: members must have two-factor auth.

  • Microsoft photo Microsoft

    Open source projects and samples from Microsoft.

  • Google photo Google

    Google ❤️ Open Source for everyone.

  • D3 photo D3

    Data-Driven Documents codes.