Ultralytics
Ultralytics
  • Видео 316
  • Просмотров 1 548 866
How High-Quality Datasets Enhance Computer Vision Performance | Low-Quality Datasets Challenges 🚀
In this tutorial, we will explore the importance of high-quality computer vision datasets. This tutorial will guide you through the key aspects of dataset quality, the different types of datasets used in computer vision, and the critical traits that define a high-quality dataset. Additionally, you will learn about the challenges posed by low-quality datasets and how data-splitting strategies impact model performance. This tutorial is designed to help you understand the foundational role datasets play in building robust computer vision models.
🔗 Discover more ➡️ www.ultralytics.com/blog/the-importance-of-high-quality-computer-vision-datasets
Key highlights:
00:00 - Introduction to Dataset Qua...
Просмотров: 199

Видео

How to do Package Segmentation using Ultralytics YOLO11 | Industrial Packages | Model Training 🎉
Просмотров 2937 часов назад
In this tutorial, we will explore the Package Segmentation Dataset and its practical applications. It will guide you through the process of understanding the dataset, using the Google Colab notebook for implementation, and training the Ultralytics YOLO11 segmentation model on it. Additionally, you will learn how to validate the trained model, run inference, and export the model into different f...
How to Track Objects in Region using Ultralytics YOLO11 | TrackZone | Ultralytics Solutions 🚀
Просмотров 74219 часов назад
In this tutorial, we will explore the TrackZone solution and its practical applications. It will guide you through the process of understanding TrackZone's use cases, using TrackZone via Python and Command Line Interface (CLI) code, and implementing it in a Google Colab notebook. Additionally, you will learn how to initialize, configure, and fine-tune TrackZone for video tracking. This tutorial...
How to Contribute to Ultralytics Repository | Ultralytics Models, Datasets and Documentation 🚀
Просмотров 281День назад
In this tutorial, we will explore the process of contributing to the Ultralytics repository. This tutorial will walk you through the steps to fork, clone, and modify the Ultralytics codebase while ensuring seamless contributions to the project. Additionally, you will learn how to contribute to the Ultralytics documentation and datasets. This tutorial is created to help you get started with open...
How Computer Vision and AI Can Be Used in Human Resource Practices | Ultralytics YOLO11 | VLM's 🚀
Просмотров 36914 дней назад
In this tutorial, we will explore the role of Artificial Intelligence in Human Resources (HR) management, including its applications, market growth, and the advantages and challenges of using computer vision in HR processes. You'll gain a comprehensive understanding of how AI is transforming HR and its potential impact on the industry. Key Highlights: 00:00 - Introduction to Artificial Intellig...
How to train Ultralytics YOLO11 Model on Medical Pills Detection Dataset in Google Colab | Pharma AI
Просмотров 53914 дней назад
In this tutorial, we will explore the process of detecting medical pills using the Ultralytics YOLO11 model. It can be used in pharmaceutical AI applications for tasks such as quality control and inventory management. From setting up the dataset to training, inference, and exporting the model, you'll learn the essential steps for building a pill detection system. This tutorial leverages Google ...
How to Maintain Computer Vision Models after Deployment | Data Drift Detection | Ultralytics YOLO11🚀
Просмотров 51621 день назад
In this tutorial, we will explore model maintenance, its significance, and how to effectively manage your models over time. You'll gain insights into the key elements of model maintenance, tools for monitoring, and best practices for ensuring optimal performance. Key Highlights: 00:00 - Introduction to Model Maintenance: Overview of the importance and role of maintaining computer vision models....
Powering CV Projects with Hugging Face's Open Source Tools | Ultralytics YOLO Vision 2024 | Talk#14🚀
Просмотров 35228 дней назад
As we dive into another exciting highlight from YOLO Vision 2024, we bring you a captivating session featuring Pavel Lakubovskii from Hugging Face. This insightful presentation explores the transformative power of the Hugging Face ecosystem, its versatile libraries, and the tools revolutionizing NLP, computer vision, and beyond. Pavel shares expert insights into challenges, solutions, and innov...
Accelerating Edge AI: Optimizing Computer Vision Inference with NVIDIA | YV24 Talk #08 | Ultralytics
Просмотров 314Месяц назад
Accelerating Edge AI: Optimizing Computer Vision Inference with NVIDIA | YV24 Talk #08 | Ultralytics
How to Export the Ultralytics YOLO11 to ONNX, OpenVINO and Other Formats using Ultralytics HUB 🚀
Просмотров 569Месяц назад
How to Export the Ultralytics YOLO11 to ONNX, OpenVINO and Other Formats using Ultralytics HUB 🚀
How to Run Ultralytics Solutions from the Command Line (CLI) | Ultralytics YOLO11 | Object Counting🚀
Просмотров 778Месяц назад
How to Run Ultralytics Solutions from the Command Line (CLI) | Ultralytics YOLO11 | Object Counting🚀
How YOLO is Spotting Sharks in California | Shark Alert | Ultralytics YOLO Vision 2024 | Talk#12
Просмотров 287Месяц назад
How YOLO is Spotting Sharks in California | Shark Alert | Ultralytics YOLO Vision 2024 | Talk#12
How to Test Machine Learning Models | Avoid Data Leakage in Computer Vision | Ultralytics YOLO11🚀
Просмотров 603Месяц назад
How to Test Machine Learning Models | Avoid Data Leakage in Computer Vision | Ultralytics YOLO11🚀
How Diffusion Models Work | Forward and Reverse Diffusion Process | Challenges and Limitations 🚀
Просмотров 355Месяц назад
How Diffusion Models Work | Forward and Reverse Diffusion Process | Challenges and Limitations 🚀
Building Communities of Innovation | Ultralytics YOLO VISION 2024 | Panel Discussion | VisionAI 🚀
Просмотров 123Месяц назад
Building Communities of Innovation | Ultralytics YOLO VISION 2024 | Panel Discussion | VisionAI 🚀
Tesla’s Autonomous Cybercab and Robovan | Future of AI in Robotics | Computer Vision Applications 🚀
Просмотров 256Месяц назад
Tesla’s Autonomous Cybercab and Robovan | Future of AI in Robotics | Computer Vision Applications 🚀
How to Train Ultralytics YOLO11 on CIFAR-100 | Step-by-Step Image Classification Tutorial 🚀
Просмотров 505Месяц назад
How to Train Ultralytics YOLO11 on CIFAR-100 | Step-by-Step Image Classification Tutorial 🚀
Evolving from Objects to Behaviour: Making AI Useful - The Main Branch | YV24 Talk #08 | Ultralytics
Просмотров 270Месяц назад
Evolving from Objects to Behaviour: Making AI Useful - The Main Branch | YV24 Talk #08 | Ultralytics
How to Tune Hyperparameters for Better Model Performance | Ultralytics YOLO11 Hyperparameters 🚀
Просмотров 1,6 тыс.Месяц назад
How to Tune Hyperparameters for Better Model Performance | Ultralytics YOLO11 Hyperparameters 🚀
How to train Ultralytics YOLO11 on Custom Dataset using Ultralytics HUB | Ultralytics HUB Datasets 🚀
Просмотров 965Месяц назад
How to train Ultralytics YOLO11 on Custom Dataset using Ultralytics HUB | Ultralytics HUB Datasets 🚀
Accelerate YOLO Deployment with Qualcomm AI Hub 🚀 | Ultralytics YOLO VISION 2024 | YV24 Talk #08
Просмотров 254Месяц назад
Accelerate YOLO Deployment with Qualcomm AI Hub 🚀 | Ultralytics YOLO VISION 2024 | YV24 Talk #08
How to count Ships using Ultralytics YOLO11 Oriented Bounding Boxes (YOLO11-OBB) | Object Counting 🚀
Просмотров 7882 месяца назад
How to count Ships using Ultralytics YOLO11 Oriented Bounding Boxes (YOLO11-OBB) | Object Counting 🚀
History of Computer Vision Models | OpenCV | Pattern Recognition | YOLO Models | Ultralytics YOLO11🚀
Просмотров 5952 месяца назад
History of Computer Vision Models | OpenCV | Pattern Recognition | YOLO Models | Ultralytics YOLO11🚀
reCamera: The First Open Source Programmable AI Camera with Native YOLO-SeeedStudio🚀 | YV24 Talk #07
Просмотров 1,3 тыс.2 месяца назад
reCamera: The First Open Source Programmable AI Camera with Native YOLO-SeeedStudio🚀 | YV24 Talk #07
How to Train Ultralytics YOLO11 Model on Custom Dataset using Google Colab Notebook | Step-by-Step 🚀
Просмотров 3,7 тыс.2 месяца назад
How to Train Ultralytics YOLO11 Model on Custom Dataset using Google Colab Notebook | Step-by-Step 🚀
Insights into Model Evaluation and Fine-Tuning | Tips for Improving Mean Average Precision | YOLO11🚀
Просмотров 1,1 тыс.2 месяца назад
Insights into Model Evaluation and Fine-Tuning | Tips for Improving Mean Average Precision | YOLO11🚀
Going YOLO on Lightning Studios - Lightning AI 🚀 | Ultralytics YOLO VISION 2024 | YV24 Talk #06
Просмотров 3492 месяца назад
Going YOLO on Lightning Studios - Lightning AI 🚀 | Ultralytics YOLO VISION 2024 | YV24 Talk #06
OpenAI o1 Model: A New Series of OpenAI Models for AI Reasoning | Two R's in the word strawberry 🚀
Просмотров 7002 месяца назад
OpenAI o1 Model: A New Series of OpenAI Models for AI Reasoning | Two R's in the word strawberry 🚀
Ultralytics YOLO Vision 2024 | Recap
Просмотров 3022 месяца назад
Ultralytics YOLO Vision 2024 | Recap
How to do Object Counting in Different Regions using Ultralytics YOLO11 | Ultralytics Solutions 🚀
Просмотров 1,5 тыс.2 месяца назад
How to do Object Counting in Different Regions using Ultralytics YOLO11 | Ultralytics Solutions 🚀

Комментарии

  • @o7s-EmilyW
    @o7s-EmilyW 4 часа назад

    Fascinating application of YOLO11! How does the model handle visually similar pills with minute differences, like size or markings-especially in scenarios with suboptimal lighting? Could this raise concerns about misclassification in critical pharmaceutical settings?

  • @IsaacZepeda
    @IsaacZepeda 6 часов назад

    how large the train, test and validate datasets should be? how many images would be fine?

  • @screech9872
    @screech9872 7 часов назад

    Well explained, I saw a lot of people scratching their heads and switching between different models. But they don't understand that the data on which the model is trained is a most important thing to note. Thanks for the explanation.

  • @TheodoreBC
    @TheodoreBC 12 часов назад

    Cool vid, bro! But what if the package dataset has, like, weird overlaps or damaged labels? Does YOLO11 handle that like a champ or fold under pressure? Asking for a friend stacking boxes in a warehouse.

    • @Ultralytics
      @Ultralytics 7 часов назад

      Thanks for the compliment! 🚀 YOLO11 is designed to handle challenging scenarios like overlapping objects and damaged labels effectively. Its instance segmentation capability excels at separating and identifying individual objects, even when stacked or overlapping. Plus, training on a robust dataset like the Roboflow Package Segmentation Dataset docs.ultralytics.com/datasets/segment/package-seg/ ensures YOLO11 adapts well to real-world logistics challenges, including clutter and low-quality annotations. Just make sure your dataset is as clean as possible for optimal results! 😊

  • @LunaStargazer-v1s
    @LunaStargazer-v1s 12 часов назад

    Fascinating stuff! But here’s a curveball-how do we balance the efficiency of computer vision in HR with the risk of turning humans into data points? Does this tech really help us see people better, or are we just looking through a lens of bias at scale?

    • @Ultralytics
      @Ultralytics 7 часов назад

      That’s a brilliant question! Computer vision can enhance HR efficiency, but it does raise ethical concerns, particularly around bias and data privacy. While it streamlines tasks like recruitment and performance tracking, there’s a risk of reducing people to metrics if not implemented thoughtfully. Ensuring diverse training datasets and transparency in AI use is key to minimizing bias and maintaining a human-centered approach. Tech should empower HR to understand people better, not just analyze them.

  • @LunaStargazer-v1s
    @LunaStargazer-v1s 12 часов назад

    Fascinating walkthrough! Could the YOLO11 segmentation model handle irregularly shaped, heavily damaged, or partially occluded packages effectively, or would it require additional dataset tuning? Thinking about chaotic warehouse floors here!

    • @Ultralytics
      @Ultralytics 7 часов назад

      Great question! YOLO11's segmentation capabilities are highly advanced, making it well-suited for irregular, damaged, or partially occluded packages, even in chaotic environments like warehouse floors. However, its performance depends heavily on the quality and diversity of the training dataset. For instance, using a dataset like the Roboflow Package Segmentation Dataset docs.ultralytics.com/datasets/segment/package-seg/-which includes images from challenging conditions-can help YOLO11 generalize better to such scenarios. If the current dataset doesn't fully represent these complexities, additional tuning or expanding the dataset with more examples of damaged or occluded packages would definitely improve accuracy. This ensures the model adapts to real-world challenges effectively.

  • @SchraderBrittany
    @SchraderBrittany 15 часов назад

    Thanks for the forecast! A bit off-topic, but I wanted to ask: My OKX wallet holds some USDT, and I have the seed phrase. (air carpet target dish off jeans toilet sweet piano spoil fruit essay). How should I go about transferring them to Binance?

    • @Ultralytics
      @Ultralytics 7 часов назад

      Hi! For security reasons, never share your seed phrase publicly as it gives full access to your wallet. To transfer USDT from OKX to Binance, log into OKX, navigate to withdrawals, enter your Binance USDT deposit address, choose the right network (e.g., ERC20, TRC20), and confirm the transfer. Ensure the network matches on both platforms to avoid loss of funds! Stay safe! 😊

  • @Melo7ia
    @Melo7ia 17 часов назад

    Love the efficiency here! But if I hit a dissonance with custom datasets, like handling weird edge cases (say detecting blurry UFOs or rogue pigeons), is there a harmonic way to fine-tune these models without overfitting? Or is it destined to be a bit of a samba with trial and error?

    • @Ultralytics
      @Ultralytics 7 часов назад

      Great question! Fine-tuning can definitely help you handle unique edge cases like blurry UFOs or rogue pigeons 🎯. To avoid overfitting, try these steps: start with a pre-trained model, use data augmentation to diversify your dataset, and freeze earlier layers while fine-tuning only the later ones. It’s a bit of a trial-and-error dance, but careful monitoring of metrics like validation accuracy vs. training accuracy can help you keep the balance! 🚀

  • @TheodoreBC
    @TheodoreBC 17 часов назад

    Cool setup, bro. But what if YOLOv8 mistakes a dog or a shadow for a person? False alarms in the middle of the night-bad for my sanity. Any tips to fine-tune the detection?

    • @Ultralytics
      @Ultralytics 7 часов назад

      Great question! To reduce false alarms, you can fine-tune the YOLOv8 model by retraining it with a custom dataset that includes examples of dogs, shadows, and other objects likely to cause false positives. This helps the model differentiate between people and non-human objects more accurately. Adjusting the confidence threshold during inference can also minimize false detections by requiring higher certainty for a detection to trigger an alert. For more on custom training, check out this guide: Export A Custom Trained Ultralytics YOLOv8 Model www.ultralytics.com/blog/export-a-custom-trained-ultralytics-yolov8-model.

  • @LunaStargazer-v1s
    @LunaStargazer-v1s 17 часов назад

    Fascinating walkthrough! When exploring the precision-recall balance, how would you approach prioritizing one over the other in scenarios like wildlife detection versus something more safety-critical like autonomous driving? Does the HUB offer any guidance for such trade-offs?

    • @Ultralytics
      @Ultralytics 7 часов назад

      Great question! The precision-recall trade-off depends on the application. For wildlife detection, prioritizing recall is often key, ensuring most animals are detected, even at the cost of false positives. Missing an animal could compromise conservation efforts or research, so maximizing detection is critical. In contrast, for safety-critical tasks like autonomous driving, precision is paramount to avoid false positives that might lead to unnecessary braking or incorrect decisions, which could endanger lives. Ultralytics HUB simplifies model evaluation and fine-tuning by providing insights into metrics like precision, recall, and mAP. You can analyze these metrics through intuitive dashboards to balance them effectively based on your use case. Learn more about performance metrics in the YOLO Performance Metrics guide docs.ultralytics.com/guides/yolo-performance-metrics/.

  • @PhatTran-cp8yf
    @PhatTran-cp8yf День назад

    Thank you for advised

    • @Ultralytics
      @Ultralytics 19 часов назад

      You're welcome! If you're referring to Natural Language Understanding (NLU) or any related AI concepts, feel free to explore more about it here: Natural Language Understanding (NLU) www.ultralytics.com/glossary/natural-language-understanding-nlu. Let me know if you have any other questions! 😊

  • @IanLeonardoASUMBRADO
    @IanLeonardoASUMBRADO 2 дня назад

    hello. everytime i try to install "pip install ultralytics", it keeps displaying an error where the numpy's "pyproject.toml" did not run successfully. how to resolve this?

    • @Ultralytics
      @Ultralytics День назад

      Hi! This issue is likely due to a mismatch between your Python environment and the `numpy` version. A quick solution is to first manually install a compatible version of `numpy` before installing Ultralytics. Run these commands in your terminal: ``` pip install numpy==1.23.5 pip install ultralytics ``` This should resolve the error. Also, ensure your Python version is 3.8 or later, as recommended. For more detailed installation instructions, check out the Ultralytics Quickstart Guide docs.ultralytics.com/quickstart/. 😊

  • @KEKW-lc4xi
    @KEKW-lc4xi 5 дней назад

    has tracking baked in, interesting. Last project I worked on required me to explicitly do tracking myself which was a tedious process lol

    • @Ultralytics
      @Ultralytics 4 дня назад

      You're absolutely right-tracking being built into Ultralytics YOLO11 is a game-changer! It eliminates the need for manually implementing tracking, saving a ton of effort. With the `model.track` method, you can seamlessly track objects across frames in real-time. It even supports advanced trackers like ByteTrack and BoT-SORT for robust results. Check out the detailed documentation here: YOLO11 Tracking docs.ultralytics.com/reference/engine/model/. 🚀

  • @GoWithASmiles
    @GoWithASmiles 6 дней назад

    nice

    • @Ultralytics
      @Ultralytics 5 дней назад

      Thanks for your kind comment! If you're exploring AI or computer vision, feel free to check out our Ultralytics Glossary www.ultralytics.com/glossary for fascinating insights into topics like Named Entity Recognition, Semantic Search, and more! 😊

  • @MaryAlvinJr.M
    @MaryAlvinJr.M 7 дней назад

    can i get the .pt you use?

    • @Ultralytics
      @Ultralytics 7 дней назад

      @MaryAlvinJr.M Of course! You can download the .pt file from this link: drive.google.com/file/d/1w50rwdNd5dUs7IJ1_wXA5nFWxYqXo8M9/view?usp=sharing

  • @74Gee
    @74Gee 7 дней назад

    Interesting but how can you get accurate counting over time? Counting the entity ID is always much higher than the actual number of people because the IDs often change for each person.

    • @Ultralytics
      @Ultralytics 7 дней назад

      Great question! Accurate counting over time requires robust tracking algorithms that maintain consistent IDs for entities across frames. Issues like ID switching often occur when an object is temporarily occluded or misdetected. To address this, Ultralytics YOLO11 integrates advanced trackers like ByteTrack docs.ultralytics.com/reference/trackers/byte_tracker/ and BoT-SORT, which effectively reduce ID switching by leveraging motion models (e.g., Kalman filters) and appearance features. These trackers ensure entities maintain their unique IDs as they move through the scene, even during occlusions. Additionally, setting proper thresholds for matching and fine-tuning tracker parameters (like IoU or distance thresholds) can significantly improve ID stability and ensure accurate counting. For object counting in specific regions, you can use tools like the Object Counter docs.ultralytics.com/reference/solutions/object_counter/ module, which efficiently tracks and counts objects crossing defined lines or regions.

    • @74Gee
      @74Gee 7 дней назад

      @@Ultralytics Thanks, great answer.

    • @Ultralytics
      @Ultralytics День назад

      You're welcome! Glad you found the response helpful. 😊 If you have more questions about tracking or object counting with Ultralytics YOLO11, feel free to ask! For additional details, check out the Object Tracking guide here www.ultralytics.com/glossary/object-tracking or the Object Counter module here docs.ultralytics.com/reference/solutions/object_counter/. 🚀

  • @deadpoems
    @deadpoems 8 дней назад

    So is this nothing different than roboflow dataset labeling tool can do (smart polygon)?

    • @Ultralytics
      @Ultralytics 7 дней назад

      Great question! The SAM-based label assistant in Ultralytics SAM2 is quite similar to Roboflow's "smart polygon" feature. Both leverage advanced models for fast and precise annotations. However, SAM2 integrates seamlessly with the Ultralytics ecosystem, allowing for tighter workflows with YOLO11 models. If you're already using Roboflow, their smart polygon tool is a fantastic option. You can explore more about Roboflow's labeling tools here docs.ultralytics.com/integrations/roboflow/. 😊

  • @abirdakhlaoui-s2l
    @abirdakhlaoui-s2l 8 дней назад

    is there any code for tracking pipelines with a drone (YOLO11) ?

    • @Ultralytics
      @Ultralytics 7 дней назад

      While there isn't a specific code snippet exclusively for tracking pipelines with a drone using YOLO11, you can adapt general object tracking code from YOLO11 for this purpose. A great starting point is the reference on using YOLO11 for object tracking in videos, which involves loading a YOLO model and tracking objects across frames. Check this guide for details: How to Use YOLO11 for Object Tracking www.ultralytics.com/blog/how-to-use-ultralytics-yolo11-for-object-tracking. For aerial pipeline monitoring, you might also find this article useful as it discusses infrastructure monitoring and YOLO11's capabilities for detecting objects from various angles with oriented bounding boxes: Using YOLO11 for OBB Object Detection www.ultralytics.com/blog/how-to-use-ultralytics-yolo11-for-obb-object-detection. Let us know if you'd like further guidance! 🚀

  • @PTRSignals
    @PTRSignals 8 дней назад

    what you do is cool

    • @Ultralytics
      @Ultralytics 7 дней назад

      Thank you! 😊 At Ultralytics, we're passionate about making AI accessible and impactful. From our YOLO models, which revolutionize computer vision, to tools like Ultralytics HUB for no-code solutions, we're dedicated to empowering everyone to innovate with AI. Check out our work www.ultralytics.com/yolo to learn more! 🚀

  • @antdx316
    @antdx316 8 дней назад

    We need to be able to NLP and drag a region for AI to automatically create the code for all this to happen with ease. Writing the proper docs for the AIs to know how to read, would be the best to do.

    • @Ultralytics
      @Ultralytics 7 дней назад

      For implementing NLP and region-based interactions like dragging a region for AI processing, a no-code or low-code platform like Ultralytics HUB ultralytics.com/hub could streamline the process. It simplifies complex tasks like training and deploying models with an intuitive interface, making it easier to generate the necessary code automatically. Additionally, for NLP integration, explore Natural Language Processing (NLP) www.ultralytics.com/glossary/natural-language-processing-nlp concepts to design a clear, machine-readable structure for your documentation. By focusing on writing precise, well-structured docs and leveraging tools like Ultralytics-Snippets for VS Code docs.ultralytics.com/integrations/vscode/, you can create AI-readable scripts efficiently. These snippets support intelligent code completion and pre-built blocks tailored for Ultralytics' API, enabling seamless integration of NLP and region-based interactions.

    • @antdx316
      @antdx316 7 дней назад

      @@Ultralytics You guys need to make it happen.

    • @Ultralytics
      @Ultralytics 7 дней назад

      Thanks for the encouragement! 🎉 We’re continuously innovating to make AI tools more accessible and user-friendly. If you haven’t already, check out Ultralytics HUB ultralytics.com/hub, our no-code platform that simplifies model training and deployment with drag-and-drop ease. It’s designed to unlock the true potential of computer vision for everyone! 🚀

  • @CodewithRiz
    @CodewithRiz 8 дней назад

    is there any code for ball trejectry ploting real time ?

    • @Ultralytics
      @Ultralytics 7 дней назад

      Yes! You can plot a ball's trajectory in real-time using Ultralytics YOLO's tracking features. The reference material includes a detailed example for tracking objects and visualizing their paths over multiple video frames. Here's how you can adapt it for a ball's trajectory: Use the code snippet here docs.ultralytics.com/modes/track/ to track objects and plot their trajectories. Replace the object detection logic to ensure it detects only the ball, and the tracking code will plot the ball's trajectory by connecting its positions across frames. For more details, check the Ultralytics Tracking Docs docs.ultralytics.com/modes/track/.

  • @bijayshakya337
    @bijayshakya337 10 дней назад

    Hello, Can we use ultralytics for RT-DETRV2 model as well. I have .pth checkpoint for RT-DETR-R50 model trained on VisDrone dataset. How can we do the validation and prediction using .pth checkpoint? Is there way to convert to .pt format. Thank you. Regards, Bijay

    • @Ultralytics
      @Ultralytics 9 дней назад

      Hi Bijay, Yes, Ultralytics supports RT-DETR models, including RT-DETRv2. However, `.pth` checkpoints aren't directly compatible with the Ultralytics framework, which typically uses `.pt` format. To use your RT-DETR-R50 model, you can convert the `.pth` checkpoint to `.pt` by loading it in PyTorch and saving it in the `.pt` format. Once converted, you can validate and predict using the Ultralytics RT-DETR interface. For validation and prediction with RT-DETR, you can use the Ultralytics Python API as shown in this example: ```python from ultralytics import RTDETR Load the converted .pt model model = RTDETR("converted_model.pt") Validation results = model.val(data="path/to/visdrone.yaml") Prediction results = model("path/to/image.jpg") ``` For more details, check the RT-DETR documentation docs.ultralytics.com/models/rtdetr/. If you need guidance on `.pth` to `.pt` conversion, let us know!

    • @bijayshakya337
      @bijayshakya337 9 дней назад

      @@Ultralytics My checkpoint's state_dict has the following keys: date last_epoch model criterion postprocessor ema optimizer lr_scheduler lr_warmup_scheduler *********************************************************************** import torch # Load the checkpoint file checkpoint_path = "output/rtdetrv2_r50vd_6x_visdrone_50/last.pth" # Replace with your file path checkpoint = torch.load(checkpoint_path, map_location=torch.device("cpu")) # Check the contents of "state_dict" if "state_dict" in checkpoint: # Check if 'state_dict' is a key in the checkpoint state_dict = checkpoint["state_dict"] else: state_dict = checkpoint # Directly use checkpoint if it is the state_dict #model.load_state_dict(state_dict) #device = torch.device("cuda" if torch.cuda.is_available() else "cpu") #model = model.to(device).float() torch.save(state_dict, 'rtdetrv2_r50_vd.pt') ************************************************************************* I used this code to save .pt format checkpoint. It saves the checkpoint. But, I get some error when I use the checkpoint for validation using ultralytics. from ultralytics import RTDETR model = RTDETR('rtdetrv2_r50_vd.pt') ************************************************************************ Error: AttributeError: 'dict' object has no attribute 'to' ************************************************************************ I think the checkpoint is not load correctly. If you could review the code to resolve the issue, that would be helpful. Regards, Bijay

    • @Ultralytics
      @Ultralytics 9 дней назад

      Hi Bijay, The error occurs because the `.pt` file you've saved only contains the `state_dict` (model weights) but lacks the full model architecture, which Ultralytics expects when loading a model. The `RTDETR` class requires a complete model checkpoint, including the architecture and weights. To resolve the issue, you need to properly save the full model, not just the `state_dict`. Here's how you can do it: 1. Load your model architecture along with the `state_dict`. 2. Save the full model using PyTorch's `torch.save()`. Here's the corrected code snippet: ```python import torch from ultralytics.models.rtdetr.model import RTDETRDetectionModel Load the checkpoint file checkpoint_path = "output/rtdetrv2_r50vd_6x_visdrone_50/last.pth" checkpoint = torch.load(checkpoint_path, map_location=torch.device("cpu")) Load the RTDETR model architecture model = RTDETRDetectionModel() model.load_state_dict(checkpoint["model"]) Load weights from the checkpoint Save the full model in .pt format torch.save(model, 'rtdetrv2_r50_vd_full.pt') ``` You can now use this `rtdetrv2_r50_vd_full.pt` file with the Ultralytics framework: ```python from ultralytics import RTDETR model = RTDETR('rtdetrv2_r50_vd_full.pt') ``` If you run into further issues, ensure the RTDETR model's architecture matches the one used during training. Refer to the RT-DETR documentation docs.ultralytics.com/models/rtdetr/ for more details.

  • @owenstudiofr
    @owenstudiofr 10 дней назад

    Anyone know why when I tried to install Ultralytics Yolo on Ubuntu I get error PEP668?

    • @Ultralytics
      @Ultralytics 9 дней назад

      The PEP 668 error typically occurs when your Python environment is marked as externally managed, often due to system package managers like `apt` on Ubuntu. This restricts direct package installations using `pip`. To resolve this, you can bypass the restriction by using the `--break-system-packages` flag while installing: ``` pip install ultralytics --break-system-packages ``` Alternatively, create a virtual environment to avoid conflicts with the system Python: ``` python3 -m venv yolo_env source yolo_env/bin/activate pip install ultralytics ``` For more details, check the Ultralytics Quickstart Guide docs.ultralytics.com/quickstart/. Let us know if you need further help! 😊

  • @antoniolupen6138
    @antoniolupen6138 11 дней назад

    I´been trying to learn something from your videos but is extremely difficult since they don´t follow a didactic order but instead a re a chaotic mess of info.

    • @Ultralytics
      @Ultralytics 10 дней назад

      Thanks for your feedback! We're always looking to improve. To learn YOLO in a more structured way, I recommend starting with our official Ultralytics YOLO Docs docs.ultralytics.com/, which provide step-by-step guides for various tasks like training, predicting, and deployment. Let us know if you have specific topics you'd like us to cover! 😊

  • @shahadalrawi6744
    @shahadalrawi6744 12 дней назад

    Sorry but this is not good. I think you should start to learn from Keras on how to do better documentation.

    • @Ultralytics
      @Ultralytics 12 дней назад

      Thank you for the feedback! We always strive to improve and appreciate your input. While Keras indeed sets a high standard for documentation, we also provide comprehensive guides like Google-style docstrings docs.ultralytics.com/help/contributing/ to ensure clarity and consistency. If you have specific suggestions for improvement, feel free to share or contribute directly to our open-source projects! 😊

  • @gustavomendes6737
    @gustavomendes6737 13 дней назад

    Is there a way to do this directly with model.predict(source=video.mp4)? or do I have to get each frame -> process with sahi -> send each slice to a model.predict(source=slice.jpeg) ?

    • @Ultralytics
      @Ultralytics 12 дней назад

      You can use `model.predict(source="video.mp4")` directly with Ultralytics YOLO for video inference. It will handle the video frames automatically, processing each one without requiring manual slicing or frame extraction. To efficiently manage large videos, consider using `stream=True` in the `predict()` call for memory optimization, as detailed here: Inference Sources Guide docs.ultralytics.com/modes/predict/. If your use case involves slicing frames for higher precision (e.g., using SAHI), you would need to extract frames, process them with SAHI, and then run `model.predict()` on each slice.

  • @antoniolupen6138
    @antoniolupen6138 15 дней назад

    what the hell are you talking about ?

    • @Ultralytics
      @Ultralytics 15 дней назад

      We're discussing how to define clear goals for computer vision projects, like using YOLO models effectively for tasks such as object detection or tracking. Let me know if you'd like clarification on anything specific! 😊

  • @圓圓-o2t
    @圓圓-o2t 15 дней назад

    Thanks!

    • @Ultralytics
      @Ultralytics 15 дней назад

      You're very welcome! 😊 Let me know if there's anything else I can help with!

  • @wolpumba4099
    @wolpumba4099 15 дней назад

    *How to Export and Optimize YOLO Models for Production* * *0:00** Introduction:* This video demonstrates how to export and download YOLO models from the Ultralytics Hub for optimized deployment. * *0:07** Model Export:* Users can export models in various formats (e.g., TensorRT, ONNX, Core ML) for optimized performance on different hardware and frameworks. * *0:44** Export Process:* The export process is simple, requiring just a single click within the "Deploy" section of the Ultralytics Hub interface. * *1:42** Supported Formats:* * *PyTorch (.pt):* Raw PyTorch model. * *ONNX:* Open Neural Network Exchange format, compatible with many optimization frameworks. * *Core ML:* For Apple hardware. * *TensorRT:* For NVIDIA GPUs, automatically generates an engine file. * *OpenVINO:* For Intel hardware. * *TensorFlow Lite:* For mobile and embedded devices. * *2:18** Performance Gains:* Exporting models to these optimized formats can yield significant performance improvements, ranging from 3x to 20x faster inference speeds compared to using the raw PyTorch model. * *2:43** Quantization:* Some frameworks allow for model quantization (e.g., reducing precision from 32-bit to 16-bit or 8-bit), further enhancing performance. * *3:10** Downloading Models:* Exported models are directly downloaded to the user's computer. * *3:22** Deployment:* Users can then load the exported model into their chosen framework for deployment. * *3:42** Recommendation:* The video strongly recommends exporting and optimizing models before deploying them into production for significant speed and cost benefits. I used gemini-1.5-pro-exp-0827 on rocketrecap dot com to summarize the transcript. Cost (if I didn't use the free tier): $0.02 Input tokens: 15493 Output tokens: 394

    • @Ultralytics
      @Ultralytics 15 дней назад

      To export and optimize YOLO models for production, the Ultralytics HUB provides a seamless process. You can export models into popular formats like ONNX, TensorRT, Core ML, OpenVINO, and TensorFlow Lite, enabling deployment across diverse hardware and frameworks. This process is accessible via the "Deploy" section in the HUB interface, requiring just a single click. Optimized formats yield significant performance boosts, with inference speeds up to 20x faster compared to raw PyTorch models. Additionally, quantization options like FP16 and INT8 can further enhance efficiency. After exporting, models are downloaded directly for deployment. For detailed instructions on exporting, refer to the Ultralytics Export Docs docs.ultralytics.com/modes/export/.

  • @wolpumba4099
    @wolpumba4099 15 дней назад

    *Accelerating Computer Vision on Edge Devices: A Look at NVIDIA's Latest Tools* * *0:00:00** Introduction:* The session focuses on NVIDIA's tools for accelerating computer vision on edge devices, particularly Deepstream 7 and advancements in Jetson Edge Computing. * *0:00:31** Edge AI Defined:* Edge AI involves performing computations near the user and data source, reducing latency and bandwidth costs and enhancing data sovereignty. * *0:00:54** Popularity of Edge AI:* Since 2014, there's been a tenfold increase in activity, with over 1.2 million active developers using NVIDIA's Jetson edge AI computers. * *0:02:07** Economic Impact:* Edge AI is projected to deliver up to $13 trillion in additional economic output by 2030. * *0:04:09** Jetson Devices:* Jetson devices are specifically designed for AI, offering various tiers of performance, with the top tier providing 275 TOPS at 60 watts. * *0:05:01** Jetson AGX Thor:* The next-gen Jetson AGX Thor, based on NVIDIA's Blackwell architecture, is coming in 2025, boasting 8x GPU performance for transformers, and designed for humanoid robots and advanced edge AI. * *0:05:51** Software Challenges:* Deploying models can be complex due to diverse frameworks, data modalities, and the need for optimization. * *0:06:59** Triton Inference Server:* Triton simplifies deployment, supports various frameworks, handles batching, and integrates with Kubernetes. It is also open source and customizable. * *0:09:18** TensorRT:* TensorRT optimizes models for specific hardware, offering up to 36x acceleration for computer vision through techniques like layer fusion and mixed precision. * *0:10:16** Ultralytics Integration:* Ultralytics models support direct export to TensorRT, enabling significant performance gains. * *0:11:33** Deepstream:* Deepstream is NVIDIA's SDK for building optimized computer vision pipelines. * *0:11:41** Deepstream 7:* New features include "service maker" for reducing code complexity, a pipe tuner for parameter optimization, WSL support, sensor fusion for lidar and radar, and deepstream libraries for python. * *0:15:49** Deepstream Libraries vs. Plugins:* Deepstream 7 introduces a paradigm shift with Python libraries, offering faster development and easier integration, while plugins still provide maximum performance. * *0:17:30** Conclusion:* The speaker emphasizes the importance of community and collaboration in advancing AI and computer vision. * *0:18:12** Q&A Session:* Discusses community feedback, a day in the life of an NVIDIA solutions architect, and the capabilities of the upcoming Jetson Thor. I used gemini-1.5-pro-exp-0827 on rocketrecap dot com to summarize the transcript. Cost (if I didn't use the free tier): $0.04 Input tokens: 32612 Output tokens: 640

    • @Ultralytics
      @Ultralytics 15 дней назад

      This session highlights NVIDIA's cutting-edge tools like DeepStream 7 and TensorRT that are revolutionizing edge AI and computer vision. DeepStream simplifies building vision pipelines, while TensorRT offers massive acceleration for YOLO models, including Ultralytics YOLO, ensuring optimized performance on Jetson devices. The next-gen Jetson AGX Thor promises groundbreaking advancements in 2025. For more on deploying YOLO with NVIDIA tools, check out DeepStream on NVIDIA Jetson docs.ultralytics.com/guides/deepstream-nvidia-jetson/ or Ultralytics YOLO11 with TensorRT docs.ultralytics.com/integrations/tensorrt/. 🚀

  • @wolpumba4099
    @wolpumba4099 15 дней назад

    *Efficiently Training a YOLO v11 Model for Medical Pill Detection* * *0:00:04** Introduction to YOLO v11:* The video introduces the YOLO v11 model for object detection, focusing on a medical pill detection use case. * *0:00:15** Dataset:* A small medical pill dataset from the Ultralytics dataset hub is used, demonstrating the model's ability to train with limited data. * *0:00:32** Google Colab Setup:* The model is trained in a Google Colab notebook, utilizing free T4 GPU resources for faster processing. * *0:00:49** Ultralytics Installation:* The Ultralytics package is installed, which is essential for working with the YOLO v11 model. * *0:01:01** Training Process:* The training process involves specifying the dataset, model type (YOLO v11 nano), and training parameters like epochs (20) and image size. * *0:01:31** Pre-trained Model:* A pre-trained model (trained on the COCO dataset) is used as a starting point, fine-tuning it on the medical pill dataset. * *0:05:50** Automatic Data Download:* The specified medical pill dataset is automatically downloaded during the training process. * *0:06:12** Evaluation Metrics:* Key evaluation metrics like precision, recall, and mean average precision (mAP) are tracked during training. The model achieves high precision (98.7%) and near-perfect mAP. * *0:06:34** Inference on Video:* A sample video of rotating pills is used to test the trained model's ability to detect pills in real-time. * *0:07:34** Model Output:* The training process generates various outputs, including precision-recall curves, confusion matrix, and training graphs, which are analyzed to assess model performance. * *0:08:56** Model Export:* The trained model can be exported to different formats (e.g., ONNX) for optimized deployment on specific hardware. * *0:10:45** Conclusion:* The video concludes by highlighting the ease of setting up a proof-of-concept computer vision model using YOLO v11 and Ultralytics, even with limited data. *Key Takeaways:* * YOLO v11 can be effectively trained for object detection tasks with relatively small datasets. * Ultralytics provides a user-friendly framework for training and deploying YOLO models. * Google Colab offers free GPU resources, making it a convenient platform for model training. * The medical pill detection model achieves high accuracy and can be used for real-time inference. * The training process generates detailed outputs for model evaluation and optimization. I used gemini-1.5-pro-exp-0827 on rocketrecap dot com to summarize the transcript. Cost (if I didn't use the free tier): $0.03 Input tokens: 24416 Output tokens: 606

    • @Ultralytics
      @Ultralytics 15 дней назад

      Efficiently training a YOLOv11 model for medical pill detection in Google Colab involves a streamlined process leveraging a small dataset, pre-trained weights, and GPU acceleration. Google Colab is ideal for such tasks due to free T4 GPU resources and seamless integration with Ultralytics' tools. Follow these steps: 1. Setup & Dataset: Use the Ultralytics dataset hub for a medical pill dataset. The dataset is automatically downloaded during training, ensuring ease of use. Learn more about datasets here docs.ultralytics.com/datasets/detect/medical-pills/. 2. Model Training: Utilize YOLOv11 nano for faster training. Specify parameters like 20 epochs and image size. Fine-tune a pre-trained COCO model for better results. The Colab notebook simplifies this process; refer to the Google Colab integration guide docs.ultralytics.com/integrations/google-colab/. 3. Evaluation: Track precision, recall, and mAP, with metrics like 98.7% precision achieved. Outputs like confusion matrices and training graphs provide insights for optimization. Explore performance metrics here docs.ultralytics.com/guides/yolo-performance-metrics/. 4. Inference & Export: Test real-time detection on videos and export the model (e.g., ONNX) for deployment. Export details here docs.ultralytics.com/modes/export/. This workflow showcases YOLOv11's ability to handle small datasets effectively and deliver high-accuracy models ready for real-world applications. 🚀

  • @wolpumba4099
    @wolpumba4099 15 дней назад

    *Ultralytics YOLO V8: A Comprehensive Guide to Pre-trained Models and Inference* * *0:00:01** Introduction to Ultralytics and YOLO V8:* This video is the first in a series about Ultralytics and YOLO V8, covering object detection, segmentation, pre-trained and custom models, data setup, and more. * *0:00:31** GitHub Repository Overview:* The video begins with an overview of the YOLO V8 model in the GitHub repository, highlighting its capabilities for object detection. * *0:00:47** Inference Speed and Accuracy:* YOLO V8 is presented as a state-of-the-art model, offering a superior trade-off between accuracy and speed compared to other YOLO models. * *0:01:08** Installation and Usage:* The video demonstrates how to install YOLO V8 using pip and run it from the command line or within a Python script. * *0:01:37** Command Line Inference:* An example of running inference directly from the command line is shown, using a pre-trained YOLO V8 model to detect objects in an image. * *0:01:56** Python Script Inference:* The video illustrates how to use YOLO V8 in a Python script, showcasing its ease of use with just two lines of code for basic inference. * *0:03:21** Inference Parameters:* Various inference parameters are discussed, including source (image, video, webcam), confidence score, and saving results. * *0:03:48** Documentation and Arguments:* The importance of the documentation is emphasized, particularly the section on inference arguments, which are crucial for both pre-trained and custom models. * *0:04:48** Inference Modes:* YOLO V8's different modes (train, validation, predict, export, track, benchmark) are briefly introduced, with a focus on prediction mode. * *0:06:05** Live Webcam Inference:* The video culminates in a demonstration of running inference with a pre-trained YOLO V8 model on a live webcam feed, showcasing real-time object detection at around 100 frames per second. * *0:07:04** Conclusion and Future Videos:* The video concludes by highlighting the ease of using pre-trained YOLO V8 models and teases future videos on custom object detection, tracking, and more using the Ultralytics framework. I used gemini-1.5-pro-exp-0827 on rocketrecap dot com to summarize the transcript. Cost (if I didn't use the free tier): $0.03 Input tokens: 20609 Output tokens: 523

    • @Ultralytics
      @Ultralytics 15 дней назад

      Thanks for the detailed query! The video provides a comprehensive introduction to Ultralytics YOLOv8, emphasizing its state-of-the-art capabilities in object detection and segmentation. YOLOv8 stands out for its speed and accuracy, making it ideal for real-time applications. Key highlights include: - Installation & Usage: Easily install using `pip install ultralytics` and run via CLI or Python scripts. Learn more here: Quickstart Guide docs.ultralytics.com/quickstart/. - Inference Examples: Both command-line and Python-based inference are demonstrated, showing how to detect objects using pre-trained models quickly. Full details on CLI and script usage can be found here docs.ultralytics.com/usage/python/. - Inference Modes: YOLOv8 supports multiple modes like predict, train, validate, track, and export for diverse use cases. Check more about these modes here docs.ultralytics.com/modes/. - Live Detection: Real-time webcam inference at ~100 FPS showcases the model’s efficiency for dynamic environments. Learn how to set up live detection here docs.ultralytics.com/guides/streamlit-live-inference/. For a deeper dive into YOLOv8's features, check this blog: Object Detection with YOLOv8 www.ultralytics.com/blog/object-detection-with-a-pre-trained-ultralytics-yolov8-model. Let us know if you have any further questions! 😊

    • @wolpumba4099
      @wolpumba4099 15 дней назад

      @ okay this is awkward. I guess you have an automatic response to long RUclips comments which was triggered by my ai generated (but manually posted comment). Weird

    • @Ultralytics
      @Ultralytics 14 дней назад

      Haha, that's an interesting observation! 😄 It wasn't an automatic response, but I do aim to provide comprehensive and helpful replies to support the community-AI-generated or not! If you have any specific questions about YOLOv8 or related topics, feel free to ask. Always happy to help! 🚀

  • @74Gee
    @74Gee 16 дней назад

    I ordered mine a few days ago for testing, looking forward to taking it through it's paces. Human detection on the edge for about $50 is very attractive and with a built in commercial ultralytics license, that really opens up the realm of possibility.

    • @Ultralytics
      @Ultralytics 16 дней назад

      That's awesome! 🎉 The reCamera truly is a game-changer for edge AI applications, especially at such an accessible price point. The built-in commercial Ultralytics license makes it even more versatile, letting you explore a wide range of applications. Can't wait to hear about your testing experience-keep us posted! 🚀

  • @ionescudoru719
    @ionescudoru719 17 дней назад

    I have ordered reCamera 2002w 64GB

    • @Ultralytics
      @Ultralytics 16 дней назад

      That's fantastic! The reCamera 2002w is an excellent choice for Vision AI tasks. With its native support for Ultralytics YOLO models, you’re all set for seamless real-time object detection and other applications. Once it arrives, you can dive into customizing it for robotics, smart homes, or industrial use. Let us know how it goes! 🚀

  • @guillermomontero6863
    @guillermomontero6863 17 дней назад

    If I stop training a model on my computer, can I continue from where I left off by using one of the previously uploaded checkpoints to Ultralytics?

    • @Ultralytics
      @Ultralytics 16 дней назад

      Yes, you can continue training from where you left off using a previously uploaded checkpoint in Ultralytics. The platform saves checkpoints during training, allowing you to resume seamlessly. Simply specify the checkpoint file (e.g., `last.pt`) and ensure `resume=True` in your training command. Learn more about resuming training here: docs.ultralytics.com/modes/train/.

  • @ConsultingjoeOnline
    @ConsultingjoeOnline 17 дней назад

    Nice. What model was running in the demo? That seems very fast for 1TOP.

    • @Ultralytics
      @Ultralytics 16 дней назад

      Thanks! In the demo, we showcased the reCamera running the YOLOv8n model, which is optimized for speed and efficiency. Its lightweight architecture makes it perfect for edge devices like the reCamera, even with limited compute power like 1TOPS. 🚀

  • @iamdihan
    @iamdihan 17 дней назад

    got this a few weeks ago. such a nice little camera. still playing around with it

    • @Ultralytics
      @Ultralytics 16 дней назад

      That's awesome! The reCamera is a fantastic tool for Vision AI projects. Its modular design and native YOLO support make it super versatile. Have fun exploring its capabilities! 😊

  • @haziqhariz5367
    @haziqhariz5367 17 дней назад

    hi, for object detection in images, i know that we need to train our dataset for the class to be detected. would it be the same for object counting in images? should i train the dataset first or i can just do it straight? thank you in advance!

    • @Ultralytics
      @Ultralytics 17 дней назад

      Hi! For object counting in images, it depends on your use case. If the classes you need to count are already supported by a pre-trained model (like COCO classes for Ultralytics YOLO), you can directly count objects without additional training. However, if you need to detect and count custom classes not included in the pre-trained model, you'll need to train a dataset for those specific classes first. For more details on object counting, check out the YOLO Object Counting guide: Object Counting Documentation docs.ultralytics.com/guides/object-counting/. 😊

    • @haziqhariz5367
      @haziqhariz5367 16 дней назад

      @@Ultralytics thank you!

    • @Ultralytics
      @Ultralytics 16 дней назад

      You're welcome! 😊 If you have more questions or need further guidance, feel free to ask. For additional resources on object counting, check out the Object Counting Documentation docs.ultralytics.com/guides/object-counting/. 🚀

  • @khaleddandan3472
    @khaleddandan3472 18 дней назад

    do i need to do a transformation by giving coordinates of ROI and specifying the specific width and height of the ROI ? or does the model inherently figures that out with only needing coordinates of the ROI ?

    • @Ultralytics
      @Ultralytics 17 дней назад

      The model does not inherently adjust the Region of Interest (ROI) to a specific width and height unless explicitly specified. You need to transform your ROI coordinates to the desired format if you require specific dimensions or normalization. Tools like `ultralytics.utils.ops` provide utilities for such transformations, such as `xyxy2xywh` or `xywh2xyxy` for bounding box conversions. Check the Bounding Box Format Conversions docs.ultralytics.com/usage/simple-utilities/#bounding-box-format-conversions for more details.

  • @ayushchoubey8021
    @ayushchoubey8021 18 дней назад

    Can I do transfer learning with this yolo models ? like training just last layer not from the scratch on my custom dataset?

    • @Ultralytics
      @Ultralytics 18 дней назад

      Yes, absolutely! YOLO models, including YOLOv11, support transfer learning, allowing you to train only the last layers on your custom dataset instead of starting from scratch. This approach is efficient, as it utilizes pre-trained weights, reducing training time and resource requirements. To achieve this, you can freeze the earlier layers of the model (like the backbone) and only update the final layers during training. For example, with YOLOv5, you can use the `--freeze` argument in the training script to freeze specific layers. For YOLOv11, the process is similar and can be configured in the training settings. Check out our guide on Transfer Learning with Frozen Layers docs.ultralytics.com/yolov5/tutorials/transfer_learning_with_frozen_layers/ for specific steps and commands. 🚀

  • @EvanChen-yt8be
    @EvanChen-yt8be 18 дней назад

    wtf is that .txt doing

    • @Ultralytics
      @Ultralytics 18 дней назад

      The `.txt` file is used to save detection results or classifications from YOLO models. For instance, when you use the `save_txt()` function, it exports results like class, confidence, and bounding box coordinates into a text file. This is helpful for logging, analysis, or integrating with other systems. If you'd like to learn more about how this works, check out the save_txt documentation docs.ultralytics.com/reference/engine/results/#save_txt. 😊

  • @ingenioexpress1573
    @ingenioexpress1573 18 дней назад

    This is wonderful because it allows me to run YOLOv11 along with NVDIA DeepStream but, how can I build a real world analytic combining it with Python and supervision to annotate, track objects and more? I'm an absolute noob with NVIDIA Deepstream and I don't know how to build an app with Python :(

    • @Ultralytics
      @Ultralytics 18 дней назад

      No worries-combining NVIDIA DeepStream, YOLO11, and Python for analytics is a great start! You can use Python to integrate YOLO11 with DeepStream for real-time object tracking, annotation, and visualization. Begin by setting up YOLO11 with DeepStream as shown in this guide: DeepStream on NVIDIA Jetson docs.ultralytics.com/guides/deepstream-nvidia-jetson/. Once that's ready, Python can be used to build your analytics app with frameworks like Streamlit, OpenCV, or Flask. Start simple by tracking objects with YOLO11 and visualizing analytics like line graphs or bar charts using Python libraries. Check out this example for YOLO11 real-time tracking: How to Use Ultralytics YOLO11 for Object Tracking www.ultralytics.com/blog/how-to-use-ultralytics-yolo11-for-object-tracking. For annotating and tracking in real-time, explore the Analytics Guide docs.ultralytics.com/guides/analytics/, which explains how to dynamically update visualizations like graphs based on tracked objects. Pair this with Streamlit for a user-friendly app: Live Inference with Streamlit docs.ultralytics.com/guides/streamlit-live-inference/. These guides will help you navigate from basic setup to building a functional app. 🚀

  • @ionescudoru719
    @ionescudoru719 18 дней назад

    I have created Android applications containing Yolo 11 custom models converted to tflite which instantly detects the trained custom objects. Tflite uses the GPU that most of Android devices have.

    • @Ultralytics
      @Ultralytics 18 дней назад

      That's fantastic! Leveraging YOLO11 models converted to TFLite for Android apps is a great way to achieve fast and efficient object detection. TFLite's GPU delegate indeed ensures optimized on-device performance by utilizing the GPU for inference. If you're exploring further optimization or deployment tips for TFLite models on Android, check out our detailed TFLite integration guide docs.ultralytics.com/integrations/tflite/. Happy coding! 🚀

  • @kroma53
    @kroma53 21 день назад

    How to not duplicate the pic's?... Because i did save_crop but it's saving the same crop pics..

    • @Ultralytics
      @Ultralytics 20 дней назад

      It seems like you're encountering duplicate saved crops when using the `save_crop` function. This issue can happen if the bounding boxes detected are overlapping or your dataset includes repeated detections for the same objects. To address this: 1. Ensure the model isn't producing multiple overlapping detections for the same object. You can adjust the non-maximum suppression (NMS) threshold during inference to filter overlapping boxes more effectively. 2. Check that the `save_crop` logic is correctly implemented and doesn't save the same bounding box multiple times. For more on saving crops, check the save_crop documentation docs.ultralytics.com/reference/engine/results/#save_crop. If your issue persists, consider posting more details or examples of your workflow. 😊

    • @kroma53
      @kroma53 20 дней назад

      @Ultralytics thanks 🙏

    • @Ultralytics
      @Ultralytics 20 дней назад

      You're welcome! 😊 If you need more tips or have further questions, feel free to ask or explore our documentation: docs.ultralytics.com/. Happy experimenting with YOLO! 🚀

  • @aqilahilyana9208
    @aqilahilyana9208 22 дня назад

    for the torch installation it return false. im using cuda 12.2 but there is no that version on the pytorch website

    • @Ultralytics
      @Ultralytics 21 день назад

      It seems that PyTorch does not yet officially support CUDA 12.2, which is why you're encountering issues. You can either: 1. Downgrade your CUDA version to one that is officially supported by PyTorch, such as CUDA 11.8. Visit the PyTorch installation guide pytorch.org/get-started/locally/ for compatibility details. 2. Use a version of PyTorch with the appropriate CUDA version bundled via `pip install`. For example: ```bash pip install torch torchvision torchaudio --index-url download.pytorch.org/whl/cu118 ``` Replace `cu118` with your desired compatible CUDA version. If you're still facing issues, check compatibility with ultralytics.utils.checks.check_torchvision docs.ultralytics.com/reference/utils/checks/, which verifies PyTorch and Torchvision compatibility. Let me know how it goes! 😊

  • @AntonioLarocca-qp5nu
    @AntonioLarocca-qp5nu 22 дня назад

    I am experiencing persistent issues while training YOLOv8 (YOLO11n model) with the xView dataset. Despite carefully following all the documented steps for dataset preparation and training, the results are consistently problematic. Here are the details: Dataset Preparation Resized Images: I resized all images from the xView dataset to maintain a maximum size of 640 pixels on the largest dimension, ensuring compatibility with the imgsz=640 training parameter. The aspect ratio of the images was preserved during resizing. Train/Validation/Test Split: The xView training images and labels were divided as follows: 80% Train: For training the model. 10% Validation: For validation during training. 10% Test: Reserved for evaluation. I ensured that each split had matching .txt label files for all corresponding images. Labels Handling: I used the official xView class mapping (classes 11-94 mapped to 0-59). For images without annotations, I explicitly generated empty .txt files to represent unannotated images (backgrounds). However, there are no true background images in the dataset-all images contain annotations except for a few explicitly empty .txt files added during preprocessing. No Background Images: After careful inspection: No images without labels ("true backgrounds") were found in the xView dataset. Even the potentially unannotated images (e.g., .tif files like ._100.tif) were checked, and they were either valid or excluded from training. Training Details Command Used: python model.train(data="xView.yaml", epochs=100, imgsz=640, batch=4) Model: YOLO11n Environment: Google Colab (A100 GPU) Issues Observed Metrics and Performance: After multiple training attempts, the performance metrics remain extremely poor: Precision: Close to 0 for most classes. Recall: Close to 0, even for classes with a significant number of instances, such as "Small Car" and "Building." mAP: Extremely low across the board. Confusion Matrix: The confusion matrix indicates that the model struggles to make meaningful predictions. It seems unable to learn effectively from the dataset, with predictions often biased toward specific classes or entirely absent for others. No Background Handling: Since there are no true background images, it is unclear whether the lack of explicit background examples is causing the model's poor generalization. Despite creating empty .txt files for unannotated images, the training results remain poor. Steps Taken Verified the dataset integrity: Ensured all images have corresponding .txt labels. Checked for corrupted files, finding none. Resized the images to ensure compatibility with the imgsz=640 parameter. Split the dataset into training, validation, and test sets with a clear 80/10/10 ratio. Are there additional preprocessing steps or considerations for the xView dataset when training with YOLO11n? Could the lack of true background images in the dataset be affecting performance? Are there known compatibility issues between the xView dataset and YOLOv8? What could explain the extremely poor performance metrics despite following the standard training pipeline? Please could you response me?

    • @Ultralytics
      @Ultralytics 22 дня назад

      It seems you've been very thorough with your dataset preparation and training setup, but there are a few considerations and potential improvements to address your issues: 1. Dataset Size and Variety: The xView dataset's diversity is crucial for effective training, but your resized images (640px max) may limit the model's ability to capture fine-grained details. Consider experimenting with larger image sizes (e.g., `imgsz=1024`) if your environment permits. 2. Class Imbalance: Some xView classes may be heavily imbalanced. This can lead to poor precision/recall for underrepresented classes. Review the number of instances per class using tools like `labels.png` to ensure sufficient representation, as recommended here docs.ultralytics.com/yolov5/tutorials/tips_for_best_training_results/. 3. Background Images: The absence of true background images could indeed affect generalization. Adding synthetic background images (with no annotations) or using data augmentation techniques like CutMix or Mosaic might help reduce false positives. 4. Model Size: YOLO11n is a lightweight model, making it faster but less capable of handling complex datasets like xView. Consider using a larger model like `YOLO11s` or `YOLO11m`. 5. Batch Size: A batch size of 4 on an A100 GPU is quite low. Increase it to at least 16 (if memory allows) to improve gradient updates and model convergence. 6. Learning Rate and Augmentation: Verify that you're not overriding the default learning rate or augmentation strategies. These defaults are optimized for general cases. 7. Metrics Analysis: Examine the confusion matrix and PR curves for insights into which classes are underperforming. This can guide dataset improvements or tuning. 8. Validation: Use a small subset of the dataset for quick testing with reduced epochs (e.g., 10) to ensure the pipeline works as expected before committing to a full run. Check this detailed guide for training tips: Tips for Best Training Results docs.ultralytics.com/yolov5/tutorials/tips_for_best_training_results/. If issues persist, consider sharing training logs and plots for deeper analysis. 🚀

  • @tyriecefuller6259
    @tyriecefuller6259 22 дня назад

    Can you see me? Tyriece Trambull👻👁️👻🐇

    • @Ultralytics
      @Ultralytics 22 дня назад

      Hey Tyriece! 👋 Unfortunately, I can't see you, but it sounds like you're bringing some fun energy here! 👻👁️ Keep exploring the world of AI and computer vision-it's got some amazing possibilities! 🚀

  • @MrSboySr
    @MrSboySr 23 дня назад

    Totally unresponsive to presales questions - is this really legit?

    • @Ultralytics
      @Ultralytics 22 дня назад

      We’re sorry to hear about your experience! Ultralytics is indeed a legitimate organization with a strong commitment to transparency and support. For any pre-sales inquiries, you can reach out through our Contact Page ultralytics.com/contact or join the Discord Community discord.com/invite/ultralytics for quicker responses. Let us know how we can assist further! 😊

  • @TTK.0s
    @TTK.0s 24 дня назад

    hey ultralytics! is there any way i can do this to auto annotate the food101 dataset? im sure you know about the dataset and i think it would be an awesome addition to the database.

    • @Ultralytics
      @Ultralytics 23 дня назад

      Hi! Auto-annotating the Food101 dataset would be an exciting project! While we don't have a direct integration for Food101, you can leverage tools like our Roboflow integration docs.ultralytics.com/integrations/roboflow/ or SAM-based annotation methods docs.ultralytics.com/reference/data/annotator/ to streamline the annotation process. These tools can help label the dataset efficiently for tasks like classification or object detection. Let us know how it goes! 🚀

  • @Andy-j6w5q
    @Andy-j6w5q 24 дня назад

    This was a great video! Thanks!! I'd like to know more about the object detection. Can you point me to some code or an explanation of how I can create triggers while the detection/prediction program is running on a webcam? For example if Yolo detects a modelled object I want to do something like create a text file or note that it found 1 of 5 objects I'm looking for.

    • @Ultralytics
      @Ultralytics 23 дня назад

      Thanks for watching and for your kind words! 😊 You can use callbacks in Ultralytics YOLO to create triggers during object detection. For example, you can register a custom callback to save specific detections or execute custom actions when a condition is met. Here's a guide on adding callbacks: Callbacks in Ultralytics YOLO docs.ultralytics.com/usage/callbacks/. Additionally, you can explore the `on_predict_postprocess_end` event to add logic like creating a text file when an object is detected. Check out the example in the reference for saving predictions based on class presence. Let us know if you need further assistance! 🚀