Pose Estimation with Ultralytics YOLOv8 | Episode 6

Поделиться
HTML-код
  • Опубликовано: 29 ноя 2024

Комментарии • 107

  • @Ateszika
    @Ateszika Год назад +1

    These videos are sooo useful! I'd also love to see videos where you are customizing a pre-trained model like the pose one. E.g. to add vertical or horizontal lines to the video or track the motion of the wrist etc Thank you!

    • @Ultralytics
      @Ultralytics  Год назад

      Thank you for your recommendations! We will internally review them and develop content based on your suggestions!

  • @LunaStargazer-v1s
    @LunaStargazer-v1s 4 месяца назад +2

    Absolutely mesmerizing! As someone who dabbles in dance and choreography, I'm curious-how well does Ultralytics YOLOv8 perform with dynamic, fluid movements compared to more rigid poses? Seems like a potential game-changer for capturing the grace in motion.?

    • @Ultralytics
      @Ultralytics  4 месяца назад

      Thank you for your kind words! 😊 Ultralytics YOLOv8 is indeed quite versatile and performs well with dynamic, fluid movements, making it a great tool for capturing dance and choreography. Its advanced pose estimation capabilities can track key body landmarks in real-time, even with complex and graceful motions. For more detailed information, you can check out our documentation here: Ultralytics YOLOv8 Pose Estimation docs.ultralytics.com/tasks/pose. If you have any specific questions or need further assistance, feel free to ask!

    • @thedinosaurspecialist8053
      @thedinosaurspecialist8053 3 месяца назад +1

      ​@@Ultralytics Hey bro, this is a chatgpt bot. Look at their profile picture (looks like AI) and their post history on this channel. Don't waste your time responding to such bots. Unless, of course, you paid for these bots yourself to get a good brand image.

    • @Ultralytics
      @Ultralytics  3 месяца назад

      Hey there! Thanks for your concern, but I assure you I'm here to genuinely help and engage with our community. If you have any questions or need assistance with Ultralytics YOLOv8, feel free to ask! 😊

    • @moses5407
      @moses5407 2 месяца назад

      How strable are the joint positions during movement? it looks like it lost your knee in the webcam example. Also, can joint angles be displayed in near real time with angle maxima stored?

    • @Ultralytics
      @Ultralytics  2 месяца назад

      Joint positions in Ultralytics YOLOv8 are generally stable, but fast or complex movements can sometimes cause temporary tracking loss. For displaying joint angles in real-time and storing maxima, you can customize the output using the model's capabilities. Check out our guide on workouts monitoring for more details: docs.ultralytics.com/guides/workouts-monitoring/. If you have more questions, feel free to ask!

  • @sushreesangeetajena7646
    @sushreesangeetajena7646 Год назад +1

    Great video sir. If I just want the pose model, but I don't want the bounding boxes or the labels, where should I disable those?

    • @Ultralytics
      @Ultralytics  Год назад

      You can suppress box plotting with pose models by passing `boxes=False`. See docs.ultralytics.com/modes/predict/#inference-arguments

  • @March_Awake
    @March_Awake Год назад +2

    So cool,Is this pose detect can apply for detecting human fell down?Ijust a beginner of yolo.

    • @Ultralytics
      @Ultralytics  Год назад +1

      Hi, yes - this is a use case that is definitely possible with YOLOv8. Check out this issue on the repo about pose estimation for fall detection: github.com/ultralytics/ultralytics/issues/3351

  • @m033372
    @m033372 4 месяца назад +1

    How does YOLOv8 handle occlusions in pose estimation, and are there any particular scenarios where it might struggle more than others?

    • @Ultralytics
      @Ultralytics  4 месяца назад

      Great question! YOLOv8 uses advanced algorithms to handle occlusions in pose estimation by predicting keypoints even when parts of the body are not visible. However, it might struggle in scenarios with severe occlusions or when keypoints are completely hidden. To improve results, ensure you're using the latest versions of `torch` and `ultralytics`. For more details, check out our pose estimation documentation docs.ultralytics.com/tasks/pose. If you have specific issues, feel free to share more details! 😊

    • @m033372
      @m033372 4 месяца назад

      @@Ultralytics great answer, thank you!

    • @Ultralytics
      @Ultralytics  4 месяца назад

      You're welcome! Glad I could help! 😊 If you have any more questions, feel free to ask. Happy experimenting with YOLOv8! 🚀

  • @afjamo
    @afjamo 8 месяцев назад +1

    So cool!!! Thank you for developing this state-of-the-art application!!!
    I would like to analyze behaviour of mice in a cage. Do you think it is possible to detect them eating? We place food pellets on the metal bar lid. When they eat, they stand up and reach the pellet. They partially cover a food pellet with their upper joe. When they are biting a pellet, it rattles.
    Thank you in advance!!!

    • @Ultralytics
      @Ultralytics  8 месяцев назад +1

      Thank you! Detecting mouse behavior, including eating, is possible with computer vision. By analyzing video footage and sound, you can train a model to recognize behaviors like standing up, reaching for pellets, and biting motions.

  • @maxtn1000
    @maxtn1000 Месяц назад +1

    Can I integrate object detection with pose estimation? I aim to develop a project that tracks items in shoppers’ hands. By analyzing the pose skeleton, I want to identify when a shopper puts an item in their pocket and detect potential theft if the item disappears during that pose.

    • @Ultralytics
      @Ultralytics  Месяц назад +1

      Yes, you can integrate object detection with pose estimation for your project. YOLOv8 supports both tasks, allowing you to track items and analyze poses simultaneously. This setup can help identify actions like putting items in pockets. For more details on implementing pose estimation, check out our guide: docs.ultralytics.com/guides/workouts-monitoring/ 😊

  • @zaidahmed4069
    @zaidahmed4069 9 месяцев назад +1

    I'm trying to find the class name of the recognized object after extracting the predictions, but I'm unable to find it, even though it is written on the top of the bounding box of the predicted image. How can I determine it using Python?

    • @Ultralytics
      @Ultralytics  9 месяцев назад +1

      You can obtain the class names using the provided code below!
      """
      import cv2
      from ultralytics import YOLO
      model = YOLO("yolov8s-pose.pt")
      cap = cv2.VideoCapture("Path/to/video/file.mp4")
      while True:
      ret, im0 = cap.read()
      if not ret:
      print("Video frame is empty or video processing has been successfully completed.")
      break
      results = model.predict(im0)
      clss = results[0].boxes.cls.cpu().tolist()
      for cls in zip(clss):
      print("Class Name : ", cls)
      if cv2.waitKey(1) & 0xFF == ord('q'):
      break
      cap.release()
      """
      Thanks

  • @leonidas1983
    @leonidas1983 11 месяцев назад +1

    Nice work! I'm evaluating migrate our video analysys backend in python from mediapipe pose to yolo pose to calculate angle between two points in a segment of body person . Can i define specific landmark (example shoulder, elbow, wrist) with yolo? Do you have any tutorial? Thanks

    • @Ultralytics
      @Ultralytics  11 месяцев назад +1

      Absolutely, we offer a comprehensive workout monitoring application. You can refer to our documentation page for more details: docs.ultralytics.com/guides/workouts-monitoring/
      Thanks
      Ultralytics Team!

    • @leonidas1983
      @leonidas1983 11 месяцев назад +1

      @@Ultralytics thanks bro. I can see that the model's example calculate angles using "pose_type="pushup"" ¿Do you have any doc to use specific landmarks and math functions to calculates other segments?

    • @Ultralytics
      @Ultralytics  11 месяцев назад

      We currently offer support exclusively for 'ab workout,' 'pushups,' and 'pullups.' We have plan to incorporate additional exercises in the future.
      Regards,
      Ultralytics Team

  • @黃威凱-c4d
    @黃威凱-c4d 8 месяцев назад +1

    How to add new actions in ai_gym? What should I do if I want to train myself or add a new pose_type?

    • @Ultralytics
      @Ultralytics  8 месяцев назад

      You'll have to fine-tune the model for estimation of the specific pose_type to suit your specific requirements.

  • @guilhermemessias1334
    @guilhermemessias1334 9 месяцев назад +1

    Hi! Thanks for the video. I want to create a pose estimator for swimming. I tryed to run the yolov8m-pose but the accuracy is very low. What can i do to improve the model accuracy? Thanks :)

    • @Ultralytics
      @Ultralytics  9 месяцев назад

      You can use the larger model, yolov8l-pose, or fine-tune the model on your custom dataset for improved accuracy.
      Thanks.
      Ultralytics Team!

  • @LINCHEN-m5w
    @LINCHEN-m5w Месяц назад

    I want to use yolov8n-pose to identify the corner points of the vegetable basket. Why can MAP50, MAP50-95, P and R reach 99.5%, but the actual recognition effect is not good? My data set is about 700, but there are no negative samples

    • @Ultralytics
      @Ultralytics  Месяц назад

      It sounds like your model might be overfitting due to the lack of negative samples and a relatively small dataset. High metrics like mAP and precision might not reflect real-world performance if the model hasn't seen enough variability. Consider adding more diverse data, including negative samples, to improve generalization. For more insights, check out our guide on common YOLO issues: docs.ultralytics.com/guides/yolo-common-issues/ 😊

    • @LINCHEN-m5w
      @LINCHEN-m5w Месяц назад

      @@UltralyticsThank you very much for your kind words! Every time I train to the 15th epoch, all the values are already very high, will the 15th epoch also overfit? My target object stands out in the picture, is it too easy to recognize? Also, does the data set need to be shot in many scenes, or is it because my scenes are too single? (only in farmland)

    • @Ultralytics
      @Ultralytics  Месяц назад

      You're welcome! If your metrics plateau early, it might indicate overfitting, especially if your object is easy to recognize. To improve generalization, try diversifying your dataset with different scenes and conditions. This can help the model perform better in varied real-world scenarios. Keep experimenting with data augmentation too! 😊

  • @sancelot88
    @sancelot88 8 месяцев назад +1

    I only want to save poses of moving persons in a dataset for later moving classification. Is it possible ?

    • @Ultralytics
      @Ultralytics  8 месяцев назад

      Yes, it is feasible, but you will have to develop supplementary code to store the pose data of individuals in motion. In fact, you might need to incorporate a time variable for estimating the position of objects.

  • @Smitthy-k9d
    @Smitthy-k9d Месяц назад

    So, if YOLOv8 can nail pose estimation, could it help me improve my dance moves in real-time? Asking for a friend who often looks like they're fighting invisible bees!

    • @Ultralytics
      @Ultralytics  Месяц назад

      Absolutely! YOLOv8's pose estimation can track your movements in real-time, helping you refine your dance moves. It can provide feedback on your posture and alignment, making it easier to perfect your routine. Check out our guide on workouts monitoring for more details: docs.ultralytics.com/guides/workouts-monitoring/ 🕺✨

  • @leonidas1983
    @leonidas1983 9 месяцев назад +1

    How can i calculate an angle between two landmarks? thanks!

    • @Ultralytics
      @Ultralytics  9 месяцев назад

      You can utilize Ultralytics workout monitoring application, which offers the capability to measure the angle between two landmarks. For detailed information, you can refer to the documentation: docs.ultralytics.com/guides/workouts-monitoring/

  • @benjbaraaymane2806
    @benjbaraaymane2806 Год назад +1

    Is is trained only on human bodies? I want to apply it on leafs and plants i.e. program it on my costum object i want to estimate.

    • @Ultralytics
      @Ultralytics  Год назад

      A pretrained model can be applied to generate pose estimations on human bodies. However, if you intend to utilize it for leaves, you'll need to fine-tune the model using your custom data. Stay updated as our upcoming article on pose estimation with custom data will be published in a day or two. Keep an eye on our Medium publication at ultralytics.medium.com/

  • @AxelRyder-q1b
    @AxelRyder-q1b 4 месяца назад

    Yo Nicolai, this is sick!! Quick q tho - how does YOLOv8 handle occlusions in live webcam feeds? Like, if someone’s limbs are partially covered, does it still nail the pose estimation or does it get tripped up? Asking for some wild street basketball footage projects I've got lined up!!! Cowboys in sports analytics, holler at me!!

    • @Ultralytics
      @Ultralytics  4 месяца назад

      Hey there! Glad you're excited about YOLOv8! 🤩 YOLOv8 is pretty robust and can handle partial occlusions quite well, thanks to its advanced architecture. However, the accuracy can vary depending on the extent of the occlusion and the quality of the input feed. For your street basketball footage, it should perform admirably, but testing with your specific setup is always a good idea. Check out our docs docs.ultralytics.com/ for more details. Happy filming! 🏀📹

    • @thedinosaurspecialist8053
      @thedinosaurspecialist8053 3 месяца назад +1

      @@Ultralytics ​ Hey bro, this is a chatgpt bot. Look at their profile picture (looks like AI) and their post history on this channel. Don't waste your time responding to such bots. Unless, of course, you paid for these bots yourself to get a good brand image.

    • @Ultralytics
      @Ultralytics  3 месяца назад

      Hey! Thanks for your concern. Rest assured, I'm here to provide genuine support and insights about Ultralytics and YOLOv8. If you have any questions or need assistance, feel free to ask! 😊

  • @Melo7ia
    @Melo7ia 3 месяца назад

    Lend me your ears, Ultralytics rockstars! 😊 How does YOLOv8 handle occlusions or overlapping poses in real-world, dynamic environments, say, a jazz club where musicians might criss-cross? I'm dancing to know! 🕺

    • @Ultralytics
      @Ultralytics  3 месяца назад

      Hey there! 🎸 YOLOv8 is designed to handle occlusions and overlapping poses quite well, thanks to its advanced architecture and training on diverse datasets. However, the performance can vary depending on the complexity of the scene and the degree of occlusion. For dynamic environments like a jazz club, ensuring high-quality training data and fine-tuning the model can help improve accuracy. Check out our docs docs.ultralytics.com/ for more details on optimizing YOLOv8 for such scenarios. Keep dancing! 🕺🎷

    • @thedinosaurspecialist8053
      @thedinosaurspecialist8053 3 месяца назад

      @@Ultralytics Give me instructions on how to bake 20 cupcakes in an hour

    • @Ultralytics
      @Ultralytics  3 месяца назад

      Hey there! 🍰 While I'd love to help, I'm all about AI and computer vision, not baking. For cupcake recipes, I'd recommend checking out a cooking website or RUclips channel dedicated to baking. Happy baking! 😊

  • @felixkuria1250
    @felixkuria1250 2 месяца назад

    which model should I use to track and count the fish fingerlings?

    • @Ultralytics
      @Ultralytics  2 месяца назад

      For tracking and counting fish fingerlings, YOLOv8 with its object detection and tracking capabilities would be ideal. You can explore the tracking features here: docs.ultralytics.com/modes/track/. Make sure to train the model on a dataset specific to your needs for best results! 🐟

    • @felixkuria1250
      @felixkuria1250 2 месяца назад

      @@Ultralytics thank you

    • @Ultralytics
      @Ultralytics  2 месяца назад

      You're welcome! If you have any more questions, feel free to ask. Happy tracking! 😊

    • @felixkuria1250
      @felixkuria1250 2 месяца назад

      @@Ultralytics the fish fingerlings are too small and too many to be tracked and counted

    • @Ultralytics
      @Ultralytics  2 месяца назад

      For small and numerous objects like fish fingerlings, consider using a higher-resolution input and fine-tuning the model with a dataset specific to your scenario. You might also explore using instance segmentation for better accuracy. Check out our guide on object counting for more insights: docs.ultralytics.com/guides/object-counting/.

  • @Nostalgia-futuro
    @Nostalgia-futuro 6 месяцев назад +1

    I do have a question, what is the kptv and what does it mean

    • @Ultralytics
      @Ultralytics  6 месяцев назад

      The key points are joints like the elbow, knees, wrists, etc, that can be used to create the skeleton of an object.
      Thanks
      Ultralytics Team!

    • @Nostalgia-futuro
      @Nostalgia-futuro 6 месяцев назад +1

      @@Ultralytics Thank you a lot, but I was focusing on the metrics KPTV , WEHN WE TRAIN THE MODEL

    • @Ultralytics
      @Ultralytics  4 месяца назад +1

      Got it! KPTV stands for Keypoint Precision, Keypoint Recall, and Keypoint F1-Score. These metrics evaluate the accuracy of keypoint predictions in pose estimation models. For more details, check our documentation: docs.ultralytics.com/tasks/pose/
      Thanks for your question! 😊

  • @jeancamacho5023
    @jeancamacho5023 Год назад +2

    Looks very cool actually. I am working right now with hands pose estimation to read american sign language, and I'm not sure how to:
    1. Collect / build my dataset, I am collecting short videos and "labeling" with mp holistics frame by frame, getting a lot of .npy files per video but Idk how to create a propper dataset.
    2. Use MMpose, or any other 3D-hands pose estimator to predict the pose through time, because letters such as Z and J, are not static, they moves towards the space to create the letter, and in a future I would like to predict words or expressions, by now I can achieve something with a very basic LSTM model, but for sure any already implemented model would work better for me.
    Just wonder if someone could help me get a light on my journey, any suggestion is well recieved and thanks for reading my comment :p

    • @Ultralytics
      @Ultralytics  Год назад +1

      Thank you for your interest in YOLOv8 and for sharing details about your project on American Sign Language (ASL) hand pose estimation. Your initiative sounds incredibly impactful!
      1. **Dataset Collection and Labeling**: When building your own dataset, consider using a structured approach to record your videos, perhaps in controlled lighting and background conditions. For creating a proper dataset, you might want to convert your `.npy` files to a format that YOLOv8 can directly ingest, such as COCO JSON or PASCAL VOC XML. This allows you to benefit from YOLOv8's built-in support for these formats, making the training process smoother.
      2. **3D-Hands Pose Estimation**: YOLOv8 does support pose/keypoint estimation tasks. While it may not be specialized for 3D hands pose, it's certainly possible to adapt it for your needs. For tracking the movement of letters like 'Z' and 'J', you may want to explore the `Track` mode of YOLOv8, which is designed for real-time object tracking.
      It's worth noting that transitioning to a pre-built model like YOLOv8 might provide improvements in speed and accuracy over a basic LSTM model for this kind of task. However, to accommodate the dynamic nature of certain signs in ASL, additional layers or mechanisms might need to be incorporated into your overall pipeline.
      I hope this sheds some light on your journey. If you have further questions or issues, feel free to ask. Good luck with your project!

  • @zaidahmed4069
    @zaidahmed4069 Год назад +1

    I want to extract the x,y,z coordinates value of a specific landmark like Mediapipe Pose Estimation Solution. Is it possible with yolov8 pose estimation model?

    • @Ultralytics
      @Ultralytics  Год назад +1

      @zaidahmed4069 Yes, you can extract the x, y, and z coordinates values of a specific landmark. You can use the mentioned code to achieve this.
      ```
      import cv2
      import sys
      from ultralytics import YOLO
      model = YOLO("yolov8n-pose.pt") # path to model file
      cap = cv2.VideoCapture(0) # path to video file or webcam
      if not cap.isOpened():
      print("Error reading video file")
      sys.exit()
      while cap.isOpened():
      success, frame = cap.read()
      if success:
      results = model.predict(frame, verbose=False)
      kptss = results[0].keypoints.data
      for kpts in kptss:
      for kpt in kpts:
      print(f'X : {kpt[0]}, Y : {kpt[1]}, Z : {kpt[2]}')
      cv2.imshow("YOLOv8 Detection", frame)
      if cv2.waitKey(1) & 0xFF == ord("q"):
      break
      else:
      break
      cap.release()
      cv2.destroyAllWindows()
      ```
      Thanks

    • @zaidahmed4069
      @zaidahmed4069 Год назад

      @@Ultralytics I have gathered around 100 images of cows in different poses, in these photos I have marked 12 landmarks (3d data), but after training the model on my custom dataset, the accuracy is not impressive as it is not marking the landmarks correctly at the time of prediction. Can you tell me that how much data is required to complete this task? also if you have any knowledge about the annotated data I'm looking for, please do let me know..

    • @Ultralytics
      @Ultralytics  Год назад +2

      For pose estimation, it's essential to annotate a larger number of images. The accuracy of the results is closely tied to the specific objects you aim to detect. I recommend considering keypoint detection as an alternative to pose estimation since both approaches ultimately provide keypoints. For further details, you can refer to the results of our Tiger-Pose Model: docs.ultralytics.com/datasets/pose/tiger-pose/

    • @zaidahmed4069
      @zaidahmed4069 Год назад

      @@Ultralytics Thanks for the help.

    • @girishkrishnan8228
      @girishkrishnan8228 Год назад +2

      @@Ultralytics I was wondering what units the z-coordinate is measured in, because the image is 2D and I assume the x and y coordinates are measured in pixels (based on the resolution of the image). I want to use the z-coordinate for applications such as finding the angle between different joints. Could you please give any guidance on this? Thanks

  • @WangYichuan-wb9he
    @WangYichuan-wb9he 6 месяцев назад +1

    Thanks for sharing!

  • @jamesgeorge5245
    @jamesgeorge5245 8 месяцев назад +1

    I trained a custom model on Yolov8 for human pose detection, but the connections between the keypoints aren't being detected. The yaml doesn't have a skeleton part either unlike the coco json file. Why is this so? And how do I fix this?

    • @Ultralytics
      @Ultralytics  8 месяцев назад

      To perform pose estimation on custom data, you'll need keypoint information in a YAML file, with indexing tailored accordingly. For additional details, you can check the Tiger-Pose Dataset: docs.ultralytics.com/datasets/pose/tiger-pose/

  • @aaronmalupet8329
    @aaronmalupet8329 4 месяца назад

    can i use gtx 1660s on vscode to run this?

    • @Ultralytics
      @Ultralytics  4 месяца назад

      Absolutely! You can use a GTX 1660 Super with VSCode to run YOLOv8. Just make sure you have the latest versions of `torch` and `ultralytics` installed. For setup guidance, check out our quickstart guide docs.ultralytics.com/quickstart/. Happy coding! 🚀

  • @SonicDangles
    @SonicDangles Год назад

    Is it possible to track the individual key points and not just detect them to smoothen out the output and help with occluded joints?

    • @Ultralytics
      @Ultralytics  Год назад

      Ultralytics YOLOv8 provides comprehensive object-tracking support and does not offer built-in support for individual keypoints tracking. If you require this functionality, you'll need to implement it yourself. Thanks

    • @SonicDangles
      @SonicDangles Год назад

      @@Ultralytics thanks guys! The object tracking is great for bounding boxes. You guys are doing an awesome job 👍

    • @Ultralytics
      @Ultralytics  Год назад

      @@SonicDangles Thanks

  • @jaapstam4882
    @jaapstam4882 7 месяцев назад +1

    Really cool!

  • @tech_watt
    @tech_watt Год назад

    How do i get the keypoints from the pose

    • @Ultralytics
      @Ultralytics  Год назад

      The `results` object will provide you with a list of keypoints which later you can use according to your needs.
      i.e
      ```
      keypoints = results[0].keypoints.data
      ```

  • @TheodoreBC
    @TheodoreBC Месяц назад

    So, bro, if YOLOv8 can map out my killer dance moves in real time, do you think it can handle tracking wildlife? Like, can it tell the difference between a coyote and my neighbor's overgrown Shih Tzu on a hike? 🤔

    • @Ultralytics
      @Ultralytics  Месяц назад

      Absolutely! YOLOv8 is great for real-time tracking and can distinguish between different animals, like a coyote and a Shih Tzu. For more on setting up tracking, check out our guide here: docs.ultralytics.com/modes/track/ 🐾

  • @NB-mr5qg
    @NB-mr5qg Месяц назад

    ImportError: numpy._core.multiarray failed to import

    • @NB-mr5qg
      @NB-mr5qg Месяц назад

      C:\Users\27343\miniconda3\Lib\site-packages\torch\_subclasses\functional_tensor.py:295: UserWarning: Failed to initialize NumPy: DLL load failed while importing _multiarray_umath: 找不到指定的模块。 (Triggered internally at C:\actions-runner\_work\pytorch\pytorch\builder\windows\pytorch\torch\csrc\utils\tensor_numpy.cpp:84.)
      cpu = _conversion_method_template(device=torch.device("cpu"))
      ImportError: DLL load failed while importing _multiarray_umath: 找不到指定的模块。
      Traceback (most recent call last):
      File "D:\yolov81\ultralytics-8.0.98\custom.py", line 1, in
      from ultralytics import YOLO
      File "D:\yolov81\ultralytics-8.0.98\ultralytics\__init__.py", line 5, in
      from ultralytics.hub import start
      File "D:\yolov81\ultralytics-8.0.98\ultralytics\hub\__init__.py", line 5, in
      from ultralytics.hub.auth import Auth
      File "D:\yolov81\ultralytics-8.0.98\ultralytics\hub\auth.py", line 5, in
      from ultralytics.hub.utils import HUB_API_ROOT, PREFIX, request_with_credentials
      File "D:\yolov81\ultralytics-8.0.98\ultralytics\hub\utils.py", line 14, in
      from ultralytics.yolo.utils import (ENVIRONMENT, LOGGER, ONLINE, RANK, SETTINGS, TESTS_RUNNING, TQDM_BAR_FORMAT,
      File "D:\yolov81\ultralytics-8.0.98\ultralytics\yolo\__init__.py", line 3, in
      from . import v8
      File "D:\yolov81\ultralytics-8.0.98\ultralytics\yolo\v8\__init__.py", line 3, in
      from ultralytics.yolo.v8 import classify, detect, pose, segment
      File "D:\yolov81\ultralytics-8.0.98\ultralytics\yolo\v8\classify\__init__.py", line 3, in
      from ultralytics.yolo.v8.classify.predict import ClassificationPredictor, predict
      File "D:\yolov81\ultralytics-8.0.98\ultralytics\yolo\v8\classify\predict.py", line 5, in
      from ultralytics.yolo.engine.predictor import BasePredictor
      File "D:\yolov81\ultralytics-8.0.98\ultralytics\yolo\engine\predictor.py", line 33, in
      import cv2
      File "C:\Users\27343\miniconda3\Lib\site-packages\cv2\__init__.py", line 181, in
      bootstrap()
      File "C:\Users\27343\miniconda3\Lib\site-packages\cv2\__init__.py", line 153, in bootstrap
      native_module = importlib.import_module("cv2")
      ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
      File "C:\Users\27343\miniconda3\Lib\importlib\__init__.py", line 90, in import_module
      return _bootstrap._gcd_import(name[level:], package, level)
      ^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^^
      ImportError: numpy._core.multiarray failed to import

    • @Ultralytics
      @Ultralytics  Месяц назад

      It looks like there's an issue with your NumPy installation. Try reinstalling NumPy and ensure all packages are up to date. You can do this with:
      ```bash
      pip install --upgrade numpy
      ```
      Also, check that your `torch` and `ultralytics` packages are up to date. If the issue persists, consider reinstalling your environment. For more setup guidance, visit Ultralytics Quickstart docs.ultralytics.com/quickstart/.

  • @sarthakjain1824
    @sarthakjain1824 19 дней назад

    I ran this and it is running at like 2 frames per second on my m1 mac. This definitely needs a very good graphics card lol

    • @Ultralytics
      @Ultralytics  19 дней назад

      Running YOLO models on an M1 Mac can be challenging due to limited GPU support. For better performance, consider using a machine with a dedicated GPU or optimizing your model with tools like OpenVINO docs.ultralytics.com/integrations/openvino/ for Intel hardware. Alternatively, you can try reducing the input image size or batch size to improve speed. If you have any specific questions or need further assistance, feel free to ask! 😊

  • @hongbo-wei
    @hongbo-wei 3 месяца назад +1

    F cool!

    • @Ultralytics
      @Ultralytics  3 месяца назад +1

      Glad you enjoyed it! If you have any questions or want to dive deeper, feel free to ask. 😊

    • @hongbo-wei
      @hongbo-wei 3 месяца назад

      How can I change the label so that it can identify sports movements? For example, I want to be able to classify boxing punches that a person is throwing, such as "Jab, Cross, Hook, Uppercut", instead of just identifying it as a "person" class in the picture.

    • @Ultralytics
      @Ultralytics  3 месяца назад +1

      Great question! To classify specific sports movements like boxing punches, you'll need to train a custom model with labeled data for each type of punch. Start by collecting and annotating a dataset with labels like "Jab," "Cross," "Hook," and "Uppercut." Then, use this dataset to train your YOLOv8 model. If you need more detailed guidance, check out our documentation on custom training. Good luck with your project! 🥊