Train pose detection Yolov8 on custom data | Keypoint detection | Computer vision tutorial

Поделиться
HTML-код
  • Опубликовано: 2 июл 2024
  • Code: github.com/computervisioneng/...
    0:00 Intro
    0:49 Dataset
    2:45 Data annotation
    10:14 Data format and file system
    13:14 Coco keypoint annotation format
    19:00 Train model
    34:19 Validate model
    45:14 Predict samples
    52:07 Outro
    #computervision #python #posedetection #keypointdetection #machinelearning #objectdetection #poseestimation

Комментарии • 211

  • @ComputerVisionEngineer
    @ComputerVisionEngineer  9 месяцев назад +2

    Did you enjoy this video? Try my premium courses! 😃🙌😊
    ● End-To-End Computer Vision: Build and Deploy a Video Summarization API bit.ly/3tyQX0M
    ● Hands-On Computer Vision in the Cloud: Building an AWS-based Real Time Number Plate Recognition System bit.ly/3RXrE1Y
    ● Machine Learning Entrepreneur: How to start your entrepreneurial journey as a freelancer and content creator bit.ly/4bFLeaC
    All my premium courses are available to the Computer Vision Experts in my Patreon. 😉
    www.patreon.com/ComputerVisionEngineer

  • @richardngweisiong2895
    @richardngweisiong2895 Год назад +21

    This is the first video that I can find to do custom dataset for pose detection and training beside the human pose. Thank you so much.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад +1

      🙂You are welcome! Glad you enjoyed it! 💪🙌

    • @gauravcuav
      @gauravcuav 21 час назад

      How to mark visibility of any particular point?

  • @yassinebouchoucha
    @yassinebouchoucha 4 дня назад

    Even after ~ one year this video is still my reference to tackle pose estimate workflow !

  • @davidpoling7748
    @davidpoling7748 4 месяца назад

    These are the best yolo walkthroughs hands down. Thanks for sharing!

  • @otaviormc
    @otaviormc Год назад +2

    Spectacular video Felipe !
    Greetings from Brazil 🇧🇷

  • @maczubel
    @maczubel 8 месяцев назад

    Excelente!!, brillante. El mejor tutorial de pose estimation and keypoints training, 1000 gracias. La simplicidad y humildad de los elegidos, de los verdaderos genios. Gracias, gracias, gracias...ya me suscribí...

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  8 месяцев назад

      Gracias por tu apoyo! Me alegra que el video te haya sido util! 😃🙌

  • @pahadiingermany3231
    @pahadiingermany3231 11 месяцев назад

    This video is great for pose detection. Thank you so much.

  • @hamzawi2752
    @hamzawi2752 Год назад

    I do not find words to thank you, you did a great video!

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад

      Thank you for your kind words, Hamzawi! 😊 Glad you enjoyed it! 😃💪

    • @hamzawi2752
      @hamzawi2752 Год назад

      @@ComputerVisionEngineer I have only one question, I hope you can answer me please, if I want only the key points without the bounding box around the object, is that possible or each object should have both bbox and keypoints. I have a project where the data is annotated with keypoints without any bounding boxes. I watched your video, you added bbox coordinate for each object. Can I train Yolov8 with keypoints detection without object detection (bounding boxes)?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад

      In order to train the keypoint detector you need both; keypoints and also the bounding box. If you only have the keypoints you could calculate an approximated bounding box by using OpenCV to find the bounding box that encloses the keypoints. 🙌

    • @hamzawi2752
      @hamzawi2752 Год назад

      @@ComputerVisionEngineer Thank you so much. Yes, I checked that Yolov8 needs both and after a long preprocessing, I managed to do it. I trained it with both bounding boxes and key points and I achieved high mAP. Thank you so much for your answer. By the way, I did not find any RUclipsrs who explained the parameters of Yolov8 or the dashboard that appears during the training (e.g., df_loss, pose_loss). If you can explain that in one of your future videos, I would be more appreciated.

  • @fifthperson9777
    @fifthperson9777 Год назад

    Amazing!!!
    Keep up the good work man💪💪

  • @codys846
    @codys846 Год назад +1

    your videos are so good and easy to follow ive really learned alot thank you.

  • @mahendrakumargohil6384
    @mahendrakumargohil6384 10 месяцев назад

    Excellent Tutorial, really very helpful. I learnt a lot and trained on my own custom dataset. Thank you so much for such a intuitive tutorial😊😊

  • @user-ej5nt6lk3q
    @user-ej5nt6lk3q Месяц назад

    This is the best project walkthrough I can find till now. By the way love the italian accent. Sounding like the godfather is himself teaching you😂

  • @Mamunur-illini
    @Mamunur-illini 6 месяцев назад

    Thanks for the great video.

  • @tanneMadOne8667
    @tanneMadOne8667 10 месяцев назад

    Pretty pretty pretty good tutorial, thank you😀

  • @rudranshagrawal7828
    @rudranshagrawal7828 5 месяцев назад

    best aadmi hai bhai maza aagya love you bhai🤎♥

  • @JasminPatel
    @JasminPatel Год назад

    Really nice work man! God bless you

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад +1

      Thank you! 😊🙌

    • @JasminPatel
      @JasminPatel Год назад

      @@ComputerVisionEngineer I really hope your channel grow, are you on LinkedIn? Let's stay connected.

  • @GabrieldeOliveira86
    @GabrieldeOliveira86 Год назад

    Great tutorial. Thank you.

  • @julianavarela4936
    @julianavarela4936 4 месяца назад

    Idolo

  • @malsharasangani6851
    @malsharasangani6851 10 месяцев назад

    Hi thank you for the video. I just want ask when we annotate is it possible to annotate videos with different positions in the same video ? Like sitting and standing and Kable them separately

  • @mobinmohammadi6562
    @mobinmohammadi6562 Месяц назад

    Thank you for your good training.
    Regarding the Python script to convert the .xml file to YOLO-pose format, is the script in the video tutorial correct or is it available in the Git repository? Because they are different. I used the Git repository but it gives an error when training the algoritm.

  • @srikantapramanik3689
    @srikantapramanik3689 Месяц назад

    The best tutorial I've ever seen. ❤
    Can you please provide us the link of the annotated dataset including the train and val images, bcz it's take lot of time to annotate all the images. so that we can use it.
    Thank you. ❤

  • @PShowmaKer
    @PShowmaKer 9 месяцев назад

    If I just want to ge the coordinate of the keypoint, what should I do?
    Execute predict(video,save_txt = true)?

  • @CheeriiohCheckmate
    @CheeriiohCheckmate 3 месяца назад

    I have a question: If I need to key point detect two objects, such as people and animals, in my label file must the object id, bounding box, and key point set of both objects be in the label file?

  • @TWUNE-oq1li
    @TWUNE-oq1li 2 месяца назад

    Hi, could I ask what the units are for the results parameters post training, for e.g. parameters such as pose_loss and kobj_loss and where I can find this information? Thank you.

  • @yassinebouchoucha
    @yassinebouchoucha 8 месяцев назад

    Good step by step project walkthrough from annotation, local hello world, cloud training, to analysing and testing final results ! Do you offer Full length Udemy like course ?

  • @IO-fz2sm
    @IO-fz2sm 11 месяцев назад

    could you please tell me the annotations process for MULTIPLE OBJECTs PER IMAGE of same class.

  • @sanjay6262
    @sanjay6262 Год назад

    Thank you sir, your videos are amazing. This is really very helpful for me. I have one request for you can you prepare some more end-to-end industry-oriented project. Specifically Mechanical Industries.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад

      Hey Sanjay, glad the videos are helpful! 😃 Sure, I enjoy end to end projects too, I may make more of those in the future. 💪💪

  • @leomatero2524
    @leomatero2524 Год назад

    thank you so much, i have been stuck by v7 for two days to train my custom dataset, the tricky part is tensor size while i changed the number of keypoints which are different with 17 kpt skeleton detection. Today! I tried v8 after watching your video, that was amazing, so convenient! Selecting a good model can do save time lol

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад +1

      Yeah, agreed! I am glad you found the video useful! 🙌

    • @leomatero2524
      @leomatero2524 Год назад

      @@ComputerVisionEngineer really useful and helpful. Looking forward to your new videos in the future! Thanks

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад

      @@leomatero2524 oh yes, it is wrong, thank you so much for noticing it! I will fix it ASAP 💪🙌

    • @leomatero2524
      @leomatero2524 Год назад

      @@ComputerVisionEngineer no worries, annotation format is always tricky 🤣

    • @Nufall
      @Nufall Год назад

      @@leomatero2524 could you please tell me what was the the issues with the annotation format?

  • @darkjudic
    @darkjudic Год назад

    Hi Felipe, how do you think if we use a larger model for pretraining will we have a better result? I.e. yolov8m-pose, yolov8l-pose, yolov8x-pose etc.
    And another question - it looks like while labeling you didn't use the variant v=1 labeled but not visible - do you think it's not important?
    Thanks.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад +1

      Hey Konstantin, using a larger model could produce better results. According to performance metrics on yolov8 website larger models achieve a higher average precision. Do you mean during my custom labeling I didn't use v=1? As far as I know CVAT doesn't support the (x, y, v) format for keypoints annotation. Since I used CVAT as my tool, I only used the (x, y) format. I wouldn't say 'it is not important', if you have visibility information for all keypoints, it's even better. But if you don't, you can still try training the model using the (x, y) format. 💪💪

  • @NathanCamillerii
    @NathanCamillerii Месяц назад

    Training/Validation images do the not need to all be the same size, correct? Thanks :)

  • @rm3204
    @rm3204 Год назад

    Hi Felipe,is it possible to share with us the annotated dataset or a subset of it in the video since annotating 600 images is a seriously quite a lot to do? Thanks for your great tutorial on the topic.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад +1

      Hey, sure, I've added a link to the data in the Readme file of this project's github repository. 🙌

  • @sanjoetv5748
    @sanjoetv5748 9 месяцев назад

    can i use this to measure distances between landmark? please answer me Thank you so much

  • @user-hw8kg5zl6u
    @user-hw8kg5zl6u Год назад

    Is that possible to make a video how to run key points detection with yolov8 in the web? By using tfjs

  • @laurahagedorn4637
    @laurahagedorn4637 4 месяца назад

    Thank you so much for this very comprehensive video! 🙌
    I was able to train a model on my custom data, but I would like to see those "lines" connecting the key points to generate a skeleton. How can I achieve that?

  • @mohd.abdulghani6172
    @mohd.abdulghani6172 7 месяцев назад

    hai, i just saw your videos, is it can be done with yolov5? i mean, can we train pose detection with yolov5? thank you

  • @stevegreen5638
    @stevegreen5638 5 месяцев назад

    in the cvat_to_coco file, the code doesn't properly close files after opening them. This caused me some issues.
    This is my correction :
    with open(label_file_path, 'w') as label_file:
    label_file.write('0 {} {} {} {}
    '.format(str((xtl + (w / 2)) / width), str((ytl + (h / 2)) / height),
    str(w / width), str(h / height)))
    ...continue

  • @sreekartammana
    @sreekartammana Год назад

    Awesome could be the small word💥

  • @ducinemaeveryday2190
    @ducinemaeveryday2190 Год назад

    thanks you

  • @SharathChandraclosetoyou
    @SharathChandraclosetoyou 3 месяца назад

    what if i have many humans in the image? will the dataset .txt file be the same?

  • @user-hl2kq7of7n
    @user-hl2kq7of7n 9 месяцев назад

    Hi, great tutorial like always, you are an inspiration for many people.
    I wonder if the "bounding box" of the classes could be an ellipse or a polygon, does YOLO can recognize this kind of annotation? In case it can, how should i edit the yaml file, beacuse i notice that in the yaml file there is not configuration for the bounding box, looks like the rectangle is the default.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  9 месяцев назад

      At the best of my knowledge, yolov8 object detection only supports rectangular shapes. You may find other technologies to do polygon detection. Alternatively, you could do semantic segmentation with yolov8, by doing so you will detect objects by producing a mask with an arbitrary shape. 🙌

  • @akhilunni100
    @akhilunni100 4 месяца назад

    Thank you for this tutorial. I have a question
    So, in the images where some keypoints are not present. What should I do ?
    How to annotate such images while maintaining the order of the annotation ?

  • @skylord_jameson5702
    @skylord_jameson5702 10 месяцев назад

    Great video, I was wondering how you were able to get the visibility tag in the labels files, as they were not included in the exported xml file.

    • @pahadiingermany3231
      @pahadiingermany3231 10 месяцев назад

      Hi there,
      Sorry that this reply is not to your question.
      I wanted to know if you manually annotated all the images in your dataset or there was a way to do annotation faster. I was working on another dataset which has 12000 train images and 2800 test images. But it doesn't have annotation.txt files for labels. Can you suggest something?
      Regards

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  10 месяцев назад +1

      Hey, In this tutorial I show you how to produce annotations using CVAT, but those are not the labels I use to train the model. The dataset I used to train the model was already annotated with the format. 🙌

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  10 месяцев назад +1

      I didn't manually annotated all the images in the dataset. Depending on the situation, there could be ways to speed up the annotation process, for example you could 'pre annotate' the data by using the predictions from other model.

    • @Paula-zi5hw
      @Paula-zi5hw 9 месяцев назад

      @@ComputerVisionEngineer Hello, and thanks for your videos, I've started with object detection for a project and I find your videos the best, most precise and helpful! However I am still struggeling a little - I used your code to convert the xml file to yolo format but as you pointed out here as well I only have format - so one question, do I need the visivility variable or is it just as well possible to just use the format, but what difference does it make then?

  • @GitPappaBaker
    @GitPappaBaker 7 месяцев назад

    One follow-up question: In the video, you annotate with CVAT, but your actual annotations have the third visibility dimension for the keypoints. What tool do you recommend that includes that ability to include the visibility dimension? I don't need it for what I'm doing this week, but I'll need it next week. I appreciate any advice you can offer that will save me from trial/error more tools.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  7 месяцев назад

      Not sure what could be the best tool to label the third visibility dimension keypoint, I haven't used any so far. The dataset I used in the video was already annotated. Sorry I can't help you with this. 🙌

    • @GitPappaBaker
      @GitPappaBaker 7 месяцев назад

      @@ComputerVisionEngineer Thanks for the reply. Looks like RectLabel is capable but the bbox annotation it is producing along with the keypoints is generated as a box around the keypoints, not an additional one around the object. Not exactly what I want so I'm pondering combining a bbox with the keypoints and then swapping the box through a script. If that tool had decent (aka: ANY) tutorial videos it would be a lot easier to use.

  • @user-bpbi76tffjo87
    @user-bpbi76tffjo87 3 месяца назад

    Congratulations! I did everything according to your guide, but I'm getting an error: "No images found in D:\keypoint_detection\leg\data\labels\train.cache, training may not work correctly.". I've tried everything, but the error persists. Is there any way I can contact you to provide the files and code so you can tell me where I made a mistake?

  • @kristinatas6100
    @kristinatas6100 11 месяцев назад +1

    Hi, I have already annotated images (composed of just non-normalized coordinates) without the bounding box. How can I convert my annotation to match yolov8?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  11 месяцев назад

      Hey, you need the bounding box as well to use yolov8. If annotating the bounding boxes is not a possibility, you could estimate it by taking the bounding box that encloses your keypoints. 🙌

  • @user-rq6wk6rn3x
    @user-rq6wk6rn3x 8 месяцев назад

    Hello, can anyone tell what is kobj_loss? Mine is zero everytime during training.

  • @ghanshyamkumbhar5953
    @ghanshyamkumbhar5953 8 дней назад

    How to add label 0 0 0 for the point which is not vissible in cvat0

  • @zaidahmed4069
    @zaidahmed4069 7 месяцев назад

    Can we extract the x,y,z coordinates of a certain landmark with yolo v8 ??
    I saw this in one of the mediapipe solutions, but that solution is not customizable on a custom dataset. So I'm looking for the alternative

    • @aditdesai4848
      @aditdesai4848 Месяц назад

      Did you find a solution to this?

  • @jordaocassiano
    @jordaocassiano Год назад

    What a high-quality tutorial, it's going to be very useful for several people. Let me just ask a question, can I follow these steps if there were more than one animal in the image, that is, different instances for the same class in the image? Would I need to modify anything in the annotations?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад +1

      Thank you! If there is more than one object in the images you would need to adjust the code in CVAT_to_cocoKeypoints.py. In the txt files with the labels, there should be a line for every object in the image; each line should follow the structure I explain in the video. Other than that everything I explain in this tutorial should work fine! 😃🙌

    • @jordaocassiano
      @jordaocassiano Год назад

      @@ComputerVisionEngineer Great, thanks. By the way, one tip I would add is to use skeleton annotation. In this format, CVAT exports in COCO keypoints (json) format, and then you can use reclabel to convert from COCO json to Yolov8 format.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад

      Oh, amazing! Thank you so much for the heads up! 😃😃😃

    • @nicolasarvanitidis7639
      @nicolasarvanitidis7639 6 месяцев назад +2

      @@jordaocassiano if i have the coco keypoints in a json format, how do i use that to make a yolov8 pose estimation model?

  • @leonidas1983
    @leonidas1983 7 месяцев назад

    Hola Felipe! Charrua? Lidero una startup que usa pose estimation... Felicitaciones por el canal!

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  7 месяцев назад

      Hola! Sí, uruguayo! Vos también? Gracias!! 😃 🙌

    • @leonidas1983
      @leonidas1983 6 месяцев назад

      @@ComputerVisionEngineer de Argentina! Felicitaciones por el canal y el contenido

  • @tarekt8571
    @tarekt8571 Год назад +1

    hi bro, how do you create mask image for parking slots in parking spot detection and counter video?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад +1

      Hey, you can use any image edition software (I used Inkscape), you need to take a frame from the video and draw white rectangles on top of it matching the parking slots.

  • @yiwang1111
    @yiwang1111 Год назад

    Thanks for the video.
    When annotating data, say an antelope, what happens if the antelope is female and it does not have antler? do we skip or do we annotate the imaginary antler?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад

      Ideally, it is better to use the x, y, v format for these situations, If you are using the x, y, v format you can label those keypoints as 0, 0, 0.
      If you are using the x, y format you could apply different criteria, one of them could be annotating the keypoints where they 'should be located' (in the case of non existing antlers I would locate the keypoints in the head). But again, this is not ideal, try to use the x, y, v format in those situations. 🙌

  • @rocketmike9847
    @rocketmike9847 5 месяцев назад

    Great video!! I have one question: is there a significant performance difference if you don't flag the visibility of the points vs doing so? Thanks!!

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  5 месяцев назад +1

      If you have the visibility information the better. But if you don't, the model is most likely to learn the structure of your objects anyway.

    • @rocketmike9847
      @rocketmike9847 5 месяцев назад

      @@ComputerVisionEngineer Thanks!

  • @EliSpizzichino
    @EliSpizzichino Год назад +1

    Have you actually tried to export directly in YOLO format from CVAT? YOLOv8 format is the same as v5

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад

      Yes, I tried. The way I managed to make it work was downloading the annotations in cvat format and then converting to coco keypoints format. If you find an easiest way to do it do let us know! Contributions are always appreciated. 😃💪

  • @Sumukhasidhma
    @Sumukhasidhma 3 месяца назад +1

    I am training the model with cvat dataset with 3 classes(bounding boxes) and 11 key points but not able to train the model it is giving the error like "C:\Users\prime\anaconda3\envs\yolo_ultra_8083\lib\site-packages\ultralytics\data\dataset.py", line 139, in get_labels len_cls, len_boxes, len_segments = (sum(x) for x in zip(*lengths)) ValueError: not enough values to unpack (expected 3, got 0)"
    can you please give reference or any data for training multiple classes and respective keypoints

    • @SpeerKING
      @SpeerKING Месяц назад

      That happen to me as i forgot to put bounding boxes. I would go to ur saved dataset in cvat and check if really all the bounding boxes and keypoints are there.

  • @tompiedefer1407
    @tompiedefer1407 11 месяцев назад +1

    Hi Felipe, I followed your tuto. However, I have a problem about corrupted images or label : "ignoring corrupt image/label: labels require 11 columns each". I have this for all my images. Do you know where does it come from ?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  11 месяцев назад

      The issue may be your annotations don't match your kpt_shape. What is your kpt_shape?

    • @tompiedefer1407
      @tompiedefer1407 11 месяцев назад +1

      @@ComputerVisionEngineer
      kpt_shape: [2, 3]
      flip_idx: [0, 1]
      I only need 2 keypoints

  • @abderrahmaneherbadji5478
    @abderrahmaneherbadji5478 Год назад

    Please could you share the trained model.

  • @is9547
    @is9547 5 месяцев назад

    Excellent tutorial - it has been incredibly beneficial. I greatly appreciate the sharing of your knowledge. I've gained much insight from it. Thank you immensely for your effort. However, I've encountered an issue when exporting annotations in CVAT 1.1 format: only one bounding box and its associated keypoints are exported, even though there are two boxes and two sets of keypoints present. Are you aware of how to address this problem? Additionally, could you provide guidance on how to train these annotations with YOLOv8?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  5 месяцев назад

      Hi, thank you! Glad the content is helpful. Not sure what could be going on, does it happen every time you try to export an image with two objects annotated on it?

    • @is9547
      @is9547 5 месяцев назад

      Thank you for your kind words. I can't do it every time I have more than two sets of annotations. And I want to know how to train more than one set of keypoints.

  • @EliSpizzichino
    @EliSpizzichino Год назад

    good but not good enought to actually bind a skeleton and make it move, so what is it usefull for?
    Also I would have used a multi-class aproach for the different parts of the body

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад

      Thank you so much for your contributions! I will definitely keep it in mind for future videos. 💪🙌

  • @sushmithas504
    @sushmithas504 Год назад +1

    Bro I recently did your emotion detection but it show modules not found utils.dataset what to do I try my best but still couldn't get it right

  • @GitPappaBaker
    @GitPappaBaker 7 месяцев назад

    Hey. Great video. I really enjoyed watching it. However, the bounding box coordinates in your image labels are now incorrect in your google drive. A quick look and comparing the antelope_10002 label from your video (13:55) to the one in your repo will show the change. I plotted the difference. The original correctly listed center with height and width, in your repo (google drive) all of the bounding boxes are listed as the start/stop corner coordinates. (so upper and lower corner of the box). I'm planning to just script a change and convert the files to the correct format but I spent a day wondering why I couldn't build your basic model so I thought I would save someone else the time.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  7 месяцев назад

      Hey, thank you so much for the heads up! I will take a look and correct it as soon as possible! 🙌

    • @user-cf7zb8pd1v
      @user-cf7zb8pd1v 7 месяцев назад

      Hello, I trained for 300 epochs according to the cloud data provided by the author and the obtained bounding boxes (bbox) are incorrect. It seems that there is an issue with the coordinates annotation file for the bbox, just as you mentioned. If it were my own dataset, following the steps provided by the author should work correctly, shouldn't it?

  • @user-hm9xy8cp7l
    @user-hm9xy8cp7l Год назад +1

    Can we have different key point numbers for different classes? as an example, I want to annotate my data which has 2 different classes. Class 1 has 3 key points and class 2 has 6 key points. Is it possible to train yolov8 on it? if yes how should we annotate data and prepare the config.yaml file? appreciate any help.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад

      I would advice you to use the same keypoints in your different classes. If this is not possible, the keypoints the classes have in common should have the same 'meaning'. Take a look at the video, some animals have antlers and others don't. The keypoints in the antlers of those animals without antlers are annotated with a visibility flag of 0 (which means not labeled). As for all the remaining keyponts, they have the same 'meaning' in all the animals (eyes, ears, legs, etc.). 💪

  • @haj9710
    @haj9710 10 месяцев назад

    Thank you for very useful video,I followed the steps you said for custom data annotations. But there is a problem, after converting from xml to .txt format using your code, how can we get the train.cache and val.cache that need to present in the 'labels' folder. Please let me how to solve this because without cache file i'm unable to train the custom dataset. Thanks in advance

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  10 месяцев назад

      Hey, the cache files are not needed. You need to locate the data (images and annotations) exactly as I show in the tutorial. 🙌

  • @rfolks92
    @rfolks92 Месяц назад

    Exporting as coco keypoints still not working as of May 2024...

  • @navidaslankhani2337
    @navidaslankhani2337 5 месяцев назад

    Is labeling each key-point with visibility flags [0-1-2] any different from labeling them with visibility flags [0-1] or [0-2]? I mean using visibility flag 1 for occluded key-points.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  5 месяцев назад

      Labelling with visibility flag for occluded keypoints could lead to a more robust training.

  • @name1566
    @name1566 Год назад +1

    I have a knowledge of Only python , would i be able to understand ?, or should I learn other things,? to start it ? please suggest

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад

      Hey, yeah I think you will be able to understand. Even without Python you would be able to follow along 90% of this tutorial. 💪🙌

  • @alantorres6371
    @alantorres6371 6 месяцев назад +1

    Hola Felipe, me ayudaron mucho tus videos en varios de mi proyectos. Sin embargo tengo un problema al intentar entrenar este modelo, no me detecta los labels y ya revisé todo mil veces y sigo sin entender por qué podría ser eso, ya que segui todo el procedimiento paso por paso y el archivo .txt tiene el formato requerido

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  6 месяцев назад

      Hola, estás usando el mismo dataset que yo en el video?

    • @alantorres6371
      @alantorres6371 6 месяцев назад

      @@ComputerVisionEngineer noo estoy usando un dataset nuevo que incluye cubos y 4 keypoints que serían las aristas de la cara superior del cubo

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  6 месяцев назад +1

      @@alantorres6371 No sabría decirte qué puede ser, pero fijate que el config todo esté bien, dirección absoluta al directorio de los datos, el número de keypoints y la cantidad de clases.

  • @eitanas85
    @eitanas85 7 месяцев назад

    What a great content, thanks.
    Is the model also capable of activity detection?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  7 месяцев назад

      Do you mean action recognition?

    • @eitanas85
      @eitanas85 7 месяцев назад

      Yes, sorry for the term abuse. @@ComputerVisionEngineer

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  7 месяцев назад

      @@eitanas85 I would do the action recognition in two steps: landmark detection + Scikit learn classifier

  • @paradisofernando97
    @paradisofernando97 9 месяцев назад

    Hi, I have a question. Is the order of the key points important even if I have a dataset with images all completely on one side? 😊

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  9 месяцев назад +1

      🤔 I would say that when training a pose detector the order of keypoints is always important

    • @paradisofernando97
      @paradisofernando97 9 месяцев назад

      @@ComputerVisionEngineer ooookk thank you very much :)

  • @shehryarashraf5840
    @shehryarashraf5840 8 месяцев назад

    your cvat to coco keypoints format convertor does not work for multiple classes. i see all my 3 classes in cvat format, but in the coco converter version i only see class-id 0 at the start of every line.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  8 месяцев назад

      Agreed. The script only works for 1 class, it needs to be adapted to deal with multi class annotation. 🙌

  • @erencerman4466
    @erencerman4466 9 месяцев назад

    Thank you for your efforts. This seems to be really the best tutorial about yolo pose.
    But I have a question: You annotated with cvat and then saved a cvat format. Then you used a (39,3) dimension during training. How did the cvat to yolo file create the third dimension. I do the same, but I cannot create the third dimension (visibility) with cvat annotation. How should I create the 3 dimensional keypoint annotation with cvat???

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  9 месяцев назад

      Hey, in this video I used cvat only to show an example of how to annotate images to train a pose detector. Cvat doesn't support the visibility flag, only 2-dimentional xy annotation is supported.
      The dataset I used to train the pose detector in this video wasn't annotated with cvat, though. The dataset is publicly available, (x, y, v) annotations are provided. 🙌

    • @erencerman4466
      @erencerman4466 9 месяцев назад

      I am struggling to find a (x,y,v) keypoint annotator, preferably opensource 😅… Could you please suggest one?❤

    • @erencerman4466
      @erencerman4466 9 месяцев назад

      ** coco annotator creates a json file, but ultralytics json2yolo conversion does not convert properly. Do I have to write a converter for coco, like you did for cvat?

    • @erencerman4466
      @erencerman4466 9 месяцев назад

      Turning a coco format json file to a yolo format was the easiest solution. These days open source community is getting weeker, every free code is used to create pennies for companies

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  9 месяцев назад

      @@erencerman4466 I struggled to find one too! I didn't find any open source.

  • @vidyazope8025
    @vidyazope8025 10 месяцев назад

    Spectacular video CV Engineercan it work in occluded environment of humans and animals

  • @hakankosebas2085
    @hakankosebas2085 Год назад +1

    what is your future video plans?

  • @narenvasantakumaar5647
    @narenvasantakumaar5647 9 месяцев назад

    Hi there! I am trying to find the xy coordinates of the keypoints on my test image, I tried your code to print keypoints but it throws on error - 'Keypoints' object has no attribute 'tolist'. Could you help me with this. I typed exactly whats in the video. I just changed the path to my trained model and test image.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  9 месяцев назад

      Hey, are you using the same version of ultralytics that I am using in the video?

    • @narenvasantakumaar5647
      @narenvasantakumaar5647 9 месяцев назад

      @@ComputerVisionEngineer I downloaded the latest ultralytics as I started working with this recently. I wanted to infer and print the keypoint values on test image, so I can compare it with ground truth value. I am not able to infer the keypoints from test image. Could you help me with the code?

    • @BettyLQC
      @BettyLQC 9 месяцев назад +1

      @@ComputerVisionEngineer, Hi thanks for your video, I used the same version, ultralytics==8.0.83, Keypoints' object has no attribute 'tolist'. Could you help me with this?

    • @narenvasantakumaar5647
      @narenvasantakumaar5647 4 месяца назад

      @@BettyLQC Use this code to predict on test images and extract the predicted keypoints coordinates:
      model = YOLO('path to your trained model .pt file path')
      results = model('image file path)) # predict on an image
      for xy in result.keypoints[0].xy[0].cpu().numpy():
      xy_values.extend(xy)
      If you want the normalised keypoint values, replace 'xy' with xyn'

  • @Hankks
    @Hankks 3 месяца назад

    I am struggling with running the training script. It can find 11 images when I run it but says all 11 are corrupted. Are the images required to be of a specific type? I am using jpg

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  3 месяца назад +1

      Jpg format should be ok. Have you tried with other images? Additionally, only 11 images may not be enough to train a robust model.

    • @Hankks
      @Hankks 3 месяца назад

      Sorry, you were indeed correct, it had nothing to do with the image file type. I have struggled to make it function on data I have annotated through CVAT, and I suspect that it might be because the converted COCO label data needs to have the 3rd value of visibillity in order to function. I annotated the data using Roboflow instead and it outputs the labels with this 3rd value for me, and it worked fine@@ComputerVisionEngineer

  • @user-hk5ss6wd7s
    @user-hk5ss6wd7s Год назад

    Hi, I want to know if I can use the keypoint to tracking an object.Thanks!

    • @user-hk5ss6wd7s
      @user-hk5ss6wd7s Год назад

      I have detected an object through the keypoints ,if I can tracking the object still using the keypoints rather than other infomation like the coordinates of bbox.

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад

      Object tracking algorithms I am aware of take the bounding box. Nevertheless, there is some research around object tracking using keypoints. If you prefer using keypoints take a look at some papers on this topic. 💪

    • @leomatero2524
      @leomatero2524 Год назад +1

      @@user-hk5ss6wd7s im trying to use the keypoints on an object to localize it, the idea is to acquire the grasping point (center of the first two keypoints, the first keypoint is used to calculate the rotation) of the object using yolo. I have tried v8 and it works pretty well to detect bboxes and the kpts inside each bbox. For calculating the coordinate of the grasping point (i assume) for your object, you need to write another simple code to calculate the center of the frame based on detected kpts. The model can predict the kpts in a specific order as you labelled them, as in this video, the order of keypoints matters. Good luck!

  • @truongnguyen4777
    @truongnguyen4777 10 месяцев назад

    In the convest xml to txt code i get index out of range (the line start with bbox)
    can u help me pls

    • @truongnguyen4777
      @truongnguyen4777 10 месяцев назад +1

      it said :
      bbox = image.getElementsByTagName('box')[0]
      IndexError: list index out of range

    • @IshaanGupta-ut8ls
      @IshaanGupta-ut8ls 5 месяцев назад

      i got same error? did you fix it?

  • @shantugowda1248
    @shantugowda1248 Год назад

    How we will get text file

  • @ferdyfebriyanto5417
    @ferdyfebriyanto5417 Год назад

    Can you make a video about how to implements DeepFace?

  • @muhammadhammadbashir8989
    @muhammadhammadbashir8989 5 месяцев назад

    I have data having two objects and every object has multiple points, (23,11) I was facing error for having heterogenous number of points. So I made them regular by putting zero padding at end of 2nd object.
    Now I am having 74, keypoints for each row, 1 id + 4 bounding box + (23*3)=69 points for both object, I have two lines.
    My yaml looks like this # number of classes
    nc: 2
    names:
    0: A
    1: B
    nkpt: 46
    kpt_shape: [46, 3]
    but model starting saying he is looking for 46*3=138 +5 in totla 143 points, so every image is getting corrupted,
    Plz help me in resolving this.

    • @muhammadhammadbashir8989
      @muhammadhammadbashir8989 5 месяцев назад

      my bad, since you have 2 objects and each having 23 points so it will be just replace in yaml nkpt: 23
      kpt_shape: [23, 3] and all other strcutre will be same

  • @namao3531
    @namao3531 9 месяцев назад

    Thanks for your video, but when I follow you I get an error:
    ValueError: not enough values ​​to unpack (expected 3, got 0)
    I tried to find the solution on github but still can't, can you help me? Thanks a lot

  • @hemanthreddy2485
    @hemanthreddy2485 4 месяца назад

    Chat to coco format not working

  • @luuminhquan8255
    @luuminhquan8255 5 месяцев назад

    can I extract bounding box from yolo v8 pose, tks

  • @user-po9mq7pv9d
    @user-po9mq7pv9d 11 месяцев назад

    I tried but the custom model could not detect anything. Is your model detect well?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  11 месяцев назад

      I had an ~ok performance as it shows in the test I did at the end of the video. Did you use the same dataset as I did?

    • @user-po9mq7pv9d
      @user-po9mq7pv9d 11 месяцев назад

      @@ComputerVisionEngineer Thanks. It didn't work well because I downloaded a part of datasets.

  • @PraveenKumarYadav-sp7wl
    @PraveenKumarYadav-sp7wl Год назад

    how do you download the images in windows with subset

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад

      Hey, if you are looking for the data I used to train the pose detector in this tutorial, take a look at the Readme file of this project's repository. 🙌

  • @Nufall
    @Nufall Год назад

    Got this error when training, could you help? ValueError: not enough values to unpack (expected 3, got 0)

    • @Nufall
      @Nufall Год назад

      I have 5 key points but the labels have 1 class name + 14 coordinate only, shouldn’t it be 15? Is there’s something wrong with the script?

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад +1

      Hey, you have 5 keypoints and you are using the (x, y, v) format for every keypoint?

    • @Nufall
      @Nufall Год назад

      @@ComputerVisionEngineer yes, i used the same script proposed on your tutorial

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад

      @@Nufall if your data is comprised of 5 points with the x, y, v format you should have 20 values for every object in your annotations file
      Class_id, xc, yc, w, h,

    • @Nufall
      @Nufall Год назад

      @@ComputerVisionEngineer i followed exactly the same steps on your tutorial, where do you think the issues might be? Thank you very much for replying!

  • @fuj7777
    @fuj7777 6 месяцев назад

    Для тех у кого ошибка "np.numpy array expected" и "images corrupted":
    Все аннотации должны быть в формате numpy по типу 0.000000, а не 0.0000000000000.
    Я исправил это добавив в скрипт "convert to coco" из видео следующее:
    format(value_here, '.6f'), для всех значений где они записываются в файл.
    Так что получившийся файл будет со всеми значениями в верном формате по типу 0.161122, 0.112343 вместо 0.23433111111112, и т.д.
    ---------------------
    For those who have the error "np.numpy array expected" and "images corrupted":
    All annotations must be in numpy format like 0.000000, not 0.0000000000000.
    I fixed this by adding the following to the "convert to coco" script from the video:
    format(value_here, '.6f'), for all values where they are written to the file.
    So the resulting file will have all the values in the correct format like 0.161122, 0.112343 instead of 0.23433111111112, etc.

  • @shantugowda1248
    @shantugowda1248 Год назад

    How we get text file in label's

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  Год назад

      Do you mean the annotations? Do you have an annotated keypoints detection dataset you can use in this project? If not, you can download the one I am using in this video!

  • @joelbhaskarnadar7391
    @joelbhaskarnadar7391 9 месяцев назад

    WARNING ⚠ no labels found in pose set, can not compute metrics without labels
    I am getting this message and there is not error loss.

    • @joelbhaskarnadar7391
      @joelbhaskarnadar7391 9 месяцев назад

      Sir I was able train it But the keypoints is not well placed I only used 80 images for training

    • @ComputerVisionEngineer
      @ComputerVisionEngineer  9 месяцев назад +1

      Try with more images. How do the evaluation plots look like?

    • @joelbhaskarnadar7391
      @joelbhaskarnadar7391 9 месяцев назад

      @@ComputerVisionEngineer ok👍🏿