COCO Dataset Format - Complete Walkthrough

Поделиться
HTML-код
  • Опубликовано: 2 июл 2024
  • A detailed walkthrough of the COCO Dataset JSON Format, specifically for object detection (instance segmentations). 👇CORRECTION BELOW👇
    For more detail, including info about keypoints, captions, etc., check out:
    www.immersivelimit.com/tutoria...
    The course is live! You can get a deal on the course and support me with this coupon: www.immersivelimit.com/course...
    👉CORRECTION: In the RLE section, I say that counts run horizontally across the image. They actually run vertically. You start at the top left pixel and count down a number of pixels and it may carry to the next column. Exact same concept, just flip the axes. Sorry about that!
    Here's the JSON file shown in the video: gist.github.com/akTwelve/c703...
    Table of Contents:
    0:00 - Intro
    2:09 - High Level JSON Sections
    2:28 - "Info"
    3:19 - "Licenses"
    4:27 - "Images"
    6:37 - "Categories"
    7:47 - "Annotations"
    8:18 - "Segmentation" (Polygons)
    11:27 - "Segmentation" (RLE - Run Length Encoding)
    17:10 - Exciting Course Announcement
    + Follow These For Updates +
    Immersive Limit Facebook: / immersivelimit
    Immersive Limit Twitter: / immersivelimit
    Subscribe to the RUclips channel too, duh!
    + Connect with Me! +
    My Twitter: / aktwelve
    My LinkedIn: / arkelly12
    + More Ways to Connect +
    www.immersivelimit.com/connect

Комментарии • 177

  • @tunaip.marques7797
    @tunaip.marques7797 5 лет назад +11

    Man, what a great video. Clear, well-edited, informative and to-the-point. Thank you for that!

    • @ImmersiveLimit
      @ImmersiveLimit  5 лет назад +5

      Thank you, Tunai! My wife (and newly minted video editor) was thrilled to hear the "well edited" part. 😊

  • @aadeshsalecha4951
    @aadeshsalecha4951 5 лет назад +66

    Hey man, you have no idea, how much you've helped me. I hope good things happen to you. Thank you so much for doing what you do.

  • @shaozhuowang3403
    @shaozhuowang3403 4 года назад +5

    Best explanation about COCO dataset I've ever seen, thank you!

  • @hiyamghannam1939
    @hiyamghannam1939 4 года назад

    Wow very nice video!! It really pinpoints all the confusions and complexities related to COCO. Thank you!!

  • @beincheekym8
    @beincheekym8 2 года назад +1

    thank you so much for the very nice video. you've put a lot of effort into making it clear, even for newbies. explaining step by step run-length encoding with a very nice example like that... kudos to you!

  • @Bapll
    @Bapll Месяц назад

    Masterfully done, my lad! Thank you so much for explaining this!

  • @TheTimtimtimtam
    @TheTimtimtimtam 5 лет назад +6

    Great job Adam ! Very helpful of you and i hope you are rewarded :)

  • @nomihsa965
    @nomihsa965 Год назад

    This is exactly what I needed. Thanks man, you are life savior.

  • @mangning1107
    @mangning1107 4 года назад +1

    cheers bro, awesome explanation and visualisation, quite logical and tidy!

  • @RajeevPareek
    @RajeevPareek 4 года назад +1

    Simple and extremely useful video. Big Thanks

  • @satyajithj1791
    @satyajithj1791 5 лет назад +2

    Great video man. Thank you so much!

  • @lettuan1982
    @lettuan1982 5 лет назад +2

    Great job! Really helpful!

  • @robmarks6800
    @robmarks6800 Год назад

    Wow, run length encoding for the maskings is so clever

  • @armanasgharpoor3192
    @armanasgharpoor3192 Год назад

    Thanks for your comprehensive video. It was very helpful.🙏

  • @user-bu2dq5tt4f
    @user-bu2dq5tt4f 4 года назад

    This is incredible job. It helped me a lot! Thanks Adam.

  • @korracos
    @korracos 8 месяцев назад

    Thank you for such an enlightening video

  • @adnanchoudhary2254
    @adnanchoudhary2254 Год назад

    That RLE stuff was good. Good work.

  • @hariharanshanmugam5026
    @hariharanshanmugam5026 4 года назад

    superb man ... this is such a wonderful explanation of the coco dataset

  • @phuctranthanh3549
    @phuctranthanh3549 4 года назад +1

    Thanks very helpful. You helped with my subject.

  • @emmali2401
    @emmali2401 5 лет назад +1

    Thank you very much!! It helps me a lot!

  • @lifanzhong9782
    @lifanzhong9782 Год назад

    You really helped me a lot! Thank you!

  • @eranfeit
    @eranfeit 6 месяцев назад

    Great explanation !!

  • @simonbrislin
    @simonbrislin Год назад

    Great video. Extremely helpful. Thankyou

  • @sarankumaar6009
    @sarankumaar6009 2 года назад

    no words to say very very tanks bro u saved my life :)))

  • @RKN998
    @RKN998 2 года назад

    thank you so much you saved me so much time you are a great person i hope great things happen to you

  • @mannversteckter5631
    @mannversteckter5631 2 года назад

    very comprehensive explain,thanks a lot

  • @jfbaezv
    @jfbaezv 4 года назад +2

    Hi Adam thanks for your video and your post, I´m trying to implement the Mask RCNN algorithm, and your video is helping me so much. Tanks so much!

  • @adwaittpandya7030
    @adwaittpandya7030 2 года назад

    Great video, this really helped, thanks!

  • @marcosandoval7260
    @marcosandoval7260 2 года назад

    Thank you so much! The video is great, you helped me a lot.

  • @sudarshannbhat6046
    @sudarshannbhat6046 Год назад

    man you saved my life

  • @mohammadamindehmolae7597
    @mohammadamindehmolae7597 2 года назад +1

    Thanks a lot, you solved my problem😉

  • @degimlnn6157
    @degimlnn6157 Год назад

    Thank you for your excellent video. It does help me a lot.

  • @user-fq5sv9xj9x
    @user-fq5sv9xj9x 2 года назад +1

    What a nice video !

  • @shreyajain1224
    @shreyajain1224 4 года назад +1

    thank you so much, it helped me alot

  • @vandangorade6738
    @vandangorade6738 2 года назад

    Great! thanks man!

  • @YoannBuzenet
    @YoannBuzenet 2 года назад

    really really clear, thank you a lot

  • @ziangzhao4584
    @ziangzhao4584 2 года назад

    Thank you very much !!

  • @Antoniolavoisier1
    @Antoniolavoisier1 Год назад

    Very clear explanation man, thanks. Also i've seen the playlist of "rendering dataset in blender" on your channel, i've dan a quick look you a're training model there is it COCO dataset that you use in that playlist? I couldn't find out it yet.

  • @Harry-jx2di
    @Harry-jx2di 2 года назад

    Thx for your video bruhh!

  • @Syzygyyy
    @Syzygyyy 3 года назад

    Very helpful, thanks

  • @zaidshair1992
    @zaidshair1992 2 года назад

    Excellent video! Quick question, what about Object Tracking? Are they expected to be custom attributes? or is there a standard COCO format for object IDs as well?

  • @donm7906
    @donm7906 4 года назад +1

    thank you very much!

  • @CheeceerLP
    @CheeceerLP 3 года назад

    Thank you so much!

  • @barth05
    @barth05 4 года назад +1

    I hope the course works out, man. Good stuff.

    • @ImmersiveLimit
      @ImmersiveLimit  4 года назад +1

      Thanks Douglas! The course came out about a year ago actually and we're coming up on 500 students. 😀 www.immersivelimit.com/course/creating-coco-datasets

  •  3 года назад +3

    Thanks!!!! I would love to see a reference on the COCO page to this video. I've just spent 2 days trying to decipher this.

    • @YoannBuzenet
      @YoannBuzenet 2 года назад +1

      Glad I wasn't the only one :D

  • @daehyeonkong1762
    @daehyeonkong1762 4 года назад

    you are so kind thankyou thankyou

  • @peizhenshi2085
    @peizhenshi2085 4 года назад +1

    excellent! very helpful

  • @sanderg9106
    @sanderg9106 3 года назад +3

    Great video! although i do recommend people watching to use 1.5x or 2.0x speed, considering this is still easy to follow at that speed.

    • @hackercop
      @hackercop 3 года назад

      thats what i did

  • @user-ey9ip8gg7n
    @user-ey9ip8gg7n 9 месяцев назад

    It was very helpful thank you very much! and I want to know how to visualise this data?

  • @abidan88
    @abidan88 5 лет назад +1

    thank you , sir!!

  • @lowi9871
    @lowi9871 5 лет назад +2

    Thanks for the video! One question, does the file name in the image segment has to be the relativ path from the json file to the image, like "/.jpg" or is only the filename needed?

    • @ImmersiveLimit
      @ImmersiveLimit  5 лет назад

      It should just be “.jpg”. Whatever code is consuming the JSON is required to know the folder path. Every neural net I’ve seen does this a little differently unfortunately, so you may have to dig into the code.

  • @minghsuanlujol
    @minghsuanlujol 3 года назад

    Perfect

  • @lokeshwaranmanohar2967
    @lokeshwaranmanohar2967 8 месяцев назад

    Hey, Thanks for the video. I appreciate your effort. Is there a script available that convert COCO (with RLE) to yolo format?

  • @rixdalerheptads1505
    @rixdalerheptads1505 2 года назад +1

    Slight correction on the RLE Segmentation part: The RLE does not count pixels from the top left corner in row-major order but rather in column-major order. That means you don't count the pixels to the right side of the image but to the bottom of the image.

    • @Mrmactippsgiver
      @Mrmactippsgiver Год назад

      do you know why? Wouldn't this result in a bad cache access pattern?

  • @andihaki
    @andihaki 2 года назад

    for tensorflow facelandmark / keypoint dataset, which dataset does it use?

  • @ds_sjpark
    @ds_sjpark 4 года назад

    thanks

  • @alexruinen9474
    @alexruinen9474 2 года назад

    Thank you very much! I was wondering which annotation tool you are using? I hope you are still seeing this! I've been looking for an open source annotation editor where I can input already existent RLE in coco JSON format.

    • @ImmersiveLimit
      @ImmersiveLimit  2 года назад +1

      I believe in this one I was just using examples from the COCO dataset. For my own datasets, I don’t create them manually, so I don’t have a tool I can recommend

  • @johannesfein6063
    @johannesfein6063 Год назад

    Hello,
    I want to convert a COCO dataset annotation to a binary mask. The problem
    I have is that after conversion with pycocotools method annToMask nearly all the labels have the same colour and the segmentation network recognises it as one label. Is there a fix to that so that every label has a different colour. Please help

  • @luzukondzuta1523
    @luzukondzuta1523 Год назад

    insightful, thanks. how do i merged the COCO annotations for trucks and persons with your custom annotations for forklifts? The COCO database doesnt have forklifts, so i have annotated the forklifts in images with persons and trucks. but COCO has trucks and person, so i want to merge that.

    • @ImmersiveLimit
      @ImmersiveLimit  Год назад

      You’ll need to write code (I would use Python) to open up the annotation json and add the new items.

  • @leehangwee5566
    @leehangwee5566 4 года назад

    Referring back to your explanation of RLE, would you then be structuring the value of the "counts" key in terms of semantic segmentation rather than instance segmentation? Because in your apple example, the mask cant differentiate the instances of the 4 apples from the mask. Does this apply to the elephants too? (because I could still see the instances on the elephant picture) Thank you!

    • @ImmersiveLimit
      @ImmersiveLimit  4 года назад

      In the official COCO dataset they are not consistent. I think in a perfect case every instance would be separate, but because humans were annotating a lot of images and probably being paid per image completed, we have some cases where they just got lazy. The elephant one is a strange case where they did all but one elephant as a separate instance. Maybe the annotation tool had a limit?

  • @zaheerbeg4810
    @zaheerbeg4810 3 года назад

    nice

  • @minkim8358
    @minkim8358 5 лет назад +1

    Hi , tks for your kind explanation and i have a one question.
    How can i get the segmentation from my own image?? should i use some annotation tools?

    • @ImmersiveLimit
      @ImmersiveLimit  5 лет назад +1

      For now, I recommend checking out github.com/jsbroks/coco-annotator. Unfortunately it takes a lot of time to do this for hundreds of images, but I’m close to launching a Udemy course that can help. Stay tuned for that!

    • @minkim8358
      @minkim8358 5 лет назад +1

      Immersive Limit tks alot i will follow up:)

  • @AleksandrSerov-rn2cn
    @AleksandrSerov-rn2cn 5 лет назад

    In the instances_val2017.json there are links to the images and there are annotations for that images stored separetely. How can i put things toghever like you did in the video? I mean combine raw image and annotation.

    • @ImmersiveLimit
      @ImmersiveLimit  5 лет назад

      You have to look for annotations that match the image id. In my example I did it manually, but you could do it with code. Does that answer your question?

  • @whitekyurem1
    @whitekyurem1 5 лет назад

    Hi. What is the difference between the captions_train2017.json file and the instances_train2017.json file? (or instances_val2017.json file)

    • @ImmersiveLimit
      @ImmersiveLimit  5 лет назад

      Check out this page on my website, I explain the different types of annotations, including captions. www.immersivelimit.com/tutorials/create-coco-annotations-from-scratch

  • @JasperHatilima
    @JasperHatilima 5 лет назад +2

    On the "area" that is found in annotations, I don't understand how it can be a representation of number of pixels that are shaded in the object. I mean, I have noticed decimal values like 702.105 or 2049.9325 when I am expecting integer values for "number of pixels"....how would 0.105 of a pixel be represented?

    • @ImmersiveLimit
      @ImmersiveLimit  5 лет назад +3

      I oversimplified it slightly with my explanation. It's technically the area inside the polygon. Even if polygon vertices were all on exact pixels (in COCO they're not), fractional areas would be possible. For example, if you had a right-triangle with points (0,0), (5, 0), (0, 5), the area would be 12.5. Hope that helps!

    • @JasperHatilima
      @JasperHatilima 5 лет назад +1

      @@ImmersiveLimit thanks a lot. It now makes perfect sense as an actual area than "number of". Your vertices example actually clears my doubts completely! I am following your material closely and I am getting a lot of help! Thanks!

    • @ImmersiveLimit
      @ImmersiveLimit  5 лет назад

      Awesome, glad I could help, Jasper. 🙂

  • @wothetzer
    @wothetzer 3 года назад +1

    Nice video!
    Is there an option to make holes in polygon annoatations?

    • @ImmersiveLimit
      @ImmersiveLimit  3 года назад

      Yes. It’s useful to know that the polygons are converted to image masks during training, so any polygon or set of polygons that fully covers the object will work. A single annotation can have multiple polygons, so if you were annotating a donut 🍩 you could think of it like cutting the donut in half and creating two polygons, one for each side.

    • @wothetzer
      @wothetzer 3 года назад

      @@ImmersiveLimit Thanks for the quick response!
      But I still getting a "false" binary image.
      In my case there is only one annotation (so only one array with coordionates). The annotation has mutiple holes in it.
      When I'm trying to convert the annotation with "annToMask" I still get a binary image where the holes are also filled with the same color as the annotation...

  • @houssamsiyoufi3205
    @houssamsiyoufi3205 3 года назад +1

    If i want to use this course to create my own datasets, will it work on yolact++, or just on Mask R-Cnn

    • @ImmersiveLimit
      @ImmersiveLimit  3 года назад

      I’ve used it for both and it worked well. Same COCO dataset format, just different configuration, obviously. 🙂

  • @peterkidd9829
    @peterkidd9829 2 года назад

    What data labelling software was used for the image masking/annotations before being exported as a json file?

    • @ImmersiveLimit
      @ImmersiveLimit  2 года назад

      You can read the COCO website for info on that, I think it’s on GitHub somewhere.

  • @dimzog
    @dimzog 3 года назад +1

    Hey, thanks for your perfect explanation! I can't figure out if coco formatting supports time series in any way? All I can think of is comparing coco["images"][idx]["date_captured"], coco["images"][idy]["date_captured"], or maybe setting all images in the same time-window (idx) with different "file_names" e.x ..[idx]["file_name_t0"], .., [idx]["file_name_tn"]?

    • @ImmersiveLimit
      @ImmersiveLimit  3 года назад +2

      Thank you for the compliment! I think you are on the right track. The official COCO project doesn’t support it, so you would have to add your own. You could add a frame number or something to the image data in the annotations json and then just consume that with your training algorithm.

    • @dimzog
      @dimzog 3 года назад

      @@ImmersiveLimit Yes, figured so.. Do you have in mind any deep learning time-series repo that utilizes COCO formatting scheme? I could definitely use something ready like Detectron2, but RNN based. Again, thanks!

    • @ImmersiveLimit
      @ImmersiveLimit  3 года назад +1

      I haven't looked for any, sorry!

    • @dimzog
      @dimzog 3 года назад +1

      @@ImmersiveLimit Alright, farewell!

  • @silentstorm4405
    @silentstorm4405 Год назад +1

    Ok so how do we actually convert a folder of images into these json files?

  • @chuan-hsienho7538
    @chuan-hsienho7538 4 года назад

    Excellent video!
    "The course is live! You can get a deal on the course and support me with this coupon"
    I can't find any coupon! Please comments.
    Does it include only annotation for segmentation and bounding box or does it include also keypoint and panoptic?
    Thanks!

    • @ImmersiveLimit
      @ImmersiveLimit  4 года назад

      Sorry I didn’t see this right away, but it looks like you already bought the course. Thank you!

  • @alirezag6603
    @alirezag6603 4 года назад +1

    does id in category start from 0 or 1, do we have background category too?

    • @ImmersiveLimit
      @ImmersiveLimit  4 года назад +1

      I believe it starts at 1 because 0 is “background”

  • @chukypedro818
    @chukypedro818 4 года назад

    Is there any difference between annotating my images for object-detection-api using Labellmg . and using the COCO Dataset JSON Format.
    I saw on your blog post, how you converted foreground and background images to 11k Training and validation data-set. I really like to achieve the same result. Can I convert the JSON file to xml..?

    • @ImmersiveLimit
      @ImmersiveLimit  4 года назад

      I’m not familiar with the other format, but you could most likely write a custom Python script to do the conversion or find one on GitHub

  • @angelgabrielortiz-rodrigue2937
    @angelgabrielortiz-rodrigue2937 2 года назад

    Awesome video. I was wondering about the coco_url. My understanding is that this is the image location where the model should look in. What if I want to create the annotations for a dataset I have on my personal computer, how would the model know where to look for the image? How would I write that in the annotations?

    • @ImmersiveLimit
      @ImmersiveLimit  2 года назад +1

      In that case just leave coco_url out. None of the neural net libraries I’ve used actually use it. I think it’s mostly for reference.

    • @angelgabrielortiz-rodrigue2937
      @angelgabrielortiz-rodrigue2937 2 года назад +1

      @@ImmersiveLimit okay, I understand. But still, how would the model know where to look for images that I have on my computer? Is there a way to specify that in the coco format?

    • @ImmersiveLimit
      @ImmersiveLimit  2 года назад

      You’ll have to read the code to see where it loads the images. It probably has a config variable the points to a local directory.

    • @angelgabrielortiz-rodrigue2937
      @angelgabrielortiz-rodrigue2937 2 года назад

      @@ImmersiveLimit I see. So that would be more dependent on the model training process than the dataset creation process?

    • @ImmersiveLimit
      @ImmersiveLimit  2 года назад

      Correct. The trainer is what looks for images, and in every case I’ve seen, you have to download the dataset to your computer. It would be too slow to fetch them all from the web.

  • @asmabenbrahem6
    @asmabenbrahem6 3 года назад

    Is this the required format for the annotations to work with matterport/mask R-CNN repo ? I think the only accepted format for this repository is the one got with vgg annotator where there are regions, regions_attributes in the json file annotations ... maybe it is a stupid question but please can someone answer me ?

    • @ImmersiveLimit
      @ImmersiveLimit  3 года назад

      Yes it is. You can check out the tutorials on our website, we have a couple that may help. www.immersivelimit.com

    • @asmabenbrahem6
      @asmabenbrahem6 3 года назад

      @@ImmersiveLimit Thak you very much, I found the right tutorial for my task.

  • @yasminebelhadj9359
    @yasminebelhadj9359 3 года назад

    Hello, can you please make a video about textvqa ?

  • @siddharthakosti
    @siddharthakosti Год назад

    hey brother loved this explanation, can you tell me how you converted this coco data format into json data format?

    • @ImmersiveLimit
      @ImmersiveLimit  Год назад

      Coco format is already in json, I didn’t convert it

    • @siddharthakosti
      @siddharthakosti Год назад

      @@ImmersiveLimit sorry my question was wrong, the question is when I am converting coco data using json.dumps(data, indent=4), the segmentation lists are coming in vertical format, and taking a lot of space. How you converted them to horizontal

  • @camilleb7385
    @camilleb7385 3 года назад

    Hello, thank you for your video. I have one question about the segmentation part. If we have an object with a hole like a cup handle for example. How to specify that the hole need to be empty ? I mean, we will have two contours and the hole contour area is empty. How represent this case?
    (Hope the question is clear, sorry English is not my language)

    • @ImmersiveLimit
      @ImmersiveLimit  3 года назад +1

      All contours are eventually converted to pixel masks, so you can either do RLE format or you could have two contours for an object, for example one of the cup, one of the handle. The result should be the same.

    • @camilleb7385
      @camilleb7385 3 года назад

      @@ImmersiveLimit thank you for your answer I will try

  • @AhmedSalimDev
    @AhmedSalimDev 3 года назад +1

    how we calc. the segmentation area please and many thanks for ur effort

    • @ImmersiveLimit
      @ImmersiveLimit  3 года назад

      Count the number of pixels in the segmentation.

  • @ramzanali9999
    @ramzanali9999 4 года назад

    which software or application had you used in this tutorial to perform annotations?

    • @ImmersiveLimit
      @ImmersiveLimit  4 года назад

      I didn't create these annotations. You might want to look at this tool though: github.com/jsbroks/coco-annotator

  • @BhuvaneshArvindK
    @BhuvaneshArvindK 2 месяца назад

    sir what is dice coefficient for this?

  • @alirezag6603
    @alirezag6603 4 года назад

    what does id in annotation represent? should the id in annotation
    be unique among the dataset?

  • @viablevikas2750
    @viablevikas2750 2 года назад

    I have annotations in the text files. How should I proceed?

  • @misunp79
    @misunp79 3 года назад +1

    First of all, GREAT job!! Thank you sooooooooooo much. BTW, is it possible to share the sample_annotations.json file with us?

    • @ImmersiveLimit
      @ImmersiveLimit  3 года назад

      I managed to dig it up. I thought I might have lost it! gist.github.com/akTwelve/c7039506130be7c0ad340e9e862b78d9

  • @kalaana_yt8864
    @kalaana_yt8864 Год назад

    hey I unable to download coco dataset from the website.. please help me

  • @pseka_f5557
    @pseka_f5557 3 года назад

    Damn Bet I can finally code a labeled sorter

  • @AnushaCM
    @AnushaCM 3 года назад

    Can I use this dataset for Object classification? How can I use it if there are two object categories?

    • @ImmersiveLimit
      @ImmersiveLimit  3 года назад +1

      No, you cannot use it for classification.

  • @watcher7351
    @watcher7351 4 года назад

    hı, ı have a dataset and annotion file, json format. json file include object location. but ı dont know import my deep learning module. ı can import images but cant label. help please

    • @ImmersiveLimit
      @ImmersiveLimit  4 года назад

      Hi, have you taken any courses on deep learning and computer vision?

    • @watcher7351
      @watcher7351 4 года назад

      @@ImmersiveLimit yes all courses use coco,mnist,cifar etc. dataset, or use tensorflow object detection apı. but ı want to prepare own dataset and use my model.
      but my label files include box location etc. ı cant import. ı took your udemy course, can ı use your course for this problem

  • @joe_hoeller_chicago
    @joe_hoeller_chicago 4 года назад

    Does annotations[category_id] coincide with the coco.names file?

    • @ImmersiveLimit
      @ImmersiveLimit  4 года назад

      I'm not sure what the coco.names file is, but category_id should be consistent with whatever is in the categories section

  • @houssamsiyoufi3205
    @houssamsiyoufi3205 3 года назад

    Can we find out the segmentation of the test images annotation (meaning the output coordinates)?

    • @ImmersiveLimit
      @ImmersiveLimit  3 года назад

      Sorry, I don’t understand your question. Can you clarify?

    • @houssamsiyoufi3205
      @houssamsiyoufi3205 3 года назад

      @@ImmersiveLimit You were able to get the segmentation of the instances of an image in the code (x and y) regardless if they are an iscrowd (sequence coding method or not), but these are the train/validation images annotations, can we find out the segmentation or the instance coordinates ( x and y points) of the test images outputs of regions of interest lets us say a car

    • @houssamsiyoufi3205
      @houssamsiyoufi3205 3 года назад

      Let us say we tested our image, can we find out the x and y coordinates of the output or not

    • @ImmersiveLimit
      @ImmersiveLimit  3 года назад

      Oh, I think I understand. So, the output of your neural net (Mask R CNN for example), would give you a pixel mask that will have imperfections. You would want to use scikit-image and shapely to create your list of points. I do something similar in create_coco_annotations() here github.com/akTwelve/cocosynth/blob/master/python/coco_json_utils.py, however I don’t have to clean up the mask first. You’d have to figure that bit out for yourself.

    • @houssamsiyoufi3205
      @houssamsiyoufi3205 3 года назад

      @@ImmersiveLimit Thank you a lot, so now I can extract the test images coordinates (x and y points ) and use the segmentation.

  • @hackercop
    @hackercop 3 года назад

    Please zoom in on vscode, otherwise great video you tought me a lot

  • @thechallenge293
    @thechallenge293 3 года назад

    Hi pleas how can i resize an already labeled dataset? Plus what is area?

    • @ImmersiveLimit
      @ImmersiveLimit  3 года назад

      Do you just want the same categories but less images of each? If so, you just have to load the JSON with Python (like I do) and remove some of the images from the list. Area is how many pixels are in the mask.

    • @thechallenge293
      @thechallenge293 3 года назад +1

      @@ImmersiveLimit No i mean resize each image to lets say 550×550. If i resize the images i have to change the points inside the annotation.json file. Is there a script or tool to do this? Anyway in the end i manually created a script for resizing labelme data since i used labelme for the annotations. The i converted using labelme2coco.py script. But a script made for resizing coco style annotations and images would be nice.

    • @ImmersiveLimit
      @ImmersiveLimit  3 года назад

      Got it, I think typically resizing is done at training or inference time automatically rather than modifying the dataset, but thanks for the suggestion.

    • @thechallenge293
      @thechallenge293 3 года назад

      @@ImmersiveLimit I was trying to use yolact in my case. I am new to yolact so i had to do it manually. Make a tutorial aboit yolact if u can! Thanks you

    • @ImmersiveLimit
      @ImmersiveLimit  3 года назад

      We do have two YOLACT tutorials on our website, just not on RUclips

  • @joe_hoeller_chicago
    @joe_hoeller_chicago 4 года назад

    Have you converted COCO-JSON to Darknet-YOLO format?

    • @ImmersiveLimit
      @ImmersiveLimit  4 года назад

      I tried at one point, but it was so much work to get Darknet-YOLO set up for training that I gave up.

    • @tiancailin
      @tiancailin 4 года назад

      github.com/qqwweee/keras-yolo3 . You can run ' python coco_annotation.py ' to convert it

  • @bharadwajsiddela4137
    @bharadwajsiddela4137 4 года назад

    How to download COCO DATASET for windows

  • @nikhillingam4630
    @nikhillingam4630 4 года назад

    where are captions with respect to images ?

    • @ImmersiveLimit
      @ImmersiveLimit  4 года назад

      They are in a different JSON file that includes an image id, caption id, and caption.

  • @LouisChiaki
    @LouisChiaki Год назад

    I am surprised that bicycles in COCO dataset are annotated in such a rough way....

    • @ImmersiveLimit
      @ImmersiveLimit  Год назад

      They didn’t pay the annotators enough for that level of detail is my guess

  • @ChitokuYato
    @ChitokuYato 4 года назад +1

    Thank you for the Jupyter notebook!
    I made a fork of your gist and added the keypoint visualization.
    github.com/chitoku/coco_viewer/blob/master/COCO_Image_Viewer.ipynb

    • @ImmersiveLimit
      @ImmersiveLimit  4 года назад

      That’s awesome, nice work! I’ll take a look when I get a chance

  • @zod1ack_csgo_clips425
    @zod1ack_csgo_clips425 4 года назад

    Can you explain how to apply 'ignore_regions' to our image?

    • @ImmersiveLimit
      @ImmersiveLimit  4 года назад

      I'm not sure what you mean, is that part of the COCO dataset?

  • @sportsclips650
    @sportsclips650 2 года назад

    آپ کے ایک کلک سے میری غریب زندگی بدل سکتی ہے اللہ پاک آپ کو دنیا کی تمام خوشیان عطا فرمائے اور آپ کے والدین کا سایہ تا دیر تک آپ پر قائم رہے آمین ثمّ آمین 🙏🙏🙏🙏mgoos

  • @Lion_McLionhead
    @Lion_McLionhead 2 года назад

    Too bad these guys can't create a smaller, unique YOLO model for each category. Most of us just want to detect humans.

    • @ImmersiveLimit
      @ImmersiveLimit  2 года назад

      I see the value for sure, but I’m not sure it would work as well with a smaller model.