Machine Learning with Synthetic Data |

Поделиться
HTML-код
  • Опубликовано: 4 сен 2024

Комментарии • 117

  • @hongdroid94
    @hongdroid94 4 года назад +4

    Last year, I saw your augmented reality lecture and found out that you only make very useful videos! I'm currently a RUclipsr who makes Android classes, and I'm still interested in your videos! Thank you for always making good videos.

  • @geri4367
    @geri4367 4 года назад +37

    Thanks for reminding me why I dropped ML and focused 100% on gamedev (:

  • @marcojoao
    @marcojoao 4 года назад +4

    You can improve using enhance image technique, and transfer learning to improve accuracy. The EIT will just stretch, rotate, drift and flip the images, and the transfer learning will help you to go at 95% accuracy

  • @UnofficialEngineering
    @UnofficialEngineering 4 года назад +2

    Fact: there is not another RUclipsr out there as innovative as Hallberg.

  • @pecke86
    @pecke86 3 года назад +1

    sick man ! i was wondering about this process for a very long time. You just created the gate of the metaverse !

  • @surajitsaikia1653
    @surajitsaikia1653 3 года назад +3

    Very amazing work. You can also rig or animate the character to generate a larger dataset. For instance, while animating change the camera views and collect the data

  • @Augmented_AI
    @Augmented_AI 4 года назад +3

    Well done bro! Great work!

    • @MatthewHallberg
      @MatthewHallberg  4 года назад

      Thanks man!! It finally worked...somewhat lol.

  • @ilirvg
    @ilirvg 4 года назад +1

    Amazing!!!
    Never seen your tutorials before and I usually do not Subscribe on RUclips but you just got a new Subscribe (and a fan)

  • @richardbeare11
    @richardbeare11 3 года назад +1

    Lol I had this exact problem with the bounds! I used basically the exact* same approach you did with the mesh sweep.
    Also, I think that's sweet how you tuck your laptop in.

  • @magefront1485
    @magefront1485 3 года назад +1

    You can use colab which has a GPU better than 2080Ti, just put all files in google drive. Colab comes with TensorFlow installed by default.
    Synthetic data on machine learning is definitely doable, there are some papers using blender to generate these training datasets.
    To address the poor results, by default all convolution-based networks tend to learn the texture of the image instead of the shape, maybe unity's texture is not that photo-realistic. It's very difficult to train a model from scratch, the common approach is to do a transfer learning, where you take a pre-trained model of a large dataset like imagenet, then unfreeze the top and bottom layer, tuning on your data.
    With 5000 images to train from scratch, you won't get good results unless it's a super simple classification, like 28x28 handwritten digits.
    Since it's synthetic data, might be better to just do a pixel-wise label instead of a bounding box label.

  • @minhdungdo4541
    @minhdungdo4541 2 года назад +1

    Amazing Amazing project. With outstanding informative video, I can't hardly thank you enough but still. Thank you very much for the helpful video. Would love to try this.

  • @iliesouldmenouer4976
    @iliesouldmenouer4976 4 года назад +2

    " i have no idea what that is, And i don't even want to know " -
    MatthewHallberg 2:30
    .
    .
    .
    .
    Much love and support for you bro , such a great content

  • @cheesiangleow4782
    @cheesiangleow4782 4 года назад +4

    Great video ! What apps did you use to scan object to 3d models?!

  • @wanderstudi
    @wanderstudi 4 года назад +2

    Idea: maybe try the 3d scan in a completely diffuse lighting setup. So you do not have hard shadows on the textures of your 3d scans. I think that could mess up the recognition of the shapes. Or did you try that already? Anyway , apart from the bulbosaur(?) (not a pokemon expert here) recognition seems quite good already.

    • @MatthewHallberg
      @MatthewHallberg  4 года назад

      I did try that I just didn’t film that part I still did get some shadows though it might have helped if I had a box with the correct lighting setup and then I could spin the object inside.

  • @DInfinity3
    @DInfinity3 7 месяцев назад

    Super!!

  • @denzilstudios7072
    @denzilstudios7072 2 года назад

    that kickflip

  • @johariawang7713
    @johariawang7713 4 года назад +2

    i use vuforia app to scan the 3d object and use it to create augmented reality app just like this one... i think it provide better experience and faster workflow..

    • @MatthewHallberg
      @MatthewHallberg  4 года назад +1

      It definitely will but my goal here was to create a generalized tracker that recognized more than one style of the same object... I just failed lol.

  • @AnkitSingh-wq2rk
    @AnkitSingh-wq2rk 4 года назад

    Hey Matthew I had a question is there any possibility for hand detection in unity ? (without that paid plugin of OpenCV) I had a project in my mind where you could interact with models shown by vuforia's image target using bounding box coordinates and 3d model coordinates ... i have done some workaround by scanning the hand through openCv in an external script and then sending the coordinates as packets over UDP to unity every frame .... but the main problem is it is not feasible for mobile phones :(

  • @jackcottonbrown
    @jackcottonbrown 2 года назад

    That is commitment. A+ for effort.

  • @abudriaz9678
    @abudriaz9678 4 года назад +2

    Always loved watching your videos❤

  • @bram_adams
    @bram_adams 4 года назад +1

    Your hard drive must be massive to hold all that data!

    • @MatthewHallberg
      @MatthewHallberg  4 года назад

      Haha MASSIVE. And you know what they say about guys with big hard drives...

  • @_wise_one
    @_wise_one 4 года назад +1

    You could have used Google colab, they give 12gb ram and GPU/TPU for free. You can keep it running even if you close the browser or shut down the computer

  • @eessaaabrahams9124
    @eessaaabrahams9124 3 года назад +1

    should have used a bunch of different pikachu and bulbasaur images instead of the same ones, then your algorithm would learn that they come in different forms, shapes and sizes

  • @mrsbtheo.a.p2635
    @mrsbtheo.a.p2635 4 года назад

    Impressed. Keep up the good work. Appreciate your insight.

  • @hamzzashaffi
    @hamzzashaffi 4 года назад +2

    Great video as usual!
    I'm also planning to start ML sooner :)

    • @MatthewHallberg
      @MatthewHallberg  4 года назад +1

      thank you! Yeah do it so you can teach me lol.

    • @hamzzashaffi
      @hamzzashaffi 4 года назад +1

      @@MatthewHallberg guess what? I haven't started it yet lol.

    • @MatthewHallberg
      @MatthewHallberg  4 года назад +1

      Hamza Shafi haha me either

    • @hamzzashaffi
      @hamzzashaffi 4 года назад

      @@MatthewHallberg ayy bro

    • @tharaniv6267
      @tharaniv6267 3 года назад

      @@MatthewHallberg Hey what app u used for photogammetry(scanning 3d model)

  • @saifking7580
    @saifking7580 3 года назад +2

    What is the name of mobile 3d images app

  • @jerinpwilson5288
    @jerinpwilson5288 4 года назад +4

    @MathewHallberg Hey nice project.. what app did you use for photogrammetry? To make the model...please reply.

    • @tharaniv6267
      @tharaniv6267 3 года назад

      I also have this doubt

    • @abhishekgoyaldev
      @abhishekgoyaldev 3 года назад +1

      @@tharaniv6267 Have you even watched the full video? 11:16

    • @tharaniv6267
      @tharaniv6267 3 года назад

      @@abhishekgoyaldev thanks

  • @DMTravelCinematography
    @DMTravelCinematography 2 года назад

    How do you draw the bounding box around the object it is detecting in-house object? I tried using renderer.bound but it does not work. Do you have samples?

  • @jonalex
    @jonalex 4 года назад +1

    Was recently asked to estimate the work need to make a mobile app that could detect specific (known) objects. I remembered watching this video a while back.
    Can some of this be combined with your previous ML video?
    Would more computer power be a partial solution to this?
    Would more lighting variations enrich your data set?
    Thanks so much for this video. You're my go-to guy for all those off the wall projects.

  • @rizzbod
    @rizzbod Год назад

    sooo coool

  • @xXMaDGaMeR
    @xXMaDGaMeR Год назад

    very cool vid

  • @karandeepdps1
    @karandeepdps1 4 года назад +1

    You need to remove the green bbox outline from images and then it will work.

  • @rogueyoshi
    @rogueyoshi 4 года назад +1

    you should join the Two Minute Papers Discord!

  • @Pathorian
    @Pathorian 2 года назад +1

    Wondering how hard this would be using 2D Data instead of 3D Data

  • @jeeteshsingh209
    @jeeteshsingh209 4 года назад +1

    Kudos to ur efforts man! 💯💯

    • @MatthewHallberg
      @MatthewHallberg  4 года назад +1

      Thank you I tried lol, almost didn't make this video for sure.

  • @DrRehanZafar
    @DrRehanZafar 2 года назад

    Great video

  • @Kaushik-eo4ll
    @Kaushik-eo4ll 4 года назад +1

    First like and first comment broo ❤️😘

  • @Jack-oq7rg
    @Jack-oq7rg 3 года назад

    is it possible to feed ReferenceImageLibrary remotely ? like download the images from remote server with the 3D prefabs to place

  • @utsavgupta4630
    @utsavgupta4630 4 года назад

    Literally loved this .....😍😍

  • @manojmadushanka9356
    @manojmadushanka9356 4 года назад

    hard fan from sri lanka you AI man

  • @maknien
    @maknien 4 года назад +2

    Did you have the green borders in all your rendered training images (7:57)?

    • @MatthewHallberg
      @MatthewHallberg  4 года назад +1

      That was just for the video I didn’t actually train like that lol. I am dumb so that’s a valid question but not that dumb haha.

    • @maknien
      @maknien 4 года назад +2

      @@MatthewHallberg Good :D I thought for a while should I even ask... But for the geometry randomizing, you should definitely check Houdini. I've been using that for all kinds of synthetic ML data stuff.

    • @MatthewHallberg
      @MatthewHallberg  4 года назад

      @@maknien INTERESTING, checking that out now thanks.

  • @evelynjunco4574
    @evelynjunco4574 4 года назад +1

    Hi Matthew, I have been following your tutorials for quite a while. I’m trying to do object detection of real objects with an iPhone. I know iPhones, Unity and window don’t always work well but still want to give it a try. I can use photogrammetry for real objects but I’m not sure on how to do machine training once I have these images. Any suggestions would be appreciated!

  • @swannschilling474
    @swannschilling474 3 года назад

    Awesome!!

  • @argmentum22
    @argmentum22 2 года назад

    prototype with 2k pictures... make sure the item you want to identify is 100% in your frame in your training pics. Don't put two items in your picture - the AI will potentially try to think there's supposed to be two; like lenses in glasses. check the the type of tensor flow model is right for what your trying to do.. some are very quick but have a lower success rate.

  • @PlasmaSabre
    @PlasmaSabre 3 года назад

    Have you heard of fractional factorial testing? Might allow you to change multiple variables at once and run far fewer tests.

  • @ashokkillo
    @ashokkillo 4 года назад

    nice video with details.. thanks for showing.. whats ur laptop specs?

  • @elidorvarosi9643
    @elidorvarosi9643 Год назад

    Would reccomend to just use some cloud instance instead of a laptop GPU as it will usually start to throttle after a while. Also, most of those frameworks works better on Linux due to the fact that virtually all the server do run on whatever linux distro was chosen by the user(usually Ubuntu or Debian like distros, sometimes Red-Hat based etc)

  • @clifflin7149
    @clifflin7149 4 года назад

    great video

  • @cintianakano5339
    @cintianakano5339 4 года назад

    Good job!

  • @dudenarima2528
    @dudenarima2528 3 года назад

    you should use for loop inside for loop inside for loop... instead of random
    :D idk anything about ml too

  • @hosammohamed7107
    @hosammohamed7107 4 года назад

    great stuff bro, i was searching to make something like wanna kicks app. Do you have any idea how to track the feet and put the shoes on legs like that? i know it's AR but i don't have any insights about how to do it :(

  • @samvidjhaveri634
    @samvidjhaveri634 4 года назад

    Just FYI, Google has a new API named AutoML. It is much easier than their old Vision API which is hard to use. But it cost $$$$.

  • @brunoomardorivalgutierrez1141
    @brunoomardorivalgutierrez1141 4 года назад

    Hi Mattehw, How can I do to enter a augmented reality house but first put the house in a specific place and then scale it?

  • @ZiyueZhang0924
    @ZiyueZhang0924 4 года назад

    you are so great

  • @ko-Daegu
    @ko-Daegu 4 года назад +1

    Am so confused when can I start learning about ML ...
    Do I need like probability theory and statistics class first after that intro to AI and read few books after that I can jump to ML
    What introductory courses would you recommend
    Side note : I’m coming from 2+ years with java and about few months with python and am learning flask now (just for the heck of it )
    Do you think am ready to start doing stuff with tenser-flow or scikit maybe ???
    And which one to start with ??

    • @MatthewHallberg
      @MatthewHallberg  4 года назад +1

      I literally have no idea on that one I never took any courses just started playing with tensorflow and following tutorials which is probably why I have no idea what I am doing in this video.

    • @annabelgroenenberg9448
      @annabelgroenenberg9448 4 года назад

      You have a very good machine learning course on youube by Andrew Ng. It's the ultimate starting point. You don't have to know a lot of probability stuff but know the basics like false positive, false negative etc...

  • @waleedough
    @waleedough 2 года назад

    I am new to the package of perception from Unity and my experience it's just for game designing, please if anyone has experience in such thing like this, I need help

  • @vasusraj
    @vasusraj 4 года назад

    How to place 3d model on detected object in unity3d with tensorflow.

  • @robosergTV
    @robosergTV 2 года назад

    dude, use Unity Perception package..........

  • @djone7672
    @djone7672 3 года назад

    Can someone please explain what type of deep learning method this process uses? is it CNN?

  • @waleedough
    @waleedough 2 года назад

    I wanna do keypoints annotations, anyone who could help me on that please???

  • @aashutoshdabhade4325
    @aashutoshdabhade4325 3 года назад

    one like just for ur Efforts! I can feel the pain.

  • @Caio-Mendez
    @Caio-Mendez 4 года назад

    I need help with the coronavirus ar app

  • @AllanPichardo
    @AllanPichardo 3 года назад

    Your model was overfit because you only had a few thousand images. Next time, print the summary from your model and take notice of the total number of parameters in the model. Then try to get 2 to 3 times that many image samples. You should get decent results then.

  • @samgrogan8815
    @samgrogan8815 4 года назад +1

    I think you overfitted to your training data which may be why the first model didn't work. Cant be sure though.

    • @MatthewHallberg
      @MatthewHallberg  4 года назад

      I was definitely getting that vibe but I don’t know how to tell for sure.

    • @annabelgroenenberg9448
      @annabelgroenenberg9448 4 года назад +1

      @@MatthewHallberg Look at the loss function via tensorboard. If your training data loss curve jumps to a high accuracy while your validation curve is way behind, you're probably overfitting

    • @samgrogan8815
      @samgrogan8815 4 года назад

      @@MatthewHallberg Really depends on your setup. But I think with tensorflow there is a callback in the keras library for early stopping and you just tell it to stop if its not seeing inprovement after so many steps and that can help avoid over fitting.

  • @Unpopular_Facts
    @Unpopular_Facts 3 года назад

    someone knows of any tutorial

  • @manaskumar2544
    @manaskumar2544 4 года назад

    You don't neet 20000 epochs to train the model with such less data,.. it's over fits the model and the model can not predict all the labels with equally accuracy!!!, You have to train untill the loss is stable,.. thats it!! When your detecting every pichachu in the world,.. you need different types of pichachu with different background images,..... Don't worry about the huge data.., all you need is 1000 images each!!,.. and epochs not more than 500

  • @tronpig
    @tronpig 4 года назад

    Add jarvis to the play store!!!!!

  • @camdenparsons5114
    @camdenparsons5114 3 года назад +1

    two words: transfer learning :0

    • @MatthewHallberg
      @MatthewHallberg  3 года назад

      Yeah this is transfer learning lol did you watch the video?

    • @camdenparsons5114
      @camdenparsons5114 3 года назад

      @@MatthewHallberg yeah haha I watched it. I was sugesting that you use a pretrain model and retrain the last or last few layers to save many hours on training. thats how the watson and google object detection APIs work

  • @sahilshaikh3454
    @sahilshaikh3454 4 года назад +1

    Why... 😑

  • @tylersnard
    @tylersnard 3 года назад

    You have to use a GPU to speed it up. Try training on Google Colab. Also, tensorflow sucks, use PyTorch :)

  • @stephanverbeeck
    @stephanverbeeck 4 года назад +1

    Great vid, too bad you have no real computer :-)

  • @joemoulton1823
    @joemoulton1823 4 года назад +1

    Lol "It was so hard to figure out how to convert bounding boxes from world space to screen space...". Really?

    • @MatthewHallberg
      @MatthewHallberg  4 года назад

      Joseph Moulton I wish it was that simple lol Unity has a function for that.

    • @joemoulton1823
      @joemoulton1823 4 года назад +1

      @@MatthewHallberg Instead why don't you manipulate the camera to look at the model and then window the frame? Then take the snapshot

    • @MatthewHallberg
      @MatthewHallberg  4 года назад

      That’s interesting never thought of that, I guess the only problem would be getting multiple objects in the same image cause that helped the model a lot.

  • @haraldgundersen7303
    @haraldgundersen7303 4 года назад

    Hope some loaded dude gives u a quantum computer... You might even find a way to prevent future AI to wipe out humanity...