Apple visionOS Object Tracking Is Here! BUT DOES It Work Well?

Поделиться
HTML-код
  • Опубликовано: 12 янв 2025

Комментарии • 95

  • @dilmerv
    @dilmerv  6 месяцев назад +3

    💻 You can also find the demo shown on today’s video via GitHub from github.com/dilmerv/VisionOSObjectTrackingDemo
    I also recommend going over all the resources available from the video description.

  • @user-vany
    @user-vany 6 месяцев назад +8

    This is a very cool idea, it was interesting to make something like the face masks that we applied. In real time.

    • @dilmerv
      @dilmerv  6 месяцев назад +2

      Sweet, I can’t wait to see what you create! Thanks for your comment.

  • @bencrystalofficial
    @bencrystalofficial 6 месяцев назад +5

    This is incredible! Can't wait for 2.0 to drop more formally :)

    • @dilmerv
      @dilmerv  6 месяцев назад

      Thanks for your feedback, I agree this is going to be huge and super helpful!

  • @anoethen
    @anoethen 14 дней назад

    Thanks for the great video. That was exactly what I was looking for.🤩

    • @dilmerv
      @dilmerv  13 дней назад +1

      I am very glad you found it helpful, thanks man!

  • @renatos.9923
    @renatos.9923 6 месяцев назад +5

    Thanks for these detailed instructions! 😀

    • @dilmerv
      @dilmerv  6 месяцев назад

      You are very welcome! Thanks for your feedback.

  • @yifansun9953
    @yifansun9953 3 месяца назад +2

    Cool Video! I just tried to use a CAD model to train the reference object. However, it didn’t work. I am curious about how to config the usdz material for a CAD model if we do not want to use the scanned version which is not as accurate as the CAD😅😅

    • @dilmerv
      @dilmerv  3 месяца назад +1

      I don't believe CAD files are supported, Apple normally only supports USDZ files only. Take a look at this suggestion as it seems to may be an option: forums.developer.apple.com/forums/thread/745911

  • @xygu-u4o
    @xygu-u4o 3 месяца назад

    Very cool feature. I have noticed that there is another newer session called SpatialTrackingSession which can provide tracked object's world position, can the one you used also output some similar data?

  • @PhilipHerlitz
    @PhilipHerlitz 2 месяца назад

    i'd love to see a detailed video on Unity Polyspatial and how to use Dynamic Lights and Shadows on AVP builds. Its been a super headache, with no clear direction. Especially for bounded volume experiences.

    • @dilmerv
      @dilmerv  Месяц назад

      Thank you for your feedback! I am about to make a new video and this will be very useful.

    • @PhilipHerlitz
      @PhilipHerlitz Месяц назад

      @@dilmerv We finally got shadows to work, but litterally had to create an empty project, and use all the defaul settings, and the sample lighting scene, just to do a build test and see if it was even possible. From there, were were able to make small adjustments. We found we liked Hard shadows better than soft shadows.

  • @kentarotakahira9514
    @kentarotakahira9514 3 месяца назад

    Thank you, Dilmer. It's very easy to follow. Is it possible to combine this object-tracking feature with the Unity AR project? (Like Ar foundation etc.)

    • @dilmerv
      @dilmerv  3 месяца назад

      That’s a great question, I don’t believe it is available yet but I know the Unity team always stays up to date with moving native functionality to AR Foundation which in this case I agree it should be available with Unity PolySpatial.
      I will keep an eye on it and once they add support then will make a new video.
      Thanks for your comment!

  • @erikmeilner8928
    @erikmeilner8928 5 месяцев назад +1

    @dilmerv thank you for your great job. I have a problem. Could you please help me? I want to attach a reality view with a usdz & an attachment(SwiftUI view) to the object tracking anchor how to do this?

    • @dilmerv
      @dilmerv  5 месяцев назад

      That's a pretty good question! let me look into it and see if I can put something together.

    • @erikmeilner8928
      @erikmeilner8928 5 месяцев назад

      @@dilmerv This would be awesome

  • @vmjgcm
    @vmjgcm 2 месяца назад +1

    Hello Dilmer! Thank you for the video :) I'm studying for VisionOS object tracking these days, and your video is being so much helpful. I already finished ML for an object and put the 'referanceobject' in the 'Referance Objcts' folder. But when I ran the project on Vision Pro, it was saying that, "Start Tracking 0 Object(s)". Am I missing some steps?

    • @dilmerv
      @dilmerv  2 месяца назад +1

      I remember having such problems when the app didn’t have the correct permissions. Did you get a popup when you launch the app to give the app permission for spatial understanding? Also thanks for your feedback on the video!

    • @vmjgcm
      @vmjgcm 2 месяца назад

      @@dilmerv Yes, I did! I solved the problem I guess. Thanks for your answer! Ah, one more questions please, is there ways making the tracking speed faster? Or need we wait for next visionOS, ARKit version updates?

  • @juliancastiblanco4924
    @juliancastiblanco4924 4 месяца назад

    Great video, you never disappoint. I have a question tho. How do you access Guided Capture iOS on your phone? The link in description is unavailable

    • @dilmerv
      @dilmerv  4 месяца назад +1

      I believe Apple updated it, here is the new link: developer.apple.com/documentation/realitykit/scanning-objects-using-object-capture but to be honest I need to test it, however feel free to check it out. Thanks for your feedback on the video!

    • @juliancastiblanco4924
      @juliancastiblanco4924 4 месяца назад

      @@dilmerv Thank you for answering me so quickly. You seem like a very kind person 😄

  • @paulalexandrupop3709
    @paulalexandrupop3709 5 месяцев назад +1

    What's the limit on the number of reference objects that you can track?
    Couldn't find that anywhere in the docs.

    • @dilmerv
      @dilmerv  5 месяцев назад

      I don’t believe there is a limit imposed but a device limitation in terms of performance. The files are quite large so if you were to add 10 or more I could see the binary be more than 1gb so that should give you an estimate of what’s possible.

  • @ZiyueZhang0924
    @ZiyueZhang0924 Месяц назад +1

    Hey Dilmer, is the object tracking feature only supported in apple native development, and not supported by PolySpatial?

    • @dilmerv
      @dilmerv  Месяц назад

      Hey thanks for your comment. As of now yes, but I know Unity always tries to stay in parity with platforms features, so perhaps it will be added in the near future. I recommend posting in the Unity forums to see if that’s in the roadmap. Also, another way could be to add a native plugin, here is an example of one docs.unity3d.com/Packages/com.unity.polyspatial.visionos@2.0/manual/InteropWithSwiftUI.html

    • @ZiyueZhang0924
      @ZiyueZhang0924 Месяц назад

      @@dilmerv Thanks for your reply !

  • @justinchang4799
    @justinchang4799 5 месяцев назад +1

    Does the tracking generalize well or is it specific to the physical object you scanned?
    Specifically I'm wondering if you can scan 1 car (compact sedan) but have the app work for all cars (compact sedans)?

    • @dilmerv
      @dilmerv  5 месяцев назад +1

      That’s a great question, currently the reason it takes so long is mainly because the machine learning algorithm tries to track the exact object in question, and it is pretty good about it. I don’t believe you could do what you are asking unless Apple provides additional data, such as how similar it could be on some type of percentage through their SDK.

  • @bencrystalofficial
    @bencrystalofficial 5 месяцев назад +1

    Do you happen to know if I update to the Beta if there are errors, is it possible to downgrade by factory resetting the device? Or will it factory reset back to the same beta?

    • @dilmerv
      @dilmerv  5 месяцев назад

      Factory reset will reset it back to Beta with a clean version, you could download the current prod version from Apple dev portal and use Apple Configurator app to restore it back to the prod version. Great question!

  • @kkkaran786
    @kkkaran786 4 месяца назад

    Can we use this as native library and plug it with Unity? to send the bounding box pose?

    • @dilmerv
      @dilmerv  4 месяца назад

      I don’t believe that’s supported right now from the Unity side of things. But I could see that being integrated in the future within AR Foundation / object tracking features. For now, I would stick with native Xcode.

  • @anirudhnv7389
    @anirudhnv7389 6 месяцев назад +1

    What if I already have the 3D model of the object I want to track as an fbx or cad model. Can I use that to track the real world object?

    • @dilmerv
      @dilmerv  6 месяцев назад +1

      If the model is very very high quality & photorealistic then yes it should work. Also, right now I believe you can only train with USDZ files so if you convert it to that format then things should work.

  • @GuilhermeGouveia-nx9rp
    @GuilhermeGouveia-nx9rp 6 месяцев назад

    Hi Dilmer
    First of all, thanks for your content !
    Do you think that The Object Tracking will be accessible from Unity side similar to Vuforia model targets

    • @dilmerv
      @dilmerv  6 месяцев назад +1

      Thanks for your feedback, and yes I am sure Unity will be making this available with PolySpatial just like they do add support for all ARKit features available for visionOS today.

  • @antdx316
    @antdx316 6 месяцев назад

    Does it way more battery doing that as it constantly looks for the objects and needs to know their positions?
    I assume if you have to make it search for more than 2, it will drop the FPS heavily and be unusable then the app crashes?

    • @dilmerv
      @dilmerv  6 месяцев назад +1

      Normally when I run CV processes on mobile phones I can tell that it is pushing its capabilities, but with this, the device performs so well, I didn’t feel like it was causing any kind of performance degradation, but I only did it with 2 objects. I will be testing this more to find out its limits, thanks for your feedback.

    • @antdx316
      @antdx316 6 месяцев назад

      @@dilmerv yea, need to look for memory leaks too
      Being able to store that stuff on a cloud to use would be good.
      Would be cool if it activated certain things if within a geozone.

  • @NikolajStausbl
    @NikolajStausbl 6 месяцев назад +1

    Have you tested if this works and tracks as well in an iOS app? Or is it VisionPro only?

    • @dilmerv
      @dilmerv  6 месяцев назад

      Based on what I read and saw in the code this is only for visionOS. ARKit for iOS provides object tracking and it came out a few years back, here is some info: developer.apple.com/documentation/arkit/arkit_in_ios/content_anchors/scanning_and_detecting_3d_objects
      Thanks and great question!

    • @NikolajStausbl
      @NikolajStausbl 6 месяцев назад

      @@dilmerv Yeah i remember :) have used it for a few different projects as well. But the scanning process is extremely limiting and adds a lot of randomness to the result so am really hoping this new way become the iOS way as well :) thanks for your answer!

    • @0GRANATE0
      @0GRANATE0 6 месяцев назад

      @@NikolajStausbl on iOS is it just 3d object detecting or also 6DoF tracking? is it fast?

    • @NikolajStausbl
      @NikolajStausbl 6 месяцев назад

      @@0GRANATE0 It is definitely not realtime and it definitely doesn’t like moving objects. Static, non-glossy, high detail objects are best.

    • @0GRANATE0
      @0GRANATE0 6 месяцев назад

      @@NikolajStausbl I was hopping for something like this: ruclips.net/video/0ORZvDDbDjA/видео.html (demo is allegedly with monocular 2d camera)

  • @antonimalachowski5262
    @antonimalachowski5262 6 месяцев назад +1

    It seems like it should be possible to reduce tracking latency on stationary objects if you integrate the tracking box with input from the gyros and accelerometers. Just a thought. Anyways, thanks for the upload!

    • @dilmerv
      @dilmerv  6 месяцев назад

      Hey thanks for your feedback, and I agree that we may have a few ways to improve it, but knowing Apple I know they will improve this very soon.

    • @uouo123
      @uouo123 6 месяцев назад

      @@dilmerv Turning on Apple Enterprise API can improve the tracking rate.

    • @robotrocks0206
      @robotrocks0206 6 месяцев назад

      @@dilmervyour 3D models are pretty big in scale compared to the PS controller and the little figure or is the UI showing it wrong? ruclips.net/video/NaB_6hLzYN0/видео.html

    • @dilmerv
      @dilmerv  6 месяцев назад

      Interesting, could you send me a link to a document where I can read more about it?

    • @dilmerv
      @dilmerv  6 месяцев назад

      They're about the same size and accurate, if you look at the measurements within Create ML, it shows the figure as 22 cm ish, that's the real size. The UI shows them big mainly because I zoomed in.

  • @CodyMaverickSurfsUp
    @CodyMaverickSurfsUp 6 месяцев назад

    what's the lag? How about accuracy?

    • @dilmerv
      @dilmerv  6 месяцев назад

      Good questions, it is slow and not instant as perhaps you've seen on iOS. For Accuracy, yes it is very very accurate.

    • @jc31bravo66
      @jc31bravo66 6 месяцев назад

      By accurate, I mean actual metrics. Do you know if Apple provides that?

  • @Polerface774
    @Polerface774 6 месяцев назад

    Is this anything even close to this in the Meta Quest 3 universe? I can't find a way to track real life objects. Do you know any solutions? Even with external hardware..

    • @dilmerv
      @dilmerv  6 месяцев назад

      Hey thanks for your feedback, but this won't work with Meta devices. Meta recently released an open source project for computer vision with Meta devices called "ocean". I recommend taking a look at: facebookresearch.github.io/ocean/docs/overview/ and facebookresearch.github.io/ocean/docs/demoapps/questapps

    • @Polerface774
      @Polerface774 6 месяцев назад

      @@dilmerv Thank you so much! What about Pico 4 or Pico 4 Entreprise? Any chance I can access the RGB sensor feed? THXX

  • @drmarioschannel
    @drmarioschannel 6 месяцев назад

    Doesn’t Polycam scan objects easily?

    • @dilmerv
      @dilmerv  6 месяцев назад

      Yes you could use polycam that’s correct, looks like they also support usdz files. I didn’t use them mainly to keep the workflow with Apple tools but you are more than right to have that as an option.
      Thanks for your comment.

  • @tevirpvr
    @tevirpvr 6 месяцев назад

    This program will work on quest-3 or it is supported only on vision os?

    • @dilmerv
      @dilmerv  6 месяцев назад +2

      Currently, this is mainly for visionOS but I remember that recently Meta implemented Apple’s spatial video, so perhaps they may leverage this tech in the near future. Great question!

  • @sock501
    @sock501 6 месяцев назад

    The best thing this can do for the market is convince Meta and Valve to remove the bullshit privacy restrictions on their headsets’ forward facing cameras. That is the only reason indie devs didn’t beat Apple to this feature.

    • @dilmerv
      @dilmerv  6 месяцев назад

      Yeap, I just can’t wait for Apple to remove the enterprise limitation on camera access, and having other manufactures also allow it would be amazing.

    • @quikbale
      @quikbale 5 месяцев назад

      Someone actually did, and it’s a lot better than this. ruclips.net/video/kiSOmFVfNpc/видео.htmlsi=IsBPWxaI5FPrHdvK

  • @uouo123
    @uouo123 6 месяцев назад

    Can you identify the PS5handle and turn it into a robot with one click?

    • @dilmerv
      @dilmerv  6 месяцев назад

      You can identify the PS5 controller during the anchor added event and swap the models yes, the code currently tracks the visualizations in a dictionary so you could easily tweak the code to do so.

    • @uouo123
      @uouo123 6 месяцев назад

      @@dilmerv cool!

  • @albertsitoe7340
    @albertsitoe7340 6 месяцев назад +1

    So instead of Apple building human humanoid robots. They’ve just created a training platform for humanoid robots.
    And all they have to do is plug in the OS and electric motors.

    • @dilmerv
      @dilmerv  6 месяцев назад +1

      Yeap, that's what they do. They create the foundation and then go big on building the top layers later on.
      Thanks for your comment.

  • @AZITHEMLGPRO
    @AZITHEMLGPRO 5 месяцев назад

    Bro really spent hours changing a bunch of lines of code, just to put some red lines around his real world objects😂

    • @dilmerv
      @dilmerv  5 месяцев назад

      I know 😅 crazy how what may look so simple could be so complex. This type of thing is very easy with Unity and other game engines, but with Apple is somehow so complex.

  • @RazDevTips
    @RazDevTips 6 месяцев назад +2

    11hrs to train a medium quality model? LOL so much for M processors, would love to be able to use this with Nvidia cards

    • @dilmerv
      @dilmerv  6 месяцев назад +2

      I was very surprised how long it took, honestly this is still in beta and I wouldn’t be surprised if it was drastically improved when it comes to prod or in a future version, but I agree, it would be very interesting to be able to run it against an NVIDIA GPU. Thanks for your comment!

    • @Rudgged
      @Rudgged 6 месяцев назад

      You’re comparing an SoC to a dedicated GPU….

    • @qwertzuiop875
      @qwertzuiop875 6 месяцев назад

      The M1 Pro is a tiny chip...

  • @ChiltonWebb
    @ChiltonWebb 6 месяцев назад

    That's disappointingly slow. ARKit is capable of much faster speeds. I wonder what's up with that.

    • @dilmerv
      @dilmerv  6 месяцев назад +1

      @@ChiltonWebb I am very curious to see how this improves once it gets out of beta. But, I agree it is definitely capable of a lot more!
      Thanks for your comment!

    • @ChiltonWebb
      @ChiltonWebb 6 месяцев назад

      @@dilmerv Yeah I think you did an earlier video about ARKit with object tracking a few years ago, and it was butter smooth. But hey, this is Apple, they don't really half ass this stuff. So I expect it will only get better before it's finalized.

  • @isiaqagboluaje7203
    @isiaqagboluaje7203 6 месяцев назад

    18fucking hrss hell noo

    • @dilmerv
      @dilmerv  6 месяцев назад +1

      Looking forward to drastic time cuts in the near future!