iPhone Virtual Production Short Film with Behind the Scenes
HTML-код
- Опубликовано: 15 сен 2024
- Short film shot 100% on iPhone with Skyglass app where all VFX are done on the iPhone in real-time with no post processing. Skyglass uses Unreal Engine in the cloud and your cameras tracking info to provide incredibly realistic 3d results.
Watch the original video in Widescreen here: • Sheridan Holbrook and ...
🚀 Want to be notified when the app publicly launches on the app store, join the waitlist here! www.skyglass.c...
✨ Join us on Discord to see more videos shot by our users / discord
✅ Follow us on Social:
Instagram: / skyglassapp
TikTok: / skyglass_app
Facebook: / skyglassapp
Website: www.skyglass.com/
i did alot of work with this app in the beta. I can tell you guys its NOT going to be used for PREVIz only but for Master Production. It is just so goo. The moment you light your foreground properly, You won. Massive shout out to skyglass as i cannot wait to get on the master copy.
I think it's great where you are going. Especially for things like pre-viz and overall blocking. Don't mind the critics!
This is what I was trying to do and somebody finally portrayed it 🙌 thank you ✨
I'm excited for you to try it out! Go get it on the App Store! apps.apple.com/us/app/skyglass-movie-effects-camera/id1626905605
This could be a great solution for low budget VP that makes this process accessible to education where we simply don’t have the budget for even a decent green screen and tracking setup.
Whilst the negative comments made so far a little unnecessary, they do raise a few good points.
I think the main issue is the quality of the backdrops. It is very difficult to create believable environments for VP in UE. Even with all the amazing features of 5.1, I don’t think people realise just how difficult it is to create realistic backgrounds in UE. It is capable but a lot of smaller studios just don’t have the knowledge or use marketplace content, most of which is 50-80% there but still require work to make them look appropriate for VFX.
I currently use Cam Track AR but this looks like it could be better!
Can’t wait to give it a try!
Excited for you to try it! You can get it on the App Store now! (US and Canada only for now) apps.apple.com/us/app/skyglass-movie-effects-camera/id1626905605
Wow!
thanks!
Is it possible to export the footage and Unreal scene with camera tracking separately to do color correction in something like After Effects? So far I haven’t been able to export anything with AI color correction, so I can’t evaluate how well it’s going to work for me. I’d like to have both as separate layers so that I can unify the colors and possibly relight in Lightroom as well. Thank you!
Yes, you can export all of the layers separately and do your own compositing in AE. The camera tracking USD file can also be opened in UE5.
Our High Quality Export servers were down for about a week unfortunately, so please try it again. AI relighting should be working again for all new clips. Old clips you captured may have corrupted files during that period.
Will this app only be available for iOS or is an Android version in the works as well?
Skyglass is iOS only for now, but Android support is on the roadmap.
@@skyglass_app waiting for android version
When will it release? Any time period?
We're planning to launch the app in the next couple of weeks. Hopefully beginning of April at the latest.
Pretty sure, there must be a way to get rid of the green screen spill also realtime with the iPhone, that would be amazing too!
We do have that tool already built into that app. It's called "DeSpill" and it actually works really well! Any green spill you're seeing here still is because of how close we were standing to a small green screen and it can be hard to get that last bit out with out color grading.
So cool😊
Thanks! The app is out now. Download here 👉 apps.apple.com/us/app/skyglass-movie-effects-camera/id1626905605
Can you guys please do somwthing for android devices
Dear All, just discovered this and it looks promising. We are working with kids (short movies, musicals, fairy tales). Question is will users will be limited by scenes from developer library or I can use my own scenes from Unreal Engine. If it's possible to use my own assets - it'd be really great tool.
Hey!! Glad you found us. This is perfect for education and those just getting into virtual production.
Yes, on our Studio plan (coming soon) you will be able to upload custom user environments from UE. Right now, you have access to our library and you can create your own environments with AI.
Download it now! apps.apple.com/us/app/skyglass-movie-effects-camera/id1626905605
And actually, IF you where able to track the environment with the lidar sensor, then you could potentially ALSO move the camera inside the virtual environment and THAT would be a major breakthrough!
yeah, we are already tracking where the camera is in virtual space. you can completely move through any of these UE 3D environments. We aren't using LiDar to do it though, we're using AR Kit which doesn't require LiDar.
@@skyglass_app wait, you can move already??? with ARKit? ARKit tracks also 3D space not just facial features? that's amazing! In your short you only did pans and tracks I assume you had not yet implemented the tracking feature? I look forward to seeing it in action!
@@AlexKongMX We'll definitely post more videos soon, but even better, you can try it for yourself! apps.apple.com/us/app/skyglass-movie-effects-camera/id1626905605
Be sure to try out one of our realistic 3D environments (the ones with the blue "unlocked" icons) to get a sense of the spatial tracking.
I cannot install it in Australia App Store.
It will be available in Australia and the rest of the world in the next few months. We're still prepping to launch on more servers.
Can you please make it available in Saudi arabia, thank you
we'll be adding support for more countries in the coming months! thanks for sticking with us.
@@skyglass_app thank you, but I will be around waiting to try this app
Nice attempt. Which Iphone are You using??
This was shot on iPhone 14 Pro
Sir how to make like this pls make some tutorial sir
Here's a quick walkthrough of how to use the app with a green screen! we have other tutorials on our channel you can watch as well! hope it helps ruclips.net/video/6eVFtZ5KyiI/видео.htmlsi=7Eee5DQdq45c7_ot
Any word yet on pricing? When will the product be available?
Skyglass is live on the App Store now! apps.apple.com/us/app/skyglass-movie-effects-camera/id1626905605
It is free to try, with Creator ($4.99/mo), Indie ($12.99/mo), and Studio ($24.99/mo) subscription levels enabling various features.
❤
This is promising - however, just read that over the site blog that Skyglass highest resolution and frame rate is 1080p HD at 30 fps; should I assume it does support lower fps? 24fps??? I'm interested for my iPhone 14 Pro Max setup.
Unfortunately 24fps is not currently possible as Apple only supports 30 or 60 fps while tracking.
@@skyglass_app Got it! Very similar to FXHome's in that matter. I've been using, (a lot), Camtrack AR, and I've come up with a steady workflow for virtual locations. However, I'm incline to Skyglass and its cloud support. I'm curious, Camtrack AR can't take advantage of Apple's full range of color depth, (limited to 8bits); will this be the case with Skyglass also? It will be great if it's higher. Let me know. and thanks for your quick reply!
Not sure this is a good selling commercial for the app in question. There is a severe lack of blending the characters to the background. Anyone with half a brain can tell its shot on a green screen - which is NOT the goal you should be aiming for when shooting green screen.
we're adding more features in app to allow better color matching between foreground and background shots. we're also adding the ability for our more pro users to color grade and do full post processing on the separate laters in any NLE of their choice.
Fair point but given the budget nature and it’s early state of development, this isn’t a bad way to communicate the product at all.
❤