NVIDIA Just Made AI Photogrammetry 1,000x Faster [Instant-NGP]

Sdílet
Vložit
  • čas přidán 14. 02. 2022
  • NeRF, ADOP, Plenoxel, now instant NGP?! we are at the peak of NeRF research. I did not expect finding a better solution for neural radiance fields would be such a trending research topic, and it's just getting started. This field of AI is going to be incredible. As of the time of publishing, another eye catching NeRF research just came out, I can't keep up with these.
    Instant Neural Graphics Primitives with a Multiresolution Hash Encoding
    [Official Project Page] nvlabs.github.io/instant-ngp/
    [Official Paper] arxiv.org/abs/2201.05989
    [Official GitHub] github.com/NVlabs/instant-ngp
    [Installation Tutorial] • Instant-NGP Windows In...
    [Installation GitHub] github.com/bycloudai/instant-...
    This video is supported by the kind Patrons & CZcams Members:
    🙏Andrew Lescelius, Panther Modern, Sascha Henrichs, Jake Disco, Demilson Quintao, Tony Jimenez, dicefist, Shuhong Chen, Hongbo Men, happi nyuu nyaa, Carol Lo, PyroBot
    [Website] bycloud.ai
    [Discord] / discord
    [Twitter] / bycloudai
    [Patreon] / bycloud
    [PayPal] paypal.me/cloudweather
    [Music] Zeeky - Springfield
    [Profile & Banner Art] / pygm7
  • Věda a technologie

Komentáře • 131

  • @LeBeautiful
    @LeBeautiful Před 2 lety +55

    this gotta be my new favorite channels of 2022. AI everything, incredible.

  • @nefwaenre
    @nefwaenre Před 2 lety +1

    i was waiting for you to cover this!!! i'm sooooo fascinated by it!!

  • @block-buster1039
    @block-buster1039 Před 2 lety +3

    Excellent video as always. Id be curious to know if photogrammetry photographing techniques would improve your results. Do you have a gimble? Right now im using 8k camera phones and it takes all day to render just a single scene/subject. I got your last tutorial on this subject to work id be curious to see if my own results would improve

  • @dmitrisochlioukov5003
    @dmitrisochlioukov5003 Před 2 lety +4

    Thanks for downloading this shit I was about to do it myself but could not be bothered with all the steps xD
    The technology truly looks revolutionary when compared to common photogrammetry, but a shame the Mesh Export is so muddy!
    Hopefully they could fix this soon ;)

  • @kwea123
    @kwea123 Před 2 lety +37

    4:45 I'm sure you did something wrong. With my experiment (on linux) the lego truck is extremely crisp and the mesh is almost perfect.

    • @bycloudAI
      @bycloudAI  Před 2 lety +15

      hmm interesting, i’ll check with the pytorch implementation if that’ll be fixed, just windows things ig
      would u mind sharing ur results too?

    • @trollenz
      @trollenz Před 2 lety +2

      Yes please share the results !

    • @MangaGamify
      @MangaGamify Před 2 lety +1

      Did you have a google Colab?

    • @kwea123
      @kwea123 Před 2 lety +6

      @@bycloudAI I think it's highly possible it's a windows problem... czcams.com/video/wh28c_8NKUc/video.html&ab_channel=AI%E8%91%B5

    • @sinanrobillard2819
      @sinanrobillard2819 Před 2 lety

      I tried on windows and also got a pretty good mesh! Did you try to leave it on training like ≈30s?

  • @brainlink_
    @brainlink_ Před rokem

    Thank you so much for this wonderful video!

  • @astar484
    @astar484 Před 2 lety +20

    nvidia engineering always manages to amaze me wow

  • @Lakus231
    @Lakus231 Před rokem +1

    this would be perfect for videos/movies to watch in VR and could also solve the double vision problem when tilting the head, from which all VR videos are suffering from

  • @lopiklop
    @lopiklop Před 10 měsíci +2

    These are the things that make AI exciting, perfecting techniques or revitalizing old ones. It's like anything that you intuitively think that technology could be able to do, AI is able to help realize. I've always thought about upscaling, and stuff like this, decades before it existed, and "AI can do it. " (The people designing AI). I guess, also this kind of stuff can be scary too because now robots have eyes. Or whatever.

  • @wuntonsoup
    @wuntonsoup Před 2 lety +2

    This is an awesome video man!

  • @user-nq5hy7vn9k
    @user-nq5hy7vn9k Před 2 lety +8

    So Making 3D VTuber models from 2D ones would be much more easy with such AI

    • @nilaier1430
      @nilaier1430 Před 2 lety +2

      Well, for creating a 3D mesh or scene you still need quite a handful of images from a different angles, but still the same object without any major changes. So, I guess, it's better to just make a model instead of drawing minimum 12 pictures from every 45°

    • @youtubedeletedmyaccountlma2263
      @youtubedeletedmyaccountlma2263 Před rokem

      @@nilaier1430 We just need some genius AI researcher to be able to reconstruct meshes with all existing data. A user will just need references meshes and telling the program what kind of object to reconstruct.

  • @Zung92
    @Zung92 Před 2 lety +7

    With cloud rendering this gonna be big change in Game, Real Estate, Filming productions and more

  • @jeanchristophesibel
    @jeanchristophesibel Před 8 měsíci

    Thanks for the short presentation, it is nice to obtain summary like this. I can't find neither in the papers nor in the github the videos you show from 3:40 to 4:00. From these, I understand that we can run the soft with videos as inputs instead of pictures only. Could you give the source for these materials please ?

  • @krysidian
    @krysidian Před 2 lety +9

    I always love when AIs start to get responsive and you get to see them work in basically real time. Makes working with them so much more intuitive!

  • @asteriondaedalus6859
    @asteriondaedalus6859 Před rokem

    Hi. Thanks for this review. You mentioned you had limitations due to your hardware? Can you please include your hardware specs in the descriptions? Just to give people a baseline.

  • @derekborders9647
    @derekborders9647 Před 2 lety +5

    As a fellow windows user, I highly recommend taking the time to set up WSL2 and get familiar with Linux. Linux can be obnoxious in a lot of ways, but it’s great to be able to work with it when needed. Bash also feels cleaner to me than cmd.

  • @isbestlizard
    @isbestlizard Před rokem +2

    This looks like so much fun! The next step will be parrallelising these so supercomputers can render worlds with hundreds or thousands of nerf objects, and generate more from text prompts. Like come on lets get proper ready player one stuff going :D

  • @captureinsidethesound
    @captureinsidethesound Před 2 lety +1

    I wonder if the results are better with RAW or ProRes video considering the program would have more data information to work.

  • @AlexCarby
    @AlexCarby Před 2 lety +1

    Recreate those CG still shots from Fight Club with this!

  • @AlanCrytex
    @AlanCrytex Před 2 lety +29

    What the hell, we can literally make a 3D model out of our existing pics & videos quick, that's insane

    • @bycloudAI
      @bycloudAI  Před 2 lety +19

      Or even 3D scene. You can have a whole ass room or even a stadium like their official demo. It's mind-blowing

    • @AlanCrytex
      @AlanCrytex Před 2 lety +7

      @@bycloudAI Exactlyy! I'm looking forward to seeing this being improved as the time goes on, everything seems to be close to realtime too and it's terrifyingly amazing

    • @fnytnqsladcgqlefzcqxlzlcgj9220
      @fnytnqsladcgqlefzcqxlzlcgj9220 Před 2 lety +3

      @@bycloudAI im wanting to use this for temples in india, make a VR pilgramage lead by sanyasi's and everything

    • @aflac82
      @aflac82 Před 2 lety +2

      It alredeady existed a long time ago, its called photogrammetry. Whats huge now is the number of imputs and the speed of the result.

    • @sayrith
      @sayrith Před 2 lety

      But how does this compare to photogrammetry?

  • @SEXCOPTER_RUL
    @SEXCOPTER_RUL Před 2 lety +3

    this might be a stupid question, but could this be used on a vr headsets stereo passthrough camera like the valve index for instance, and run in real time rendering the scene in full 3d?

    • @Alechk4
      @Alechk4 Před rokem

      The render time of a single frame in FullHD is around 15 seconds. Not doable for now...

  • @Bezio.
    @Bezio. Před 2 lety +2

    Mmm, how does this remind me of the braindance in Cyberpunk. Camera flight in the area outside the eyes. Very cool!

  • @mityashabat
    @mityashabat Před 2 lety +2

    I may be cherry picking but you were saying that Ray Marching is used.
    Correct me if I'm wrong: rendering of Neural Radiance fields happens via Volume Rendering.
    From what i heard, there are methods that jointly optimize Radiance Fields and Signed Distance Functions for solid surfaces. And Ray Marching can be used to render surfaces from SDF estimations.

    • @bycloudAI
      @bycloudAI  Před 2 lety

      im not exactly an expert but this is what I found
      en.wikipedia.org/wiki/Volume_ray_casting

  • @TheBoringLifeCompany
    @TheBoringLifeCompany Před rokem

    yepp I've suffered same hours of setting up issues. The results are rewarding. RTX 3090 is the reference GPU for these workloads. Btw, how to export the video from the scene fly rather than from OBS?

  • @FUTUREDESIGN3000
    @FUTUREDESIGN3000 Před 2 lety +1

    Great Image and Render Samples but if you slow downs the narration somewhat I feel that it will help us 3D neophytes absorb this new important information more easily!!! Thx!

  • @Uhfgood
    @Uhfgood Před 2 lety +1

    So take aspects from both methods, so you don't have that weird point cloud, or that strange texture disappearance thingy.

  • @GabryDX
    @GabryDX Před 2 lety +1

    Hi, thanks for your work. As a reference can you tell us what GPU are you using? Thanks, really appreciated.

    • @christophebrown978
      @christophebrown978 Před 2 lety +1

      2:04 he says he has an RTX 3090

    • @GabryDX
      @GabryDX Před 2 lety +1

      @@christophebrown978 Probably I've missed it, thank you!

  • @whatohyou01
    @whatohyou01 Před rokem

    Now add this function with stable diffusion then we can input 2d photo->3d model with deduced human joint skeletons->move to pose it as however users want->-choose prefered art style safetensor models->convert it back to 2d photos.(fingers/toes maintained since it refers 3d models with finger joints).
    If it's possible then it will also be made into animations/movies when you provide sample movies to the character.

  • @Cziczaki
    @Cziczaki Před 2 lety +1

    Is there any software for AMD Graphics? I would like to turn some images into 3d models too.

  • @UFOgamers
    @UFOgamers Před 2 lety +6

    Can you imagine running all Google Maps pictures on this and getting a game like map of every city on earth...
    Then you use it as a video game terrain, and voilà, you get the most realistic game ever...

    • @jimj2683
      @jimj2683 Před 2 lety

      That is one of my biggest dreams along with aging reversal. Imagine GTA 7 with the entire Earth in perfect detail and using materials from UE5 to make things destructible (using AI).

    • @Richienb
      @Richienb Před 2 lety

      czcams.com/video/8AZhcnWOK7M/video.html

    • @captureinsidethesound
      @captureinsidethesound Před rokem

      @@jimj2683 , it would require servers with real-time map streaming because the cost to download the entire Earths maps locally would be extremely expensive even if half the texture maps where reused for basic foliage and buildings similar to how Ubisoft does it with some titles.

    • @jimj2683
      @jimj2683 Před rokem

      @@captureinsidethesound That is exactly what Microsoft flight simulator 2020 is using.

    • @captureinsidethesound
      @captureinsidethesound Před rokem

      @@jimj2683 , its also rendering only from a top view in which isn't that demanding on hardware, even on cell phones. They are using height maps that only give a "3D" look from a distance.

  • @AnthonyGarland
    @AnthonyGarland Před rokem

    Thanks!

  • @CyberWolf755
    @CyberWolf755 Před 2 lety +5

    I just want this tech available on non-RTX hardware in the future and not come with a catch of having to buy a Nvidia GPU for this niche.

  • @krinodagamer6313
    @krinodagamer6313 Před 2 lety +1

    this is wild!!!!!

  • @Cool-wh6ov
    @Cool-wh6ov Před rokem

    Have you tried output your video into images of tiff format

  • @Mobay18
    @Mobay18 Před 2 lety +6

    Can you please try this on a image set of a Hollywood actor? It would be soo cool to make 3D models of characters.

    • @EddieBurke
      @EddieBurke Před 2 lety

      That not how this works lol

    • @Mobay18
      @Mobay18 Před 2 lety

      @@EddieBurke Why would it not work on faces?

    • @sayrith
      @sayrith Před 2 lety

      That's illegal. Or at the very least, unethical.

    • @Mobay18
      @Mobay18 Před 2 lety +6

      @@sayrith Why is it unethical? 3D models is art the same way as a 2D drawing is.

  • @Draaack
    @Draaack Před 2 lety

    Loving every video, AI is gettin sweet

  • @Instant_Nerf
    @Instant_Nerf Před 2 lety +1

    How would .. or what would a 360 scene provide? By this tech ..

  • @polynightingale3969
    @polynightingale3969 Před rokem

    Hi did you able to enable dlss function ?

  • @sayrith
    @sayrith Před 2 lety +2

    How does this compare to photogrammetry?

  • @desu38
    @desu38 Před 2 lety +1

    Ngl, of all artifacts, the clouds are probably the coolest.

  • @readmore8974
    @readmore8974 Před 2 lety +1

    Have you looked into the mesh issues yet? The guy who said that it worked fine on Linux posted their results. I'd really like to use this in my Windows machine but I'm not gonna bother if I'm going to have issues like that.

    • @hdl468357
      @hdl468357 Před 2 lety

      where is the "worked fine linux results",can you share the link? thx

    • @readmore8974
      @readmore8974 Před 2 lety

      @@hdl468357 "instant ngp demo" on youtube, you have to scroll down a bit. The channel is "AI葵". I wish I could link it but I hope this helps.

  • @user-by6vj9fj2k
    @user-by6vj9fj2k Před 2 lety

    3:21 is this Tbilisi?

  • @DiegoAlanTorres96
    @DiegoAlanTorres96 Před 2 lety

    That's really f*ckin nuts

  • @artemtalanov3205
    @artemtalanov3205 Před 2 lety +1

    love this Nerf - NERV vibe, God in his heaven, Everything's right with the World 🙃👽👾

  • @sierraecho884
    @sierraecho884 Před 2 lety +1

    holy fuck. I love photogrammetry but this is on an other level. 2sec. you must be wrong. I got to try that out.

  • @MangaGamify
    @MangaGamify Před 2 lety +2

    Can anyone share their Google Colab version?

  • @hdnasauploads8863
    @hdnasauploads8863 Před 2 lety +1

    Can you export from this?

  • @wolfofdubai
    @wolfofdubai Před 2 lety

    Where can I download it?

  • @blinded6502
    @blinded6502 Před 2 lety +4

    Didn't they increase performance 10x recently?

  • @TiagoTiagoT
    @TiagoTiagoT Před 2 lety +1

    Get an additional HDD (maybe even an external one) and install Linux there, even if just to run Linux-specific apps like you mentioned here.

  • @_casg
    @_casg Před rokem

    So like I can’t get the obj looking normal ?

  • @amsrremix2239
    @amsrremix2239 Před rokem

    So these are like really fancy photo scans ? These are really cool - but I’m sure the topo is still really messy . So you can’t really rig or do dynamics with any of the stuff …. Super beautiful though . I can’t wait for the day where you can photo scan a person with auto correct topo

  • @useraccount4201
    @useraccount4201 Před 2 lety +5

    05:14 Amogus

  • @kukukachu
    @kukukachu Před 2 lety +1

    I mean, this is no surprise as 2023-2024 is the new new age of technology. If you're impressed with this, just wait till you see what we actually get.

  • @kumarmanchoju1129
    @kumarmanchoju1129 Před 2 lety +1

    Make rendering of "The Matrix" using movie footage

  • @petergedd9330
    @petergedd9330 Před 2 lety +1

    Tip, leave gaps between words

  • @METTI1986LA
    @METTI1986LA Před rokem +5

    Its basically 3d scanning...

  • @WayneStakem
    @WayneStakem Před rokem

    Imagine when they implement this to Google Earth. Put your VR goggles on and start walking around town.

  • @NoName-br8pb
    @NoName-br8pb Před 2 lety +1

    How do these models handle photos of people?

  • @ncmasters
    @ncmasters Před 2 lety +3

    pls make video how to install. I tried but its hard to understand

    • @ncmasters
      @ncmasters Před 2 lety +1

      oh its in the description :O

  • @0GRANATE0
    @0GRANATE0 Před rokem

    any tutorials on this?

  • @mcn8o
    @mcn8o Před 2 lety +2

    You should take a look at the Concert Creator AI. It generates a realistic looking virtual piano player

  • @jadenkarim5367
    @jadenkarim5367 Před rokem

    Fellas pack your bags we're moving to Mars probably next month!

  • @wonjaeseo1994
    @wonjaeseo1994 Před 2 lety

    TO THE MOON

  • @sillyfella2009
    @sillyfella2009 Před rokem

    New 3D saul😳😳😳

  • @themadears
    @themadears Před rokem +1

    Can you make tutorial please

  • @justjake5389
    @justjake5389 Před 2 lety

    just imagine this technology with google streetview :D

  • @yosha_ykt
    @yosha_ykt Před 2 lety

    in future probably ai will create entire city only with google street

  • @Instant_Nerf
    @Instant_Nerf Před 2 lety +2

    When extracting google maps into blender ..when close up…The textures and meshes are so bad. If they improved those by just a little we could make movie scenes out of them. Sure from above with lighting and other fx it looks good.. but if you want to tell a story .. you want to be ground level from the camera angle.. and that right now is not possible.. I dont know why they just didnt use drones to capture the images when creating 3d maps.. they used satellites images and airplanes to take photos from different angles. I just think the quality from a drone especially with todays cameras.. and lidar.. would have been something special. But maybe that will come too. They can just send out multiple drones and auto map everything .. really fast .. It wont take as long as when the first started. The other thing we can do is replace some of the ground, trees, flowers, even watter effects from a ripped google map .. into blender.. etc. that also makes it nice for eye candy. Id like to make a documentary and I need footage.. I have all my aerial scenes.. just need lower scenes as if filmed from a camera crew.. if anyone has any suggestions about how to go about it. let me know. Going to these locations that I need footage from is not an option for me. Would the above tech be feasible for my project ?

    • @railgap
      @railgap Před 2 lety

      you're concerned about resolution, but then you propose LIDAR as a solution. Stay in school, kid.

    • @Instant_Nerf
      @Instant_Nerf Před 2 lety +1

      @@railgap I dont think you know what lidar does. Has nothing to do with resolution, you misunderstood what I was asking not providing a solution. so look at youreself before you make yourself look like and idiot trying to bring others down.

  • @nocnestudio
    @nocnestudio Před rokem

    This volume illusion.

  • @dediamkaming
    @dediamkaming Před 2 lety

    Star Trek/Doraemon technology gone true

  • @Labelye
    @Labelye Před rokem

    Could be apply NERF on photos of telescope James Webb

  • @adcsil3710
    @adcsil3710 Před rokem

    1:42 how the heck you have chinese-cantonese english accent suddenly

  • @MidhunRaj_I_am_here
    @MidhunRaj_I_am_here Před rokem

    So nref can't convert to a 3d Mesh

  • @InfiniteComboReviews
    @InfiniteComboReviews Před rokem

    You know, 3D game devs hate retopoing and unwrapping models, and it seems like programmers and coders seem hell bent on replacing every aspect of designing stuff EXCEPT those parts.

  • @cholasimmons
    @cholasimmons Před 8 měsíci

    Still think we're in base reality?

  • @nownezz
    @nownezz Před 5 měsíci

    DOOR STUCK

  • @ushakovkirill
    @ushakovkirill Před rokem

    Gamarjoba

  • @cherubin7th
    @cherubin7th Před rokem +1

    WTF. Just install Linux like every free person.

  • @javieroatway4059
    @javieroatway4059 Před 2 lety

    Love your video! !! You are missing out - P r o m o S M!

  • @SSingh-nr8qz
    @SSingh-nr8qz Před 2 lety

    WITCHCRAFT!!!

  • @lod4246
    @lod4246 Před 2 lety

    hi second comment

  • @bruhmoment23123
    @bruhmoment23123 Před 2 lety

    Third

  • @EwanCodeTalker
    @EwanCodeTalker Před 2 lety

    lol evafans are so cool

  • @Trapanism
    @Trapanism Před 2 lety

    Please review LALAL.AI's new engine. Maybe you can compared the results of the previous engine from your first video to the new one. You can also still use the old engine as an option, so maybe you can review new samples too.