Regarding full body motion capture, Move AI might be worth a look. It used to be iPhone only, but they just announced an experimental version for HD cameras and Android devices.
I’ve been playing with this. I used all the ipads in the house for set up 🤣 as could get hold of 3 Phones 😅 Thank though ! What were your results like?
There will be something for android i’m sure of it. Some kind of bridge. I quoted the tech lead on that bit as I didn’t want to be the bearer of bad news haha
Been always a big fan of you, and i think you would be a great candidate to cast Cleopatra if you ever decide to make an unreal short based on your own look.
Thank you for all of your videos! You explain so efficiently. Learning here is a relief, compared to a lot of other tutorials are 4x longer and don't communicate nearly the same amount of info. Loving all of the Unreal, character, animation and AI content! Seriously, thanks.
Thanks for you described our feelings about Metahuman. I wish the next update becomes to make such a tecnology more useful and quick access. As if we have a cake but it is fixed at the other side of the table and we can pick up just a piece of choclatte on it each time
big problem with Android phone is that they used to have ToF camera, but for some reason they have been getting rid of that (damn Samsung....) and you need ToF censor to do this
A bunch of the recently new possibilities with AI art piqued my interest, and I expect many opportunities with full 3D animations to be just around the corner. Prompt based images from stable diffusion are nice, but you may end up with different faces every time. and you need to fix wonky problems, like hands and other weirdness. With additions like controlnet, esp. openpose and masking you really get more control. With LoRa you can train your own models and may combine them. But it's all just not ideal. One of your videos, getting an image into blender, either with KeenTools FaceBuilder plugin (or DAZ FaceGen / Face Transfer), and then get it into metahumans unreal could be a huge step forward. But with how fast development is going, maybe there will be huge breakthroughs that will directly link a set of images (front/45 angle plus face mesh) to either a 3D asset or a pose or face for a 2D artwork directly. Maybe the tech will go so crazy as to confer add-ons like clothes and hairdoes, first from image to image (in different poses/angles) then to a 3D asset. Now you do loose some consistency, with different angles and higher resolution introducing small changes. Perhaps metahuman can produce output an AI model can be trained on and similar altered files could then be read back into metahuman? You never know, every week there seems to be something new!
I know a lot of dev is going behind the scenes using image to mesh…. It’s coming. It’s a lot to keep up with 🤣 But yes I also believe 3D AI will work together in harmony very well. It’s exciting to be in these early stages and seeing all the development almost in real-time. I agree prompt based images were okay… but we want to do something much bigger ✌️✨
@@promptmuse yes, I'm such a beginner, but I have a start now with sd with controlnet in a google collab now and it works. I haven't made an image I'm please with yet, but will try again this weekend 😃.
Always wondered how is it that while animations are being captured from a real person they still look so weird/uncanny, now i got the answer, the person's face while doing said expressions looks extremely weird/uncanny
Big issue with metahuman is that it creates EGGHEADS. I have 0,025mm accuracy 3D scanner (metrology grade scanner for reverse engineering) that makes acurate 3D models. And i tried to use that model with metahuman. ANd metahuman fucked it up as it created EGGHEAD. Washes out many medium and large details. Using metahuman sliders to make model look right is also dumb. They basically transfared zbrush "move topology" and "mov elastic" brushes in to sliders. And what is the point of scanning when metahuman fucks it up? long road a head with metahuman.
Yeah totally agree with you. Tbh I think metahumans are only good for the head….you are also limited with their damn neck thickness. It’s a huge give away and gives the meta human “look”
This is what they have stated. Hopefully this isn’t the case, as they said it’s been made for that….which could translate as it maybe unstable on other devices but still work. That’s an assumption on my behalf.
Thank you. There's one more thing I have to add to the list-buy an iPhone:) I assume that in a couple of months there will be various extensions in stable diffusion that will already be very good at generating video and improving masks.I assume that in a couple of months there will be various extensions in stable diffusion that will already be very good at generating video and improving masks. Good luck and success to all
Yeah I hear ya! Try the forum for latest announce. Also, on this page there is the latest documentation on the top horizontal tool Bar forums.unrealengine.com/categories?tag=metahuman
@@promptmuse What tool do you use? I am making only a low-poly stuff for an older game, but I am still curious. Good luck with your channel, you are doing great job ;)
for a company that develops games 99% for windows you'd think they would release their apps on android as most windows users have android phones... oh well what a shame.
I’m sure they will develop something. Someone ( A kind open source dev developed) Developed a Live link android version for the current pipeline. So I can image in someone will do they for the release. 👍
Visuals are good for games indeed, but AI is more important. I do not mind fighting stick figures as long as they do not act like DOOM\Diablo NPC's. Hope some chatGPT AI can be implemented for such task.
Hello Prompt muse. You are very pretty but I was just curious why is one of your eyes different to the other one. Sorry if it is rude to ask ☹ I am just curious. Thanks.
I didn't even know you covered UE5 topics. I like Unreal, but I am writing a new, lightweight engine that's faster, better graphics and integrated with machine-learning at the hardware level 😅
Meta human is super annoying. CC is better I think but when I imported the characters from that software I kept getting deformed hands and legs in the trial version. Did that happen to anyone else? I really wanna buy it for its flexibility.
I cannot stand Apple. Will never buy another. There's almost double of android users out there over apple. These devs making apple users first or exclusively apple must had a boat load of money thrown at them. It makes no sense to make a app that's exclusive to a smaller market using unreal engine which anyone can use. The devs must be some serious fan boys or something. I enjoy android but I wouldn't exclude a huge market like that either.
thank you for covering this! you are my go to for meta human DIY etc. Thanks!
Thanks! It’s going to marry up nicely with AI as well … well here’s hoping 😂👍
Regarding full body motion capture, Move AI might be worth a look. It used to be iPhone only, but they just announced an experimental version for HD cameras and Android devices.
I’ve been playing with this. I used all the ipads in the house for set up 🤣 as could get hold of 3 Phones 😅 Thank though ! What were your results like?
@@promptmuseYou're very welcome. I'll let you know as soon as I get some time off. I'm scrounging all the cameras I can find.
Thank you for taking the time to condense all this information, more exciting content ahead no doubt!
Top content! Thank you for sharing these news.
No problem at all! The full length official one is in my description. 👍🫶
I wish we had an Android version 😔
You can buy a relatively cheap older iPhone just for the live link app
Lol I’m picturing a robot that’s unhappy about the biocentricism of the metahumans app.
There will be something for android i’m sure of it. Some kind of bridge.
I quoted the tech lead on that bit as I didn’t want to be the bearer of bad news haha
@@promptmuse Wonderful!😊
@@promptmuse Can you link the video where you quote the tech lead? i.e., the one you included a still image that says, "Voice of Raffaele Fragapane"?
Love your in-depth tutes. Thanks.
Loved this update. Thanks!
Thanks so much 🫶 Glad you loved it !!
Been always a big fan of you, and i think you would be a great candidate to cast Cleopatra if you ever decide to make an unreal short based on your own look.
Haha Maybe. If I had time to do a short !
Thank you for all of your videos! You explain so efficiently. Learning here is a relief, compared to a lot of other tutorials are 4x longer and don't communicate nearly the same amount of info. Loving all of the Unreal, character, animation and AI content! Seriously, thanks.
Thanks for you described our feelings about Metahuman. I wish the next update becomes to make such a tecnology more useful and quick access. As if we have a cake but it is fixed at the other side of the table and we can pick up just a piece of choclatte on it each time
empath senses tingling u good?
big problem with Android phone is that they used to have ToF camera, but for some reason they have been getting rid of that (damn Samsung....) and you need ToF censor to do this
Its slow and laggy for my pc but cloud runs good
A bunch of the recently new possibilities with AI art piqued my interest, and I expect many opportunities with full 3D animations to be just around the corner.
Prompt based images from stable diffusion are nice, but you may end up with different faces every time. and you need to fix wonky problems, like hands and other weirdness. With additions like controlnet, esp. openpose and masking you really get more control. With LoRa you can train your own models and may combine them.
But it's all just not ideal. One of your videos, getting an image into blender, either with KeenTools FaceBuilder plugin (or DAZ FaceGen / Face Transfer), and then get it into metahumans unreal could be a huge step forward. But with how fast development is going, maybe there will be huge breakthroughs that will directly link a set of images (front/45 angle plus face mesh) to either a 3D asset or a pose or face for a 2D artwork directly.
Maybe the tech will go so crazy as to confer add-ons like clothes and hairdoes, first from image to image (in different poses/angles) then to a 3D asset. Now you do loose some consistency, with different angles and higher resolution introducing small changes. Perhaps metahuman can produce output an AI model can be trained on and similar altered files could then be read back into metahuman? You never know, every week there seems to be something new!
I know a lot of dev is going behind the scenes using image to mesh…. It’s coming.
It’s a lot to keep up with 🤣 But yes I also believe 3D AI will work together in harmony very well.
It’s exciting to be in these early stages and seeing all the development almost in real-time. I agree prompt based images were okay… but we want to do something much bigger ✌️✨
@@promptmuse
yes, I'm such a beginner, but I have a start now with sd with controlnet in a google collab now and it works. I haven't made an image I'm please with yet, but will try again this weekend 😃.
@@teambellavsteamalice The good news is we are all beginners haha With how rapid development is 👍 Have fun !!
New Iphones have front facing depth sensor. On most Android phones is just a regular camera.
7 hours... only 8 minutes were interesting.... it's a revolution.
I'm a little lost on the Hellblade example. Was that an example using Metahumans or the Unreal animation tools? Or both?
This is all great in metahuman if they allow you to do body morphs. Other than that it is just a great face puppet.
4:40 me and my boyfriend got into an argument over who’s phone went off lmaooo
Cascadeur is pretty solid for body animation 👌
Looks good overall, but it seems like the lips are never fully closing during speech.
Any idea when meta human will come out of early access?
thnx god there is some dev which is not interested in UEFN as well.
Always wondered how is it that while animations are being captured from a real person they still look so weird/uncanny, now i got the answer, the person's face while doing said expressions looks extremely weird/uncanny
🤣😂😅 When I use facial mocap and look I myself I forget my own mouth moves. But yes some serious facial exaggeration 😁
Great stuff.
Thank you!
It's getting hard to keep up.
Big issue with metahuman is that it creates EGGHEADS. I have 0,025mm accuracy 3D scanner (metrology grade scanner for reverse engineering) that makes acurate 3D models. And i tried to use that model with metahuman. ANd metahuman fucked it up as it created EGGHEAD. Washes out many medium and large details. Using metahuman sliders to make model look right is also dumb. They basically transfared zbrush "move topology" and "mov elastic" brushes in to sliders.
And what is the point of scanning when metahuman fucks it up?
long road a head with metahuman.
Yeah totally agree with you. Tbh I think metahumans are only good for the head….you are also limited with their damn neck thickness. It’s a huge give away and gives the meta human “look”
Are you going to do the second part of blender scenario from images for music videos
I know I keep saying it’s coming 😩 It is ! 🫶I might do it on a youtube live
@@promptmuse Keen to see this too!
Any plans to do live streams?
Oh yes for sure! Really wanting to make some time for some 3D and Stable diffusion tutorials on live streams 👍
Wauuuuu...
Aren't the new Self Tracked full body trackers of HTC Vive not a breakthrough for cheap accesible body motion capture?
I’d love to have the opportunity to test all of the products against each other and see how they compare !
wait...would u pls clarify in 2:48 u mentioned iphone11 upwards..so the iphone10 (X/XR/XS) doesnt work?
This is what they have stated. Hopefully this isn’t the case, as they said it’s been made for that….which could translate as it maybe unstable on other devices but still work. That’s an assumption on my behalf.
Thank you. There's one more thing I have to add to the list-buy an iPhone:) I assume that in a couple of months there will be various extensions in stable diffusion that will already be very good at generating video and improving masks.I assume that in a couple of months there will be various extensions in stable diffusion that will already be very good at generating video and improving masks.
Good luck and success to all
Can someone point me to where metahuman puts up its update writeup? Digging thru the site is a torture of misdirection.
Yeah I hear ya! Try the forum for latest announce. Also, on this page there is the latest documentation on the top horizontal tool Bar forums.unrealengine.com/categories?tag=metahuman
Thank you for the video! I wish metahumans were available for unity
theres a tutorial on youtube to export to blender or 3d software so then you can export it again to unity
@@leslyskugga where there is a will… there is a way 😉.. but it goes against their usage policy
@@bumstudios8817 just like ia art taking someone else work as reference to develop an image, you can just change your model or map texture and use it.
Ultra photo-realistic midjourney images and unreal engine CZcams creator actor. I need someone to help me here
Will iPhone x work with it ? IPhone 10 also have depth camera in it
To be honest, the character in that 1 year old Unity demo looks much more realistic than those presented by Epic Games.
I agree, I actually make my own models in an external model but use the meta human rig ….kinda like what they did in this demo
@@promptmuse What tool do you use? I am making only a low-poly stuff for an older game, but I am still curious.
Good luck with your channel, you are doing great job ;)
Can this be use in companion with Midjourney characters?
Yeap, you can just texture map midjourney skins to your mesh 👍
helping with your timestamps, the brilliant fortnite update is at 7:35
Thank you. I'm going to hopefully cover that in a separate video as it's pretty insane
for a company that develops games 99% for windows you'd think they would release their apps on android as most windows users have android phones... oh well what a shame.
I’m sure they will develop something. Someone ( A kind open source dev developed) Developed a Live link android version for the current pipeline. So I can image in someone will do they for the release. 👍
Visuals are good for games indeed, but AI is more important. I do not mind fighting stick figures as long as they do not act like DOOM\Diablo NPC's.
Hope some chatGPT AI can be implemented for such task.
Hello Prompt muse. You are very pretty but I was just curious why is one of your eyes different to the other one. Sorry if it is rude to ask ☹ I am just curious. Thanks.
🤣 No idea, I think I smudged my makeup but just carried on regardless. But I have a very asymmetrical face anyhow. Not rude to ask, I do not mind :)
Won't it work on iphone XR
I hope it does, we will soon seee
I didn't even know you covered UE5 topics. I like Unreal, but I am writing a new, lightweight engine that's faster, better graphics and integrated with machine-learning at the hardware level 😅
I've been using UE for a while now for 3D work. Love to know more about you development. This sounds really neat.
@Prompt Muse follow us, I made some "test" videos/edits that went well and we will be posting content soon like no one has seen 😄
Thanks for all the AI news.
You are welcome ! I enjoy creating the content :)
What about using Android phone? Not everyone has an iPhone.
I’m sure there will be some way to control with alternatives. Surely they would leave that gap in the market. 👍 We will see !
@@promptmuse Thanks. Oh, I forgot to say I love your channel.
@@promptmuse Thanks. Oh, I forgot to say I love your channel.
@@donelkingii3738 Thank you so much !! 👍✨
Wut duh...
JUST MARRY MEEEEEEEEEEEE
Ima Android User
I believe there will be something for Android users...they just cannot leave a gap in the market like that. Makes no sense.
man i only have android. why give all this to iphone thats bogus.
Meta human is super annoying. CC is better I think but when I imported the characters from that software I kept getting deformed hands and legs in the trial version. Did that happen to anyone else? I really wanna buy it for its flexibility.
Dude you feel like an AI sometimes with no expressions.
I feel like I’m turning into AI haha 😂
nope.. i can c she have a mesmerizing engrossing expression that create peaceNlove
Awesome tool. But only for the rich of course who can afford an iPhone with a LiDAR sensor. I guess I can count my lowly iPhone 8 out.
I cannot stand Apple. Will never buy another. There's almost double of android users out there over apple. These devs making apple users first or exclusively apple must had a boat load of money thrown at them. It makes no sense to make a app that's exclusive to a smaller market using unreal engine which anyone can use. The devs must be some serious fan boys or something. I enjoy android but I wouldn't exclude a huge market like that either.