FINALLY: Consistent Characters in AI Video! & MORE (Latest in AI)

preview_player
Показать описание
We kick off with Lucid V1, an AI-generated Minecraft demo playable on consumer-grade hardware. Then, we dive into the groundbreaking advancements in AI video generation with updates from Vidoo AI and Dimension X, followed by Google’s recapture project. We also talk about high-quality visual effects from auto VFX, virtual try-on videos with FashionVDM, and the cutting-edge image generation models by NVIDIA and Black Forest Labs. Lastly, I provide updates on OpenAI's new O1 models and discuss what’s next in AI advancements. Tune in for a glance at the future of AI and its incredible potential!

▼ Link(s) From Today’s Video:

-------------------------------------------------

▼ Extra Links of Interest:

Let's work together!

Thanks for watching Matt Video Productions! I make all sorts of videos here on Youtube! Technology, Tutorials, and Reviews! Enjoy Your stay here, and subscribe!

All Suggestions, Thoughts And Comments Are Greatly Appreciated… Because I Actually Read Them.

00:00 Introduction to the Exciting World of AI
00:27 Lucid V1: AI-Generated Minecraft
04:42 AI Video Generation Updates
16:38 OpenAI's New O1 Model and Future Directions
21:49 High-Resolution Image Generation
25:43 Conclusion and Upcoming Announcements
Рекомендации по теме
Комментарии
Автор

BTW to join the rest of the community, get the latest on AI, and learn in depth about AI join my FREE discord server! discord.gg/mattvidpro

MattVidPro
Автор

I think the answer to the consistent characters problem could be designing an image generator specifically made for doing template grids of characters, that has many angles, poses, and expressions. And also a way to retake that same template and make a way to edit just clothes, hair styles, even scars. With that, then we would need an AI image/video generator that can take many templates of characters for making them interact with each other. Finally, it would be great to have also templates for places, like a home, a forest, an office. Even objects, like swords, specific clothes, meals, etc. I think that could be the easiest way for the end-user for making consistent media storytelling.

isaacbegue
Автор

For consistent characters as someone who has worked with A.I. 3D models since the beginning, you need to create a 3D model as a hard reference and then everything else could be generated normally. Otherwise your always going to have warping with people in motion. It would also fix the hand warping issue 💯!

jakehenne
Автор

@8:50 I think what's most impressive is the consistency of "Nirvana" under the jacket even though its covered up several times.

cobaltblue
Автор

Exciting times!!!
The pace of improvements is incredible

BarnabasLawson
Автор

The groundbreaking thing about the O1 model, in my understanding, is its ability to scale over time. Imagine this: our brains are capable of quick responses, mostly about things we have remembered at some point, like fundamental mathematical operations (e.g., multiplication tables). Current LLMs work similarly. We ask a question, and it immediately provides an answer.

But our brain is also capable of solving larger problems if we take more time to think about them.

The O1 model works in a similar way. Instead of giving a fast answer, it "thinks" about the problem to reach a conclusion.

The breakthrough here is that the O1 model is capable of scaling over time, rather than just over parameters.

And, as Sam Altman said, they want to extend the time it spends thinking about a problem to hours, then days, and even up to weeks.

That's the amazing thing.


---

Could an engineer build a car in one week?

No.

Could a thousand engineers build a car in one week?

Probably.

Could one engineer build a car in ten years?

Yes.

sesamring
Автор

I find the concept of "world models" really intriguing. If a sophisticated enough world model were to exist, it might look a lot like the world we live in....

koroyaku
Автор

5:25 This part is about being able to combine people, objects, and environments together in a video.

chariotsx
Автор

I am eagerly waiting for your Suno V4 demo. Few already got an early access. I hope you do too.

Africa_Nexus
Автор

Thanks Matt, God bless everybody ❤4rmZambia 🇿🇲, the comp update of vidu looks interesting for sure and yes they have kind of solved it for sure 🎉

Copa
Автор

This video is incredibly insightful! The advancements in AI video generation are truly promising. I can't wait to see what the future holds. Thanks for the content. Did you ever try AI VR experiences? it's wild

eypacha
Автор

The ultimate in consistency will be for the AI to build the scene and characters in 3d, and then animate for video. The video would then have an additional pass to smooth the 2d image that's generated and remove "artifacts". Once the 3d assets are created, they'll always be available to use again later. It's important to recognize that the image generators we currently have DO NOT do this. They generate 2d images frame after frame, which presents the potential for morphing and distortion as the video progresses from frame to frame. That doesn't happen with 3d assets.

HostileRespite
Автор

Now just imagine generating an image in Midjourney, making it talk with Runway Act One, then add camera orbiting in DimensionX!

XerazoX
Автор

14:37 Hey bro this is the real deal, this tool will do very well. Can you imagine how many people will buy more clothes plus they'll feel more satisfied with all their purchases. The more people get to see this Ai tool the more it will get popular. 15:05 People will love to see their size and body shape when using this Awesome tool. 16:23 People will love to see whether certain clothes color suit them, and if certain outfits agree with their body. 20:49 And to add to this concept, if the Ai Chatbot can ask you the user relevant and related questions to your main question then it can develop a deeper and clearer context to return the accurate answer to you.

aliettienne
Автор

That thumbnail was misleading . Consistent characters he only mentioned For a couple of minutes in the video And that was the main reason why I clicked on the video., and that was only for a few facial expressions . What about constant body and costume design Camera angles and lighting

dwainmorris
Автор

12:28 Here's my imagined use. You have a shaky video, stabilize video without having to zoom in, then autofill the edges so it looks whole.

rmt
Автор

This AI gaming stuff should be done with 2d platform games first for its reduced processing and I’m sure that would be FIRE!!!

goodtimesdays
Автор

Most humans do mind modeling of other actors very early on mom/dad and the generalized minecraft imagination to play out possible scenarios is what reseches have been studing as world models to predict physics, and should work in conjuction with some internal monologue like o1 has

SinCityRadio
Автор

Could you have imagined that when you started this channel playing Minecraft, you'd be playing in a real-time, AI-simulated Minecraft world 11 years later?

aexetanius
Автор

“Holy shit! This guy’s taking Roy off the grid! This guy doesn’t have a social security number for Roy!”

a_random_voice_in_the_void