Stable Diffusion in Code (AI Image Generation) - Computerphile

preview_player
Показать описание
Mike Continues his look at AI Image Generation with Stable Diffusion


This video was filmed and edited by Sean Riley.


Рекомендации по теме
Комментарии
Автор

If anyone is stuck with the code. The "i" should be a "t" in this line in the loop:
```
latents = scheduler.step(noise_pred, i, latents)["prev_sample"]
```

BernardJollans
Автор

I've been using Stable Diffusion to _deCGI_ images. Take a screenshot from a game, run it through SD with a low noise rate, give it a detailed description of everything in the picture and it produces pretty solid photo recreations of the images. Also, often, it gets possessed by Eldritch gods and spews out monsters.

DampeSN
Автор

The very concept of embeddings is amazing to me. It's literally "organize concepts themselves into points in space, where similar things are closer together, in many many dimensions; now you can do arithmetic on *the meanings of words, phrases, and sentences.* " Want to add the meaning of "horse" and the meaning of "male"? Well, just add these vectors together and the resulting coordinates will point right at "stallion"!

They amaze me so much that, when I watched Everything, Everywhere, All At Once for the first time, I completely geeked out when I realized their description of the organization of the multiverse is effectively a well-embedded latent space 😅

IceMetalPunk
Автор

I've been playing with Stable Diffusion (specifically the "InvokeAI" fork because I don't have 10gb VRAM), and I've found out that spamming the end with keywords like "realistic, 4k, trending on artstation, 8k, photorealistic, hyperrealistic" has more effect on how good the output image is than I thought.

YSPACElabs
Автор

this guy makes sense. I want more of him teaching SD and how it works.

bustedd
Автор

I'm sorry but, "unlock your face with your phone" just cracked me up..

paulspaws
Автор

Mike asked himself what the use case for mixing two prompts is.
I used this only yesterday, to produce a photorealistic painting of an owlbear from DnD...

So it has practical uses!

morphman
Автор

I love it how you simplify and explain this heap of complexity that is in generative models like this. You gave me the impulse to play around with it, inspite of being pretty complicated code due to the depth of the abstraction. It's a lot of fun to fantasize about something and have the model come up with a visual representation.

byteborg
Автор

"there are questions of ethics, there are questions on how it's trained. Let's leave those for another time"

well, if that doesn't just sum up the tech industry.

paultapping
Автор

SD is just outstanding. It can mimic the other projects and the 1.4/1.5 models will be public domain. You can't beat that.

jeffwads
Автор

I really liked the stable diffusion that came with the webui that you could install on your own computer, to avoid quotas or subscription costs, and it provided easy to use UI as well. With inpaint feature inside the UI as well. Shoutouts to people who make those applications from the rough code for regular people to use.

Yupppi
Автор

"Simple, you just chip away all the stone that doesn't look like David."

christopherg
Автор

I would like to see a version of the code where it shows the result of each step, so you can see the noise getting reduced with each iteration

simplesimon
Автор

I like how your channel has adapted to the advent of the machine learning boom we are experiencing

lucamatteobarbieri
Автор

Love Mikes explanations, somehow he manages explain so complicated stuff in so simple and understandable way.
It will be interesting to know Mikes opinion om Midjourney as it's seems like the winner for now among the picture creation AIs.

jenka
Автор

The current version of the reference notebook is already deprecated due to Hugging Face's API changes :)
You try to operate on "image", which is now a DecoderOutput class:
image = (image/ 2 + 0.5).clamp(0, 1)

It is fixed by unpacking its tensor attribute with its sample method:
image = (image.sample / 2 + 0.5).clamp(0, 1)

thomasnicolet
Автор

So amazing ❤ I love stable diffusion
Playing around the few last weeks

serta
Автор

This video finally explained the code to me in a simple way! Now im less confused!!! Amazing extra documentation from you guys

aorusaki
Автор

Mike is a legend, truly great videos with him

angeleeh
Автор

Immediately recognized the book on Dr. Ponds desk - Prof. Paar was one of my teachers when I studied IT sec. Nice to see it outside of Germany too!

DeKubus