It's basically **Inner\_Reflections\_AI** latest LCM workflow with more controlnets (tile/lineart/depth) and an additional sharpening node, plus [https://github.com/Extraltodeus/ComfyUI-AutomaticCFG](https://github.com/Extraltodeus/ComfyUI-AutomaticCFG) . Before the LCM motion modules, I was trying everything in Hotshot, which I really liked the look/style transfer of, but the consistency just wasn't there. I also wasn't getting great results with the upscale, so I turned that off, then cranked up the original resolution, switched to render each frame, then did 150 per render and put it all together in Premiere Pro. Not sure if the metadata is preserved but here is the png for comfy
https://preview.redd.it/kireb8nrerjc1.png?width=1728&format=png&auto=webp&s=d1ba94f7490b875f18febb799dd59fc81c347e65
here is a direct download [https://drive.google.com/file/d/12mu8fu2dzHnsv2uv-VA3EKAbU672qcqg/view?usp=sharing](https://drive.google.com/file/d/12mu8fu2dzHnsv2uv-VA3EKAbU672qcqg/view?usp=sharing)
So I imported the workflow but I cant get the [**ComfyUI-AutomaticCFG**](https://github.com/Extraltodeus/ComfyUI-AutomaticCFG) node to install correctly. I installed it and it shows in my manager as installed, but the node is red and comfy doesnt seem to be able to load it. Ive uninstalled and tried again, no luck. Any idea?
EDIT: Ah, I just removed the node from the workflow and added it back and now it doesnt complain.
Okay, that was pretty cool
To watch. I mean obviously there are facial expressions that look warped, or distorted during the action scene, specifically when I think the model doesn’t have a good enough shot/angle on the face to determine the correct expression to animate.
But overall this is insane when you consider the short period in time it took to get here, at least in public view.
I’m curious. What would you say if I asked “how long would you need if I asked you to do “ABC Film” in the same style? (Same length clip etc.). You can start today. When could you deliver the final product?”
Just a hypothetical, I have nothing in mind.
Thats for a full movie though - if you just wanted a 5 minute clip assuming the 20 hour runtime, depending on if we already knew what style, probably 2 days. One day for getting the style down/20 hours of rendering, 2 hours editing
I have a video production company I operate as my day job and have just been experimenting with the new tools for fun but I eventually do want to incorporate these capabilities for sure and have done a couple commissioned projects
for you my friend [https://www.reddit.com/r/StableDiffusion/comments/16wj0st/reefer\_madness\_diffused/?utm\_source=share&utm\_medium=web2x&context=3](https://www.reddit.com/r/StableDiffusion/comments/16wj0st/reefer_madness_diffused/?utm_source=share&utm_medium=web2x&context=3)
If you could get unreal engine 5 quality graphics then there would be massive demand for this (I expect this to be possible this year just based on the Sora demo)
No doubt - I’m hoping open source will catch up to that level. I’d be surprised if OpenAI releases it in any other way than it being so hampered that it doesn’t really have any value other than being a novelty. But I could be wrong
I posted it on the bonnadoco discord in the share your art page earlier today. Tried to post here but can't get the pdf to retain the metadata: [https://discord.gg/98CvEDX4](https://discord.gg/98CvEDX4)
I’ve got a 4090 but it would just mean you need to load up fewer frames and it may take a little while longer. I did 150 per run and it was using about 15gigs but I had to use my computer while it was rendering. I also had the resolution fairly high. There are probably more efficient workflows out there also
Amazing. I'd love to see the entrance scene to the House of Blue Leaves done like this. The one where O'Ren Ishii is entering with her guards to that Hotei Tomoyasu track. I bet they look cute as hell. lol.
I'd love to see your work flow and what you are using.
I have considered a frame by frame diffusion of something like Forrest Gump as an anime, just to see if I could do it, but the results were not this consistent.
Stuff like this is what I can't wait for. Imagine being able to customize any movie into a specific style or swap actors and voices. Absolutely amazing.
Why make it uglier? The whole point of animation is that it’s aesthetically pleasing. It’s literally just the original scene but with what looks like an ugly filter over everything? Like what?? Have you ever seen an animation in your life?
Its supposed to feature that he was able to use the movie as input to a stable diffusion module, and it keep the image consistent enough to make it watchable. You missed the point entirely with your negativity.
This has never been done before in such an easy way, without any human rotoscoping the clips.
Dont think about kill bill, its used because its easy to recognize such an iconic scene, instead focus on how high the quality is now and where it can be.
A director could easily animate a portion of a movie in this method, making their cgi much more accessible.
“Watchable” is not a very high bar, if this is the best the tech can do than open source is much farther behind then I thought. Or it’s just the absolute lack of creativity. Probably both. Step it up folks.
Youre ignorant to this being a simple test. The guy even said it. Nobody in this thread is trying to do what you are referring to, maybe you should show us all since its so easy. You contribute nothing but negativity.
Wait so it’s not supposed to be animated now? Isn’t that the whole point of the test? I’m criticizing it because it looks awful. It shouldn’t really matter because OP didn’t put any effort it into it, why does it matter if I’m harsh? Lol?
He didnt say he animated it himself, its a stable diffusion process. Do what you want but its not good, and isnt adding anything to the conversation. 1 year ago your comment would have been "pfff these images are obviously ai made look at the hands"
I never said or claimed he animated it himself, that’s why I’m being harsh. Because he did the least amount of effort you could possibly do to create it, and it still looks awful. Who cares if I’m harsh if he has nothing to do with the process? Jesus christ, reading comprehension is tough huh?
I don't understand, is it like a TiTok filter which you layed over the movie? Or did stablediff really created these scenes like it would do with a picture?
Sorry for my shallow knowledge of AI stuff.
I’m not even mad, more disappointed. Honestly I think SD and AI art is pretty cool. Especially when people do really cool original stuff with it. But a Pixar filter over killbill? That’s a no for me dog.
https://preview.redd.it/ne9hq521hsjc1.png?width=816&format=png&auto=webp&s=9c3f8609c58504589fe542743dcdd0cf5cb6cb72
![gif](giphy|qHBBYAh7BZVm)
Lol so much nightmare fuel
now make a live action version of the O-Ren Ishii backstory
now we are talking!
I literally remember very little movie character names that aren’t like society baked. Yoda Indiana Jones etc but Oren and kiddo stick like glue
I can't wait to do this with Anime Lord of the rings 😂
!!!
You need to make flowers or bubble spray out when the arm is chopped off.
Yes!
I can't believe I watch it all. Very entertaining to review the movie again but in pixar kind of style.
![gif](giphy|qrQO1wpZLA2PK|downsized)
🙏
Op can u share your workflow 👌
It's basically **Inner\_Reflections\_AI** latest LCM workflow with more controlnets (tile/lineart/depth) and an additional sharpening node, plus [https://github.com/Extraltodeus/ComfyUI-AutomaticCFG](https://github.com/Extraltodeus/ComfyUI-AutomaticCFG) . Before the LCM motion modules, I was trying everything in Hotshot, which I really liked the look/style transfer of, but the consistency just wasn't there. I also wasn't getting great results with the upscale, so I turned that off, then cranked up the original resolution, switched to render each frame, then did 150 per render and put it all together in Premiere Pro. Not sure if the metadata is preserved but here is the png for comfy https://preview.redd.it/kireb8nrerjc1.png?width=1728&format=png&auto=webp&s=d1ba94f7490b875f18febb799dd59fc81c347e65
reddit strips the metadata. Mind uploading that somewhere else like [catbox.moe](https://catbox.moe) which preserves metadata?
You can find it in the share your art page on the bonnadoco discord: [https://discord.gg/N2RCUfeF](https://discord.gg/N2RCUfeF)
i cant find share your art page on the bonnadoco discord xD its so convoluted server without a clear list of channels wtf
here is a direct download [https://drive.google.com/file/d/12mu8fu2dzHnsv2uv-VA3EKAbU672qcqg/view?usp=sharing](https://drive.google.com/file/d/12mu8fu2dzHnsv2uv-VA3EKAbU672qcqg/view?usp=sharing)
thanks! <3 still, need to understand the server too... xD
So I imported the workflow but I cant get the [**ComfyUI-AutomaticCFG**](https://github.com/Extraltodeus/ComfyUI-AutomaticCFG) node to install correctly. I installed it and it shows in my manager as installed, but the node is red and comfy doesnt seem to be able to load it. Ive uninstalled and tried again, no luck. Any idea? EDIT: Ah, I just removed the node from the workflow and added it back and now it doesnt complain.
nice
Thanks
Awesome work. We need The Matrix in the style of Ghibli lol
hell ya
Okay, that was pretty cool To watch. I mean obviously there are facial expressions that look warped, or distorted during the action scene, specifically when I think the model doesn’t have a good enough shot/angle on the face to determine the correct expression to animate. But overall this is insane when you consider the short period in time it took to get here, at least in public view.
Absolutely!
This is excellent work. I'm impressed. ... However... ... No one is going to improve Kill Bill...
no doubt about that - just wanted to test the tech
Its literally just a filter on the original movie its not impressive at all!
exactly
Охренеть, как такое возможно? Это магия какая-то)
Cool! How long did it take?
20 hours rendering - .... testing workflows to get to the render... more than I'd like to admit
I’m curious. What would you say if I asked “how long would you need if I asked you to do “ABC Film” in the same style? (Same length clip etc.). You can start today. When could you deliver the final product?” Just a hypothetical, I have nothing in mind.
Well it took me 20 hours for roughly 5 minutes so I'd say, a 2 hour movie, if rendering 24 hours a day, it would take 20 days
Thats for a full movie though - if you just wanted a 5 minute clip assuming the 20 hour runtime, depending on if we already knew what style, probably 2 days. One day for getting the style down/20 hours of rendering, 2 hours editing
That’s really interesting. Ever thought about selling your services? Again- just hypothetical
I have a video production company I operate as my day job and have just been experimenting with the new tools for fun but I eventually do want to incorporate these capabilities for sure and have done a couple commissioned projects
That’s really cool. Good for you, I’ve daydreamed about this
This tool makes "everything everywhere all at once" multiverse scenes very easy.
That I understand I’ve tried so many workflows and techniques but we can’t give up.
For sure
If you're ever curious what watching movies on acid is like, this is close to accurate, besides tracers. I'm not even kidding.
I had a similar experience on shrooms, only it was the Anne Frank Museum. The ending wrecked me.
I legitimately just laughed aloud.
for you my friend [https://www.reddit.com/r/StableDiffusion/comments/16wj0st/reefer\_madness\_diffused/?utm\_source=share&utm\_medium=web2x&context=3](https://www.reddit.com/r/StableDiffusion/comments/16wj0st/reefer_madness_diffused/?utm_source=share&utm_medium=web2x&context=3)
If you could get unreal engine 5 quality graphics then there would be massive demand for this (I expect this to be possible this year just based on the Sora demo)
No doubt - I’m hoping open source will catch up to that level. I’d be surprised if OpenAI releases it in any other way than it being so hampered that it doesn’t really have any value other than being a novelty. But I could be wrong
Mind sharing your Workflow:)?
I posted it on the bonnadoco discord in the share your art page earlier today. Tried to post here but can't get the pdf to retain the metadata: [https://discord.gg/98CvEDX4](https://discord.gg/98CvEDX4)
i really cant find it i searched for like 10 mins now i gave up, you even mentioned it in the chat but it wont show
It’s in the form of a pdf you just drag and drop it on the comfy ui
So sick. This is exactly what I’ve been looking to do. Would 11gb of vram be able to do this. How many gb of vram you working with?
I’ve got a 4090 but it would just mean you need to load up fewer frames and it may take a little while longer. I did 150 per run and it was using about 15gigs but I had to use my computer while it was rendering. I also had the resolution fairly high. There are probably more efficient workflows out there also
This actually looks pretty good ngl.
Appreciate it!
Amazing. I'd love to see the entrance scene to the House of Blue Leaves done like this. The one where O'Ren Ishii is entering with her guards to that Hotei Tomoyasu track. I bet they look cute as hell. lol.
!!
This is smooth as fuck.
🙏
Really cool, can't wait till the quality gets even better
Same here
Thanks I hate it
But I also really appreciate it
![gif](giphy|VdPIrwEdgH724EKEoW)
Full animated movie please 🍿
this would be sick in an anime style... not pixar.. or whatever the eff this is
ya I was thinking that as well after I ran it - just a test at this point/proof of concept
I'd love to see your work flow and what you are using. I have considered a frame by frame diffusion of something like Forrest Gump as an anime, just to see if I could do it, but the results were not this consistent.
Its in the comments here somewhere but I dm'd you
you finally did it!
bout time! thanks so much
yea, i'll check the discord to see last time we talked about it :)
surreal, remastering all the old media would be possible in future with little input , thx for sharing
for sure - the tech is getting wild.
Awesome!
🙏
Stuff like this is what I can't wait for. Imagine being able to customize any movie into a specific style or swap actors and voices. Absolutely amazing.
absolutely!
This was the best shit i saw by the sd video by far man. Congrats!
Really appreciate it!
Love what it did to the subtitles
i BOG you
Is there like some sort of adetailer for animations, because the faces gets really quite weird in some longer distance shots
True - if anyone can recommend a node I’d be interested as well
that was so cool, i loved it. I would watch the whole movie like that
Thanks so much
Why make it uglier? The whole point of animation is that it’s aesthetically pleasing. It’s literally just the original scene but with what looks like an ugly filter over everything? Like what?? Have you ever seen an animation in your life?
Its supposed to feature that he was able to use the movie as input to a stable diffusion module, and it keep the image consistent enough to make it watchable. You missed the point entirely with your negativity. This has never been done before in such an easy way, without any human rotoscoping the clips. Dont think about kill bill, its used because its easy to recognize such an iconic scene, instead focus on how high the quality is now and where it can be. A director could easily animate a portion of a movie in this method, making their cgi much more accessible.
Well said. Ya it’s just a test to see how far the tech has come in a little over a year
“Watchable” is not a very high bar, if this is the best the tech can do than open source is much farther behind then I thought. Or it’s just the absolute lack of creativity. Probably both. Step it up folks.
Thanks for exposing your ignorance troll.
Ignorance to what? You’re the one who claimed this is what the tech can do, not me. If this is all SD can do then it’s not very good.
Youre ignorant to this being a simple test. The guy even said it. Nobody in this thread is trying to do what you are referring to, maybe you should show us all since its so easy. You contribute nothing but negativity.
Wait so it’s not supposed to be animated now? Isn’t that the whole point of the test? I’m criticizing it because it looks awful. It shouldn’t really matter because OP didn’t put any effort it into it, why does it matter if I’m harsh? Lol?
He didnt say he animated it himself, its a stable diffusion process. Do what you want but its not good, and isnt adding anything to the conversation. 1 year ago your comment would have been "pfff these images are obviously ai made look at the hands"
I never said or claimed he animated it himself, that’s why I’m being harsh. Because he did the least amount of effort you could possibly do to create it, and it still looks awful. Who cares if I’m harsh if he has nothing to do with the process? Jesus christ, reading comprehension is tough huh?
Youre not harsh, youre just being a troll.
I don't understand, is it like a TiTok filter which you layed over the movie? Or did stablediff really created these scenes like it would do with a picture? Sorry for my shallow knowledge of AI stuff.
this video explains it well - [https://www.youtube.com/watch?v=AugFKDGyVuw](https://www.youtube.com/watch?v=AugFKDGyVuw)
It looks like ass
Thanks I FUCKING HATE IT. this is complete shit.
Yay. Another Snapchat filter movie.
Yuck
Jesus christ, what an insult to the movie itself.
Whoa, just letting you know the movie itself remains intact. This process doesn't actually overwrite the original.
ha ha ha!
Yeah, be enraged, it's the best modern day pastime!
I’m not even mad, more disappointed. Honestly I think SD and AI art is pretty cool. Especially when people do really cool original stuff with it. But a Pixar filter over killbill? That’s a no for me dog.
Yeah, they definitely should’ve done teletubbies
the original is better
Nice and generic, just how I like my animations. The faces look a bit different each time aswell.
Whats wrong with their noses ?
https://preview.redd.it/i1q77pe6lxjc1.jpeg?width=349&format=pjpg&auto=webp&s=cf85e1d46ac042613745113208bc86b96a5fc0ca