Video consistency in stable diffusion can be optimized when using control net and EBsynth. In this tutorial, I’ll share two awesome tricks Tokyojap taught me and introduce a free generator for creating a grid from your images.
It’s a crucial step to achieve flicker free and consistent AI animation. Tokyojap uses the text2img tab instead of the img2img tab in stable diffusion, which allows for incredible style transfer in high resolution quality.
He calls it the temporal consistency method. It’s the best flickerfree AI video technique I’ve come across in 2023, made possible with stablediffusion, controlnet, and EBsynth. Get ready to take your videos to the next level with these powerful tools!
0:00 Introducing Tokyojab’s 2 hacks
0:38 Showcasing Tokyojab’s videos
2:30 Start Tutorial
3:36 Export video into img sequence
5:25 Making the grid from 4 images on Sprite sheet packer website
6:17 Settings in Stable diffusion and Control net
7:07 The 3 civit.ai models Tokyojap uses + the installation
7:48 The VAE (variational auto encoder) + the installation
8:39 Prompting in Stable Diffusion
10:17 Cutting the grid into 4 images with ezgif sprite sheet cutter
11:01 Create the images in EBsynth
11:57 Stitching the images together in davinci resolve 18
13:03 What is in the 2nd Tutorial
-Tokyojab instagram: https://www.instagram.com/tokyojab/
-Tokyojab Reddit post:
The 3 civit ai models:
-Art & Eros
Installing the vae model, on the hugging face website
Sebastian Kamph’s Installation of Stable diffusion automatic 1111 webui
Creating the grid and cutting it:
EBsynth Free Software
DISCLAIMER: No copyright is claimed in this video and to the extent that material may appear to be infringed, I assert that such alleged infringement is permissible under fair use principles. If you believe material has been used in an unauthorized manner, please contact the poster.