Stable Diffusion Deforum: Going Viral on TikTok & Instagram | Arnold Oberleiter | Skillshare
Search

Playback Speed


1.0x


  • 0.5x
  • 0.75x
  • 1x (Normal)
  • 1.25x
  • 1.5x
  • 1.75x
  • 2x

Stable Diffusion Deforum: Going Viral on TikTok & Instagram

teacher avatar Arnold Oberleiter

Watch this class and thousands more

Get unlimited access to every class
Taught by industry leaders & working professionals
Topics include illustration, design, photography, and more

Watch this class and thousands more

Get unlimited access to every class
Taught by industry leaders & working professionals
Topics include illustration, design, photography, and more

Lessons in This Class

    • 1.

      Introduction

      2:04

    • 2.

      What is Stable Diffusion

      2:01

    • 3.

      The Tecnologie behind Stable Diffusion: The Diffusion Model

      3:08

    • 4.

      Why we need Stable Diffuison

      1:54

    • 5.

      What are Models, Leonardo AI as example

      3:35

    • 6.

      What is the Seed, Midjourney as example

      3:10

    • 7.

      What are Frames

      1:41

    • 8.

      Prompt Engineering in Stable Diffuison

      4:37

    • 9.

      Finding Inspiration on Lexica and Leonaro AI

      2:42

    • 10.

      Join the Deforum Discord Community

      0:52

    • 11.

      Local Installation of Stable Difusion with Python

      1:55

    • 12.

      Rundiffuison: A Cloud solution that works

      1:45

    • 13.

      Colab Notebook and Google Drive

      5:38

    • 14.

      The Pricing in Google Colab

      2:32

    • 15.

      The Path Setup for the Models and Output

      1:46

    • 16.

      How Deforum works

      2:11

    • 17.

      Overview and Quick Guide

      2:52

    • 18.

      The Basics of Animations: Angle and Zoom as example

      4:11

    • 19.

      Kamera movement in 2D and 3D

      3:22

    • 20.

      What does the formula in the brackat mean

      4:18

    • 21.

      Noise schedule and strength schedule

      4:51

    • 22.

      Coherence and cadence make the videos smoother

      2:17

    • 23.

      UPDATE: Version 0.7.1 is here!

      1:41

    • 24.

      The prompting cell and negative prompts

      5:30

    • 25.

      Load and Image settings and the resolution

      2:58

    • 26.

      Sampling Settings, the seed and the steps

      3:10

    • 27.

      Batch settings and prompt weights

      3:41

    • 28.

      The last settings befor the rendering

      3:26

    • 29.

      Render your first video

      6:17

    • 30.

      Letting people age and the resume animation

      9:24

    • 31.

      Your learning experience

      0:40

    • 32.

      Init Image: start with your own foto as keyframe

      9:51

    • 33.

      Start with your own videos as starting keyframe

      6:53

    • 34.

      Video Input: Lay prompts over videos

      9:14

    • 35.

      The firs 3D video

      8:52

    • 36.

      ChatGPT as personal assistent

      8:08

    • 37.

      Use other and specific models

      3:43

    • 38.

      Keystring generator for faster workflow with deforum

      11:45

    • 39.

      Audio keyframe string generator synchronize music and animation

      12:08

    • 40.

      Complete Projekt

      14:52

    • 41.

      What you shold create and where you can get inspiration

      3:02

    • 42.

      Learning experience in 3D

      2:43

    • 43.

      Kaiber a solied and easy alternative

      3:16

    • 44.

      Loopsie and disco or wrab diffusion

      1:55

    • 45.

      Stable Diffusion [Wrapfusion] in Google Colab

      15:49

    • 46.

      What is next: Create and share your animations

      0:57

    • 47.

      Thank you

      0:50

  • --
  • Beginner level
  • Intermediate level
  • Advanced level
  • All levels

Community Generated

The level is determined by a majority opinion of students who have reviewed this class. The teacher's recommendation is shown until at least 5 student responses are collected.

220

Students

1

Project

About This Class

Would you like to learn how to create viral videos using the latest technology (like Stable Diffusion and ChatGPT), in the world of video creation for social media, Instagram, and TikTok?

Or do you simply want to become an expert in Deforum and showcase cool videos to your friends?

Then this course is perfect for you!

In this course, we will delve into the technology of Stable Diffusion and how to utilize it to create videos that can go viral on social media.

We will extensively explore the possibilities of Deforum Stable Diffusion and together, discover how this technology can be used to produce engaging videos for platforms like Instagram and TikTok.

The course provides a comprehensive introduction to using Deforum for video creation.

We will learn the fundamentals of this technology and compare how it differs from other tools. Additionally, we will become familiar with the user interface and structure of the platform.

A key focus of the course is on creating effective instructions and learning how to adapt inputs to achieve better results. We will also cover all essential aspects of parameters, including camera settings, camera angles, different perspectives, zooming, and proper prompt engineering for Stable Diffusion.

You will learn how to optimize camera settings to evoke the desired mood and impact in your videos. We will explore various shooting angles and discover how they can influence perception and visual experience.

Furthermore, we will delve into prompt engineering to achieve the best results.

You will learn how to structure inputs and instructions correctly to achieve the desired video outcome.

To fully grasp the potential of Stable Diffusion, we will also acquire the necessary technical background knowledge. You will understand how a diffusion model works, the role of frames and seeds, and much more.

Through detailed instructions, you will be able to apply what you have learned and create your own videos that can go viral on social media platforms.

This course offers you the opportunity to expand your video creation skills and develop a deeper understanding of new AI tools.

Whether you are a beginner or an experienced video producer, this course will help you take your videos to the next level and achieve the desired impact on social media.

The course also provides quick support to ensure you get the most out of your learning process.

If you are ready to unleash your creativity, conquer the world of Stable Diffusion Deforum, and viral videos on social media, then this course is perfect for you.

See you in the course

Meet Your Teacher

Level: All Levels

Class Ratings

Expectations Met?
    Exceeded!
  • 0%
  • Yes
  • 0%
  • Somewhat
  • 0%
  • Not really
  • 0%

Why Join Skillshare?

Take award-winning Skillshare Original Classes

Each class has short lessons, hands-on projects

Your membership supports Skillshare teachers

Learn From Anywhere

Take classes on the go with the Skillshare app. Stream or download to watch on the plane, the subway, or wherever you learn best.

Transcripts

1. Introduction: Hi, my name is RNA and in this course, I will teach you how to make this cool videos in the forum diffusion. So do you will not learn how to go viral on TikTok and Instagram. Although you just want to become an expert in stable diffusion and show your friends code videos, then this course is just right for you. In this course, we will learn about stable diffusion and how to use the forum to make videos that can go viral on social media. We will cover all important aspects of the parameters including camera access, camera rotations, sue, Right prompt engineering for stable diffusion and much, much more. You will learn how to properly structure the inputs to get the perfect videos to understand the full potential of stable diffusion, we also learned the necessary technical background knowledge. You will understand how a diffusion model works, what roles frames and seeds play, what key-frames during generators are, what they are, coherencies, and much, much more step-by-step instructions, you will be able to apply that learned knowledge perfectly in Brexit. Whether you're a beginner or unexperienced video producer. This course will give you the necessary tools to make good animations for the desired effects on social media. If you are ready to go viral or just want to learn, then see you in the class. There's no mercy again, this just hungry or thirsty. Parsons in different versions each do one question worsens why. We'll practice carton and you'll see the different Berman 2. What is Stable Diffusion: What is stable diffusion? The first question that comes up is maybe what's they will diffusion is stable. Diffusion is simply a diffusion model that's open-source. But what does it mean? Open-source means simply that everybody has access to it. Everybody can see the code and everybody can build on the code. The forum is one extension that somebody has made on stable diffusion. And the diffusion model is simply just something that makes your pictures, you gave it a prompt. Prompt is simply that the UPF, the diffusion model and the diffusion model gives you pictures. And we need to go to some of this background information because it will be important later in the course. Here behind me is a little Wikipedia post. And it says some basic things about stable diffusion. For us, it's only important that stable diffusion is diffusion model. You can find it on GitHub, it's written in Python and the rest you will learn in this course. So stable diffusion is diffusion model, just like meat journey and Dolly. But the big difference is that stable diffusion is open-source. Everybody has access to the code. Everybody can build on stable diffusion. And stable diffusion doesn't run in a Cloud. You need your own GPU to use stable diffusion, or you can borrow some GPU from Google or from runway. We will see all these things in detail in this course because I think it's smart to work with Google Colab. If you don't want to work with Google Colab, you need a lot of GPU on your computer, but it's possible. So what do you take from this video? From this video you'll take that stable diffusion, make pictures out of text, out of our brand. And that's it. 3. The Tecnologie behind Stable Diffusion: The Diffusion Model: As you know, stable diffusion is diffusion model. Who would have thought that? How does a diffusion model works? It's really simple if we make it simple. So just take a look at this picture. You see here a cat and here is the noise. And think of it just this way. You give the diffusion model a lot of pictures, possibly all pictures that you can find on the internet. You will give the diffusion model, cats, dogs, and everything that you know about and you will describe it. In example of this picture, we will say maybe a gray cat with a pink nose. And you describe everything that you see in the background where the focus is. Maybe also some camera details with what camera you have shot this and so on. And on this picture, we will add a little bit of noise, just like here, but with always the same text. And you go on with that until you have only noise left. In this process, the machine will learn how a cat, we'll look how this text will add up to something like that. And after that, if you give the machine only in the noise that text, the machine knows what this text means. This text means that you will see a cat. And if you do this long enough and often enough, the machine will basically understand how things look, what text will lead to what pictures, what prompt lead to what pictures. And that's important because we need to describe in the Bronx exactly what we want to see. We say in the prompt, only our cat, we will get a random cat because the machine doesn't know what Kat we want to see. We need to be specific. We want to have a gray cat that sits on the beach. For example, the sun is rising, the ocean is blue. And we can add some prompt magic eight k high detail and so on. We will look at the prompting later. But just a moment back to the diffusion model. I describe it like this. Just think of it this way. You'll look and this guy, and you'll see some clouds. Maybe your girlfriend stands right behind you. And Jesus, can you see the apple in the clouds? And you don't see an apple. But then she points at the clouds and say, Just look at it. It looks like Can I pull? And then you start to realize, you'll see on that ball in this guy, obviously there is no apple, but your girlfriend gave you the prompt. You are now primed to see an apple. Your brain knows how Apple's look like, and your brain now see the apple in the sky, even if there is no apple there. And that's exactly how a diffusion model works. Just broke down in simple, simple terms. But that's enough for us. 4. Why we need Stable Diffuison: Why do we need stable diffusion? I already told you that their style eat meat journey. And there's also Leonardo ai. And you can make really, really good images on these diffusion models. Leonardo ai is one diffusion model that runs on stable diffusion. So there's one example you need it to run other things on it. And the forum is one thing that runs on stable diffusion. And with different room you make this cool videos. So if you just want to make normal pictures, you will most likely use my journey dolly or Leonardo. Leonardo AI works also unstable diffusion and it's really, really easy to use. We will take a brief look at Leonardo also in this course. But we will use Leonardo most likely chest to steal some prompts. We can also steal prompt. We can use prompts, we can tweak prompts, and we will use them in the forum to make our videos. We will use these prompts, mostly from Leonardo and lexica. If you just want to make pictures used alley, meet, journey or Leonardo. But we need stable diffusion in the background. Stable diffusion is the backbone of a lot of cool features. Leonardo runs on it, the forum runs on it, and there are tons of other things that run on stable diffusion. You can even use other models and your own models to improve the quality of your pictures. We will see all of that in detail later in the course. So basically, we need stable diffusion to run things on it. And the forum is one thing that runs on stable diffusion and devoured them is awesome 5. What are Models, Leonardo AI as example: In the last video, you heard that unstable diffusion. You can run models. But what are models for that? We go in Leonardo and we'll take a brief look. I'm here in Leonardo and I'm on the fine-tuned models. But what does it mean? It means simply that stable diffusion is trained on specific pictures. So stable diffusion, in this case, the version 1.5 of stable diffusion, was trained on a specific data set of pictures to make pictures that look something like this. So if you have something in mind that looks similar to pictures of this, you should use a model that's specifically trained for this. You can also see that there's a training resolution. So this model is specifically trained to make pictures that look like this in our resolution of 640 times 832 pixels. And you can go on and on and look at different models. The Leonato diffusion model, for example, is drained, stable diffusion 2.1. It's made for resolutions of 1,024 by 1024 pixels, and it is known for incredible shaving and tie contrast. You make pictures that look like this. If you have something in mind that look like pixel art, you should use a model that looks like pixel art. If you want to make bad Lexus, maybe it's the best idea to take models that are specifically trained for bad Lexus. I think you get the idea. If you have something specific in mind, you can take a specific model for that. All this dancing style, this demons, and so on, that you'll see trending on TikTok. It would be maybe smart to use a model that is specifically trained to make this undead persons and the leonardo you have also another cool, cool feature. You can also make your own models. For example, here I have made two models. One model that I have trained specifically to make colorful flowers, and I have trained another model to make close-ups of insects. And I think in the future, you can also link your own models to default room, I think. And I hope Leonardo AI will give his API. And then you can link your own models, your own terrain models to D40. That will be a game changer If you could even train your girlfriend's face on a model and then link it to the forum and make videos. This specific person. So if you have a model that's trained on specific phases, you will get consistent output that will be important later, we will take a look how we can use models, and we will also look where we find models. We need to find models that we can download and upload in Google Drive. After the upload, we can use the models in default room. Most of them all these are for free on Hugging Face, and that's what we will not learn in this course. Later, we will download models, we will upload them in Google and we will use them in the ballroom. 6. What is the Seed, Midjourney as example: What is the seat and why we need it? The seed is simply something like this starting noise. That first signal that you're on diffusion model gets. If you give the diffusion model as specific seat, you can simply work more precise. You can tweak prompts and with the same seed, you will always get similar results. Just take a look at MIT journey and Leonardo first. This right here is a nice picture. We can see that prompts here, that negative prompts that will also be important later. And that right here, That's the seat. If we, for example, copy the seat and use the exec, same prompts, we will get the exact same picture. But if we use the same seed and simply tweak the brand a little bit, we will get a little bit different results. But thanks to the same starting noise, thanks to this seat, we will get really precise pictures. The picture will look just like the other or similar to the other. They can look at Germany and I will show you something. This dog's these two pictures right here. The right picture is simply simply a picture that I have made and the picture on the left is basically the same picture with the same seat, but I changed the prompts. So you can see you get pretty, pretty similar pictures, but they are a little bit different because we use the same seed, the same starting noise. We can tweak little, little things. This picture really similar to the other picture because of the same seat. But they can look at this picture. This picture right here, that's basically exactly the same picture as the previous one that I showed you. Because I use the same seed and the same prompt. So that same seat and the same prompt will give you exactly the same picture. And that's important. If we use us eat maybe a fixed date, we can work really, really precisely. We can tweak just little things. The same seed with the same prompt will give you basically the same picture. We want to have really, really consistent outputs. We need to work like that. And if we want to change chest little small things in the prompts, we will also get really consistent things. If we work with a fixed date, if we use random seeds, the outputs are more random. That's okay. That's totally fine. And I do that often. But if you want to fine-tune things, fixed it, make sense, or at least you should know the receipt. Later, we will learn more. But now you have a good basic understanding. What does the this and what it does and why we need it? 7. What are Frames: What our frames, for some, it's maybe common sense, but I need to make sure that we are all on the same page. Frame is simply one specific picture of a video. If you click pause my video and you'll see simply that one little picture of me. That's one frame of my video. You'll see something on a video camera like 30 FPS or 60 FPS. It simply means that the camera makes 30 or 60 pictures per minute. And all these pictures make up one video. Just think back on the first cartoons. What did people do? They use something like this. They draw some lines on the pages and they did this. All these pictures made a video and every page here is one frame example right behind me. Let's take this picture. I have made this picture in my journey, but you can imagine that this big check in real life would be in motion. So if you have a video of this, all these workers would work and not just stand here really lazy. But if you have justice that the picture, it's one frame, more frames made up a video. 12 FBS is 12 frames per seconds made up in a video. So just common sense that our frames and we need them 8. Prompt Engineering in Stable Diffuison: Prompt engineering in stable diffusion. As you know that Brown is just a text that you give stable diffusion to make your picture. And I already told you that it's important what text you give stable diffusion. Let's make a quick example here in Leonardo, right now, we work at stable diffusion version 1.5. Here we can choose our models. We take stable diffusion 1.5 for this example. Now here you can type in your prompt. Let's make an ugly little prompt. That's not perfect. The prompt is woman and press Generate. Now you have a picture of a woman, but the picture is not optimal. Because the prompt is not optimal, you need to be specific. You need to describe exactly what you want to see. Now I show you a good prompt. I'm thinking like this would work. A beautiful woman, masterpiece, realistic. 29 years old, beautiful face, dark hair, wearing a tight white shirts, studio lighting, cinematic light, eight K resolution. Whom? Human hands. Elegant, dynamic, highly detailed, and so on. Just be specific, GIF, Exactly, What do you want to see. Let's create this one. Now we have a picture that is more specific, but you may be see that the picture is not perfect, the hair is not perfect, the eyes are not perfect a lot than Nike's, maybe a little bit long. So we add a negative prompt. The negative prompt is the thing that you don't want to see in the picture. And in this example, we don't want to see a close up blurry eyes to face this plastic deformed, blurry but anatomic bad eyes, crossed eyes and so on. Just gave examples. What do you not want to see? Things like plastic, ugly, blurry, two phases, two hands, ugly, limps, text, and so on. That are all things that work is good because diffusion models, they usually don't make your texts. So in the negative prompt, you give the things that you don't want to see in the forum. It's not that important to use negative prompting because the videos are fast. So you don't see exactly what's going on. But It's important that you know, it's possible to use negative prompts. And if you want to make normal pictures, for example, in stable diffusion, you need to take negative prompts. Let's make this picture and you will see how much better the quality will be generate with the negative prompt. And now you see we have a real masterpiece. So negative prompts work really, really well in stable diffusion. First picture is a woman, and I must admit the woman is prettier than I would have thought normally, if you take a prompt that only says, Woman, you don't get really good results. After that, we go to prompt. But the last thing that missed in this prompt, whilst the negative prompt and I got prompt with a good negative prompt, you get really, really good results. Just look at this. Now we have a picture that looks really, really nice because of stable diffusion derived prompt and the right negative prompt. So in short, just describe exactly what you want to see. Use compositions. Use the resolutions. Maybe use cameras, use angles, use compositions. Describe exactly what you want to see. And in the negative prompt, describe what you don't want to see. Things like ugly blurry, two hats, two faces, deformed, ugly limps and so on. That are the classic things. Also long-neck, long body, just type in what you don't want to see. So basically, they'll stable diffusion while the woman z and what you don't want to see. We will need this later in the lesson. But I must admit the negative prompts are not that important thing in the forum, but you need to know them and we can copy a lot of prompts. And in the next video, I will show you where you can get some inspiration for your own prompting. 9. Finding Inspiration on Lexica and Leonaro AI: Getting inspiration and maybe also some prompts. In this lecture, I will show you where you can get inspiration for your own prompts or where you can copy them. You should go to showcases. One showcase is Leonardo, and I liked this showcase really, really much. You can take simply pictures that you like, maybe something like this. And you'll see that prompt and the negative prompt, this things you can copy. Another showcase is lexica and tear is basically the same thing. And you can search while you will not see. Maybe you want to see yeah, demon, just search it and search for the picture that you like. Click on it and press Copy prompt. If you want to copy them. You can also take likes right here. And after that, you can go to your likes. And you will see some picture that you already like. And you can simply copy the bronze of these pictures. Maybe you like this venom and you can simply copy the prompt if it's right for you. Same thing in Leonardo. Same thing in leonardo, you can simply search for something, maybe a zombie, and you will get mixtures, press on them and copy the prompts if you'll think they are nice. And now I want to show you just a quick, quick what the z thus press on this picture. We have already seen that, but I want to show it again because that's important. You can hear re-mix exactly this picture by pressing on this button. Now we generate this. As you see, we have two pictures that are pretty similar to this one. But what's the difference? The pictures are similar but they are not the same. But what if we simply copy also the seat? Just copy the seat. Will let the bronze and everything else the same. Now we use a fixed date. We activate this eat, we booked the seat right here and we press generate again. Now you'll see the picture is even more similar. This picture right here, you can say that is pretty much the same picture or stays right here. So the seed is really powerful in my mind, you can see the colors and everything they match, whether to this picture than the previous one. Now you know where you can get inspiration from. Use Leonardo, use lexica, both of these prompts, negative prompts. You can use both of them. And you also seen the seat in action 10. Join the Deforum Discord Community: Join the deformed community on this card if you want to, always be up to date, Of course, you can also join the community and you will see immediately what's going on in the forum. Simply press this right here. After that, you will have this default room in your discord. And of course, here is everything you need to know. They make announcements. Sometimes if there are cool things that are happening, They have all the official links right here for you. They have a quick user guide. They have some tutorials, and you can simply chat with other users. My mind, it's really worth to join that community on this cart. Of course, you will always be up to date in this course, but a community is always nice. 11. Local Installation of Stable Difusion with Python: How you can get the forum or and stable diffusion. One method is to install it locally, but that's not the method that I liked the most. But if you have a really, really strong graphic card with 16 GPU or more, you can do that. In my mind, it's maybe best if you have a little, little bit experiencing biotin. I just go to a simple and quick, quick overview how to do it. But I must admit that's not the thing that I do. After that, I show you two videos, how you can use the ballroom without a lot of GPU. And in my mind, it's the better way. So you need to download for Windows. You need an account in Hugging Face, and you'll need to download that automatic 11. 11. You need Biden installed on your computer. And you need the Hugging Face account and attacking phase. You need to download the stable diffusion models. You can start with the model 1.5 anthem all the point though. And after that, you can also include other malls if you like. So if you have a really, really strong graphic card and maybe some experience in bitumen, you can download all of this and granite locally on your computer. The next two videos are for you. If you are just like me and you're a little bit lazy, don't want to install all the things and run all on the Cloud. I also have to say if you install this locally, all your GPU will get used to render your videos. So as soon as you start rendering, you will most likely not do a lot with your computer, expect their rendering. The next solutions are in my mind that better solutions. And with the next dilutions we will work in this course. 12. Rundiffuison: A Cloud solution that works: The second option that you can consider is random fusion. Random fusion borrows you there GPUs and you can use them to render the videos. They have all installed. And you can simply click and run it or you can simply go to app dot run diffusion.com and just dry it. They have different GPUs for you. They have HE abide ramps, they have 16, and they have 24. Use the subscription that is perfect for you. They borrow you there GPUs for $0.50 an hour up to two bucks and $0.50 an hour. And it's right. You need to pay that per hour, and it doesn't matter if you render or not. So if you work in stable diffusion, you will need to pay at least $0.50 per hour. That can add up, that can be expensive, but everything is installed also automatically 11, 11. So you can also work in default room. You can simply borrow their GPU. Everything is installed, everything is ready, but you need to pay $0.50 an hour. And in the next video, I will show you Google Colab. In my mind, That's the best experience. You can get the GPUs from Google Colab. They are really, really cheap. You can connect and disconnect really fast. So when we try to prompt and make our decisions, how we make the video, we can simply disconnect, we connect to render, and it's really, really cheap. In my mind, Google Colab is the way to go and that what we will learn in this course. But if you want to try around diffusion, feel free to try it 13. Colab Notebook and Google Drive: The Google Colab Notebooks, in my mind, that's the way to go. Google Colab is really nice for default room and everything you need is this copy and a bit of space in your Google Drive for the models, I think everybody has a Google account. So simply copy this notebook in your Google Drive and do everything that's necessary. I must admit that it's possible that you need to make a subscription. Subscription costs $10 a month. You can try it if you like it, go if you really hate it and don't want to work with the forum, just cancel it. And if you have some luck, you can try it for free. It depends a little bit how many GPUs Google Colab has free right now. This right here is the interface, if you wanna call it interface. But we will take a close look at everything that's necessary to make the right decisions for the forum. Even if it's a little bit overwhelming right now, you will understand really, really quick how all of this work. If you wonder why my page is black, they'll simply two tools, settings, and make the theme either adaptive, if you hold Windows, is dark or take here the dark mode. I have, my whole window's a dark mode, so that works fine for me. Right now We work in the forums, stable diffusion version 0.7, every now and then, new versions pop up. But it's simple. Just copy and new page of Google Colab notebook and you will have the newest version. The next thing that you have to do is simply to copy this in your drive. And don't worry that these words are in German. I'm from ethylene speak German, but in your case, in standoff in Google Drive, copy it and there will simply be the word copy in Drive. So press to copy in Drive. If you go to your Google Drive, you will most likely have this order now, Colab Notebooks. If you make your double-click on it, you will most likely also have just one copy of this thing right in here. And you can make copies of this copy just to make sure that we have almost always a fresh notebook here in the default settings. If you double-click on one of these, you will simply land in your copy. You can rename the copy. And I will do that right now and make this example. And we can simply save this as example. After that, you need to click Run and Run all. Now everything will get installed. Then models will download automatically in your Google Drive. And after that, everything is ready to go. You need to go through this process just one time. You need to download all models. All everything needs to be installed and everything needs to be in place. After a few minutes, you need to give permission to your Google Drive. Just take your Google account. Gave permission. This whole notebook is designed that way that everything will run automatically. So first this cell, then this cell, and every cell after that, we'll run automatically. And I think that will maybe take you some five or 10 min until every cell is done. If we go down a bit, here are, for example, the prompt cells that are the prompts that you will later gift to the forum to make your animations. And right now we have chose to animation mode, none. That is the default setting. And if we take none, this prompt, so they sold their normal prompts will gather around and these two pictures will get created automatically as soon as this notebook has done its thing. You also see that animation prompt right here. And with the animation prompts, we will make our videos. And for the animations we have to choose something other. If we take none, we will get pictures. If we take to the 3D video input or this right here, we will get other outputs. The animation prompts and the videos are for other electrons. If you scroll down right here, now we will get our first pictures because we have chosen none. We will get two pictures here of the prompts that are in the default settings. And you'll see it right there. These are our first two pictures. Because we chose none. We get these two pictures. So a beautiful lake and the portrait of the woman. Everything is set, everything is ready to go. These two pictures will now be in your Google Drive. If you go to your drive, there will be a folder AI. If you double-click on it, you will have here most likely that folder stable diffusion. If you click on it, you will have this folder and there will be a folder stable thumb. In the folder, stable fund, you will have your pictures right here. And right now there are a lot of other pictures that I already made with stable diffusion. But in your case, if it's your first time, you will only have these two pictures that we created right now. Everything is that everything is ready to go. And in the next video, we will take a look at the pricing. If you want to make more animations and guilt animations, most likely you will have to make a little subscription 14. The Pricing in Google Colab: Let's take a quick look at the pricing. If you are a new Google Colab notebook, you will see that they gave you a video card. So the Tesla T4 is right now the video card that we got. And if you're lucky, you get this for free. But if you're not lucky or there are a lot of people that are also want to have good video card you need to make as subscription. You'll simply press on this right here. Right now you can see that I have 60 h left and I will pay 2 h bare. Our basically just press on this right here, and you will see the following. You can make as subscription for Google Colab Pro, and it will cost, in your case, most likely $10 a month. You can also buy the Colab ProPlus body. In my mind, you really don't need this. They collapse zero gives you 100 h. And in 100 h you can do a lot of stuff. You have 100 h to render your videos. And with this graphic card, we will pay 2 h/h. So we have 50 h to render our videos. And that's really a lot. 50 h. You can render a lot of videos and you can always disconnect. If you don't wanna render. If you try to prompt and you'll make your settings, you can disconnect and it will don't cost you your money. Don't cost you a European beautiful units. As soon as you are connected, you will pay 2 h hour and you can press simply on this right here. If you want to disconnect, click simply the dumpster right here and cancel it. And right now we don't pay nothing. We have our 60 h left and we will pay nothing. We can make our adjustments right here, and we can simply connect as soon as we will now create something. So we might world $10 a month are unacceptable price because you get your GPUs, your computer will not run hot. You can do other stuff with your computer and 50 h of pure rendering Israeli a lot. It's much cheaper than the run diffusion. It's much cheaper than other solutions 15. The Path Setup for the Models and Output: The best setup. In this video, we will take a brief look where your models are and where your pictures and your animations will get saved. Path set up. For me, there are only three things important, the model path and the output path. And you can see the model path goes like this. Content drive My Drive, ai models and the output path goals content drive my drive a stable family. If you go into your arrive and simply see all the things right here. So drives my drive. Now we go into AI, drive My Drive AI. Now stable diffusion or models, these folders I have created on my own, we will do that later. You have most likely these two folders models right here, there are your mother's, most likely you will have just one model right here. And under stable diffusion, this right here, stable fun, there are your outputs. So really, really simple. Models are right here and the outputs are there. We have also the model setup and we can take our own models right here. We will learn how to upload our own models and how to use them later. Model checkpoint, you'll need a model that you have in your drive. For us, it's most likely this right here, and on the default settings, it will be right this later we will maybe choose a custom model to make some more pictures. In my mind, That's it with the models. We hold this right here. Really, really simple 16. How Deforum works: What is the forum diffusion? With the forum, you simply create frames. You know what frames are. We can create frames. We can create 50 frames, 100s or 1000s frames. We make all frames that we want to see and we can adjust, ever reframe how we want to see. We can include seeds and other stuff. And we can also include camera access, rotations and everything that you can possibly think of. But the only thing is that we need to adjust all these things accordingly to what we want to see. We need to be specific. In the following lectures, we will take a close look how that works. So the forum is basically just stable diffusion. So diffusion model that creates pictures or frames of a video. As soon as we have all our frames that we want to have, we can blend them into a video and we can use pictures of our cell, for example, to start with our first frame. After that, we can transform this picture in everything that we want to see with this diffusion model. So to summarize this, the volume diffusion is an extension that runs on stable diffusion and it can create forams for you. You can make these frames how you want to see them. You can input your own videos. You can input frame or a picture of yourself to start with, and you can deform your frames, your pictures. And everybody works in the forum like this. You give the parameters that you think are right. You click Render. You see if it's cool. If the output is good, alright? If the output is not good, you tweak parameters a little bit and your render again, that is how you work in the ballroom. And in the coming lectures, we will take a closer look at everything. Let's start with a brief overview in the next video. 17. Overview and Quick Guide: That the forum quick guide and overview. In this lecture, I will give you a quick, quick overview that you have something in your mind. Because after that lecture we will look at everything in detail. I have found this quick guide really, really helpful. And as soon as you are done with this course, you can make animations just like this with rotations in 3D and so on. But first we have to take a look at the basics. You can make your animations in to the end in 3D. And it's important to understand that if you take none, you can also make just the pictures. But we look now at the 2D and 3D settings where you have the emotion side things in 2D. The first thing, it's the Zoom that right here is assume of 0.99 and it zooms out. And the same thing, assume of 1.01. It's in place, zooms in, as you see in this animation. That 2D angle rotates your camera. 2d translation x will move your camera to the left or to the right. Positive values, we'll bring it to the right and negative values will bring it to the left. That translation y, we'll move your camera up and down in the same manner. Positive, bring it up and negative values bring the camera down. If you like. With this value, you can also zoom in specific corners. But that's something that I usually don't use that much. Besides that, you have also the to the perspective flip and that looks like this. Now we come to the 3D motion settings. Translation x is basically the same thing as in to the ends. Always translation, y, translation, that is in 3D, the same thing as the zoom into the bacteria work with higher values. There, free rotation, rotates the camera up or down. 3d rotation y to the right or to the left. And the 3D rotation, rotate the camera positive values clockwise, negative values counterclockwise. That was a quick twitch twitch overview. And it would make sense to open this quick guide in a window and use it as soon as you start typing in your own commands. Because it can be a little bit tricky if you start with this. So just open this quick guide and leave it in a corner where you can see it. And now we dive into more detail 18. The Basics of Animations: Angle and Zoom as example: Animations and set things in the forum. In this video, you will get the basics, that stuff right here. We have all done. Now with the sad things. It works like this. Animations. You know, with none we make our pictures and 2D, 3D or video inputs are four videos. For now we worked with 2D to make it a little bit easier. The next thing are the max frames. How many frames do you want for your animation? Let's say we want to have 200 frames for our animation right now. Border replicate, we can simply leave this editors. The first thing right here on our emotion parameters is the angle. And you know, from the quick guide, the angular rotates your animation. The first thing right here, that's the frame. So we start at frame zero. And the thing in the bracket is the value. So we don't rotate at all. If we put one right here, we rotate the whole video. So from frame zero in the value of one. And we can animate this. If we make a comma, we can do something like this. For example, from frame 52 points, we want to see rotation that goes counterclockwise, so a minus of two. And we can animate this further from frame 100, we want to have a rotation that is again clockwise, but, but this time free, that's a fast rotation. From frame 150. We want to have, again some things that I think, you know, what do you get from frame zero to 50, we have a rotation that goes clockwise until frame 50. At frame 50, we start rotate counterclockwise at faster speed. At frame 100, we start rotate clockwise again and we rotate really fast. And DataFrame 150 after 150, we are static with the zero. That works in all of this. You can animate every frame, for example there. So we start here at frame zero again, we have assume that zooms in, as you know from the quick guide, we make a coma. And let's say from frame 50 ohm, we wanna have assumed that does absolutely nothing. So this right here is static. From frame 100. We want to have assumed that does something other. For example, let's say we want to zoom in but really slow. And don't forget to put your commerce. And at frame 150, we do, let's say negative 0.99. And that's how we animated this thing right here. We did the angle first rotation clockwise, then counterclockwise, clockwise, but faster. After that, the last 50 frames and DataFrame 200 because we build tier 200 statics. The Zoom. First we have a pretty fast to my DataFrame, 50, 50-100. We are static. 100 to 150. We have a Zoom again, but a bit slower. And 150 to 200 we have a negative x2. That are the basics. You can animate every single frame of this thing on every single parameter that was angled and Zoom. And in the coming lectures we will take a look at everything. And the basics always stay the same. You work with frames, two points, and then in the bracket is the value. How big the value is? You should know from the quick guide 19. Kamera movement in 2D and 3D: Translation, x translation y translation c. And of course, their rotations. These things really need translation. So, you know it also from the quick guide, the x-axis is simple. What does the x-axis? The x-axis moves simply the camera. It moves the camera to the right or to the left. Positive values will move the camera to the right. Negative ones will move the camera to the left. And we can work here with values from negative eight to positive eight. In my mind, that's more than enough. And same thing is for the y-axis. The y-axis moves the camera up if you use positive values. Same thing here where you can go up to eight and negative values will move the camera down. I recommend you to play with values that are two or three to start with. You can go up until then, but in my mind, 10,000 bring the best results. So basically x2, this, y, lose that. And you already saw their rotations and the rotation make the most fun. I find the best way to explain it is this. So the first thing is take your right hand and try to do something like this on the picture. Translation x will rotate your camera like this. Translation y like that. And translation set will rotate your camera through this. I hope this all makes sense. And I think pictures do that better chop. So now I will blend you again the pictures from the quick guide in my mind, if you see something that that's always the best thing, just look at the pictures, how they rotate, and you will understand how the camera settings will work. Positive values again versus negative values. You can use both of them. And remember they're free the rotation right? Now. We don't use them. We work with 2D, but later we will use them that are basically the rotations and the camera movements. And don't be shy. If you do something wrong, it's no problem. You'll make some adjustments and that, and maybe it's even better the way you didn't want it first. So I think now you have a basic understanding of the translation x, translation y anti-free the rotations and as you know, translation, that is the Zoom. But in 3D, under rotations, you work with smaller values. The values in the bracket should not be higher than three. We will look at that later. So translation x, translation y, we worked with a little bit higher values. That, that is the zoom and the rotations simply rotate. I think now you have a basic understanding what these camera access to. But here is a formula in the bracket. And at this formula, we will look at the next video. 20. What does the formula in the brackat mean: Formula in the bracket, what does it do? You can also use formulas to animate your friends. You don't have to type every frame yourself. You can also use formulas and you can even use KEY string generators. And also at the generators, we will take a closer look. But first, what's this formula? This formula simply animate your camera movement. But how exactly will this formula animate it? We can ask the chat GPD, what will this formula do on our x-axis? It will simply make waves. So it amplitude a wave with the value of ten. So in this case, the wave will reach a maximum of plus ten and that minimum of minus ten. And you know, that's what we do in the forum. But how we can make it even simpler, we can go into Google and simply copy our formula. And now you see it. We move on the x-axis constantly, from minus ten to plus ten, minus ten to plus ten minus ten plus ten. So on the x-axis, we simply go in waves all the time. And with such formulas, you can make a little bit of time segment. And you can put everything that you want, the nose you can make through casinos for example. You can also make this Tanga and if you want, and you can animate this further. You can also say maybe casinos and then dimes, dimes, something other times this right here, but this time, see you knows, you can play with these brackets how you want. You can put it in Google and look how the camera behaves, but you don't have to do this. I normally don't use that much. But if you like, you can use formulas to make faster animations. If you want to have a specific camera movement, maybe a wave that goes all the way. You can try this. And of course you can also adjust the values. So instead of free, maybe make here an eight and you will see the camera will behave completely different. That derived here, we'll give a really, really cool animation in my mind, you can simply play with these formulas how you want. And if you find something cool, just copy and paste it right here in your bracket. And you can even make a coma and animate from frame 100 completely different. Let's say first you will not have this animation. And from frame 100, you want to have the cool animation that we did right here with the camera movement that behaves a little bit silly. That's it. So first, we have the camera movement from frame zero to 100 in normal waves. Later on, we have some really funny, funny camera movement, or you can simply delete all of these and booked zero. If you put zero, we will be static. But make sure to put this also in brackets. Of course, after our frame, we need to put this formula in a bracket. And this right here, we don't need that space and that's it. So I think now you can understand how these translations and rotations work. You can put everything on every camera movement. You can animate the frames by hand, by KEY string generators or by formulas. And for the key string generators that for later lessons. And you can always put zero if you want to be static or a simple letter in the bracket. For example, on translation, why we want to move into speed of three positive. And that's it. 21. Noise schedule and strength schedule: Noise and strength, that are two things that I will not explain your next. Now you can see their perspective and the perspective flips. You have seen them in the quick guide. I personally don't use them, but this right here, this is important. I also don't use them that much because it normally works only in 2D. But the noise casual is really important and so is the strength. What is the noise? As you know, stable diffusion is diffusion model. It creates the things out of the noise. And if you take a lot of noise, it can make the big chairs faster. If you want to make an animation where this Zoom is really fast, it would make sense to give stable diffusion a little bit more noise. Stable diffusion creates your frames, are your pictures out of the noise. And if the Zoom is too fast, stable diffusion doesn't keep up with generating new stuff. If you make it a little bit noisier, stable diffusion will have an easier time keeping up, creating new stuff. So are really fast too. You could use a higher noise. And the next thing is the strength. What is the strength? Here? You have a strength of 0.65. And remember, you can also animate all of this. So let's say you want to have the noise from frame 100. You want to have it a little bit higher. So 0.04 from frame 100, you can animate all of these just like we have seen. And same thing is with this strength or this strength says the following. You create a frame and the next frame will be 65 per cent, similar to the previous frame. If you put here 0.80, for example, the foramen will be 80 per cent similar to the previous frame. If you take 0.12, it will be 12 per cent similar to the previous frame. So you can play with these values a little bit. If you take it too high. If you take it one, then video will not animate because all frames will look similar. If you take 0.9, for example, you will have a pretty, pretty consistent outputs. And 65 is the default setting that works really well for most cases. But let's just make a quick example here with the noise. Let's take our Zoom. First, we have assumed that a little bit faster. And then after frame 50, we have a really slow or near static soon. So it would make sense to increase the noise a tiny bit. First. And then after frame 50, we could use the standard noise of 0.02. And the strength that works for me just fine. 0.65 works for all frames in our case. So just remember, if your resume is fast, give it a little bit more noise. You don t have to do that, but you can try. And the strength, you can increase or decrease it just like you like, just remember that first frame and the second frame and so on. They will look similar to the previous frame. They will look, in this example, 65 per cent similar to the previous frame. You can play with that. I recommend 0.4 to 0.8 because higher values and lower values spring, sometimes uglier results, but I recommend playing with this. And there is also a relationship between the noise and the strength. If you use higher strength values, you can use less noise. They can look at this example. There is a lot of math behind of this. Enter basic suggestions are that if you increase the strength to 0.9, you can or should use your own noise. If you go down to 0.7 strength, you can use a noise of 0.02 and the strength of 0.5, you could use a noise of 0.04 that are rough estimates that are starting points and you need to play with them. And we have the right noise for our strength. In this example, we have 0.65 strength and our noise is 0.0, 3.0, 0.02. And that is right in between of these two 22. Coherence and cadence make the videos smoother: You may ask yourself what's next, and I will tell you that coherence is next. This stuff right here. I like to leave this as it is because in my mind, these changes little to your outputs. You can also use a mask and the anti blur. And I personally also don't use this much. But the next thing right here that's important in my mind is the coherence. You can use the color coherence and you can simply try what's best in your style. I have found that HSV is in most cases relatively good. But you can also work with lab or the other ones. These differences are just minor. They make a little bit of other colors. And this thing right here, that is important, diffusion cadence, I use here nearly every time. One, if you use one, that if your model will diffuse or generate every frame, if you use two, it will only generate every second frame, and it will blend A-frame in-between the two frames. And if you take eight, of course, that will take a lot of render time. And it will blend eight frames in-between the two frames that will be diffused. Just try to play a bit with this. If you take higher numbers, you should get us move for video. But in my mind, it's best to work with one or two, but trust, simply try it and see what you get out of it. I, personally, I use one in this guide right here. They describe decoherence like this. So of course you can play with all of this LAB HSV and so on. And the example of the cadence five is this. One foramen will get diffused and there will be blended or drawn for non defused frames in-between of debt. So the total output is five. If you take eight, the total output is eight and only one frame will diffuse. Just simply try to play a little bit with this. I personally, I use just one. Normally 23. UPDATE: Version 0.7.1 is here!: The AI space moves really fast, and for that reason, I'm also fast with my updates. Here in the forum difusion, there was a small update. We are now in version 0.7 0.1 The community said that the prompt box was a little bit too complicated, and for that reason, the prompt box is now a bit simpler. Let's just take a look. Every other stuff right here is exactly the same. But the prompt box got a little upgrade. If we scroll down, you can see right now there is only prompt negative prompts. And the negative prompts will work. Just as always, you don't need animation prompts anymore. You can type in your simple prompts right here, even if you want to make pictures or video that makes no difference. You can also include your normal prompts right here. You can include the hashtag to exclude something. Everything else stays the same. So you just to take into account that and the negative prompts work without animation prompts or something. It all depends on what you use right here. If you take non, you make pictures and your videos you make in 2d3d or video input. If you scroll down a bit, here was patch settings and over patch settings there was also prompt settings. Also the prompt settings are gone to make it easier. You also don't need to make prompt settings. Everything gets easier, everything is more. Right now, simply try it out. 24. The prompting cell and negative prompts: The next thing I want to show you is the prompting cell. And in the prompting cell, you will enter your prompts. If you scroll down a bit, you will get to the prompting cell. And here are your normal prompts. Here you can simply type all the prompts from what you want to make a picture. Every time you make a new one, you can simply put this right here and create your prompt. For example, a dog or whatever you want to see. And now we will make free pictures of these three prompts. But just if we take this right here, we need to choose none. If you want to activate this bronze, if you want to activate one of these prompt, simply put this right here in front of it. And when they are agreeing, they will get totally ignored. But here I want to show you that animation prompts. That animation prompt will get activated if you choose something like 2D, 3D or video input. If you choose something from the other, these prompts will get activated. How this works. Now, from frame zero up to 20, you will see our beautiful upward trending on art station. From frame 20 to 30, you will see the banana, and from frame for the 240, you will see your coconut. And from 40 until the last frame, you will see this thing right here. Same thing right here. If you want to activate one of these, just put this before the prompt and it will get green and it will get totally ignored. So in this example, we have our banana until frame 41, last little example that Apple will be an apple until frame 20. And from frame 20 on the apple will deform in a banana, and that will happen in a few frames. Your strength up there will determine how fast the apple will change into a banana. Frame 20 in this example would be 65 per cent an apple. But the frame after that, it will be simply a banana. So I think you get the idea from frame zero to 20. You have an apple. From 20, you've heard that you have the banana. Banana, banana has to transform. And that's based on the strength that you gave them all. 65 per cent of the first frame will be an apple, but the frame after that you will have nearly only that banana. If you increase the strength, it will go slower. If you decrease it, it will be faster. And you can change and go on with what you like. You can also make the banana, for example, 200 frames long or how long you want. And you can of course, make just u1 and say, let's say from frame 60, you want to see a cat. You can do whatever you like with this prompts. And as you know, in stable diffusion, we can also work with negative prompts. And the negative prompts work like this. Let's say you have the cat. Let's add a little bit. Prompt magic eight K, high-quality, vibrant color. And after that, you do it like this. Minus, minus neck. That are now our negative prompts. And you'll just type in the things that you don't want to see. Ugly, blurry, extra limps, deformed. Answer one here or you can simply put what you don't want to see. And of course, you can simply copy the prompts from Leonardo. This right here is just an example. Just go in Leonardo, copy the positive prompt, type in minus, minus neck, and then copied and negative prompt. And you will have really good prompts for stable diffusional. One morpheme. You can also include the negative prompts like this. You can simply type of income. Dash, negative prompt equals bracket, make a new one, and just type something like this. Ugly. Also these you can put wherever you want it. You can also say from frame to frame 20, you don't want to see blurry. You can do whatever you like also with this negative prompts. But I must admit I nearly always use this negative prompts because with this, I have some issues sometimes, so I delete them right now. That whilst the prompting cell, if you choose none, you will get your pictures. If you choose to 3D or something others, you will get your animation prompt. You can type on every frame, what do you want to see? And you can include negative prompts. Just simply put minus, minus neck and type in your negative prompt. That is the fastest solution. And if you don't want to see something, booked a hashtag in front of it and it's gone 25. Load and Image settings and the resolution: Load and image settings at these two things, we will look in this video. Underneath the brown box, there are the load side things. These first things right here, I usually just leave them as they are, but this will be important. Image settings. What does this mean? Width and height. It's simply the resolution. So we get that picture, 512 by 512 pixels. Why we have here $500 by 512, because our model was trained on this resolution. So in theory, we get the best output if we use our resolution of five-twelfths by five-twelfths. But most of you will most likely make pictures or animations for Instagram. For Instagram, we have to take other resolutions. There is no way around it. For instant TikTok, there are other resolutions. Five-sevenths is 61,000.24 works optimal for n star and TikTok. So if you want wanna make things for your phone, you should use debt resolution. And I will also link to a PDF where you can control what resolution for what you should take. But for our example, right now, we want that simply roll with 512 by 512. Because for us, it makes no difference. For me right now, it's only practical to take lower resolutions because the render time will be faster. And this right here, Bit Depth output. You can leave that on eight or you can increase it and 32. But what does this mean? I meant by example. You can also double-click on this right here and you'll see the code. And if you are a coder or you can do a lot of stuff right here. I personally, I'm not a coder, so I double-click again and it goes away. But if you are a coder, I'm sure you can do a lot of stuff here. But back to the bit depth output. If you increase that, you simply get more colors, you get deeper colors, you'll get more colors. You will be more precise with your outputs, but the render time will go up. And I think you can or should start with eight, safe some render time. And I think the resolutions good enough with eight. So in simple terms, what we learned in this video, bit depth output, leave it at eight. If you want to increase it, you will be more precise and get more colors. And the image settings, you need to take the right image settings. For me, 512 by 512 works wonders because it goes faster to render in this course. But for you, maybe 576 by 1024 with work really good. If you want to make things for TikTok can then stop 26. Sampling Settings, the seed and the steps: That sample site things. And now we get to the seat. You can see it is right here under the sample settings. There is the seat. Right now we have a seat of minus one. What does this minus one mean? It means that every picture will get on you random seat and you can't track it. So every seat will be completely random. The first seed Cook be five, the next seat Cook be 53,000 or 2 billion because the seed is really, really big. There are a lot of different seeds out there. If you want to be more specific, you should put here our seat. What do you use is completely irrelevant. It could be one, it could be ten. It could be, for example, 888. It's just important that you can track your seat because eventually you will need it. Under the seat you have the sampler. There are different models. Some of them are a bit faster, some of them are a bit slower. I like that default settings are the first one right here. But for now we just work with the default settings. The next step right here is steps. But what does this number mean? It means that the AI, that stable diffusion will run 50 times over your picture. So it controls 50 times. If your picture is right or not. You can use here higher or lower numbers, but something 40-70 works pretty well. And I must admit I use mostly the default settings because the default settings work really, really well. After that, you'll have the scale. And just like in leonardo, you can use a higher or lower scale. If you use a higher scale, stable diffusion will lean heavier into your prompts. Something 5-15 is normal. I usually also use here the default settings. But in some animations, scale of ten makes sense. If you want to have really specific prompts as Kayla, sudden is maybe a bit too low. I personally, I work 7-10 most of the time. And of course, back to the steps just for a moment. If you increase the steps to 70, for example, your render time will go up. Safety will most likely be enough in most cases. And these things right here, until the prompt settings just leave them as they are. The default settings worked fine in most, most cases. So we'll leave them. We need to go on with our course. In the next lecture, we will look at the prompt settings. So in this lecture, we took a look at this seat. Use a random number, but you should use a number. Don't use minus one because our random seed is not optimal scale, you'll seven to start and use 50 steps to start. That default settings are fine here 27. Batch settings and prompt weights: Prompt and bedside things. And I know there's a lot to talk about, but that's the way it is in the ballroom. So right now we are in the prompt settings. You could use prompt weight things, and that's one thing that I will show you right now. You can use something like two points and two or three or something, and you will get more strength on that specific word. For example, a beautiful green apple. And if you put two points and two, the green will be twice as much importance as the apple. And if you put three, it will be three times more important than the apple. But I must admit I most likely don't use that at all. But you can, if you want to use it, if you make a prompt and you'll think one specific area of dead prompt should be more neglected or stronger, or you can simply put two points and R2 or R3 behind it. And right here in that prompt settings, you can activate it. So leave it as it is. Normalized. Prompt waits, just leave it and Locke waited. So prompts, I don't use them at all. Now, down to the batch set things, this number right here is simply how many variations you get. We believe that I do own right now, the batch side things. That's where your pictures will be saved later in your drive. So let's just make an example for this and simply call it banana file format name. I use mostly the seat. If you take really, really long prompt right here, your name will be crazy. Behavior. The default setting is either, what does either mean right now we have a set of 888. Our first frame will be rendered at 888. If you take their second frame, will be rendered at the seat of 889. After that, the third seat, 890, 891, 892, and so on. With the ITRC, you get a really pretty consistent outputs because they see it alternates only one at a time. If you use that fixed it exactly. You have the fixed date you have on every frame, the seat 888. And that makes your output even more stable, more similar. If you take a random seat, exactly, a random seed will be applied. If you want to make something that's really funny or really don't need any specific look. You can also use random. After that you have leather and that simply goes, I think two steps up and then one down and two up and down and so on. And an alternate that something in-between of leather and random, I think, but I must admit I use most of the time iter and fixed that are the seats that I use normally. I don't know why we even have a leather. I don't know it, man. So for this example we use either seat, either number. You can also increase the item number. So in our example, the item number is one. That means every frame increases the seed pi one. And we don't want to make grits. And right now that's it for the prompt settings. Just remember how your seat will behave and where we are. Images will be saved 28. The last settings befor the rendering: In this video, we will look at the last settings and in the next lecture we will go render our first video. So now to the left side, things in IT settings, what our image settings. If you want to take your picture, you to start your animation, just like in TikTok, we will need to use them, but we get later to this. For now, we'll leave it alone. Under the init settings there, There's also the mask file. And if you want to mask a specific thing out of a video, you can also do that. But right now we don't use any mask or any innate setting exposure. In contrast, I only used the default settings right here because I have found that there is little to no difference if we make here something. And same thing with match conditional settings. We leave this all at the default settings. And I have to say all the rest, we simply leave it at the default settings. If we scroll down, we have here our two pictures, our output pictures, our video will appear here, but only if we disconnect this right here is keep video for ran all. If we click on it, we will skip the video. If we unclick this, we will make our video. But you can also run all the other things. Leave this as it is, and make the video afterwards. And if you work with Blender, for example, or After Effects, you can also use all your frames and simply put them in Blender or After Effects. Most of the time, I do that in Blender. But for this course, we will take this step right here. So we uncheck this box and T or you can make your frames per second. So if you want to have the video with two other frames per second, just leave it at the default settings. 24 frames per seconds are also something that we see often and you can also use 30 s. Just use a normal random number. If you have a camera, you can also use the same frames per second to make your videos if you want to blend them together. And here you have only the paths left. So wherever your stuff will be saved, our video will be saved under Dr. MyDrive, a ICE table diffusion and stable fan. And the same thing for our video later on. The last thing right here is the max frames. How many frames max do you wanna render your video? We have here through handwriting, the default settings, and I think our animation has also non-modern through hundreds. So yes, that works. If you put here 250 and you make 250 frames, you will also need to increase your video right here, or the last 50 frames will not be rendered. You can increase this how much you want. That was it that are the last settings. In the next lecture, we will render our first video and we will see if we have luck, if we can use the free credits from Google Colab or if we have to subscribe, I already have a subscription, but you can try it for free. I wish you the best luck. 29. Render your first video: It's render time. But first, let's take a brief look, wild with it. This right here. If you work normally in stable diffusion, it's all set and done. But we need to play all cells after that because in the course we are a little bit slow with our adjustments. That animation we take to the max frames, we have a 200, the angle, the zoom into translation x. We did that all together, so we rotate clockwise and counterclockwise, clockwise then static answer one. The next thing is the prompt. Prompts go until frame AT. And we have first our apple banana than our cocoanut. But I will not increase the frame number here a bit. So maybe we take these 230 right here to 60, these to say 100.150. We wanna have our cat resolution, 512 by 512. And all the other stuff is okay. And I disconnected this right here, so I uncheck that box and we make a video with 12 frames of it. And we need all in all 300 frames. And that's it. You click and run and run all. Because most likely you are also a little bit slow on your first time. So we need to run all our cells, also these cells, if you make your normal animations, you can also run just the animations than the prompts. And lastly, the Load Settings and then you are ready to go. If you click Run all, you can see right here, if you have luck or not, I already have the Colab Pro for me. It's no problem. I'm now connected and I now pay who hours per hour of render time. If you are lucky, you can do it for free. And if you are not lucky, you have to make your subscription right now. So now we only have to wait. We need to give permission to our Google Drive as soon as this is all done. And then we get our animation. As soon as this appears, just gave normal permission. Take your Google account and accept. And now our first frame start to render. And I must say this takes a bit of time. We are right now of four from 200. And you'll see also the Zoom in action. We zoom in right now and the Zoom in, my mind is maybe a little bit strong, but that's totally fine. That video is only, for example, and I will see you as soon as all 200 frames are rendered. Take with me or less, look at our motion parameters. So the angle we rotate clockwise and counterclockwise a bit faster, then again clockwise and even faster. And the last frames are static. The Zoom first, it is fast, then it's that thick. Then we have a slow Zoom and Zoom will go backwards. And the translation x will do really, really funny stuff. Let's just see how all of these will look in a video that's 200 frames long and that will be made in 12 frames per second. And there we have it. Now, let's just see what we do. First, we should zoom and behave in waves. And then it should be pretty ugly relatively soon. As soon as we get our other settings. Yeah, just like that. X camera movement is not really nice. But look at the cats. The cats are relatively good. If we just delete the camera movement of the x-axis, we will have a pretty nice video. I think. We've started with an apple go to banana, standard coconut. Then we start really shaking. And the last things are our cats. The cats have vibrant colors and everything is added is just this camera settings on the x-axis. They are really not optimal. They are way too strong. And of course you can also save your image just pressed for a point right here, and save it to your computer and it will be automatically downloaded. But you don't have to do that because that is all also in your drive. Just go into your Google Drive, you know that path. So AI, stable diffusion, this right here. And we have made the folder banana for our outputs. So in the folder banana, there are all our 200 frames. And if you want that, blend them together or let's say in After Effects or in blender, you can totally do that. And if you go down to the last frame, so that's our last frame. You will also have this text document and your video right here. So you have your video and all your frames already in google Drive. And by pressing on the three dots, you can also download it directly on your computer. And of course you can also download these frames or IT, or every time you wanna do that. And in the next video, I will show you a cool trick that you can do with this frames in your Google Drive. And that's just the beginning. We can make whatever we want. We can animate this however we want. We can start with our own big tours, we can start with our own videos. We can make whatever we possibly can think of. The only limitation is your own imagination. You can make add very thing in this tool. And in the next lectures, we will go any further and we make animations with our pictures, with our videos, we will take the video input and so on. After the course, you will know everything. So see you in the next videos. 30. Letting people age and the resume animation: Maybe you saw some videos how a person gets older on the forum diffusion video. In this lecture, we will take a look how we could possibly do that, because that's really, really easy. We start with the basics. So the first thing we wanna do is Chose to the, in this example, I want to make this animation 100 frames because I be honest here, I want to save some time. We don't need The angry or any rotation. So we put zero here. We don't want to have any Zoom. We want to have a static picture. So we also delete all of this book, no Zoom right here. And the camera behaves moment of all, access should also be zero. So we simply delete all of this and both are, so here are quick zero translation y of course, also at 0.3, the things don't matter right now because we used to the video, the strength. We can increase that a little bit. Let's just roll with 75 right here. Color coherence. Let's play with HSV for this example. We don't need any video input. Their resume animation, that is one thing that I want to show you right after this. The prompts, of course we need to change them. So we start at frame zero. Let's just say with a beautiful woman, 20 years old, high-resolution trending on art station. We can delete that for this example, we booked a little negative prompt. So the mind is minus neck, ugly, blurry, limps. Let's just make this a close up. So we start at frame zero. Beautiful clothes. Alpha woman, 20 years old, high-resolution enter negative prompt is ugly, blurry and extra limps. Let's just say extra. We can simply copy this prompt and both data right here in every prompt cell that we want. Let's just make this until here. Are a woman should be 30 years old, right here, 40 years here. And let's just say they should be 50 years old right here. We will not change also the frames. So 24 days, 44, this, 64, this. And let's just also made D right here and make a woman also 60 years old. And you can play with that how you want they killed. Also be a little bit older or let's say something like this. You can play with these things. The load settings are fine. They see it doesn't matter in this example, scale, It's all fine. Maybe we lean a bit heavier into our prompts and there's one that batch name. Let's just say woman, safe behavior. You can use that fixed seed for this, but I roll with either because people change if they get older. So either works fine. In this case, we don't use any image settings. The last thing is the video input. So we make just 100 frame video to add frames per seconds. Worked fine for me. And we uncheck this box. So that video really I've created immediately and we simply run, all right now, or if you have all installed and your drive is connected, you can simply run chest this box right here, animations, after that, prompts and after that the lobes and things. But in my case, I need to run all this right here is our animation. So our a woman starts at 20 years old and she got older until she is 70 years old. And we have not the prettiest results. I think here in the middle ground there resolution is not the best. That can be because of this seat or it can also be because of the prompts. There are two solutions that you can take right now. Maybe. I think this is because of the strength right here. So if you want to make here better outputs, you can either decrease the strength of it. So our strength with zeros 75 and maybe stable diffusion leans too heavy into this strength. So it makes a bit ugly results. So we can simply decrease the strength debate. Let's just say 0.65, 40.60. That's totally fine for this right here. And you can also do the following. Let's take our example, but let's assume that we really like our first frames. And let's assume our last frames did something specialist, something that we don't like. Let's just make our family thing right now. Let's say we don't want to have a woman on the last frames, but it should be a man. In the age of 55. The woman decided that it wants to be a man right now, are beautiful close up of a man, 70 years old, high resolution I also included this shot with our canon. So we now have also included the camera, the rest of the frames, let's assume, let's just assume we really liked them and how we can do if we wanna make only the last eight different new. That's a really simple. We go into our Google Drive, AI, stable diffusion. We go into our woman example. And now we need to find the frame that we don't like. So in our example, we can start that frame 78 or 79, or even AT if we want, let's assume we want to start at frame 79. What we can do possibly only to start with this frame, just click on it. Copy this until the underscore here before the other stuff, we simply copy this. We go back into Colab and we use this right here. Resume animation. First, we checked this box and we put our frame right here. And now we simply do the following. I decrease the strength of it, and I used this resume. So play in the settings, playing the prompts. So we'll remember until frame 78, we assumed we had a good output. Play in the prompts. And now play the load side things. And we will restart our animation path from foramen 78 or from frame 79. If you think your output is good and DataFrame AT just do that. But I think our output is not the best. So I uncheck this box and I, and I wanna run all prompts you because our animation is not perfect. Now we get our final output. I think it's a bit better right now. The frames are more clear. I think. The last thing is demand. You can see if we take a camera like we did with our man, the output gets even better. Our prompts right now are really simple or not good, but the prompt chest with they can get more detail. You'll see that first prompt, they are not optimal, they are not perfect. But just with the add of the, of the camera, we could make better outputs. And we can even change this woman into a man with a simple prompt. And remember, if you like your first frame, just use your resume animation and you will start your animation from this specific frame that you built here if you check this box. And we should also write better prompts. That is just for the example, we can simply copy prompts from Leonardo or from lexica. And I'm sure we can get better outputs than this. But for this example, I think it's perfect. You'll see what happens if you start with not-so-good prompts. And the last prompt is a bit better because of the camera, you'll see the output Israeli pattern if we use a camera, for example. And that was it. You can make people older with some prompts and some tweaks. You don't like your anti or output. Just run all, just tweak some examples. Just tweak a bit of parameters. That is how you work in the forum. If you don't like a specific number of frame, Let's say you like everything until frame 50. Find frame 50. Copy the number posted under the resume animation, check the box and run it again. And you will start from frame 50, 60, or the specific frame that you pull tear. In the next videos, we will get further and we use our own pictures, our own videos to start our animations. So stay tuned. 31. Your learning experience: Your learning experience. I recommend you to learn in the forum and to learn you have to use it. So you should make your own animation right now. Try to type in what do you want to see? You can make a static picture. So put zero in every camera movement. Describe your prompts just booked in what you want to see and play with the seat and the strength. Play yourselves and just let me know what you got. I'm sure you got something cool. 32. Init Image: start with your own foto as keyframe: Start your animations with your own big tours. I already told you that that's possible. You can start your animations with your own videos and with your own big charts. And in this lecture we will take a look how you can start your animation with a picture of view. First, of course, you need a picture of you. So let's just take a screenshot. I think I have found a really natural one. So the next thing is you need to upload this into your Google Drive. Simply go into your Drive in AI and make a new folder right here. So let's just make this Screenshot folder. For example. In this screenshot folder, you need to upload your pictures. And in our case, it's this picture. As soon as the picture is uploaded, you need to go back into Google Colab. Now, we animate this thing. So what we will not make AI tools here to the video, I want to make it, let's say 120 frames. We will. Now we don't want to have any angle, so we don't want to have our rotation for this one. I actually want to have a little bit of zoom. Let's just say we start with Zoom that goes in. But then after frame, let's say after frame 60, we want to have a little bit of a negative Zoom. So that is the right this, we don't want to have any movement on the camera access free. This stuff is not important right now. That strength, I roll with 0.65. We don't need any of the other stuff, but right now we need to make our prompts. So we start at frame zero with what prompt we want to start. We use some frame from lexica. First of all, I will not be a man with a beard because my beard is not pretty and it's not that big. So copy prompt back right here. At frame zero, I want to turn same plate into a man we've appeared. After that. I want to transform into, let's say, into a zombie with this right here, I will not be a zombie with some flowers. So copy prompt, and let's put it right here. And the last thing I want to be is this right here, so I will not completely die and be as Collette. The other prompts, we activate them with our hashtag, and now we have to choose our frames. So at frame zero, I start with this portrayal of a man that is totally fine for me. We have a zooming until frame 60, I think. And after that we start to zoom out until frame 55. I wanna be the zombie and the metals Collette, I won't be after frame AT as soon as we start really to zoom out, I want to transform into this scale that and I think that's it right here. The load side things, we need to take a resolution that works with my picture. I have the picture in 16 to nine, So we need to take a resolution that works with that. If you struggled with the resolution, just take a look right here. And I think this right here works good for me. So 854 times 480. So we don't render all the day. So this right here is 16 to nine. That is okay. That C doesn't matter. We'll leave this seat. Steps 50. Does Kayla want to put it back down at the default settings of seven? This right here is all okay. The batch name drugs for me we want to have are either seat and this right here, this is now embarked in it. We need to check this box marked enter strength right here. This tells you what the first frame does. The first frame of our new animation in this case, would replicate only ten per cent of my picture. But I want to increase that to 90 per cent. That gift smoother transitions. So my picture is of course that the first frame, and after that, we increase it until 0.9. So that first starting frame of our animation is 90 per cent similar to the picture that we get. Stable diffusion. This image, that's also important. We need to connect our picture to this. So we give the path, what do we need to do is simply click here on this folder. We need to connect our drive. So we press this right here. After that, this cell will appear and you'll need to click play on that gave permission. Now you'll derive showed up B are here and if it doesn't, just press Refresh right here. So now we need to give the path. So Dr. My Drive AI screenshot and that is our picture. So press on it Copy path. And we want to take this path and bolted right here as our innate image. And that's basically it sold. That first frame will be replicated in 0.9 of the strength of my picture. And after that, this strength right here will be used. So this 0.65 that we already put in, our brands are okay and everything else works just fine. So now it's time to play. So play the sad things, played a prompt, played a load settings, and now you have it. It's, your output folder is not connected because in the course it takes always a little bit of time. We need to run all again. So simply run and run all. You need to connect everything if you're works normally and don't make your course how you do that? You are already connected. Now that right here is our video. And let's play. You see I transform relatively quick in the man with the beard. Then I get to the demon and to the skillet. And also the zoom behaves exactly how it should. But if you are not happy with your outputs, of course you can tweak it. If you start with an init image, maybe you want to have an image that's more similar to that. Now what do you have to do is the following. Just increase your use in it to one. That means stable diffusion will replicate your picture one-to-one. So your picture will be the starting frame. After that, we could also choose to tweak the prompts a little bit or something other. I think I will not rotate a little bit on the last frame. So I animate this from frame AT as soon as we get this kel that we will now start to rotate a bit, frame 82 points. And the rotation should not be that fast or rotation of one is enough for me. Zoom, I wanna make it a little bit stronger and in both directions a little bit stronger. That is how you work with that. So now we have 120 frames, a little bit of a rotation from frame AT, we start to zoom in a little bit faster and to zoom out is also a bit faster. I want to increase the strength, also just a tiny bit, maybe to 75%. I think the prompts, they work guilty in our example. So what we do now is play the animation, play the prompts, and play the load side things because the resolution was fine. Everything worked relatively good. And now you can see the transformation begins again, and I see you as soon as this is finished. So we now make a video of this later video. And I think now we get something called the first frame is exactly me. So stable diffusion chance to replicate at me. And the animation goes like this. The Zoom is a bit faster. We start to rotate and also the backwards Zoom is a little bit faster. I think the animation isn't bad, and it starts with our own picture. It starts with our own face. We transform first into a man with a beard, then into an-an that with flowers. And lastly into a scale that this is how you work with the ballroom. You make something go render. If it's cool, Perfect. If it's not cool, go tweak your parameters and render again. That is how you work here. If you don't like the last frames, go find the range that you don't like, put it on the resume animation and start the animation from this frames. And this is the trick. You just need to try it. And in the next video, we will see how we can start with our own videos. So just the videos that you saw trending on TikTok can then stuff. So see you in the next one. 33. Start with your own videos as starting keyframe: Let's start to animate our own videos. And there are several ways how you can do that, but the first steps are always the same. You need a video for him or herself. So let's start with something like this. Right now. I'm in shortcut and I simply corrupt the video just like I need it. So we have this video right now to start with. My video is relatively short, but that's totally fine. And you'll see, I'm just grabbing the camera. And that right here, That's our last frame. We need that last frame. If you work with shortcut or da Vinci, just simply export your last frame if you work on their own because you can totally do this all on your phone. Just take a screenshot of this last frame. I already did it and that is my screenshot of the last frame. This screenshot we need to upload into our Google Drive. So let's go into AI and we bought our screenshot into the Screenshot folder. The next thing we need to do is, of course, go to Google Colab. And the steps right now are nearly the same as we did with our older video. But our key frame right now is the last frame of our video. Now let's animate this thing. Also. This time we worked in 2D max frames, 120 ROK angle. I want to rotate, but I want to rotate faster in this one. The Zoom first two main and then zoom out. I think that works really fine for us. This stuff right here works all the great thing my mind, strength 65 now to the prompts. Always wanted to be a Greek gods. So that's our first prompt. So copy prompt. We go back. The first thing I want to be is the Greek god. Copy. After that, I have found again as zombie. And the last thing is a full body portraits of random. And I think that frames work fine, but I just want to tweak these things. So from frame 60, I will not be this horrible thing. And the venom. So 0-50, I'm a Greek god. 50-90, I'm a zombie. And 90-120, I'm venom. The other things are deactivated. The resolution works fine with this. All the other things works also fine because we have already safe that last time. So now we just need to give the path of our innate image. And that's basically just the same process. So use in it, of course. Then press this order. Mount your drive again. If it's not already mounted, drive. My derive AI of course, screenshots. This screenshot right here is the screenshot that we need. So Copy path bolted under the init set things. And we are ready to go. So run all our round, just the cells that you need. But in my example, I need to run all again because I am not that fast right here. Now let's see what we got. And you need also to check this right again and are using it. The strength right now is at one and it replicates nearly exactly my photo. But you can also use 0.60, 0.70, 0.8, or in most cases 0.9. That's the thing that I would recommend you to use if you want to animate your own video. But one works, also good for me right now, let's just see what we get. We should first domain then start rotate after that, Zoom Out. Yes, we are the Greek god after that, the zombie. And lastly, we are abandoned. I think the video is really, really cool right now. I need to download this to my computer or you can also of course use it from your drive. Now we take our video and make it into shortcut. We have to put the two things together. Okay, And now I did a really stupid mistake. Just locate it. I took the wrong picture to start the animation. I took the wrong screenshot for our folder, but we need to change this. I will change it right now. The only thing I changed is that strength right here and the strength down. The strength. I want to increase it a little bit because I think that it's better and the strength on the US in it, I also decrease to 0.9. 90% of my first frame will get replicate that. And I simply use now another image. And I hope this time it's the right one. Nice. That's the new video. So we actually start now with 0.9 of the strength. I transform into the Greek god than that zombie. And lastly, the venom we start to rotate, that's fine. So download it and into shortcut. I already added a little bit of music. I'll let, I'll show you that later. As you see now the video is really good I think, or it's better. Let's just see how the translation works. Now that are my last frames. So perfect, I think that's perfect. That translation is really smooth. And I transform into the Greek god really smoothly. I think that works fine for us. One last time I want to see it. After that, I simply export it with the music and you can listen to everything. And as you know, you can tweak everything that you want if you don't like it, just tweak it and don't be such an idiot as me. Use the right screen shot to start your animation. And that's the video that we got with the music 34. Video Input: Lay prompts over videos: Video Input in the forum. You can put your video in the forum and overlay prompts. And that's really, really easy. Just take a look with me. The first thing we need is a video. If you have a video from you, find, if you don't have a video from you, we can also go backslash.com, search for a video that we will now overlay. You can download your videos here for free. I think I found a good video. I want to roll with this video for this animation. So there are two persons that are dancing and I think they need a really good prompt over it. I already downloaded it. You need to upload it into your Google Drive. I recommend to make a folder in the AI folder that these named video inputs. In the video inputs there is the video that I uploaded. I call it Video Input default room. So we don't make any mistakes this time. We go into Google Colab and now we need to animate this, but it's really, really simple. All we have to do to set the animations from 2D to video input. If you choose Video Input here, that said things, all doesn't matter. So we can leave our max frames as they are. And also the emotion parameters will get totally ignored. All we have to do is to give the path. So we need to take our video input and gifted video path, and that's easy. We just click on the sidebar. As always, we need to mount our Google soon as your drive is, that you need to find the path. And as always, Dr. My Drive AI, in this case video inputs, Video Input, the forum copy path. And we need to put our path right here beneath this extract frame. If you put here a one, every single foramen will get x-direction. And every single frame we'll get that prompt over it. If you type too, every second frame, we'll get to extract that, then we'll replaced with a prompt. And you can also take free. You can take whatever you want. For this example, we go with one. I want to have every single frame of this. But if you want to save some render time, just put two right here. All the other things you can leave at the default settings. What's important are the frames. I want that this connects all of these frames. So we built our hashtag gets always and we will not run with the same prompt for every single frame. And I think this cyber punk from lexical should work well. So what Cyberpunk we will not take? Let's just take this one right here. I think that's relatively good. So we have a big prompt and a negative prompt. So copy prompt, put it right here, minus, minus neck. And we also take the negative prompt. And as you see, that's a really, really big prompt, but that's totally fine. There are various enough space for this prompt. Image settings is 16 to nine. And I think our pixels here work well. And I'm not totally sure, but it can be that this bigs let's get totally ignore because our video input already has the resolution. That seat, we can use this seat right here. The c doesn't matter that much in that example, a scale of seven works fine for me. And deceit behavior I want to take either because a little bit of diversity is fine for this video. The batch name, I will not have this right here. Cyber punk dance. But the init settings, they are now important. That's drank is the most important part of this thing. If you take one right here, we will get exactly our video. So stable diffusion will not run any prompts. If you put 0.9, 90% of the original video will be original and only ten per cent will wait into the prompts. I think things between 0.4 and 0.8 work really good. And we want our role with 0.8 for this example just to look what we get. And we want to go down here because this right here is also important. I want to have the same frames as my video animation has. Their video has 25 frames per second. You can simply look at this at the outset things and I think we are ready to go. So now it's time to render all and we will see us as soon as this is rendered. This right here is our video that's in Google Drive I think that strength is a little bit too low and the bronze, the bronze are not perfect. We did the prompt of a 4k barrier, and that looks a little bit silly in my mind. But you get the idea. I think I want to run this again. Let's just take a look at the original video, for example. The original video is this. And you'll see the prompts. They get a polite, but we don't do a great job. We need, we need other prompts and a little bit higher weight. I actually typed this prompt myself are beautiful, highly detailed cyber punk. And so beautiful, colorful masterpiece, award-winning picture, vibrant colors, clouds on the ground, the cyber punk standing on clouds. And I think that prompt will run better. I hope it at least because you need to try this enter strength, I want to increase it a bit in its settings. We need to decrease actually this strength to get more of the prompt. So 0.7 or 0.65 even. Let's run with 0.7 and we see what we got for this example. Just to save some render time, I want not just render every second frame. And let's see what we got with this one. You can also read it right here. So export one every two. We export every other frame. And that should be faster. Now that is our next video. And I must say, I like this more. As you can see, it's just in my mind a little bit better but also not optimal. And I have to say, I want to run this again and again with other prompts and with a fixed seat. So see you as soon as all it's rendered. And there we have it. There go to old zombies from walking that database shop for me in this example. Just like a set, you need to play with your settings. You need to play with the strength. You need to play with different set. You need to play with different stuff. But I think the video right now does a relatively good job. And I want to save this. So simply say if it onto your computer and just let's take a look at the original video. And you will see the difference is really, really big right now. So that's the original dancing video. I think the video is good, and that is our video. The Bronze data, relatively good job. And you can see all the dancers and the poses are relatively the same. We have found, we have found the right strength. And every thing for this. Let's just see what we did. Of course I took another problem. These are walking dead zombies. I also increased the scale to eight. I took a fixed seat. The strength, We'll see your 0.78. And that's basically it. And we did a video with 25 frame and I x-ray that every frame, not every second frame. So I think the best output we got right now. And of course you can still tweak these. You can still make this how you want. Try different prompts, try different settings, and just experiment with it. You can really do a lot with this link embark, for example, did the whole music video with this technique. They animated their whole music video with stable diffusion 35. The firs 3D video: I want to be the greatest. Everybody on the patient, I look around, I feel like everybody is the fake is somebody guess every day and I'm hoping one day I blow up in a basement. Same in the top is so big and I don't think that I think is amazing. Way for my day when a play and sold our souls were 1,000. Our first 3D animation. And for that, the first thing we need is of course, our video and our key frame to start with. And I want to do some fancy **** with my hoodie. So just a moment. So I think I got something here. I take the last key frame and make a screenshot of this. I already uploaded it in Google. We go back to our example. And now let's animate with age three the right here. We make this animation 350 forams, the angle and the zoom. We don't need this because we work in 3D. We have translation, translation, x translation, that for the Zoom and translation, Really why my goal is basically to zoom to the object and move to the right. But I always want to see my object. Then when we are here, we also rotate a bit back and also zoom a little bit back. I hope this works. I really think it works. You can see here how the camera excess behave. So that's right here is the x-axis, that's the zoom and theorists that 3D rotation way. And in later lessons I'll show you exactly how you can write this and how you can write this really, really fast and easy. But right now, don't worry about this. If you want, you can copy this. I can also link it. What are the next steps to do that color, I think HSV works fine here. I think the noise is fine if we just leave it at 0.02 for every frame, their strength, I want to maybe, yes, I want to roll with this strength and the restaurant theories of fine. The next things are the brands. And now let's see how we find our prompt. I actually found some prompts and some prompts I wrote myself. So from frame zero, I want to be a black metal skeleton rapping. I think that's nice because I have my hoodie after that, that foramen 100, I will not transform into the goals. Try that. And I think that's also nice because first we have the wrappings, Catlin, and after that the goals dry. That is basically nearly the same thing, just with some fire on it. And I had that prompt right here in my midtone account. At frame free handwrite, we want to transform in a zombie from walking dead. That prompt we already had in play. I just like the prompt and I want to see how we got to this. If we transform after the goals try to into a zombie, I don't know how we behave, but we will see. And I actually also added a few frames after frame 350. I will not be a zombie and I just put this brown TIGR from lexica. Our whole animation is now 400 frames long because I increase this right here just a bit to 400. So we have the last 50 frames as another zombie, I want to just transform again. The next thing I did is the seat. We just use this seat. The image side things are 16 to nine shots with a random normal pixel number, not too big so that we don't render all day long. There's kale, I increased it to eight. The batch name is cyber punk dance. We just use the same folder because it makes no difference that the Z behavior, I just write it with a fixed seat, but that doesn't give the results that I liked. So I took ITER because sometimes if you take a fixed seat, stable diffusion can lead to heavy into the prompts and the outputs where we're just ugly. I must admit it. I use are in it image. So I checked this box, the strength, I increased it to 0.9, and I simply put the path to my drive right here. And as you see, I took the right key frame because I named it who the keyframes. So no issues right there. We don't use a mask at all and everything else is ready. So now we go render and as soon as everything is rendered, I will see you again. We have now nearly all of our frames. So we are now at frame for 893 or 400, but the animation stopped. So we need to go to Google and find our frames. Both decks here, right down to the resume. And we did that play again. After that, I created the video So uncheck this box. I took 30 frames per seconds because our video has a lot of frames and I don't want to have a video that's too long. We go into our folder and here are all our frames. And also the last thing right here is the video. I downloaded the video, and let's take a look. What do we do? So the first frame is me, of course, and it's 90 per cent me. But you will see I will change relatively quick into the skillet and ds. And I think that rotation also do. Thus. He's chop. We get into the goals for either into the zombie right now, yes. Nice. We rotate. I think the video is relatively good. So we take this video and we need to put it in short cut to our other video. Now I want to see how the transition is. Nice. I think it's really good. So you'll see I turn immediately into these wrappings, Collette. And I think we did a really, really great job in transforming the first frames. They are all relatively Gotye think. So let's just play the whole video. I think we do a really good job in this one. So, yes, all we need to do is maybe put a little bit of music over this. And I think we are ready to go. Our animation is really, really good. I think maybe we could change the prompts a bit. Maybe the goals try that from the beginning would be the better option, or maybe node zombies at all. But I think we did a relatively good job with this. And you have seen how easy you can create these videos and you can create every thing that you can imagine. There are no limitations. So just be creative. Play with the prompt, play with the settings, play with different stuff. Maybe put a little bit of music under it and you are ready to go. If you start your own animations, do maybe something where you have a movement and out of the movement started randomization. You can make this or you can clap or something, and you take your specific key frame and then you transform it. Or just like I did with my hoodie. And then you start transform yourself in every thing that you want. And there are really no limitations. And over the next videos, you will learn how you can animate this camera movement easier. We can use keyframe, generate the rest. We can use jet GPT. We can just simply use different tools and make our own workflow out of it. So stay tuned and try to animate your own cool videos. I want to be the greatest. Everybody on the patient, I look around, I feel like everybody is the fake is I make every day and I'm hoping one day I blow up in a basement statement at the top is so big and I'm going to shade that I think is amazing way for my day when I play and sold our souls were 1,006 36. ChatGPT as personal assistent: In this video, I will show you how I use GPT as a personal assistant to make my animations. First things first, we need to prime church EBT. So I told him You are a pro when it comes to diffusion models, camera access, angles, and rotations in 2D and 3D helped me to make incredibly good videos in default room diffusion model that creates individual frames that can be made into a video. I will post you some information here. You read it all and answer. Okay. Only first information. I posted simply the first information from one of our Quick Guides. And I told him, but wait, I have more. Next information. I posted all the animation set things from one of the quick guides. More information. 3d animations like things, again, from one of the quick guides. And then I gave Jack GPT an example. So this is an example of settings I like. And I posted simply the 3D rotations and what they did. I asked, Would you say, you can help me write animations in 3D with the forum like in this example. And check GPT basically said, Absolutely, and the more information we give GBT that better he get. So you can go on and train Chet UP te, further in this chat. Or you can simply ask the following. I have as seen in mind, all in 3D, object sits on that chair right in front of the camera. The camera zooms in and then rotates to the right wherever other object appears. And chatty be Degas, it feels practically the following. He knows that it's zooming in is the translation set in 3D. So he does a really good job. And the rotation right is defray the y-axis anti basically just give me all the things that I really need it. So we can simply put this in stable diffusion and I think that should work. Or I know that works because I hear a lot of animations this way with Jesse Beatty. And you can do that too. Just gave chat D the right information and he will provide you the right answers. That is, how you work with jet GPT. You must be specific. You must give chat UP te the right information. After that, you simply need to tell chat GPT, what do you want to see? Another example, Let's make it a bit more complicated. For example, next scene, 400 frames long, gave me only the mask file object in front of the camera. The camera zooms slowly in, goes to the right, but rotates in a manner that the object is still to see in front of the camera. After frame 200, the camera rotates to the left there zoom is faster and our other objects is to see until the last frames. And there we have it and we have even our quick explanation what we do. So of course, we need to select 3D, our frames to 400 angle and zoom with doesn't need it. After that direct translation x, I think we need to tweak this. So the translations that is the Zoom of course, then the rotation for EY -15 and after that plus 15, that might be a bit high, we can or should decrease that maybe to ten. But we could simply copy and paste this stuff into the forum. And let's just see what we get. And if you're not happy, remember, just tweak your outputs. It can be really that easy. And now I also told check GPT that he can or should rotate from frame 220. 20 also did a great job on their rotation that. So let's just try this thing. I called it all into my Google Colab notebook. Now it's time for us. We have all the translations, all a set. We need 300 max frames and we will not simply built the strength, the strength thes. Okay, in this way, I think we can even increase it just a tiny bit. We need to use proper prompts, of course. And like I said in the last video, I want to start with a ghostwriter right here. And I actually want to be the ghost rider. And then the last frames until we really start to change camera angles and start to zoom in. And that is basically frame 200. So the last 200 frames, we wanted to be something else or something else will appear in our animations. We will see what we get. I think this male zombie drinking our coffee should appear. So we basically just copy this right in here. And we change the frame number to 200. The other frames we this active weight with our hashtag. As always, we need to make the right resolution. And I think I want to increase the height just by five pixels because sometimes it's not perfect with, with our format. Maybe it works, maybe it doesn't. We will see this scale is alright, Everything here is fine. Our batch name. That's all Okay. In my mind, I see it works. Also. We need to use a neonate, the strength of 0.9. We need to give the path, again drive my derive AI screenshots. I use the who the key frame because I think that with the who The wasn't that bad is the idea. We built it right here and everything else is fine. So we need to run this thing and we will see if it works or not. And I got some results. I wasn't quite that tab. I tweaked it. I spoke with jet GPD, and after a while I got this and I think that result is pretty good. Just take a look with me. So we started just to goes dry there and we zoom a bit. We rotate, debate with our cameras. And now comes the rotation and the transformation of the prompts. And I think all in all, it's not bad. Again, one last time. And again, we need to tweak things, we need to try things we need to render, render again, run again, speak with GPT. That is how we work here. As you can see, I just spoke a bit with JEPD, so the objects that bears were relatively quick. I wasn't happy I got new outputs and so on. You just need to talk with Jesse Beatty, a combination of trial and error, sketchy PT and keyframe generators, which we'll talk about later, are the key here. You need to try, you need to explore, you need to render, and you need to tweak. That is how we work here in the ballroom. Use GPT, use also maybe the quick guide takes the quick guide if you're not sure how you rotate or how a camera behaves, just look quickly at the quick guide and you will know it. After that, you can tell chat GPT what was wrong, what he can make better, what you can make better. And maybe later how you can adjust this with key-frame generators. And you will make really, really good videos. I'm sure about that. Just explore yourself 37. Use other and specific models: I already told you at the beginning of the course that you can use your own models in stable diffusion. But how you can do that in the ballroom, It's actually simple. The first step that you need to do is to find the specific models. And I see it all the time. Vapor like dance and stuff. They like demons. They liked this things for the animations. So we need to find a dungeon model. Adoption model that I see relatively often is this right here. It's from cv.com. You simply press this button here and download it to your computer. And you'll see the file is relatively big. But if you downloaded it, you need to upload it to your Google Drive. And the same thing on the Hugging Face. You can find a lot of different models on Hugging Face. That's one model that I see that most people use for pictures that make good outputs, for faces, for example, they make relatively realistic faces. With this analog diffusion model, you need just to click on this link and it will start downloading. As soon as you download that all the models that you like, you need to go into your Google Drive. You need to go into the folder AI. You go into them all this folder. Here, you'll simply need to put your model. Right now, I have three models here. So simply upload the models that you like to use for your animations. And it makes no difference what mountains you upload. You can use any model right here. And then you just go back to your Google Colab. And backing Google Colab, you'll need to go to multiple setups. Model checkpoint. You need to choose custom. So simply click on Custom and then everything you have to do is on the cast them checkpoint path. You need to give the path to your model. And as always, Dr. Mydrive, in this case models and the models will appear that you will not take. Simply press again on them all the copy path and that path right here in it. And it's all set and done. If you have a model that makes warriors, for example, really good, you should use them all for warriors. If you wanna create some warriors and you can go on with this, how you like it. And in these models you can also find some prompts. Sometimes. For example, if we go into this tangent model, you can click on it. And in the model you will find a lot of example. And this example is this right here, and that's the prompt right here. And the negative prompt, you can simply copy this thing into your Google Colab and that's basically it. And if you like, you can also copy the seat if you want to have this exact thing and you're ready to go, just simply find the brands that you like. And you can copy them into your Drive. And you can make exactly the stuff that you can see here into these models. And this is how you can use custom models inside of Google Colab. It's really easy. Just find them all others that you'll think that makes the best outputs for your specific needs. Then you need to download them, upload it into your Google Drive, use custom model and then give the path to your model and you're ready. And if you have latch, the guy that has created them all, Let's also some cool prompts for you. So just try it out. 38. Keystring generator for faster workflow with deforum: The website right behind me, that's a key string generator. And with KEY string generators, you can, of course create key strings. But how exactly that work, you will know it after that video. First things first, you need to go to this website. And here you can animate everything that you can possibly imagine. You'll have to type in what key frames do we want to see how long your animation should be there? Animation, for example, start at frame zero right here, and it goes and DataFrame 100, you can increase that. How much do you like maybe until frame 1,000. And here you can see following. We start at minus one and ends at plus two. You can see this right here. This is plus two and this is minus one. And now you can set the values that you need. You know, if you're working on translation x, translation y, and also the translation set for the Zoom in 3D. You can work with values that go from minus ten to plus ten. So if you want to create animations for this free access, it would make sense to take values from minus ten to plus ten and simply gave their key frame generator these values right here. Let's assume you want to create an animation that is 150 frames long. And now you can simply make the animation how you like it. Just think what x is. Do you want to rotate or what access you want to animate? Let's start with translation x. Translation x does this. So what you want to see from frame zero and frame 150 on the x-axis. Let's assume you want to start, of course, at zero. So simply click kids €1. And then we want to, for example, make the animation until plus two-and-a-half. But it should go back, and it should go back even further than it started. So it goes to minus one. You click at minus one. And now you have right here that KEY string for all these animations. And you can go wild with this. If you click, for example, edit control point, you can pull on this and you can animate this, how you like it. For example, we like this thing right here, and that's now our x-axis. And of course you can't just copy this string if you press copy string. And now we put it on our x axis and they are released from frame zero we start, we should start maybe also the animation at zero. But after that, you can see how the thing behaves. It behaves exactly like we told him. Soda, x-axis, short first, move to the right. And it moves to the right in a really funny manner. After that, it moves to the left and it goes even further to the left. Then we have started. And you can do this with every axis. Let's just make an example. The y-axis, I like to hold them still. I don't wanna see anything on the y-axis, but I want to have translations that I want to have their Zoom. And you can also of course, make the Zoom maybe like this. You want to start zooming in at 04:00 A.M. 50. The frame should go from 50 until 150. And you know that translation that so their Zoom, you can also go with values from minus ten to plus ten. So we leave the exact same thing right here. So now let's animate the Zoom. The Zoom starts at frame 50, and we just simply wanna zoom in maybe the highest value, 2.2, 0.8. For example, we start at frame zero and the Zoom should behave like this. We want to have a really, really funny Zoom. You can even see that frames right here. So this right here is frame 50. And at frame 100, for example, we want to have a really funny Zoom. So we zoom in in n. Well then we make a quick zoom out. Then we statically zoom in until we are a bit higher. So now just copy the string and we want to post it under the Zoom. Completely delete the Zoom, and copy the Zoom right here. And remember we should start maybe a zero here. And the first forams we want to also have static. We also animate the first frames right here. So we start with zoom, that Maybe completely static, and that's it. Now on to the next things. We have translation x, translation y. We want to have that translation that we have as Zoom, and it starts at frame 50. Now we have the rotations. And on their rotations we work with smaller values. So it would make sense to change this. I want to animate the whole axis, so 0-150. And the values are like to make them from minus free until plus free because that or the max values with we will now work. And now you have to think a bit. What do you want to see? You'll have your object. And the first thing is our x-axis that do something like this. And from frame 50, we start to zoom in. And zoom behaves also a little bit funny. So it would make sense to animate the rest of the access with things that you want to see. Remember, if you are unsure about the rotations, just take a quick look at the quick guide. That for 3D rotation, x moves or rotates that camera around the x-axis. The rotation y rotates it around the y-axis. And of course, they're free the rotation that Rhodes the camera over. I will not make something like this. Maybe we can start at frame 50 with our 3D rotation x, so that we may be doing the following. We are static. We zoom in and as soon as we start to zoom, we go a little bit upwards. We start, of course it's the euro. We go maybe until frame 60 or something like that. We go static or nearly static, maybe even a bit down. But after the 60, I want to rotate the camera upwards. It should rotate upwards, maybe a value of 1.6. I think we want to rotate upwards relatively quick. And I want to also animate this just a bit. I want to have funny thing right here. It should behave like this. So copy string and just booked it at our rotation for e, the x. The next thing that we want to see is how we want to behave on their rotation free the way. Remember that 3D rotation y rotates like this, and I think we should rotate also a little bit in that direction in our example. And we should take this also. Maybe we could even take this same settings right here and also put them under our translation. Why? I wanted to see how we behave if we simply put this also right here. So in theory, we should do something like this. Our object is that then we are static first and go just a little bit with the axis. And as soon as we zoom, we go upwards and to the right. Rotation radiates that. So basically also their rotation in 3D, we want that chest. Keep that. Now to the next side things. So the max frame at 150 because our animation is 150 frames long, three the angle and Zoom, no values, that's all fine. The 3D rotation, of course we need to delete this thing. We don't need any fancy perspective flips because that's also for 2D. We don't want to have that strength and noise is fine. I think maybe we decrease the strength little bit, but the noise is fine. Coherence with lab is also fine. We don't need any video input. I also don't want us start with our picture of me. Write this basic example. We will not just make normal things, our prompts. So we want to start off course at frame zero. We start with our beautiful colorful donkey, vibrant color masterpiece trending on art station. And it should transform into an extremely disturbing horror film photography of walking dead zombies. So again, I just want to play with zombies. And why I chose frame 70. I chose friends 70 because at frame 50 we start to zoom in. And from frame 60 on, we rotate with our camera. Maybe this disturbing zombies can appear behind the donkey, or we have just to see how this warm behaves. Image settings are completely fine. The seed makes no difference. Scale and steps is find the Euler sampler is also fine. We will not call this donkey. The seat is either. That is fine for me. We don't use our name on it. And I think all the rest is fine. So now simply go render. So I click Run, run all and I see you as soon as this all is rendered. Right now I'm in Blender and let's just see what we got. So while our donkey first is nearly static, we start to move to the right. The Zoom hits in. And the other Bronze Star start to change. And we rotate to the right and upwards in the 3D motion. So just like we told him, but I'm not really happy with the outputs. We need to rotate maybe a little bit slower. We need to maybe activate. Notice Zoom is fine. Maybe be killed. Start with the movement of the x-axis a little bit sooner but slower. And we need to simply fine tune this fine-tune performance. If you want to, fine tune acts as if you want fine tune everything that you want to fine tune. I will skip that we think in this video because that is up to you right now. You know how to create everything, how to use KEY string generators. And if you get an output, you just simply need to tweak it. Tweak it until you've got something cool 39. Audio keyframe string generator synchronize music and animation: Audio, two keyframes, string generator. This thing blew my mind as I explored it the first time. There is the possibility to use a specific music or a specific soundtracks and to extract the key frames from specific parts of this music. For example, just download the music, which of course, you will need to have their permission to use. You need to use copyright-free music, maybe from the YouTube audio. Take for example. You can download it and you can upload it in a door where you extract a specific part of that song. For example, I already downloaded the song and you can go to Lao Ai. Ai is my favorite tool for extracting these things. And it works really easy, but you have to make your subscription if you want to do more here, there is a possibility to test this thing. And I think you can test it for about 10 min. And after 10 min, you need to make your subscription. If you don't want to make any subscriptions at all, you can use phone Nick mind, and on phone Nick mind. It's basically the same thing. You just upload your music and then you can extract what do you want now here, let's just make an example here on Lalla AI. But it works in every tool nearly the same. You can also run local things on your computer. The first thing you need to do is, of course, upload your song. So we simply press select new file. We can drag and drop our file right here. As soon as this is upload that it starts, then you need to select what you want to extract. And in most cases that rums work best. So let's extract the ramps for this specific example and we'll make it don't refer to seconds. So that is not too big. As soon as everything is done, we need to go to our audio to create forams during generator. And here, of course, we need to upload our file. We can drag and drop our files of only the drums right here in it. And you'll see it. We have 34 s of our song, and we have here our keyframes. All keyframes are here listed. The first thing you need to check is the frames per second. You need to know how many frames you want to create. For example, if you take 24 frames per second right here, you will get an animation that has 800 frames. If you take 12 frames per second, animation is of course shorter. It is only 408 frames because the soundtrack is 34 s. And if your time is 34 by 12, you get the 408 frames. If you time it per 20 for you, of course you've got a lot more frames. The next thing right here, that's the function. We work our example with the zoom and the default settings works perfectly well for the Zoom. But what does this formula mean? Let's just make a quick example. So we have one plus x to the power of four. What does this mean? For example, if we go into the forum and let's take the Zoom for an example, the Zoom is a multiplier. Of course. This right here would be a static. Zooms on, Zoom at all. And you can multiply it by this value and you will zoom slowly in if you work on other things like maybe the y-axis or the x-axis. But I like to zoom the most. You need to change this. You could change this to only x to the power of four, because x is always that ramp and to the power of four. But because Zoom is our multiplier, we need to work with this one plus x to the power of four. So we simply multiply this, and every time that RAM hits, we will get a zoom. You can see it right here. For example, at frame 403, we don't have any zone 404. The we don't have any Zoom, but at 405, we have our Zoom. And that will be exactly when the drum hits. And if we work with our Zoom, of course, you can also change this x to the power of two. And you maybe think that x to the power of two makes smaller adjustments in the zoo, but that is not right in this formula, smaller is bigger. And I tried to explain to you why. In our example, x is smaller than one. And if you multiply x with something and x is smaller than one, a higher value to multiply means a lower number. Let's just make a quick example. So that right here is our example. If the dram doesn't, it said all we have zero If the drum hits really strong, we have 0.3. If the drum hits just a little bit, we have 0.1 and our formula will go like this. One plus 0.3 to the power of four equals 1.081. And if you take that same formula, but with a lower multiplicate, or we get the following. One plus 0.3 to the power of two equals to 1.09. So as you can see, the value of the Zoom is higher. If the multiplicate there are this thing right here is lower. So you just need to remember, if you take the power of two, your output will be stronger than to the power of four. Because we work with values that are smaller than one. That was basically the explanation. Let's assume we want to have a strong hit. So we work with to the power of two. And because we work in the Zoom, we can leave the rest the same cell chests ended up. And you can see it also right here. For example, on frame 405, we have our Zoom of 1.15. If we take it to the power of four, our frame 405 is 1.02. So a smaller value. If we take e to the power of free, of course, it will get right in between. And I think I want to work with the power of free in this basic example, because I think that's who makes the most sense for me. And now we are ready. We have uploaded our, so we choose a frame rate of 12 frames per second. We chose this formula because we work in the Zoom. The Zoom is a multiplier and we took it to the power of four. We have our 408 frames and we simply press Copy String. Now we go into our collab. We booked it right here into the Zoom, and you can see we are basically ready to go. I will not work with our 2D animation right here. So we simply choose 2D. We take our frames, our max frames are 408 frames, as you know. Border replicate works good for me. The angle we put that zero, and of course we put also the other translations at zero because I don't want to have any movement at all for the translation x, translation y. Let's make something finally arrived here. We want to rotate to the whole video, but the Zoom hits only when also the armpits. The other x's will get ignored because we use 2D animation. Noise and strength is right with this. The next thing we need are the prompts. And I want to simply delete this prompt or this equity weighted. And I want to have the same prompt for every frame of our animation. Just did ramp should animate this thing. Our prompt is an epic battle ground safety, basically in a cyber punk style. And of course, I just copied this from lexica. Our image settings are fine because in the course I think this resolution works fine. Scale is okay, the steps are okay. We don't deal is of course an image, so we don't need this. The next thing that is important to me is this right here, the frames per second for our video, we need to choose the same rate for frames that we had in our keyframe generator. If you take here 12, you also need to make the animation with 12 frames, or it doesn't work at all. If you just take one frame more right here, the whole animation not work, it will work, but it is not at the exact same time as the ramp hits. So now the only thing we have to do is to render all the max frames at 500 right here work fine for me because the other frames don't get animate that. So run, run all and we are ready to go. So everything worked well and I just downloaded the file and uploaded it into shortcut. I took also the audio file. And let's just see what we get. I wanna play the whole song right now. And we're really nice. Our drums and our animation are perfect. It works really, really good. And it will also relatively simple to create. We just took our audio file, we accelerated their drums, then we had to upload it into our keyframes string generator. We need to choose the right perameters, so forth, the zoom, it's a multiplier and we need to take one plus x to the power of 23 or four. It depends a bit how strong you want to have your resume. You need to take a lower number if you want to have a higher value at the zoom, we need to take the right frame rate and then we are ready to go. Just download your animation uploaded into a video editor and just synchronize it with your audio file. And you can see it also right here. The audio file is exactly the same length as our animation. So it worked really, really well. No need to tweak this thing because I am happy. I'm sure you can also make really good animations that are synchronized with your audio files. So good luck with that. 40. Complete Projekt: We've seen a lot of different stuff and let's just use this stuff. Let's just use this knowledge to create a full project. I have something in mind and I already started something. Just take a look with me. I sit here in my kitchen and I just try to drink your coffee and no, Normally, I don't wear glasses. That is just for the video. I nearly didn't see anything with this glasses. I put on my glasses, I drink their coffee. I see something in their papers, and I do write this days right here. That is our last frame. I want to start our animation from this foramen. Maybe I could include something right here. Something like AI is going to take over their warmth or something. And as soon as I see it, I spit out my coffee. And from here on, we want us to art my animation. So we take a screenshot of this picture and we'll load it up in Google in our life, you already know how to do that. Let's go in Google Colab and we want to make our animation. Our animation should be in 3D. And I think the right forams or something like 300. I think that is more than enough for our video like this. And maybe I even can take this video for the intro or promo for this course. We will see how this goes. The next thing we want to animate is all of this angle. Of course, I pulled us the Euro because we don't want to rotate at all. The zoom, we don't need the Zoom. You already know if we want to zoom in 3D, we need to animate that translation set. At the first axis that I want to animate is the translation x. And now we need to think what we want to see. You know, we start with our key frame like this. And we need to start to animate our prompts. I think I want to get into a zombie as always. And maybe after that we tweak how operon. So we need to take a look what we want to see. So the first thing is, of course, our access, how they should behave. And I think I want to animate my axis like this. Right here is our object. And I want to start the Zoom in translation, x and translation. Why does nothing but we start to zoom in and then the 3D rotations rotate up and to the right. I think we should change prompts as soon as we start to rotate up. I think that should look relatively nice. And I wanna take KEY string generators to make my animations faster. And just for the sake of this tutorial, I will also animate the X axis, but I think that wouldn't be necessary in our example. But just let's do it. We can always tweak if we don't like it. So the first thing we want to animate is the x-axis. I need my three hundreds of frames from minus ten until plus ten. And let's go about the x axis. I think they should be relatively even until the frame 150 at 150 starts our magic. And the x-axis should go to the right, maybe just a little bit. So something like that. And I think that is fine. Just copy string. And we put this at our x-axis translation. Why I don't need that at all. And tear I wanna start exactly at zero. That translation that, that's something we need. The next thing I want to animate is the Zoom. I think I want to have a Zoom that starts really slowly and increases a bit. So we start at frame zero. We go, I think DataFrame 150. After frame 150, I want to increase the zoom drastically, maybe 23. And it should happen on DataFrames 160. And after that, the Zoom should slowly be much, much slower until the end of our animation. So that right here is our Zoom. So copy the string and under translation set. The next thing are the 3D rotations. I want to rotate to the right at frame 150. So we go back right here. This time we need lower values. So minus free until plus free. Because you know, this axis behave much, much more. I want to also have, of course, the 300 frames. We start at zero and we should go maybe until frame 150 with zero rotation right here. But after that, we start to go upwards and maybe Until a value of 1.55 ink. That is fair. Or let's just play with one right here. Copy string. And at our x-axis, same thing I wanna do with our y-axis. And now the only thing that we need is our rotation. And I think I want to have a rotation or we can even let this deal and maybe re-run it if you, if we think we need a rotation afterwards, the next thing we need to take is that noise. I think we can start with a noise of 0.2. But at frame 150, some magic happens. So we want to increase this after frame 150 a bit, maybe to 0.0 for their strengths. I think that strength is right for us. We could work with age as v In this example, I think that makes sometimes the better outputs. We need no video input and all. The rest of this is fine. And now let's go to the prompts. In what we want to transform. I found a prompt on Leonardo and I want to transform in this. I hope that it works. So just copy prompt. We go back into call the app. We put it right here at frame zero. We also have a negative prompt, so minus minus neck. We need to copy our negative prompt and we built it right here. So that is our positive prompt and we want to change prompt as soon as we rotate. Remember we start to rotate upwards and to the right at frame 150. And also the Zoom starts to be more quick. So I want to change prompts may be at frame 160 or 170. So that may be new things up a year as soon as we zoom out of our frames. So I want to activate this prompt right here, maybe from frame 100 and let's say 165. Start with, we need to tweak if it's not good. And what we wanna see right here in the background of us, in the background, something like heme should appear. So copy prompt. So at frame 165, we want that paste is writing. And at my first prompt I want to add right here, blade girl on that arc from drinking our coffee because I spit it out my coffee. And basically that's it. Notes that things are fine. Our resolution works well. The seat doesn't really matter, but we have a seat. Euler, 50 steps, we need to lean a bit heavier into our prompt. Maybe eight. Brown said things are fine. We will not name this. Spitting my coffee. Use the CT are either seat I think works also well. We need to use an init. So we activate the innate strength is 0.9. That is fine. And now we need to connect our Google Drive and gave the path to our image. We do that as always. We click right here, we connect our Google Play. This cell gave permission. If it doesn't appear, we need to refresh. We go to derive my derive AI. Now we need to take the right screenshot, start keyframe. So cup Copy path. We need to book the bath right here. And we can go on What's next. We don't use any mask. We don't use this things at all. But I want to have 24 frames because my camera or social two with 24 or 25 frames. And I think we are ready to go. The max frames down here are to 500. I think all the rest is relatively fine. Yes, the prompts change right here. And the only thing that we need to do right now is go render. So simply run all or just run here the animations after that, ran the prompts, and now run the load side things and we are ready to go. And I see you as soon as everything is rendered. And here we have our video. Let's just take a look at transform. The transformation is good and smooth. We start to Zoom. I think we should rotate right now to the right. Also this prompt right here, they are really nice. I think now we change prompts and rotate to the right. Yes, Change prompts to the right. I think now more noise should start kicking. Yes, more noise so we can create this stuff faster. And I think the whole animation is a relatively good. And of course we can and probably should tweak that if we want. Maybe their camera or rotations are a bit too harsh at the ending and a bit too slow at the beginning. And we could simply tweak that if we like. But I think I don't want to tweak this. I think that is relatively good, so I downloaded it. We upload it in our shortcut. I tried to put it right here on my video. And I think that transformation should also be relatively smooth. Let's just take a look. Now I split my coffee and now I should transform. And you'll see the transformation is really, really smooth. And then I'm relatively quickly, this lady. Yes, I think the video is relative guilt. Let's take a look at the whole video. First with my stupid glasses and can't see anything. But that's okay. Start to split and now I transform. The ladies are illegal. The Zoom is also okay, it's not too fast. I think that's relatively guilt also the rotation. I think I liked that. It's maybe a bit too fast at the end. But I want to simply leave this as it is. We can always tweak it if we want. But I wanna leave it as it is. I think that's really, really nice. And you can see thanks to the increased noise at the ending, we can create the stuff faster. So stable diffusion keeps up with creating our new stuff. We can also make this longer or other or animate this further. We can tweak and do whatever we like with this. Now we got to a full project and I think, you know what you have to do now, Exactly. Start your own projects. And it's a relatively simple, just make a video. And it doesn't matter what we do. Start with animation, something like a cliff or something like that. And then start your animation. Take your last keyframe, upload it in your Google. Gift the path and use the innate tweak, the prompts, tweak the camera access, use keyframe generators and you can and will create relatively good stuff. We did exactly what we want to do. Our object was right in front of us. We zoomed. After that, we rotate upwards and into the right corner. And the zoom was faster and we rotate even faster. We did exactly what we want to do. And thanks to this course, you can also do exactly what you want to do. If we want to tweak this, maybe we could decrease the speed a little bit of the last frames. We could also change the prompts are little tiny bit later, and the outputs are really amazing. But I think also this output is relatively good. We should probably just overlay a little bit of music over this. We are ready to go. So let's go create your own animation. 41. What you shold create and where you can get inspiration: You may ask yourself, would you should create with the forum diffusion? And I really don't know it because you can create everything with the forum. You can create to the animation, 3D animations. You can overlay your prompts over your video's, over your own videos, Overstock videos over everything. Dan, you can make your own animations. You can start with a video from you with a video of an animal, with a picture of an animal. And you can also animate this in to the 3D, every single thing, how you want to see it. So just simply go and explore the forum. And if you really don't know what you should make, go to TikTok and Instagram. Just find some information. Go to TikTok and look videos just like that. That's maybe the most popular video on TikTok and the video that really, really got viral. I'm sure you'll find a lot of stuff that you can explore. You could also go to Instagram, maybe to their site of stable diffusion. They have also some cool videos. As you can see here. They have animated a cool free the video right here. And they have also done something, I think with oval lame prompt right in this video. And you can make every single thing that you want to see. Just go on social media and explore a bit and maybe try to rebuild it, make it better. Give it your own style, give it other prompts. Try to animate the camera rotations child's IQ. So on TikTok, I'm sure you'll need a lot of time if you want to recreate thing that you see on Instagram. If you want to recreate maybe a video like this, I'm sure we'll need a lot of time because there are a lot of details in this video. And if you maybe go to his profile page, you will see that the camera rotations of his new or videos, or maybe most of the time this same because it is rarely, rarely, rarely time expensive to animate your camera like this. But thanks to this course, you'll know how cool you can do this a lot faster. You can use chat GPT, you can use key stream generators, but you need to try it. You need to explore it. You'll need to tweak it. As soon as you have something that you really like, you can use it for more animations. Simply copy your Colab notebook and do it again over and over again. You can also create your whole music video, extract maybe the rams or something other, and create your own music video. You can make this how long you want. You can even do hours and hours of these animations. So go explore yourself 42. Learning experience in 3D: If you made it until here, congratulations. Now I have a learning experience for you. And I think that learning experience, wealth, they care a little bit to do. I think you should create your own video. Make a video from yourself. Do something like drinking our coffee. And then start with an animation or something like that. And start your own animation. You should make that animation in 3D. And you should rotate the cameras how you like them to rotate, you should use keyframe stream generators to make the process smoother. And if you want, you should even use your own models. Just think about what you want to create. Do you wanna create photo-realistic, mildest yields down our log diffusion model? Do you want to make the immense? You could also take our demand model or you can search for your own model yourself, load everything up in Google. So the last key frame of your video and them all gave the right path and start to animate as soon as everything is animated. Tweak it until you're really, really like it. Use GPT if you think you can't do it alone. But I think with the key string generators, you are really, really good to go. After that. Load everything up in your favorite editor and underlaid with a bit of music. And now you should publish it on Instagram, on TikTok, on Twitter, or maybe on YouTube shorts. Just choose what you like and remember, takes the right resolution for your platform, Instagram and TikTok, you don't use 16 to nine, but if you want to make a YouTube video, you should use 16 to nine. So just remember to use the right image dimensions. As soon as everything is done, as everything is published, hit me up with a message. You can also send me the thing. Just let me know if you can create cool stuff and I'm sure you can create it. So have fun exploring the forum. And if you need help, I'm here for you. 43. Kaiber a solied and easy alternative: You may ask yourself if there are alternatives to deform diffusion. And yes, there are alternatives. And some of them are executable, good. Others of them are, in my mind, just too pricey and the outputs are not that good. One alternative that is really, really easy to use. And it's not that pricey Khyber. If you go to Cairo or you can go to the gallery and see what other people did with Khyber. But I want to show you what I did with Khyber. So we go to my videos. I've simply created two videos because I must admit it that was for free. I don't want to pay for this because the control is not that guilt. The interface is really, really easy to use. You simply upload a picture of you and you can type in this prompt and you can type what do you want to see? So camera, for example, zoom in. And you can also include a rotation if you want, and you'll see the output. So the output is relatively good, but you don't have a lot of control, and you can only do this in 2D. So in my mind, it's really not worth it right now. But the other feature is relatively good. You can also overlay your pictures with prompts. As you can see right here, I simply uploaded a video from me where I am. But he basically just humbling around in my room. And you'll see what I did do it with overlaid prompt. A man dancing in the style of illustration, highly detailed, simple, smooth, and so on. And it was really, really easy and it is intuitive. You can just upload it. The prompts are newly created for themselves because you just have to type in two or three words. The rest is the prompt magic from Khyber. And you get your video at the end. And you can also animate whole videos. I think Lincoln Park did their music video completely with Khyber. So in Khyber or you can work, it can be a bit expensive if you need to do a lot of it, but you have not the best control. A my mind, you should work with divorce him diffusion. And there's also wrap diffusion. Rep, diffusion is another Colab notebook. And in my experience, rapid diffusion makes good outputs, but it can only overall a prompts over your videos, but you can use control net and stuff. So rep diffusion is relatively nice. But I want to dive into it because I'm not an expert in red diffusion, rep, diffusion is easier to use than the forum, but I think also you have more control in the forum and you can make cooler stuff with the forum. No other thing you can make such good 3D creations. After that, there is also one last app, and the last step I wanna show you in the next video. 44. Loopsie and disco or wrab diffusion: Loop CAI, it's an alternative, it works, but it doesn't work that goat, you have nearly no control over it. But you can use it on your phone. And it's easy if you want to create some animations with loops, They just try it. Log yourself in, pay for the bucks a month, and have no control over your outputs. I wouldn't recommend you to use loops, but I just wanna say that loops there is also an alternative. So for me, it's not really a good alternative, but you can use it if you want. But this right here, this could work diffusion. So that rapid diffusion that I already told you in the last video. We've wrapped diffusion. You can overlay your prompts over your videos. And that works also in Google Colab. But I must admit that I am no expert in rapid diffusion. If you will not drive rapid diffusion, you could maybe use this free protocol that works also, but the best and newest version you have to pay Alex on patriarchy. The most things are relatively the same as in stable diffusion DeVoe room. But not everything is the same. But I think if you try just a little bit, you can also create your first animations here with rap diffusion. And just like I said, you can overlay your prompts over your videos with rapid diffusion. And you can do that really, really good with really a lot of control. But my preferred way is the boron diffusion because of their 3D rotations and the animations that start and go to nowhere. So I think the forum is the way to go 45. Stable Diffusion [Wrapfusion] in Google Colab: Stable Be fusion. That's also a cool tool that I will show you how to use in this course, step one, of course, you know it, you need a file. We do basically the same stuff as soon as the frame comes around where we want to transform. I cut the video and then I extract the part where I want to be transformed afterwards. Of course, I go one step back, because we need that later. Now I have the file. I put it simply into a folder because we need that file later. The next step is we need to go to Patrim. We work in a Cop notebook, but the Cop notebook we will get only access if we subscribe to someone on Patrin. He give always the New West versions and he gives tips and so on if you have any problems. There's also a discord server for me. I think this is really nice. Here I am into Patrim. The name of the guy is X L E L A or most likely XS. If you read it from behind, he is really active. He gives some updates every few days. And right now we are in version 0.20 That's only a preview. We need to be a power member. If we want to have this, I'm not a power member, I just spend that $1 But we can have stable B fusion, the version 19.8 You need to download your version that you like. Sometimes the newest versions don't work that well. A version that I always liked using was the version 0.16 You can also simply press on this file. If we press on this file, you can download it on your computer. The next step is to click on the Cop notebook. I will link you everything. This right here is a normal notebook. You need to go to uploads and you need to find your file. Right now I use version 0.16 because this works always really, really nice. We upload here our version and then we are ready to go. Now this cop notebook maybe looks a bit overwhelming because it is really, really big. There's code included and so on. But this is really easy to use. We don't have to do a lot. First step, I like to go to run time. We need to make sure that we use GPU, and I like to use a stronger GPU change run time. The version 100 U works fine. Also the 100 U works fine. This works for me. If you use four U, this will take forever. Then we scroll down to set up. We need to press play on the first cell. Of course, we need to give permission because this will get access to our Google Drive. So we give permission right here. Then we need to scroll down a bit. We need to go to basic settings. What's the batch name? We just call this test. But we've do these because I think I already have a test then. What's the width and the height? Simply the resolution. My video is in 16 to nine. This resolution works fine for me. Now we need to go to Video Input Settings, and of course, we need to give our path. We click right here as always. Then we can upload our video right here, or we can even upload it in our drive. But this right here works fine for me because I just want to make this video. I don't want to have this in my drive. So this is totally fine for me if it gets deleted afterwards. As soon as this is uploaded, I called it wrap me, we press the three dots, we press copy path. We delete this path right here and we put our path here in this field. This field is the next important field, extract frame. If we put one right here, every frame will get reworked. Every frame will get extracted and reworked with our prompts afterwards. If you put two right here, video will rendered in the half amount of the time, but I want to have every frame rendered in this video. Then we scroll down if you like, you can store every frame in your Google Drive, but you don't have to do that. But I'd like to do this if we want to work in Blenda afterwards, for example, After that you can scroll down even further until you get to this code here. It's also really easy. Everything you have to do is instead of CPU, we use CPU in this field. Here is model path and control net models. These are simply the default models that we have right here. I want to make the short and easy. So we use the default models, but you can also go to Cvti.com if you need to have a specific look. You can download models for free right here on CVI. Upload them into your Google Drive, and then you can give the path to the model that you uploaded. But like I said, I want to make the short and easy. We use the default models, that's basically it. If you scroll down a bit to non GUY, it comes right here. You can also delete the prompt and use different prompts. But right now, I want to use the default prompt, because we can change the prompt afterwards. Everything we need to do right now is simply run and run. All now the Colab Notebook will do its stuff, but this takes a bit of time. If you scroll up to the top, you will see that every cell will get automatically activated. The co lap Notebook does simply stuff. It installs different stuff, it download stuff. This takes a while. This took about 5 minutes. If we scroll down, you can see the program is doing his stuff right here. So it extracted some frames. Then at some point, you will see a preview under the cell. Do the run. This is the cell that is nearly the last cell in this Cop notebook. If you like this output, perfect. If you don't like this output, you can press Stop right here. And then you can scroll up until you see here, the prompting box. Until you can put whatever you like in the prompting box. You simply include the things that you want to see in this prompt. And in the negative prompt, you want to include what you don't want to see. Text, naked, nude logo, cropped to heads and so on. This works all really good. I think this cyberbunk look, I really like this. I want to transform in such a cyberpunk, the only thing that I want to change is I don't want to be a woman, I want to be a man. Remember, you can always copy prompt. You can ask Chachi BD for stable diffusion prompts and so on. Now with press play again. Down here we have another prompt. Now the only thing that we need to do is play again right here. And then we should transform into a cyberpunk man, instead a cyberpunk woman and perfect. Now this works really, really fine for me. As you can see, now we turn into the cyberpunk man. You can also see this takes a lot of time. I think this will take maybe up to an hour because this is not accurate. Sometimes you can really wait a lot. These are right now, 142 frames, just from experience. Sometimes you need to wait for an hour, you need to drink a coffee. You can see right here all the images they will get stored into my drive, content drive drive, stable B Fusion images out. I go into my drive, now I go into AI stable B Fusion images out the test order. Here you can see the frames come. Now we have now our third or forward frame. And the frames will be collected in this folder. But we can also make a video as soon as all of this is done. I see you. As soon as the video is rendered, just make a coffee. Now we have it. This is right now, the last frame. Now we want to create our video from our frames. All the frames are now also collected in our Google Drive. We can also use the frames and put them in blender, for example. But I like the easy way I want to create the video inside this Colab notebook. That is really easy. All you need to put right here is the right frames per second. Remember, if you have a video with audio, it's important that you use the same frames per second in this video right now. Because if you use a higher or lower number of frames, the audio will not be synchronized. My video has 25 frames, and for that reason, we need to put 25 also right here. Then we are basically ready to render. We click Create Video right here, and then we go. But I already showed you how you can do that. Simply press Play and you will get your video. In this video, I will show it in Blender. We need to go into our drive, We go into our folder where all the frames are. I want to delete the first frame, and then I download all of these frames. These are 141 frames. If we scroll down, I already downloaded them and I saved them into a folder. Remember, you don't have to do that, but I just want to show this in blender because I think maybe it's valuable for some people. Then we go into blender, we need to make sure that we go to the video sequencer and it all should look like this. Then we aim for a resolution that is perfect for us, 1,900.25080 is perfect. The frame rate, I set it to 25 frames, just like I told you, and the output should be a video. Now, be press Add, and we search for our frames. We press image or image sequence. We need to find them. So these are our images. And if we press A, all of them get selected. Simply press A on your keyboard and then press Add Image strip. Now they are basically all added into your blender. Then we need to make sure that we save them in appropriate way so you can leave them into this folder. Or you can create a new folder right here or search for a folder on your desktop. Just do how you like it, maybe we do it into Paps, accept it. Then we need to adjust our frames. Right here. Frames start, one is okay, but the end is at 141. The last thing we want to do is into the left corner press, render and render animation here you can also see the process. If you like, I will see you as soon as this is rendered. This is normally relatively fast. Now it's rendered, everything will get automatically into the folder that we set the folder wraps. Just remember, you can also do this directly in the Covid Notebook. Just press play. But I think for some people it's maybe cool also to no blender. The next thing is really easy, we go back into shortcut. We include our video right here, and then we just throw it on our timeline, directly over our video. And it's basically done. You can see, as soon as I clap right here, the next frame steps in, and I will get into this cool animated stuff. Let's take one less look at this. I think this is really good. Perfect. Now we want to export this, and basically it's done. In this video, we took a look at Stable B Fusion. It's a Colab Notebook. Behind the Colab Notebook works Stable diffusion, it's relatively easy to use. Just use a video clip where you want to overlay your prompts. As soon as you have found your video clip, you need to go into the Colab notebook, but you only get access if you download it from X on Patri. As soon as the notebook is opened, just connect it, press a few play buttons. If you really like use different models from CBTI, then you simply press Run. All of course give the paths and so on. Just like I told you, this will need a bit of time. Maybe you need something like an hour or so to render this. After it's rendered, you can make a video of it directly into the Cole Notebook. But in this video, I wanted to show you that it's also possible in blender, you can download all your frames and you can make your frames into a video. Also in blender, then we just throw it into shortcut. And it's basically done, so relatively easy, but you need to do a few steps. You need to go to Patrian. It's a bit of work, but thanks to this tutorial, it's easy. Remember, as soon as it starts to render your first frame, you can always post this and adjust your prompts. You can put anything that you like into this prompt. You can also transform, for example, into a turtle or whatever you like. Just try it out. 46. What is next: Create and share your animations: What's next? What you should do with the information from this course. I will tell you, you should explore the forum. And after you explored it, you should create your animations. After you created your animations, you should share them on TikTok, on Instagram, or you can even create your own YouTube channel. This video is have the potential to get clicks, to get likes, and they have the potential to generate followers for you. So make videos and simply post them. If you have followers, you can do a lot of stuff. You can sell them things. You can show them things. And it's simply cool to have a lot of followers. Make your own AIR. Start with the volume diffusion and share it on Instagram and TikTok. And I'm sure you will have a lot of fun doing it. 47. Thank you: The last thing I want to tell you right here in this course is, thank you. Thank you that you put the time into watching this course. Your time is valuable. So I appreciate your time. And if there are any questions at any time of the course, please feel free to ask. I answer super quick. I answer every time as soon as I can. One last thing, if you will, can give me a super good writing. I would really appreciate it. And the last thing I want to say is go explore yourself. I think AI is the future and you are on the forefront of the future because you watched this course. Good luck. I hope I see you in the next one.