
Leveraging AI
Dive into the world of artificial intelligence with 'Leveraging AI,' a podcast tailored for forward-thinking business professionals. Each episode brings insightful discussions on how AI can ethically transform business practices, offering practical solutions to day-to-day business challenges.
Join our host Isar Meitis (4 time CEO), and expert guests as they turn AI's complexities into actionable insights, and explore its ethical implications in the business world. Whether you are an AI novice or a seasoned professional, 'Leveraging AI' equips you with the knowledge and tools to harness AI's power responsibly and effectively. Tune in weekly for inspiring conversations and real-world applications. Subscribe now and unlock the potential of AI in your business.
Leveraging AI
192 | Create AI images like a pro! Using New ChatGPT, Midjourney, and Open Source with Luka Tisler
👉 Fill out the listener survey - https://services.multiplai.ai/lai-survey
👉 Learn more about the AI Business Transformation Course starting May 12 — spots are limited - http://multiplai.ai/ai-course/
AI-generated visuals have become table stakes. But what if you're still playing checkers while others are playing 4D chess?
In this session, we’re breaking open the real world of visual AI — the one most people don’t even know exists.
You’ll see the 3 major paths to creating powerful visuals with AI:
- the conversational path (think: ChatGPT),
- the classic tools (Midjourney, etc.),
- the open-source frontier, where precision, layering, and total control change the game.
Guiding us through this visual odyssey is Luka Tišler — a visual AI educator, workshop leader, and founder of an academy teaching professionals how to wield tools like ComfyUI, ControlNet, and more. Luka doesn’t just play with prompts — he builds pipelines. If you want consistency, control, and pro-level results, he’s your guy.
You’ll walk away understanding the trade-offs between tools, how to make the right choice for your business needs, and why there’s a whole new level of image generation out there waiting for you.
About Leveraging AI
- The Ultimate AI Course for Business People: https://multiplai.ai/ai-course/
- YouTube Full Episodes: https://www.youtube.com/@Multiplai_AI/
- Connect with Isar Meitis: https://www.linkedin.com/in/isarmeitis/
- Join our Live Sessions, AI Hangouts and newsletter: https://services.multiplai.ai/events
If you’ve enjoyed or benefited from some of the insights of this episode, leave us a five-star review on your favorite podcast platform, and let us know what you learned, found helpful, or liked most about this show!
Hello, and welcome to another live episode of the Leveraging AI Podcast, the podcast that shares practical, ethical ways to improve efficiency, grow your business, and advance your career. This is Isar Metis, your host, and we've been talking a lot on the podcast on one of the top use cases of AI today, which is creating. Visual content for business use cases. Now, if you are creating images with AI for either fun or presentations, it has actually become really easy to do. You can do this across multiple tools, but there's still one main issue, which is consistency. And when you're creating. Images. In a business perspective, consistency is key because you have either your brand guidelines or your logo or the actual look of your product that has to stay consistent. Otherwise you cannot use, the images that you're generating. And so creating consistent imagery is. A problem that requires skills and knowledge and the right tools in order to use. Now, while mainstream tools, such as Chachi pt, or Midjourney or Gemini are good enough on the entry level, there are. Other open source tools that are actually providing a whole universe of tools of capabilities that make different things that are problematic with the available standard tools out there. And they solve for that in really beautiful ways. So. The open source world is very much more of a geeky kind of solution. People are like, oh, I don't know how to use it. So today we're gonna demystify a lot of the open source tools, specifically when it comes to creating imagery, specifically when it comes to creating consistent brand relevant product, relevant imagery. Our guest today, Luca Tisler, has been in the world of graphic design and visual content for businesses for many years across multiple roles in multiple companies. But in the past two years, he's been focusing on helping businesses, training businesses, and consulting to businesses on how to implement AI based solutions for company specific branded. Content, which makes him a perfect guest to show us through this process. I'm personally played a little bit with open source visual tools, but not a lot. And so I'm personally very excited and curious to see what, Luca has to share with us. And so Luca, I'm really happy to have you. Welcome to Leveraging ai.
Luka tisler:Hey Isar, thank you for having me again. really excited to be here and talk about, imagery and creation and open source.
Isar Meitis:Thank you. Yeah. Be before we get started. First of all, if you are joining us live, either on Zoom or on LinkedIn. So thank you so much for joining us. I'm sure all of you have other stuff that you can do on, Thursday, noon if you are listening to this after the fact. So first of all. Just so that you know, we do this every Thursday at noon on Eastern time, so you can come and join us. There's always amazing people like Luca that are gonna share use cases, and then you can chat with people in the chat and get to know other people, network, as well as be able to ask questions, which you cannot do if you're listening to the podcast, after it's been, released as a recording. Also one last thing. We are running the AI Business Transformation course for the spring version already. It's been running for two weeks. It's been amazing. There are 25 people in this specific cohort, and they're learning things such as. Generation, video generation, content creation, data analysis, data manipulation, writing, basing prompts, et cetera, et cetera, as well as how to implement AI successfully business wide, how to come up with a business strategy for AI implementation. The next course we just announced the dates is starting in the beginning of August, so we do the. Courses at least once a month, but most of them are private to specific organizations. So if your organization needs training, you can reach out to me, but if you're looking for a course, you can just sign up for yourself. Then we just announced the dates for the next course. You can go to our website or click on the link in your show notes and get to the sign up for the course. but that's it from me. Now, let's give the stage to Luca and let's talk about ai. Image generation. If you have any questions and you're here with us live, feel free to ask them in the chat. If you are listening to this after the fact and you're like, oh my God, I wanna see this as well. So it's gonna be available on YouTube as well. There's gonna be a link in the show notes for you to go to see this on YouTube. But we will explain everything that we're doing and everything that's on the screen so you guys can follow us, even if you're driving your car or walking your dog or doing whatever it is that you're doing, running on the treadmill that you're doing while you're listening to your podcast. So, Luca, the stage is yours.
Luka tisler:Thank you so much, ISAR. so just a brief introduction, my name is Luca, of course, and I've been, in business, business for the last almost 20 years. I started as a video production, but then I quickly moved to video post production. then I moved to compositing, digital compositing, VFX, animation, motion design. So everything, visual, everything that's moving, it's my domain. so about almost three years ago, I found out this little thing, on a discord called Mid Journey, and I knew immediately that this is going to change everything. half of a year later, I, resigned in my company. In the company, I quit because there was not enough time for me to learn. so I spent a lot of time learning and discovering, and of course my journey was not enough. I wanted more and more and more. So the next progression was, of course, stable diffusion. Table diffusion is the first open source AI image, creation platform kind of. it's not as relevant as it used to be because, there has been some, advancement on that area. I. yes, sorry. And also
Isar Meitis:serious leadership issues that almost got them bankrupt and a lot of other stuff. So, so on the business side, they had an amazing technological platform and they weren't doing great decisions on the business side of things.
Luka tisler:Absolutely. I agree. I wish them well. I hope their business is, back on track. and, but they did. A huge thing to the open source community because they were first that opened up the weights for 1.5 and SDXL models. So this is image model is something like large language models. people teach how to, well people teach them and they teach them by. Inputting an image and the description of that image, and they put both as a pair into the magic box. And they do this billions and billions of times. And all of a sudden now we have a machine that knows what is a car, what is a banana, what are the glasses, and can also draw them and it's getting better and better and better. I remember, the first tries were very awkward. the. The pictures resembled the stuff you prompted, but it wasn't that right? But I knew that this is going to progress. So it slowly, slowly started to progress from 1.5 to 2.0 now as the Excel, and then three and 3.5. Just before they released 3.5, there was a new kit on the block called Flux. People went berserk because it was truly a model that understood natural language. So you could start prompting as you speak. And this was amazing. This was incredible because no image generator, understood natural language. We had to talk to it. With tags, so like, um. a city, cyberpunk, dawn, a car reflections, good quality and so on. And, you know, we just had our fingers crossed to get those images as best as possible. But now, I mean, you can talk to all of them in natural language. Do this, do that. Give me this, give me that. So, the progression is amazing and, there has been also the proprietary models. The rates in quality. So now we have like imaging. We have Ideal Gram. we have many, many, many image models that are absolutely gorgeous and they produce amazing images. And the best thing is that with the proprietary models, you have, control to an extent. So, you can control your imagery, but most, by prompting. but open source has something more. And, it started with stable diffusion, and it's called control net. So those are kind of differently trained models to guide, the imagery you want it to look like. So, for example, you have to know how latent diffusion technology works. I'll, I'll keep it very short and simple. So it starts with noise, just nonsense, complete nonsense. And then it starts, putting that noise away. And with each step, it, takes a bit of noise away and inject some of your ideas. That is conditioning AKA prompting. So, imagine like you are lying on the field and staring into the sky, watching clouds, and you can see a shape of a turtle in a cloud and you decide, okay, that's a turtle, right? Latent diffusion works in similar way. So it kind of says, okay, I can see that inside. I'm gonna shape that into the image that I already know. when I go to my magic box and see what's, for example, sunglasses. So, aha, okay, we have sunglasses Now I will shape this noise with conditioning, AKA prompting, and we will get a nice result. So stable diffusion. Opened up the models so we can, create our own models. We can teach our own models. we can fine tune them and we can create Lauras about this a bit later. but now this noise has constraints. now all of a sudden you can. control it by, let's say, with the post. So if you prompt, the man has his left leg and right hand in the air, if you prompt us, you probably won't get. That's, specific results, right? Or you will have right hand or left hand. Maybe your hand will have seven fingers and so on. But with control nets, we guide this noise. So we say to the model, okay, here is the image. Of the man with in this position and via control net, we transfer the post to our image. So we get, the result that we actually want. And, post is just one of those many control nets. We have lots of control nets, and they're different for each use case. Some of them are good for, like I said, poses, some of them are good for architecture. Some of them are good for, well, a lot of use cases. and they're getting better and better and better. yeah.
Isar Meitis:What else? Yeah. So to add my 2 cents to what you just said, to connect a few dots of the things you mentioned. One is, how these models are getting trained. The way they're getting trained is they took gazillion images and then noise them step by step by step. So what the model is getting is getting the final. Photo that is an existing image of anything you can imagine. And then it gets another level with 2% noise and another level with 6% noise and another level with 8% noise all the way to a hundred percent noise. So that's how they teach the model to basically then reverse the process to de-noise from a hundred percent noise to an outcome because it has seen. Multiple noising levels of any image that you can imagine. So this is how it works. And the control nets allows you to guide the process beyond the prompt itself by giving it references that it understands whether about position lighting, reflection. Outlines of things, like literally anything you can imagine that is a part of creating an image, you can then use, think about it as a, an additional reference for the denoising process or for the creation process. It doesn't really matter how the creation happens, you're just giving it more references than the written reference that you had in your prompt.
Luka tisler:Exactly. that's a very important part and, I'm sorry I skipped it, but it is, noising and denoising process. It goes both ways. So you are 100% right. so, but not just open source has a control. Lately we've seen the development, also in proprietary models. For example, mid Journey has this amazing thing called omni reference. and you can actually. put in the platform, the image you want to recreate, and it'll recreate it very, very good. you can input your character so the consistency is kind of solved. We'll probably never have 100% consistency, but I mean, it depends what you're working with. most of AI lands on, digital media. and on digital media, we usually have smaller screens, right? And, the human eye just says, for example, Robbie Margot. Robbie Margot is very, very famous in AI community because everybody is testing with her face. And, if you see a person similar to, Robbie Margot, you will just tag her as Robbie Margot, and that's it. It doesn't have to be 100% consistent. It's good enough for our brain, but on the larger scale, it's just not good enough. Then we have to use Photoshop, then we have to use in painting, Well, we get about 80 to 85% done with ai, but the last 10 or 15% we have to work on manually. And this goes for all imagery actually. if you want to bring your imagery to another level, you will have to add the human touch in the process of, the creation of the image. So Photoshop is not going anywhere.
Isar Meitis:for now, I'll say something about the whole Photoshop thing and then we can dive into the actual examples, but Photoshop is not going anywhere for now. I agree. I think what Canva did to graphic design on the lower levels AI will do to Canva, because my gut feeling tells me that these two universes are going to merge. And you might still use Canva, but you will use AI in Canva versus dragging and dropping and using templates. You will have an idea in your head and you will request it, and it will be generated on the fly. And I think what's gonna happen is features that you currently have in Canva will migrate into Chachi, PT Gemini, mid Journey, and so on. Where you will understand layers and will understand text, you will understand grabbing a component of the image, filling out the background where you move the thing from, in a very intuitive way. And there is very small doubt in my mind that's gonna happen in the next few months, meaning this merging of design and. An AI image generation to one unified environment. And I do think that the only gap between that and the professional world is some additional tools that just don't exist in the simple tools. And I think it's just a matter of time, until they're there as well. So I'll add. Two things that are very obvious to people who are in the professional side. One is upscaling. So like you said, if you wanna print a billboard that will cover the side of a building, the resolution that midjourney gives you is just not good enough. But there are already amazing upscales today, and once it's gonna be built into Midjourney, then that problem is solved. and the other is. Really small fine things on specific textures, specific fonts versus just random fonts and things like this. masking of specific aspects. Those of you know what I'm talking about. And so I assume these universes will merge together and there's gonna be more professional tools with AI built into them. like. Photoshop, which already has, you know, their version of AI built into it. And there's gonna be the more basic tools that are either gonna be Canva or working in the actual tools themselves, such as Gemini, me, journey, et cetera.
Luka tisler:Absolutely. and we can also see the rights of agents. So you don't prompt anymore. You talk to machine. Yeah. So do this, do that, change this, change that. And I think this is one of the biggest, things that is happening in AI right now. because we won't be, Producing imagery and videos with our mouse. We will guide the model with our language, with our speech, and all the changes will be instant.
Isar Meitis:Yeah. shall we jump into examples?
Luka tisler:yeah. Yeah. Um. what do you wanna
Isar Meitis:see? I think it would be interesting to see, first of all, for people to see like control net, and what exactly and how it works. So just to give examples of things. And I think doing the same thing in mid journey would be useful as well. So showing kind of like the omni reference and how that works. And I think. Both these things will show people stuff that they may or may not know how exactly, it works, and maybe they haven't experimented with that before.
Luka tisler:Absolutely. So for example, here is the midjourney, and, you can, well, I just created this mascot, this image for, a language school that I'm working with and I used. let me see. I
Isar Meitis:used, so for those of you who are not seeing as Luque is, searching for those of you just listening, we're looking at Mid Journey, which is one of the better AI image generation tools. And we're looking at like a Yellow Frog or lizard that is teaching in a classroom, as the mascot. And it's wearing like a blue hoodie, with the logo of the school. I think what we're going to see is how it was actually created.
Luka tisler:I don't have this image by hand right now. Let me just close the Lighthouse Academy website. Okay. Subtle, subtle. I, yeah. But let's try to create something with Margo Robbie, scratch. first of all, no one prompts. From their head anymore. So everybody's using lms. So, give me an, an image prompt for a woman in red dress. posing in front of, Eiffel Tower, and we're going to make it, editorial
Isar Meitis:very quick. So again, for those of you're not seeing, we're writing this prompt in Chachi pt, and it is going to give us the prompts, better prompts with a lot more details and qualifiers, than if we wrote the prompt ourselves. By the way, there are multiple custom gps already created and available that are very good at that, that are built to give you highly detailed prompts for, image generation. But as you'll see in a minute, even just writing what Luca just wrote will give you a very long fourth sentences worth of details in a prompt that you can then paste into whatever it is that you're using.
Luka tisler:All right, so our prompt is, I won't read it because it's just too long. But anyway, we will set settings to, let's say 16 by, oh, let's go with three to four. Let's raise stylization and variety. everything is okay Version. Ooh. I experimented with Mid Journey version three. Okay. Stylization a bit higher and we are good to go and press Oh no. We will, import our image that I took from Internet of Marco. Robbie, We didn't say anything about Margot Robbie. We just said, a woman, a stylish woman. Yeah. And we dragged and dropped omni reference, into omni reference and omni strength. You can, I. Use it from zero to 1000, I will use the low value about 300. And now, oh, we can also, insert image prompts and style references. So if we don't know how to describe the style, we can just drag and drop the image of a style we like or. Use image prompt so we kind of get the composition that we want, from another image. Okay.
Isar Meitis:So again, just to explain what these are, going back to our initial conversation, when you are now prompting these tools, in addition to the written prompt, you can drop in images and use them for different purposes. So you can use an image as a reference for a person, a reference for a style, a reference for the composition of the actual image itself, a reference. and when we say style, this could be. Very broad, like cartoonish versus realistic, but it could go into way more detail, like the color palette that is going to use and so on. and you can do all of that, definitely mid journey that now has it broken up into different levels and tools. I.
Luka tisler:And here we are. We get a woman that is similar to Robbie Margo, but she is not Robbie Margo. Yeah. because our reference was set very low, but now we can use everything and we can up the omni strength, so our woman will be. similar, a bit more to Marco Robbi.
Isar Meitis:So again, for those of you who are not watching this, there's a slider next to the image that you upload and you can move it left to right between zero and a thousand. You can also enter the parameter as a parameter, but they mid journey now moved it to make it, more user friendly where you can just move the slider around, and you can control how much weight. The image that you uploaded will have in the output, and if you bring it closer to a thousand, it's gonna be very similar to the image you uploaded. again, just to explain to people who are not watching, the image that we uploaded is just a face. You can't see the entire person. the. Prompt is for a woman in a red dress. Hence we see the entire woman and it still knows how to pick the head of the person from the other image and apply it to the full body shots that we're doing right now.
Luka tisler:So the thing with Mid Journey is after the prompt, it sets a couple of parameters. And you have to be very careful. So if you change the parameters, you always have to delete your parameters at the end and use the new settings that you are giving it. Because right now we. created four more images, but the omni reference weight was left at 300. So because, preferences behind the prompt are stronger than the preferences that we use through the website. So each time when we, change our preferences, we have to delete those from the prompt and set. Everything else through the website and now omni reference is high enough and we will get, a woman that looks like Margot Robbie because we created two sets of images. But, I didn't set the omni weight, high enough. And while you are waiting, we can actually check out Chad GPT because Chad GPT has introduced, images in March, I think. And now we can ask it to, create the image prompt. So we will use the same prompt. So here is the prompt and I will ask GPT Create an image based on this prompt, and I need a wide screen aspect ratio. So GPT images are using different kind of, architecture. if we check out how, oh, there it is. And yeah, she looks much more like Margot Robbie.
Isar Meitis:Yep.
Luka tisler:So, as we said at the beginning, the latent diffusion works with adding noise. Yeah. well, Chad GPT uses different kind of structure and it's, I don't know how it works, but it's adding the details from up to bottom. Yeah. Yep. You can't communicate with Midjourney like communicating with GPT because MIDJOURNEY is not large language model. it knows images, but the images that it's creating are much, much higher quality than images that we're getting from GPT because GT's first priority is language or words. they also did a good job training it to, create images. But, it's not as good as my journey quality wise.
Isar Meitis:there's an interesting question while we wait for Chachi pity to generate the image. you know what I'll touch before I jump to the question I wanna kind like go back to what you said as far as the differences between Chachi PT and ME journey. And then we're gonna go to like, control net and stable diffusion or maybe comfy y not to scare people too much away, but pros and cons of the different systems. From a professional approach perspective, mid journey still provides better results. yes. From a day-to-day usage perspective, Chachi PT generates good enough results in many cases, and to get to these results are easier. Why? Because it understands context and you can provide it a lot more information that you can, because it's a complete conversation. It's not just, here's the prompt to create the image, and now I wanna change something in the image. I need to go and write. A completely new prompt that will start from scratch. Basically, it understands the conversation, understands the context. You can upload your brand guidelines to Chachi PT as an attachment, and you will know how to use them in the image. which majorly. Does not know how to do because it does not know how to read PDF documents. And so things like that are things that are benefits of using Chachi pt. another advantage of Midjourney is speed. Midjourney generates four images in about 10 seconds. Chachi PT creates one image in about a minute. And so if you want to iterate a lot, doing it with is gonna work. A lot faster. and as I mentioned, me, journey has built a lot of tooling around its image generation with different sliders and bars and controls of different things, and being able to reference previous images and being able to reference previous prompts and being able to add all these different things where that does not exist on the Chachi PT side. if you're a beginner, I think starting with Chachi, BT is easier. If you're a more advanced user, you can get better results with using midjourney either way. Both of them, I think for the average user, for generating images, for presentations and or basic social media stuff, both tools are definitely adequate.
Luka tisler:Yeah, absolutely. And you're spot on with adding GPT, different pieces together and it knows how to create images. with those pieces, right, you can upload the, let's say, an image of a chair, image of a wardrobe, an image of a bed, an image of a picture, a painting, and you drop. those images in Chad, GPT and say, create a room out of these images and it'll be spot on. It'll be perfect. But as you said, it has its limitations and for professional usage it's not good enough because we need to create images fast. And, Very, very, high quality. Yeah. So Chad g PT is just not good enough. also, nor Chad, GPT nor journey, cannot batch render images. Yes. So you can't just say, create 100 images based on this prompt, and I will select the perfect one, while, open source tool can
Isar Meitis:do that. So let's really jump to that. Let's jump to maybe Confu Eye just to show people what it is and I know it might scare people in the first minute, but I think we can explain what it is and how it works and maybe demystify it a little bit.
Luka tisler:yeah, this is my, workflow. It's a bit more advanced. We can check out the, where is a, so again,
Isar Meitis:for those of you who don't see the screen, what comfy UI is Think about a flow chart of a process where every step of the flow chart, you have multiple tools that you can connect that. Impact how the image will be generated. So it's not post-processing, it's actually how the image will be created. And those of you who can see, you can see there's like, I don't know, 20 different boxes each one with different components and lots of lines that com combining them. That looks like a web of different things. But each and every one of these boxes adds another layer of control over the. Process on how the image is gonna get generated. And by combining them together, you can be a lot more specific with how the output would look like. Going back to consistency, when you can control the lighting, the angle, the pose, the graphics, the style, the. Entry images, the size, the like literally every aspect of how the model actually works at the model level, you will get a lot more consistent outputs. Which going back to what Luca said, if you want to now. Mimic a photo shoot where you're gonna bring a model. Let's take your example of a, a girl in a red dress in Paris, in front of the Eiffel Tower. you're not gonna take two pictures if you're doing an actual photo shoot. You're gonna take 250 50 pictures and then pick the two best ones and then work on them in Photoshop, and then pick the final one based on that. And. You can do that in confi I with no problem with a much higher level of consistency and with a lot more control of exactly how it's gonna look like. So think about a professional photographer. You don't just take random pictures, you set up the camera to the right setup with the right aperture value, with the right lighting, with the right camera, with the right. ISO parameters, like all the stuff that you want to control in the image you will set up in your camera. You don't just randomly shoot and click. And this is kind of like what you can do in the digital world with comfy UI and open source tools.
Luka tisler:that was so beautifully said. I can give another comparison. So, if you just want to use your computer, you will probably buy Mac. Yeah. But if you want to know every nitty gritty detail, you will buy this computer by parts and you will assemble it for yourself. Yes, because you have control over each part. What part, what is the compatibility, where you will put it? And you know this thing by heart. Not just opening a computer, start working, but you know everything that's happening inside, every process. So Kafi is the same. For example, here is a note where you define the checkpoint or the model you will be using. This is the Laura so the note for specially trained small models that you can, train yourself. I. we have clip loader, we have, VAEs, we have resolutions. We have so many things that you can toy with and see what are the results. So I would say that comfy is for people that want to know more and. Most of all are very, very curious because if you don't have curiosity in your genes, under your skin, then this will be just too overwhelming. I understand people who don't want to use comfy. They just want to produce, but, well, I need a bit more. I need control. I need to know how I will. construct the image and I can change the schedulers. I can, change the samplers. I can change anything and I can influence my image based on all these parameters. now in my journey, you have a couple of parameters here. You have thousands, and you can combine them, and you can do well a lot of things. But what we see right now. It's a fairly complex, not too complex, but fairly complex workflow. So maybe it would be better if I would show you one a bit more simple. Exactly. Yeah. So for example, let's take a control net workflow, and I have a couple of missing notes. That's okay. not notes, but, yeah, so let's just switch a couple of stuff. VAE. not prune, but yeah, that's okay.
Isar Meitis:So again, for those of you who are not watching, now we're looking at a process that has four or five different nodes versus the. 30 that was on the screen before nodes are basically building blocks. So think about them like Legos. So you can build a simple Lego with a bunch of parts, or you can build a really complex Lego with multiple parts. And the more you know what the parts do, the more you can use more Legos to build more sophisticated stuff. So you don't have to start with like the 30 step process. You can start with a four step. Process and build something that will be a lot easier to construct, but still will provide you a lot more control than doing it in, let's say, mid journey on Definitely on Chachi pt.
Luka tisler:Correct. So what we did now is, okay, everything is ready, we can run it. So I inserted the image reference. and I expect my new image will have a person that stands exactly the same as the woman on the photo, but I have to give it. So,
Isar Meitis:so again, This is a tool that is not for somebody who's just wanna play around and create images. This is a professional tool that if you do this for a living, you can do this. Or like Luca said, if you're just really curious about image generation and you want to experiment with. Doing things that are beyond what the image generation tools can do, you can do that as well. And again, the cool thing here is that you literally control every single thing. And one of the cool things that we talked about several times in this episode already is control nets, right? So as an example, you can create, use a. Model of a person or a cat or an anything, and then have the output resemble it in whatever way you want. So either resemble it in the texture or resemble it in the fur that it has, or resemble it in the pose that it's standing or walking or jumping in and so on. And it will know how to do that because there's one component that you control, which is. The pulse of the person or the texture of the image and so on, where it will follow, what the control net tells it to do.
Luka tisler:Right now, control net is not working. I don't know why.
Isar Meitis:Yeah. The beauty of live demos. So Let's do a quick summary, of everything we talked about. Sure. And then we'll see if you have any final thing to add. Yeah. There are really three main channels today, right? One is the more. Professional by yet readily available Proprietary models like Mid Journey. There is the open source universe that has tools like Stable Diffusion and Flux that have multiple ways to use them, including through com v ui, which just starts another layer of control over just using the model itself. But you have to have an open source model to do that because literally what Comfy UI does is it controls. A huge number of parameters within the model itself. That literally what it allows you to do. And then there,
Luka tisler:there are, if I may interrupt you, there are two. Yeah, yeah, for sure. First of all, you need a powerful computer. Yes. You need a graphic card that has a lot of VAM, so your basic. graphic card is not good enough. it has to have at least eight gigabytes of VRA or more. The more, the better. and the second one is a steep learning curve. So it's not something that you just toy with. You have to invest time to learn how to use those tools. But when you use them, when you start using them, you become invincible. You can create anything you want. If, of course things are working.
Isar Meitis:Yes. Well, I think they're working. It's exactly the same thing as a complex machine, right? The chances of getting one thing wrong and then the output is not exactly what you wanted. The more complex the machine, the more chances you're gonna get something that is not, what you meant. and then really the other component is going to a tool like Chachi, PT or Gemini, with both can create. Good enough images for many day-to-day things, but with a lot less control, a lot less parameters, a lot less capabilities to know exactly what's gonna happen. For many daily use cases, like creating presentations or an image for a social media post, that is the easiest way to go. If you are a professional designer or if you need to create stuff that is more consistent and so on, going the open source path, we're just gonna give you a lot more capabilities. Luca, if people wanna follow you, learn from, you know, more about what you do, how you do it, hire you, what are the best ways to do that?
Luka tisler:well in social media I'm not spread around. I mainly use LinkedIn and Instagram here and there. So you can meet me on LinkedIn and, send me a message, connect to me and, if you have any questions, I will be happy to answer them.
Isar Meitis:Awesome. Luca, thank you so much. I think this was very, educational and valuable to people. I think most people. At least people that I know, and I know a lot of people who are playing with AI do not know a lot about the visual site in general, and definitely not about the open source models and how they differ and what benefits they provide. So I'm sure this was very, helpful to people. Thank you so much. To everybody else who has joined us. I appreciate you being here. I appreciate, spending the time with us, and, being active in the chat. And until next time, have an awesome rest of your week.