Bernard Maltais
Bernard Maltais
  • Видео 27
  • Просмотров 390 546
New Stable Diffusion amazing LoHA training method
I present a new training method to obtain very clean and sharp models with high likeness and flexibility. The method does not require caption files and produce better results than all previous methods used.
Here is the resulting model if you want to try it: civitai.com/models/79210/kelly-jenner
Просмотров: 10 084

Видео

How to use tensorboard to look at Stable Diffusion training logs
Просмотров 4,4 тыс.Год назад
TensorBoard is a tool for visualizing and understanding machine learning models and experiments1. It can track and display metrics such as loss and accuracy, show the model graph, and project embeddings to a lower dimensional space2. Stable Diffusion is a method for training generative models using diffusion processes. In this video, you will learn how to use TensorBoard to look at Stable Diffu...
Modeling The White Queen as a Stable Diffusion LoRA
Просмотров 6 тыс.Год назад
This tutorial will go through my new process for producing LoRAs. It should give you tips and tricks on how to create your own. Civit.ai model: civitai.com/models/10347/the-white-queen
Quickly fix bad faces using inpaint
Просмотров 24 тыс.Год назад
This is a short tutorial on how to fix bad faces in stable diffusion using the inpaint feature.
Creating a LoRA weight using kohya_ss GUI, part 2: Training the model and creating outputs
Просмотров 63 тыс.Год назад
This is the part two of creating a LoRA weight. It will explain how you go about creating your model and generating realistic outputs from prompts. kohya_ss GUI: github.com/bmaltais/kohya_ss kohya_ss LoRA extension for webui: github.com/kohya-ss/sd-webui-additional-networks Part 1: ruclips.net/video/N4_-fB62Hwk/видео.html
Creating a LoRA weight using kohya_ss GUI, part 1: Dataset creation and preparation
Просмотров 83 тыс.Год назад
This is the part one of creating a LoRA weight. It will explain how you can go about using images available via google search to create your model. kohya_ss GUI: github.com/bmaltais/kohya_ss kohya_ss LoRA extension for webui: github.com/kohya-ss/sd-webui-additional-networks Part two: ruclips.net/video/k5imq01uvUY/видео.html
Quickstart: Training a stable diffusion model using kohya_ss GUI
Просмотров 44 тыс.Год назад
Quickstart tutorial on how to train a Stable Diffusion model using kohya_ss GUI.
FSD Beta 10.11.1 Canada - Une promade de samedi matin - Complet
Просмотров 7252 года назад
Une promenade d'environ une heure par un beau samedi matin d'avril à Gatineau, Quebec, Canada. En français pour mes amis Quebecois et français de partout dans le monde.
FSD Beta 10.11.1 - Canada, Quebec - Test new camera extension and audio recorder
Просмотров 2282 года назад
Short test to see how the camera extension and new audio recorder perform. You are the judges. Test court pour évaluer comment l'extension pour la caméra et le nouvel enregistreur audio performe. Vous êtes les juges.
FSD Beta 10.11.1 20220329 end of day drive
Просмотров 1572 года назад
Conduite avec FSD Beta 10.11.1 sur routes avec de très mauvaise lignes suite à notre hiver rigoureux. Drive with FSD Beta 10.11.1 on roads will really bad markings due to the rough winter we just had.
Model 3 - Beta FSD 10.11.1 - 2022.4.5.20 - 28 March 2022 - Long drive with roundabouts
Просмотров 2322 года назад
Highlights: 0:00 Intro 10:37 Uncomfortably close to roundabout exits 13:22 Kids playing hockey in the street 13:37 Almost running over a bicycle on the side of the road 14:08 Passing a rolling car exiting from his driveway in a neighbourhood street... not cool... 15:30 Turning the wrong direction toward another car 24:03 Not positioning in the right lane to turn right after the highway exit 31:...
FSD Beta 10.11.1 Canada, from good to bad (weather)
Просмотров 3182 года назад
A drive with FSD Beta leaving with nice clear weather and ending in a snow-covered road challenging FSD Beta.
Model 3 - Beta FSD 10.11.1 - 2022.4.5.20 - 27 March 2022 - Route du Pirate, Gatineau, Quebec
Просмотров 2392 года назад
Etablissons le standard d'une route typique répétable afin d'évaluer la progression de FSD Beta à Gatineau, Quebec.
Model 3 - Beta FSD 10.11.1 - 2022.4.5.20 - 27 March 2022 - Premiere Sortie Gatineau Quebec
Просмотров 1,9 тыс.2 года назад
1iere video de FSD Beta Canada 10.11.1 prise le jour du lancement de celle-ci.
Cybertruck testing side mirrors and turning
Просмотров 38 тыс.2 года назад
www.google.com/maps/contrib/104756815649499553761/place/ChIJDQFm_N1qkYARblRN1PJYIpU/@37.9765144,-121.0400129,9z/data=!4m6!1m5!8m4!1e2!2s104756815649499553761!3m1!1e1
Tesla Cybertruck testing turning radius and visible side mirrors
Просмотров 100 тыс.2 года назад
Tesla Cybertruck testing turning radius and visible side mirrors
Tesla Gigafactory Texas Waterfall
Просмотров 3673 года назад
Tesla Gigafactory Texas Waterfall
Stuff you see driving a Tesla vol.1
Просмотров 3863 года назад
Stuff you see driving a Tesla vol.1
Comment ajuster son frunk sur sa Tesla Model 3
Просмотров 3503 года назад
Comment ajuster son frunk sur sa Tesla Model 3
YouTube on my Tesla with authentication
Просмотров 1963 года назад
RUclips on my Tesla with authentication
Chauffage de la Tesla model 3 2021
Просмотров 1943 года назад
Chauffage de la Tesla model 3 2021
Tesla Model 3 - Recharge de la batterie
Просмотров 2,4 тыс.3 года назад
Tesla Model 3 - Recharge de la batterie
HOWTO Deploy Azure infrastructure using the GC Cloud Accelerators Azure Development VM
Просмотров 1245 лет назад
HOWTO Deploy Azure infrastructure using the GC Cloud Accelerators Azure Development VM
HOWTO GC Cloud Accelerators Azure Dev VM
Просмотров 1215 лет назад
HOWTO GC Cloud Accelerators Azure Dev VM
How to add custom ringtone on android
Просмотров 836 лет назад
How to add custom ringtone on android
Halloween with Philips hue lightning strobe light
Просмотров 10 тыс.7 лет назад
Halloween with Philips hue lightning strobe light
Confused Canadian geese
Просмотров 2759 лет назад
Confused Canadian geese

Комментарии

  • @beragis3
    @beragis3 7 дней назад

    Kohya's interface changed so much since this tutorial that I can't follow it.

  • @NorwegiansCantCook
    @NorwegiansCantCook Месяц назад

    I came pretty far

  • @tim1398
    @tim1398 2 месяца назад

    How can we log more items like time_step etc?

  • @gv-art15
    @gv-art15 2 месяца назад

    Thanks a lot

  • @user-wk4np7fh2v
    @user-wk4np7fh2v 2 месяца назад

    cant even open the file browser to pick a model "TCL was probably not installed correctly" holy shit trying to do anything with this makes me want to commit suicide

  • @vivavideo-videofilmer
    @vivavideo-videofilmer 3 месяца назад

    horrible tutorial

  • @bokabanana6398
    @bokabanana6398 3 месяца назад

    hello does kohya_ss upload my training images to a server or something? i have alot output traffic if i try to train with it

  • @tharindumadushan1569
    @tharindumadushan1569 4 месяца назад

    Can you give us New Tutorial they is change everything now :(

  • @HanSolocambo
    @HanSolocambo 4 месяца назад

    2:45 So g-r-e-a-t, thanks. I can finally check my tensorboards even after a Kohya/Onetrainer training is finished. Didn't how to access those logs folders in tensorboard anymore. Cheers for that. Very helpful video (b^-^)b

  • @jameswest9552
    @jameswest9552 4 месяца назад

    Juste efficace, merci 😉

  • @ismaildugu1998
    @ismaildugu1998 5 месяцев назад

    help please it says File "D:\AI\KohyaTraining\kohya_ss\sd-scripts\train_db.py", line 118, in train text_encoder, vae, unet, load_stable_diffusion_format = train_util.load_target_model(args, weight_dtype, accelerator) File "D:\AI\KohyaTraining\kohya_ss\sd-scripts\library\train_util.py", line 4113, in load_target_model text_encoder, vae, unet, load_stable_diffusion_format = _load_target_model( File "D:\AI\KohyaTraining\kohya_ss\sd-scripts\library\train_util.py", line 4067, in _load_target_model text_encoder, vae, unet = model_util.load_models_from_stable_diffusion_checkpoint( File "D:\AI\KohyaTraining\kohya_ss\sd-scripts\library\model_util.py", line 1008, in load_models_from_stable_diffusion_checkpoint info = unet.load_state_dict(converted_unet_checkpoint) File "D:\AI\KohyaTraining\kohya_ss\venv\lib\site-packages\torch n\modules\module.py", line 2152, in load_state_dict raise RuntimeError('Error(s) in loading state_dict for {}: \t{}'.format( RuntimeError: Error(s) in loading state_dict for UNet2DConditionModel: size mismatch for down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([320, 768]) from checkpoint, the shape in current model is torch.Size([320, 1024]). size mismatch for down_blocks.0.attentions.0.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([320, 768]) from checkpoint, the shape in current model is torch.Size([320, 1024]). size mismatch for down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([320, 768]) from checkpoint, the shape in current model is torch.Size([320, 1024]). size mismatch for down_blocks.0.attentions.1.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([320, 768]) from checkpoint, the shape in current model is torch.Size([320, 1024]). size mismatch for down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([640, 768]) from checkpoint, the shape in current model is torch.Size([640, 1024]). size mismatch for down_blocks.1.attentions.0.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([640, 768]) from checkpoint, the shape in current model is torch.Size([640, 1024]). size mismatch for down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([640, 768]) from checkpoint, the shape in current model is torch.Size([640, 1024]). size mismatch for down_blocks.1.attentions.1.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([640, 768]) from checkpoint, the shape in current model is torch.Size([640, 1024]). size mismatch for down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). size mismatch for down_blocks.2.attentions.0.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). size mismatch for down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). size mismatch for down_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). size mismatch for up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). size mismatch for up_blocks.1.attentions.0.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). size mismatch for up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). size mismatch for up_blocks.1.attentions.1.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). size mismatch for up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). size mismatch for up_blocks.1.attentions.2.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). size mismatch for up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([640, 768]) from checkpoint, the shape in current model is torch.Size([640, 1024]). size mismatch for up_blocks.2.attentions.0.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([640, 768]) from checkpoint, the shape in current model is torch.Size([640, 1024]). size mismatch for up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([640, 768]) from checkpoint, the shape in current model is torch.Size([640, 1024]). size mismatch for up_blocks.2.attentions.1.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([640, 768]) from checkpoint, the shape in current model is torch.Size([640, 1024]). size mismatch for up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([640, 768]) from checkpoint, the shape in current model is torch.Size([640, 1024]). size mismatch for up_blocks.2.attentions.2.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([640, 768]) from checkpoint, the shape in current model is torch.Size([640, 1024]). size mismatch for up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([320, 768]) from checkpoint, the shape in current model is torch.Size([320, 1024]). size mismatch for up_blocks.3.attentions.0.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([320, 768]) from checkpoint, the shape in current model is torch.Size([320, 1024]). size mismatch for up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([320, 768]) from checkpoint, the shape in current model is torch.Size([320, 1024]). size mismatch for up_blocks.3.attentions.1.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([320, 768]) from checkpoint, the shape in current model is torch.Size([320, 1024]). size mismatch for up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([320, 768]) from checkpoint, the shape in current model is torch.Size([320, 1024]). size mismatch for up_blocks.3.attentions.2.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([320, 768]) from checkpoint, the shape in current model is torch.Size([320, 1024]). size mismatch for mid_block.attentions.0.transformer_blocks.0.attn2.to_k.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). size mismatch for mid_block.attentions.0.transformer_blocks.0.attn2.to_v.weight: copying a param with shape torch.Size([1280, 768]) from checkpoint, the shape in current model is torch.Size([1280, 1024]). Traceback (most recent call last): File "C:\Users\isodu\AppData\Local\Programs\Python\Python310\lib unpy.py", line 196, in _run_module_as_main return _run_code(code, main_globals, None, File "C:\Users\isodu\AppData\Local\Programs\Python\Python310\lib unpy.py", line 86, in _run_code exec(code, run_globals) File "D:\AI\KohyaTraining\kohya_ss\venv\Scripts\accelerate.exe\__main__.py", line 7, in <module> File "D:\AI\KohyaTraining\kohya_ss\venv\lib\site-packages\accelerate\commands\accelerate_cli.py", line 47, in main args.func(args) File "D:\AI\KohyaTraining\kohya_ss\venv\lib\site-packages\accelerate\commands\launch.py", line 1017, in launch_command simple_launcher(args) File "D:\AI\KohyaTraining\kohya_ss\venv\lib\site-packages\accelerate\commands\launch.py", line 637, in simple_launcher raise subprocess.CalledProcessError(returncode=process.returncode, cmd=cmd) subprocess.CalledProcessError: Command '['D:\\AI\\KohyaTraining\\kohya_ss\\venv\\Scripts\\python.exe', 'D:\\AI\\KohyaTraining\\kohya_ss/sd-scripts/train_db.py', '--bucket_no_upscale', '--bucket_reso_steps=64', '--cache_latents', '--enable_bucket', '--min_bucket_reso=256', '--max_bucket_reso=2048', '--learning_rate=1e-05', '--learning_rate_te=1e-05', '--logging_dir=D:/AI/kohya destination\\log', '--lr_scheduler=cosine', '--lr_scheduler_num_cycles=1', '--lr_warmup_steps=200', '--max_data_loader_n_workers=0', '--resolution=512,512', '--max_train_steps=2000', '--mixed_precision=fp16', '--optimizer_type=AdamW8bit', '--output_dir=D:/AI/kohya destination\\model', '--output_name=last', '--pretrained_model_name_or_path=D:/AI/sd.webui/webui/models/Stable-diffusion/dreamshaper_8.safetensors', '--save_every_n_epochs=1', '--save_model_as=safetensors', '--save_precision=fp16', '--train_batch_size=1', '--train_data_dir=D:/AI/kohya destination\\img', '--v2', '--xformers']' returned non-zero exit status 1.

  • @tristanwheeler2300
    @tristanwheeler2300 5 месяцев назад

    oldie but goldie

  • @bulgingbattery2050
    @bulgingbattery2050 5 месяцев назад

    I have a folder with 200+ images in it, and Kohya_SS says that there are not images in that folder.

    • @justinschumann8215
      @justinschumann8215 5 месяцев назад

      You need to put in a folder containing the folder with the images, aka the parent folder of the image folder.

  • @user-kh2zk2rn7i
    @user-kh2zk2rn7i 6 месяцев назад

    Hello Bernard, I followed your excellent presentation step by step, but I cannot obtain models under the "safetensors" extension! In the "Models" folder there is only one .json file... Do you have a solution to offer me? Thanking you.

  • @quercus3290
    @quercus3290 6 месяцев назад

    awesome, a tutorial that actualy tells you what you need to know. Well done.

  • @oritfx
    @oritfx 7 месяцев назад

    Thank you for the tutorial Bernard. I was wondering if it was possible to generate a smaller file (I am thinking LoRA size with dreambooth results). Do you know if that can be accomplished?

  • @jackkoveleskie2504
    @jackkoveleskie2504 8 месяцев назад

    I just had the exact same phenomenon over my house! Hundreds of confused geese. It was weird! So I googled "confused geese" and saw your video. Gettysburg, PA. 12 Jan 2024. 1615

  • @flodyto
    @flodyto 8 месяцев назад

    Hola gracias por el tutorial. Hay alguna alternativa gratuita a Gigapixel que pueda hacer lo mismo???

  • @Genteka
    @Genteka 8 месяцев назад

    This videos obsolete now, all can be done on tensor art.

  • @RodrigoLM92
    @RodrigoLM92 9 месяцев назад

    hey! I recently started following your channel and watching your videos about Kohya, LoRa and other stuff... but I'm having difficulty with some things... I would like to know if you can help me create a LoRa and in the meantime clarify some doubts that I'm having... I'm willing to pay for your time. btw congratulations on your videos... they are very enlightening and dynamic haha

  • @damasoroma
    @damasoroma 9 месяцев назад

    Hi Bernard, Thanks a million for your efforts and your tutorial, I watched so many tricks thanks to your video! I was wondering I could I create a specific part of a body I would like to focus on man's chest (muscles and hair) and I was wondering if I need to takejust chest training images or the full body pictures of a man. And what about regularization pics? Should I take just chest or face or the full body? I'm a bit confused. Thanks a million!

  • @MrCracktuber
    @MrCracktuber 9 месяцев назад

    Merci Berny

  • @tonychong8181
    @tonychong8181 9 месяцев назад

    Thank you so much for all your work! Learned a lot from this video. Subscribed!

  • @RobertMichael-jh6fu
    @RobertMichael-jh6fu 9 месяцев назад

    I have a 4090, if i don't care about speed, should i do float precision for better accuracy? whats a good basic config for highest accuracy?

  • @gnetwork
    @gnetwork 9 месяцев назад

    First, thank you for your Guide. But as other mentioned, it has been trained on a known person what makes it a bit hard to see if the values are correctly. Could you make a video with an unknown person?

  • @laperchaude
    @laperchaude 9 месяцев назад

    Merci pour le tutorial et l'outil!!

  • @DraikRoan
    @DraikRoan 10 месяцев назад

    MFW when you check OP's textual embeddings 😏 allfoursnaked corneosidedoggy sexyback bimbostyle busty_15

  • @MonkeyDIvan
    @MonkeyDIvan 10 месяцев назад

    Amazing stuff! Thank you!

  • @s3bl31
    @s3bl31 10 месяцев назад

    Dont work for me i dont know what the problem is in the preview i see a good face but in the last step i turns back to the bad face and the output is just a even worse oversharpend face

    • @marksanders3662
      @marksanders3662 5 месяцев назад

      I have the same problem. Have you solved it?

    • @s3bl31
      @s3bl31 4 месяца назад

      @@marksanders3662 Are you using a amd card? if so i think i fixed with --no-half in the command line. But idk since its that long ago and i switched to Nvidia.

  • @whyjordie
    @whyjordie 10 месяцев назад

    this works!! thank you! i was struggling so hard lol

  • @bhupendraking9370
    @bhupendraking9370 11 месяцев назад

    How to train SDXL model with large number of images? like example 30k images

  • @botlifegamer7026
    @botlifegamer7026 11 месяцев назад

    I know this might be outdated not but for some reason When I train using BF16 I can only use a batch size of 1 or I get out of memory. even at 512,512. for a test. How are you able to do 8 I have a 3090 by the way. followed your settings im using BF16 and 23 images.

  • @lemur1870
    @lemur1870 11 месяцев назад

    I have trained the lora, and I have used it for stable diffusion, but I keep on getting black images, doesnt mater what i type in

  • @adair546
    @adair546 11 месяцев назад

    Hi! @Bernard Maltais, do you have a relation whit HandyPrint?

  • @rondovk
    @rondovk 11 месяцев назад

    grateful for video but damn that audio makes me wanna die

  • @LouisCubingChannel
    @LouisCubingChannel 11 месяцев назад

    kohya_ss 和 webui 可以共用虚拟环境吗

  • @christopherquenneville
    @christopherquenneville Год назад

    Quebec Power

  • @Soulsphere001
    @Soulsphere001 Год назад

    One problem that I'm running into is that it takes over three hours to train using 7500 steps, but looking at my CPU and GPU performance shows that I'm only using around 10% of my CPU and usually only around 3% or lower on my GPU.

  • @didierlebouc2788
    @didierlebouc2788 Год назад

    Excellent presentation

  • @hotmarcnet
    @hotmarcnet Год назад

    🎯 Key Takeaways for quick navigation: 00:02 🌐 El video explica cómo recopilar imágenes de internet y prepararlas para crear un modelo LoRA utilizando Kohya SS GUI. 01:13 🖼️ Es importante seleccionar imágenes de alta resolución para el modelo LoRA, no importa la proporción. 03:19 📂 Se organiza una carpeta llamada "Jennifer Aniston" para recopilar las imágenes y otra subcarpeta llamada "image" para almacenarlas. 09:27 📸 Utiliza Gigapixel para mejorar la resolución de las imágenes y luego las guarda en formato JPEG. 24:10 🖼️ Emplea el Auto1-1-1 Web UI para mejorar la calidad de algunas imágenes, ajustando parámetros como el denoising strength. 28:03 📈 Refina las imágenes con Auto1-1-1 Web UI, manteniendo la apariencia original de los rostros para no comprometer el modelo final. 29:42 🖼️ La preparación de un conjunto de datos para crear un modelo LoRA comienza con la limpieza y mejora de la resolución de las imágenes. 32:18 📷 Después de la preparación de las imágenes, es importante agregar descripciones a cada una de ellas para enseñar al modelo sobre el tema relevante y no importantes detalles. 34:33 📝 Las descripciones se pueden validar abriendo cada archivo de texto junto con la imagen correspondiente para asegurarse de que sean precisas. 44:10 🚀 Una vez validadas las descripciones, se está listo para configurar el modelo LoRA y proceder con la fase dos del entrenamiento. Made with HARPA AI

  • @armandoaddel
    @armandoaddel Год назад

    Having issues with the Group images script, It only does 10 images at the time and only once. Windows 11 machine. Anyone else having that issue.

  • @Sabrinahuskydog
    @Sabrinahuskydog Год назад

    Your audio is fucking awful with an echo in it. We can't understand you. I just had to mute it and use CC instead.

    • @RobertJene
      @RobertJene 6 месяцев назад

      ANNNNNDDDD he's eating

  • @ThaRaiZe
    @ThaRaiZe Год назад

    These settings dont work. I keep getting CUDA out of memory issues even when using a 24 GB card. LoRa's train no problem. There's something off in these settings

  • @I77AGIC
    @I77AGIC Год назад

    45 minutes??

  • @metasamsara
    @metasamsara Год назад

    thank you, really clear and concise tutorial

  • @GooseAlarm
    @GooseAlarm Год назад

    Does anyone know why I continue to get - CalledProcessError: Command '['C:\\Users\\Name\\Desktop\\Vlad Diffusion\\kohya_ss\\venv\\Scripts\\python.exe', returned non-zero exit status 2. I can't find anything that fixes this issue and after watching so many videos on Kohya it feels like I'm the only one with this issue.

  • @NetTubeUser
    @NetTubeUser Год назад

    10:00 -- Question: Have you used the Internet since 1991 or just recently, like two years ago? Your image is smaller because, for some strange reason and logic, you think that dragging an image from Google Search into a folder will restore it to its original size which is completely wrong! You have taken the image from the temporary cache of your Internet browser! To get the right size of the image that you want to save, you need to access that image FROM the website, directly. But because you don't want to visit the whole site, you only need to use the direct link to the image. To do that: On the Google Search page, click on the image to open and see the Google preview of that image FROM the website, then right-click and select "Open image in new tab". And NOW and only NOW you have the right size image and can drag it into your folder (or right-click and save) ... not the smaller version of the image from the temporary cache!

  • @owenpaulmeyer
    @owenpaulmeyer Год назад

    🎯 Key Takeaways for quick navigation: The video is about creating a LoRA weight using kohya_ss GUI for model training. The process involves collecting images from the internet and enhancing them for model use. The speaker demonstrates searching for high-resolution images of Jennifer Aniston from the 1990s. The images are collected in a specific folder structure for processing. The speaker uses Gigapixel AI to upscale and clarify the images. Some images are processed using the Koya SS GUI tool to improve their quality further. The speaker discusses considerations for aspect ratios and cropping images. The images are being prepared for use in a model training process. Some images are discussed as potentially usable while others may require more improvement. The video highlights the workflow of collecting, enhancing, and preparing images for training. 32:18 The next stage involves adding text captions to the images to train the LoRA model. 32:49 Captions are added by pasting the image directory and using utility tools for text captioning. 33:20 The process involves configuring parameters like batch size, beams, and minimum caption length. 34:18 Captions start appearing next to each image file as the process proceeds. 35:01 Descriptions are reviewed and edited for accuracy and relevance. 35:26 Descriptions focus on clothing and essential details while avoiding unnecessary specifics. 40:10 Providing clothing details helps the model learn without enforcing specific attributes. 44:31 The process of captioning and validating images is complete, preparing for the next phase of training the LoRA model. Made with HARPA AI

  • @ysy69
    @ysy69 Год назад

    Appreciate your tutorials. From all of your use and testing of SD, would you arrive at your best preference being Lora training and not Dreambooth ?

  • @letitiadominica2588
    @letitiadominica2588 Год назад

    learned from the video, but tried 2 on the same settings and it resulted in a melted faces like the one you are showing in this video @ 80%. Is hires fix a requirement to make it work?