Training a GAN from your Own Images: StyleGAN2 ADA

Поделиться
HTML-код
  • Опубликовано: 29 дек 2024

Комментарии •

  • @KenJee_ds
    @KenJee_ds 3 года назад +6

    Very excited to experiment with this!

  • @eeeeric1966
    @eeeeric1966 3 года назад +9

    This work is incredible! I'd love to have a GAN take something like a simple art style or type of doodle... and then create more of them... I have no idea where to start though... Your 1970s comic book GAN results are amazing!!

    • @HeatonResearch
      @HeatonResearch  3 года назад

      Thank you, yes the 70s GAN was fun! Glad it is helpful.

  • @SongStudios
    @SongStudios 3 года назад +3

    Very exciting! Especially the minecraft Gan! Now I've also been experimenting with the same thing that your minecraft GAN does. Take a lot of minecraft images, squash it down to 1024 images (I'm going to train on 1024 by 1024). Train it for a while. Now I've actually gotten way more different results, I've also rotated in the dimensional Z coordinate (up & down) to add a bit more randomness. I'm also training on about 6000 Images.

    • @JohnSmith-ox3gy
      @JohnSmith-ox3gy 3 года назад

      How did it go? I remember the ballpark estimate beink 10k pictures.

    • @SongStudios
      @SongStudios 3 года назад

      @@JohnSmith-ox3gy Very good.

  • @TheAcidPrank
    @TheAcidPrank 3 года назад +5

    If I use a pretrained model, does it require the same amount of GPU Ram for inference ?

  • @makeandbreak127
    @makeandbreak127 3 года назад +6

    I love your videos, thank you for making these. I have a question - do you have a video on how to label images to create a conditional model (cond = True) with the training

    • @HeatonResearch
      @HeatonResearch  3 года назад +1

      Thanks! I have not tried that feature of StyleGAN yet.

    • @hdubbs9174
      @hdubbs9174 3 года назад +1

      @@HeatonResearch I would also be interested in this, especially multi-conditional.

  • @iamisobe
    @iamisobe 2 года назад +1

    Is there anyway to get some 1 on 1 help setting one of these up? Looking to train my own using m y images to create visuals for my DJ sets. Will pay for lessons

  • @noiamhippyman
    @noiamhippyman 2 года назад +1

    This is perfect! I finally know what I'll use all my GPUs for when I finally stop mining. I feel like I could have a lot of fun with this. Thank you for sharing how to actually use the source.

  • @Andy-rq6rq
    @Andy-rq6rq 3 года назад +2

    Definitely be interested in a WSL2 video

  • @Utsab_Giri
    @Utsab_Giri Год назад

    Hey Jeff, thank you for making this. I can't seem to be able to find your video on what to do with the .pkl files.

  • @smokinep
    @smokinep 2 года назад

    Awesome video ... Quick question though. If I train the model with my own data using the Pytorch implementation can I sell my product/services commercially ?

  • @jaydy71
    @jaydy71 Год назад

    Is this really the same Jeff Heaton that gave us Encog and those videos about it with creased bed sheets in the background and everything? Don't get me wrong, those early videos taught me a lot but things have changed :)

  • @fullbinky
    @fullbinky 2 года назад +1

    Hi Jeff thanks for the very detailed video. How would I go about running your image in WSL?

    • @HeatonResearch
      @HeatonResearch  2 года назад

      It should be runnable like any other docker image. I have not tried by StyleGAN image in awhile, it is StyleGAN1, so there could be some software rot. I will consider it for a future video.

    • @fullbinky
      @fullbinky 2 года назад

      @@HeatonResearch thanks Jeff, will give it a shot when I get a chance

  • @retaillab5641
    @retaillab5641 3 года назад +3

    Amazing Video! Can we extract specific features for these custom data and change the output generated based on tweaking the features(Eg: Fish shape/Color etc)?

  • @beanoevil777
    @beanoevil777 3 года назад

    You have WSL2 working with an Nvidia GPU? Didn't think that was possible except through an insider build...

  • @ahpacific
    @ahpacific 2 года назад

    Hi @Jeff Heaton - thank you for the awesome tutorial.
    When I run the train command - i get the error: "nvcc fatal : Value 'sm_86' is not defined for option 'gpu-architecture'" - any advice on how to address this?

  • @xristoskatsaros6409
    @xristoskatsaros6409 3 года назад

    this might be a dumb question but is there a way to use the 70s scifi model with pytorch? i'm having trouble getting tensorflow to play nice with my system, but had no problem getting a pretrained pytorch BIGgan model to generate pictures and videos. thanks in advance!

  • @uatiger1
    @uatiger1 3 года назад +1

    Great video as always. Thanks. I wonna request you to also make a video of AttnGAN. Its a text-2-image GAN. I want to learn more about it and understand how to retrain it for something other than birds or flowers.

  • @malice112
    @malice112 2 года назад

    What cloud service do you recommend for running a GAN if I don't want to invest in expensive GPU's?

  • @douglasturbinski6926
    @douglasturbinski6926 2 года назад

    How did you compile 30k comic images and then feed them into the gan?

  • @minhaoling3056
    @minhaoling3056 2 года назад

    Hi , what would you recommend if I want to generate image from binary images using only a laptop..

  • @pupupipi1259
    @pupupipi1259 3 года назад

    hello im a total noob on this subject, i feed a stylegan frames of a shortfilm i made. But the results after weeks of training on colabs with lots of accounts was just similar frames but all deformed. Could there be a way of recreating the short but in the style of the gan? Like ordering the frames and the movement to sync with the real short but all with that interpolation beetween frames? or maybe i should try another method

  • @shrimpychipofficial
    @shrimpychipofficial 3 года назад

    This worked well for me at first, but on my second try with a different dataset I'm getting an error saying "UserWarning: semaphore_tracker: There appear to be 34 leaked semaphores to clean up at shutdown
    len(cache))" on the initial training and then it just stops :/ my dataset is only about 350 images and I compressed them more and also tried on a different computer so I'm not sure how to fix this

  • @hoz85
    @hoz85 3 года назад

    Thanks a lot for the video. Can you explain 3000 kimgs mean 3 million images there are in your dataset?

  • @karolguzikowski4812
    @karolguzikowski4812 2 года назад +2

    Hi, great tutorial, thank you a lot! Could you make a tutorial about conditional StyleGAN (multiclass)?

  • @JackJulian-t3o
    @JackJulian-t3o 8 месяцев назад

    Hi Jeff, I have emailed you.. Trying to get the CoLab working again. Getting a few errors.. Any help? :)

  • @dougpaterson4490
    @dougpaterson4490 3 года назад +1

    Would love to see your comments on WSL2

    • @HeatonResearch
      @HeatonResearch  3 года назад

      I've really had no end of problems with WSL2, it is a bit of a disappointment to me. I think the fact that it has been in "beta" for years now shows the "seriousness" that Microsoft puts into GPU use under Docker in Windows. Linux has had this figured out for years.

  • @Kmysiak1
    @Kmysiak1 3 года назад +1

    Have you produced a video like this using colab pro?

    • @HeatonResearch
      @HeatonResearch  3 года назад

      Not yet, but working on that now, actually.

    • @Kmysiak1
      @Kmysiak1 3 года назад

      @@HeatonResearch awesome! Been learning a ton from your channel.

  • @tamararosenblum4693
    @tamararosenblum4693 3 года назад +1

    Hi there, Is it possible to train and create GANs using only my Mac laptop? I am not running linux. thanks!

    • @HeatonResearch
      @HeatonResearch  3 года назад

      GANs in general, yes, but not sure what software to use. StyleGAN2, no, because it requires a CUDA GPU.

    • @tamararosenblum4693
      @tamararosenblum4693 3 года назад

      @@HeatonResearch Thanks for your reply! Hmm this is what I was afraid of. I will look into who is working on Macs. Thanks again

  • @fienter9753
    @fienter9753 3 года назад

    hi Jeff ,thank you for the tutorial !
    may I ask how do I correctly resume my training?
    sometimes my training just crash,and I use --resume parameters path to my latest snap pkl file
    and the training will just restarts from 0 kimg
    and my training results became really weird(weird colors,weird shape)

  • @NIRo-gamedev
    @NIRo-gamedev 3 года назад +1

    4:12 Lex? :)

  • @maloukemallouke9735
    @maloukemallouke9735 3 года назад +1

    Great Job ,
    i am still waiting find digits in Random image Not just writed hand ones .

  • @lukeowen7636
    @lukeowen7636 2 года назад

    Hey, great tutorial! In theory, would style gan work with images that are vector style graphics and not photos?

    • @stickwithit
      @stickwithit 2 года назад

      I think if you can render the vectors as a flattened PNG with no transparency, you should be able to fit the parameters required to run the GAN

  • @fredflash7108
    @fredflash7108 3 года назад

    jeff knows what the ppl want!

    • @HeatonResearch
      @HeatonResearch  3 года назад

      Thanks! I try my best. And am always glad to hear what the ppl want too!

  • @sohannannra604
    @sohannannra604 3 года назад +1

    How could I change the last command to allocate more memory to my GPU?

    • @sohannannra604
      @sohannannra604 3 года назад +1

      I have 8 GB of dedicated GPU memory but it is saying that it is finding an error in allocating memory when I clearly have enough. Could you help me and provide a solution?

  • @RajuYadav-qm4ic
    @RajuYadav-qm4ic Год назад

    😍😍

  • @Irfanansariqh
    @Irfanansariqh Месяц назад

    Best one sir❤

  • @ComicsInk
    @ComicsInk 3 года назад

    So do the image dimension have to be powers of 2? i have a dataset of images that are all 100x100, but i'm worried it won't work.

    • @HeatonResearch
      @HeatonResearch  3 года назад

      Yes, powers of two... generally 256x256, 512x512, 1024x1024, 2048x2048... higher seems to crash the current stylegan.

    • @JohnSmith-ox3gy
      @JohnSmith-ox3gy 3 года назад

      @@HeatonResearch Could that be a problem with RAM?

  • @Ephemeralkazu
    @Ephemeralkazu 3 года назад

    How can I resume when I stopped the training

  • @KeeperOfKale222
    @KeeperOfKale222 3 года назад +1

    Maybe this is out of my league here, but how did NVIDIA make such an amazing user interface where one could drag a box over the selected area of images and it show live the new GAN results? And it was sort of in an animated flux between images:
    ruclips.net/video/9QuDh3W3lOY/видео.html

  • @ktoksu
    @ktoksu 3 года назад

    Could you explain what kimgs are and why they impact training time?

    • @HeatonResearch
      @HeatonResearch  3 года назад +2

      That is the number of kilo-images from the training set that are evaluated. The higher the number, generally, the better the result (lower FID). The size of the training set does not really impact the training time, just the quality, since the training is randomly sampling from the training data and also augmenting with random variations on the training data.

    • @ktoksu
      @ktoksu 3 года назад

      @@HeatonResearch Thanks!

  • @chinmay996
    @chinmay996 3 года назад

    Hi, learned a lot from your videos. Is there a way to do hyperparameter tuning for this model? I was not able to find much resource for this. Can you give few tips from what you have explored while training?
    Thanks

  • @HeatonResearch
    @HeatonResearch  3 года назад

    Want to do this on CoLab? ruclips.net/video/L3JLzoe-dJU/видео.html
    Want to do this on Windows? ruclips.net/video/BCde68k6KXg/видео.html
    Want to see this as an article? towardsdatascience.com/generating-your-own-images-with-nvidia-stylegan2-ada-for-pytorch-on-ampere-a80fab52d6b5

  • @cyrus5295
    @cyrus5295 3 года назад

    This cannot be done on Nvidia Jetson Nano I suppose?

    • @HeatonResearch
      @HeatonResearch  3 года назад +1

      I believe so, but not something I've tried.

  • @arianpetrovic4751
    @arianpetrovic4751 3 года назад +1

    Great Video!
    Can you go deeper in the Datasets Preparation?
    The PyTorch Version of StyleGAN allows to use a dataset.json for basic class labels.
    How I can generate this json file without typing line by line?
    What exactly are thees class labels? Its possible to bump the result quality with a labeled dataset?
    For example, I label faces with male/female or with glasses and without glasses.
    Thanks

    • @DigitalMarksman
      @DigitalMarksman 3 года назад

      Yeah, that would be very exciting to watch since I couldn't really find any real info about it over the internet (nor other people experimenting with that).
      I would be interested to have labeled data that might help with missing details on my target training data and see if that improves final model accuracy and prevents overfitting as well.

    • @arianpetrovic4751
      @arianpetrovic4751 3 года назад

      @@DigitalMarksman I have found a prelabled Flickr Dataset www.github.com/royorel/FFHQ-Aging-Dataset
      But I don't find the way to use the labels in StyleGAN2. I'm currently testing a basic pytorch script to figure out how to feed the data into StyleGAN2. If you have success feel free to share your experience, I will do it also.

  • @Monad.x
    @Monad.x Год назад

    Can you train a StyleGan on Mac M1?

    • @HeatonResearch
      @HeatonResearch  Год назад +1

      I have not been successful with that. It has CUDA specific code.

  • @harperjmusic
    @harperjmusic 3 года назад

    Anyone else having problems trying to set up nvidia-docker on wsl2?

  • @1Poxxxx
    @1Poxxxx 3 года назад

    why has noone built a GUI for Stylegan yet?

    • @HeatonResearch
      @HeatonResearch  3 года назад

      Well, it costs money, but I believe this system has a GUI on GANs. runwayml.com/. But agreed, someone should create a free GUI on top. hmmmmmm :-)

  • @geetanjalisharma1501
    @geetanjalisharma1501 3 года назад

    how to train stylegan on grayscale images ?
    Sir, actually I have implemented without any chnages in channel input but it does not work. but the problem is it will works for only few number of images i.e face140.jpg and then program terminate.
    Error:
    tensorflow.python.framework.errors_impl.InvalidArgumentError: 2 root error(s) found.
    (0) Invalid argument: input depth must be evenly divisible by filter depth: 1 vs 3
    [[{{node InceptionV3/_Run/InceptionV3/InceptionV3/import/conv/Conv2D}}]]
    [[InceptionV3/_Run/InceptionV3/Reshape/_5015]]
    (1) Invalid argument: input depth must be evenly divisible by filter depth: 1 vs 3
    [[{{node InceptionV3/_Run/InceptionV3/InceptionV3/import/conv/Conv2D}}]]
    0 successful operations.
    0 derived errors ignored.
    please help me sir.

    • @HeatonResearch
      @HeatonResearch  3 года назад

      Have not tried that. I THINK, so long as they are ALL grayscale it would work. Failing that I would convert them to color.

  • @josephstanski5180
    @josephstanski5180 2 года назад

    Only love - :o)

  • @geetanjalisharma1501
    @geetanjalisharma1501 3 года назад

    how to train stylegan on rectangular images ?

    • @HeatonResearch
      @HeatonResearch  3 года назад

      That is not trivial, it would require modification of StyleGAN. But it is something I am thinking of trying.

  • @TehViiRuSs
    @TehViiRuSs 3 года назад

    Can you do a video on U-GAT-IT please?

    • @HeatonResearch
      @HeatonResearch  3 года назад +1

      I will have to have a look at that, thanks!

  • @ptkwilliams
    @ptkwilliams 2 года назад

    The images also need to be at least 64x64 pixels

  • @moncano3399
    @moncano3399 3 года назад

    interested in a WSL2 video

  • @manamsetty2664
    @manamsetty2664 3 года назад

    Can we do it in a laptop

    • @HeatonResearch
      @HeatonResearch  3 года назад

      Yes, so long as the laptop has a CUDA 10/11 GPU. I have a whole series on using a ThinkPad P53 for this kind of thing. ruclips.net/video/EroJ-SzKXm4/видео.html

  • @niu1909
    @niu1909 3 года назад

    I'm interested in Windows WLS2!

    • @HeatonResearch
      @HeatonResearch  3 года назад

      So am I! There will be something coming from me soon on WSL2 and Win11.

  • @moahaimen
    @moahaimen 2 года назад

    IT DOESNT WORK, $ bash: nvidia-docker: command not found
    , DOES IT WORK ON WINDOWS PLATFORM

  • @felicytatomaszewska
    @felicytatomaszewska 3 года назад

    Did I hear Intel Xenon processors 0:59 🤭🤭🤫🤔

  • @generalshepherd457
    @generalshepherd457 2 года назад

    You could make some quick bucks if you unloaded some of that lenovo equipment. Then you just tell lenovo that somebody stole the computer. Let me know if you are interested.