100x Faster Than NumPy... (GPU Acceleration)

Поделиться
HTML-код
  • Опубликовано: 16 ноя 2024

Комментарии • 105

  • @MrPSolver
    @MrPSolver  Год назад +67

    UPDATE: Thanks to @swni on Reddit for the suggestion to use the `ids_pairs` array to index to get `x_pairs` and `y_pairs` as opposed to reusing the `torch.combinations` function. This reduces the simulation time required for 10000 particles to only 20 seconds (about half what is shown in the video). Code has been updated on GitHub!
    To compare NumPy and PyTorch fairly under these new conditions, I simulate 5000 particles in each case. PyTorch takes 6.3 seconds to run (remember, it also has around a 2 second overhead), while NumPy takes about 823 seconds, indicative of about a 100x increase.

    • @ibonitog
      @ibonitog Год назад +2

      Could you test CuPy please?

    • @MrHaggyy
      @MrHaggyy Год назад

      There must still be a lot of potential. A GPU calculates 1080x1920 ~2Mio RGB value per frame. You don`t need to check n^2 combinations for collision, n! should be enough because if P1 collides with P2, P2 also collides with P1. Especially something like checking for collision can be blazingly fast on a GPU. Your 3070 has over 5000 cores and each one has SIMD instructions. So you can do about 20k fp ops per clk. I would check the particles for collision when creating the pairs. You have the function anyway so it`s an easy-to-fix bug.

    • @avishah9406
      @avishah9406 4 месяца назад

      The same thing can be applied for NumPy as well. Replacing the get_delta_pairs function with
      def get_delta_pairs(x, ids_pairs): # added extra parameter
      return np.diff(np.array(x[ids_pairs[:]]), axis=1).ravel()
      The itertools.combinations function takes a long time when the number of particles increase, so using the ids_pairs which was already created, can reduce time taken, as combinations is called twice is each iteration. Using 400 particles now takes 3 seconds instead of 57 seconds (NumPy).

  • @fabiopimentel6981
    @fabiopimentel6981 Год назад +51

    Bro please never stop doing physics videos, they are amazing! I know they are not the most popular videos in your channel but they are super helpful for someone that only had one programming subject and was with Fortran :( . Greetings from the Dominican Republic! haha

  • @ConstantlyDamaged
    @ConstantlyDamaged Год назад +59

    It always catches me off guard to see non-meme videos from you. I am more into web-interacting services rather than data manipulation/science-so async is my wheelhouse rather than this stuff. Still fascinating to watch.

  • @EVL624
    @EVL624 Год назад +24

    I am a numerical physicist, and this will be very helpful for me. I am currently running all my simulations om CPU (though using MPI for parallellization)

    • @geoffreyanderson4719
      @geoffreyanderson4719 Год назад

      Please learn about nvblas and openblas and code vectorization in my other comment here today. The two keys are writing vectorized numpy or pandas code, plus activating the nvblas or openblas subsystem. Let me know if want help.

    • @jawad9757
      @jawad9757 Год назад

      You may want to take a look at CUDA C++ if you have Nvidia GPU(s) and are concerned with performance

  • @zb9458
    @zb9458 Год назад +8

    Awesome vid! Love seeing Pytorch being leveraged for its first class GPU support for things other than machine learning. If I recall correctly, someone had a blog post about using pytorch to optimize a shape for rolling (i.e. reinventing the wheel) and it used pytorch, super funny, but cool. Great video!!

  • @Impo666
    @Impo666 Год назад +7

    I absolutely love this. I'm making my own game engine (fun hobby, tbh) with OpenGL, numpy and Python, and for some time I've thought about where to simulate my physics. This is an eye-opener, and it looks fun as heck! Espec. the matplotlib animation for some "lazy" collision simulations. This vid brings me straight to my college days

  • @antonioarroyopolonio2520
    @antonioarroyopolonio2520 Год назад +3

    Amazing content.
    I had a professor when I was in the physics degree that told us about the power of GPU when coding "big numbers". The GPUs have up to 1000 more ("dumb") cores than the CPU and that can be really powerfull. I am now working on my PhD and I use python to do the work. I think that I can learn a lot from you.
    Thank you!

  • @samlaki4051
    @samlaki4051 Год назад +6

    first non-humour vid i see and its awesome! will try to learn more! thank you professor!

  • @tiberioborgesvale3213
    @tiberioborgesvale3213 Год назад +2

    Congrats! Great video! Please, don't stop, your videos are incredibly didactic! I allways cite your channel to my students in my Classical Dynamics classes.

  • @shamaldesilva9533
    @shamaldesilva9533 Год назад +7

    Dude a GPU accelarated python series would be amazing 😍😍😍

  • @gabrielzinatorosa3024
    @gabrielzinatorosa3024 Год назад +3

    Very interesting content and I really appreciate the way you show both notebooks side by side to compare the results. Thank you very much!

  • @Ilya-iu5ih
    @Ilya-iu5ih Год назад +1

    Great video, thanks! Consider using indexing by the coordinates of particles in space. The idea is that the coordinates of the particles are rounded to the size of the box, and the collision check occurs only for those particles that are inside the same box. This usually reduces the number of pairs by 90%.

  • @Nooreazy
    @Nooreazy Год назад

    I'm honored that your stuff comes up on my feed. Amazing work!

  • @tunafllsh
    @tunafllsh Год назад +2

    GPU takes advantage of linear operations. So I'm not really sure, but if you use some data structures like quadtree the complexity of the computation might drastically simplify. And you won't need to calculate all n² distances. In fact most particles are not collading with each other. One need to test it, but with that CPU might still outperform the overhead of the GPU, since there won't be that many computations.

  • @geoffreyanderson4719
    @geoffreyanderson4719 Год назад +1

    Nvblas can be used by numpy. Nv stands for nvidia. Just configure your host a bit, which is easy. Openblas can also be used by numpy, which is more common. By default, your linux is using a gnu blas which is super slow by comparison. Nvblas uses the gpu for the linear algebra operation s in your numpy code. Just be sure to write vectorized numpy code , not for loops. You don't change your application code at all, which is a big benefit for ease of maintenance. Openblas will recruit all your cpu cores and implicitly parallelize your matrix matg, greatly speeding it up, as will nvblas.

  • @bwallace1598
    @bwallace1598 Год назад

    Great educational video, mate! I'm a CS Grad student and was beginning to get to the later ML courses. Your explanation and side-by-side logic demonstration with Numpy convinced me to do a bit of research and switch from TF to Pytorch! Thanks so much!! I eagerly look forward to the next video!

  • @monkeyofwar6818
    @monkeyofwar6818 Год назад

    It's so relaxing so see someone else's explanation. I'm so tired of doing work in graduate school XD

  • @Louis-ml1zr
    @Louis-ml1zr Год назад

    Nice I've been waiting for this one ! thanks , looking forward to seeing the next ones

  • @tiddlywinks497
    @tiddlywinks497 Год назад

    Oh damn, this is what my thesis is on! Good to see that some great resources are being put out for it

  • @alexdefoc6919
    @alexdefoc6919 Год назад

    I hope you continued this series

  • @omarllama
    @omarllama Год назад +2

    12:00 why don't you simply use torch.cdist (if you have a batch of vectors, otherwise use torch.pdist) which calculates the p-norm (p=2 in your case) distance between each pair of the two collections of row vectors. This is supposed to be much faster than your code, even though I didn't test it.

  • @Ilya-iu5ih
    @Ilya-iu5ih Год назад

    torch by default includes in each tensor the telemetry necessary to calculate the derivative for the error back propagation algorithm. Use the require_grad=False parameter, this will speed up the calculation even more.
    x = torch.randn(3, requires_grad=False)

  • @tuomassavolainen218
    @tuomassavolainen218 Год назад

    Nice video, did not think about using pytorch to replace
    Numpy, but it makes perfect sense for parellelizing numpy code👍. Just a quick tip for additional speedup. Instead of comparing the distance directly you can compare the squared distance for collision detection, this avoids using the square root function which is "slow" at least compared to all the dot products, though it might not matter much for simulations of this scale.

  • @govindtuli2913
    @govindtuli2913 Год назад

    Multiprocessing library also helps to utilize all available threads. I was generating a mandlebulb and it went from 4 minutes to 1 minute when I optimized code for using it.

  • @seelooooo
    @seelooooo Год назад +3

    Nice video!
    I’ve also got a question in the part of calculating whether particles collide with each other. Is there any advantage of the video’s method compared to use:
    DIS = torch.cdist(points,points) < collision distance
    DIS = torch.triu(DIS, diagonal = 1)
    Pairs = DIS.nonzero()
    Or, they are having the same computational complexity?

    • @MrPSolver
      @MrPSolver  Год назад +3

      Never seen "torch.cdist" before! Thank you for this comment. Huge reason why I post videos like this...to learn more from the comments :)

  • @Thierry080
    @Thierry080 Год назад

    Super cool, your meme videos are hilarious but this quality content is why I subbed in the first place

  • @MyMinecrafter97
    @MyMinecrafter97 Год назад

    Very interesting, would love to see more of this!

  • @frankkoslowski6917
    @frankkoslowski6917 3 месяца назад

    You said you use a GTX 3070?
    All I can find here is a RFX 3070.
    Was trying to figure out what makes your calculations for 10,000 particles that much faster,
    compared to my GTX 750Ti that of course would crash the system with 10,000 particles,
    until an additional Tesla M4 was installed, completely bypassing the functions of the GTX 750Ti,
    resulting in
    rs,vs = motion(r, v, ids_pairs, ts=1000, dt=0.000008, d_cutoff=2*radius)
    calculations: Wall time: 6min 55s,
    which is still much slower than 48.9 sec, as your demonstration boasts.
    The animation with 10,000 particles looks awesome. Thx.
    Have not yet found the necessary control software to raise the clockspeed of the Tesla M4.
    Current driver installations supplied by Nividia have cleverly merged the two cards as one.
    Ergo.: We have now an augmented GTX 750 Ti that benefits from the original factory settings of the Tesla M4.
    Additional software needed to raise the clockspeed of the cards in their combined setup to the level of their full capacities was not found yet.

  • @s.v.8662
    @s.v.8662 Год назад +3

    Could another distribution really come up for different potentials around the particles? I thought of the Boltzmann distribution as a thermodynamic necessity due to maximization of entropy

  • @daze8410
    @daze8410 Год назад

    Just recently got familiar with multithreading so I guess this is the natural progression

  • @williamcase426
    @williamcase426 6 часов назад

    How do I resolve the error:
    [WinError 2] The system cannot find the file specified

  • @vialomur__vialomur5682
    @vialomur__vialomur5682 Год назад

    Wow this is really interesting! Thanks! Waiting for more videos

  • @cw9249
    @cw9249 Год назад

    perfect intro to torch for someone who is familiar with numpy

  • @LS-xb2fh
    @LS-xb2fh Год назад +1

    If this is all about optimization, you should probably compare the sqared distance between particles to eliminate the need to calculate the square root. :)

  • @kedaracholkar6055
    @kedaracholkar6055 Год назад +26

    GTX 3070? Do you mean rtx?

  • @Andres186000
    @Andres186000 Год назад

    Thank you so much , I was already using PyTorch for something, but I couldn't figure out how to create the equivalent of the "x_pairs" array I needed to use, thanks.

  • @dargi_amorim
    @dargi_amorim Год назад +1

    Thanks for the video ! You are awesome !
    I have a question. Is it possible to use pytorch to optimize a code with a lot of functions from scipy ? Like solving a lot of differential equations, nonlinear equations, interpolating and integrating functions all in one big code. I'm currently optimizing my code with the joblib library to run it in parallel.

  • @shinrinyuku4972
    @shinrinyuku4972 Год назад

    Very good intro video into GPU programming, it even gave me couple of ideas. One question if I may. Why wouldn't you do the simulation with event driven algorithm, since that would save a lot of resources and you can avoid overlaps of particles (ie the need to choose small timesteps). I get this is a tutorial/introduction video, but that implementation would be very interesting as well!

  • @m_sedziwoj
    @m_sedziwoj Год назад

    Maybe I add that you can use AMD GPUs but currently only in Linux (as Nvidia have CUDA, AMD have ROCm)

  • @calebwhales
    @calebwhales Год назад

    Great to see ya again mate

  • @frankkoslowski6917
    @frankkoslowski6917 3 месяца назад

    It would be interesting to know what sort of a GPU you are using to achieve the record-breaking walltime of 48.9sec.
    Compared to my 33:12.4 min for a max of 8000 particle,
    on a NVIDIA GeForce GTX 750 Ti sporting 640 CUDA-cores @ 100% of the available 2GB mermory in use,
    while offering badly interrupted video performance on all 3 monitors.
    However, after rewriting your code as a CUPY hybrid, the wall time was reduced to 11:31 min for the 8000 particles,
    while keeping GPU memory use between 33-75%, and thus well away from the videocard crash
    encountered when attempting 10000 particles using your code unmodified.
    Anyway, arangements were made today,
    to obtain a NIVIDIA Tesla M4 graphics card to be used in conjunction with the existing card as a dedicated number cruncher.
    Hopefully that will get us closer to the desired performance.

  • @ibonitog
    @ibonitog Год назад +11

    What about cupy (CUDA drop-in replacement for numpy)? Is the performance uplift comparable to pytorch?

  • @arduous222
    @arduous222 Год назад

    I frequently do so-called "model-fitting" using MCMC (or anything good enough), where each set of data consists of 1k-10k data. I wonder whether this could benefit from GPU acceleration or the overhead would be too much.

  • @btr_Z
    @btr_Z Год назад +2

    Thanks, now I finally will have one reason to tell my dad to buy me a graphics card😂

  • @antonioparesini912
    @antonioparesini912 Год назад +1

    Where billy

  • @frankkoslowski6917
    @frankkoslowski6917 Год назад

    Nice sales pitch for Microsoft Visual Studio.
    Had `cuda` up and running nicely in a previous installation of Visual Studio.
    All of that had been with C++ in mind. So Python was not really considered at that time.
    Was hoping to do the same with PyCharm and Browser-based Notebook using Python exclusively.
    That's when it got confusing to the point of dropping the idea. 😒

  • @tclin77
    @tclin77 Год назад

    Definitely a must watch!

  • @gdmitrich
    @gdmitrich Год назад

    thats so cool
    need this more in field of quantum chemistry❤❤

  • @galenseilis5971
    @galenseilis5971 Год назад

    Why was it 'bad' that some of particles were colliding in the initial conditions?

  • @nikolasfrancorios312
    @nikolasfrancorios312 Год назад

    "GPU, wich most people have acess today"
    Looks like we've got some serious worldknowing issue going on here

    • @MrPSolver
      @MrPSolver  Год назад

      Lots of free nodes you can use here that have access to GPU resources:
      colab.research.google.com/

  • @baldpolnareff7224
    @baldpolnareff7224 Год назад

    So my intuition of rewriting stuff in pytorch just for fun was not unreasonable after all!

  • @habtiephys
    @habtiephys Год назад

    I think operations on Pythorch Tensor are also faster than on Numpy arrays both on cpu

  • @maelstrom254
    @maelstrom254 Год назад

    Torch jit and torch compile is a lot faster than just torch

  • @LydellAaron
    @LydellAaron Год назад

    Sweet topic. Thank you!

  • @PV10008
    @PV10008 Год назад

    This is brilliant!

  • @c0d3w4rri0r
    @c0d3w4rri0r Год назад

    I mean you start by saying most people have access to a GPU these days and this is absolutely true. But plenty don't have an NVIDIA GPU and as I understand it pytorch doesn't support non NVIDIA gpus? might be worth re writing this with pyopencl.

  • @nieczerwony
    @nieczerwony Год назад

    It's all great but one thing. I don't need your face (person) sitting in fron do 2 screens and covering them 😂

  • @void2258
    @void2258 Год назад +1

    What program are you using here that let's you put notes in the code like this?

    • @MrPSolver
      @MrPSolver  Год назад

      VSCode and Jupyer Notebook!

  • @harikrishnanb7273
    @harikrishnanb7273 Год назад

    2 minutes into the video, what about the performance of pytorch compared to numpy in CPU? is it faster there also !? have you tried numba !?

  • @mathiaz943
    @mathiaz943 Год назад

    Does GPU mean NVDIA GPU specifically? Will we ever have libraries utilizing ANY general GPU?

    • @bryce07
      @bryce07 Год назад

      it works with any GPU

  • @TawsifTurjoeee
    @TawsifTurjoeee Год назад

    can you suggest me books for this relevant problems of laplace transform via python

  • @jorgesaxon3781
    @jorgesaxon3781 Год назад

    PLEASE MAKE A TUTORIAL ON HOW TO HANDLE BIG INTEGERS (>64INT) ON THE GPU 🙏

  • @TZ-nd1cm
    @TZ-nd1cm Год назад

    What are libraries that must be imported?

  • @pauldirac6243
    @pauldirac6243 Год назад

    When I try to run your code, I get the error message: No module named 'torch' What am I doing wrong?

  • @mikejohnston9113
    @mikejohnston9113 Год назад

    I can't get accepted into your discord. In the two lines ani.save()... I get a file not found exception. I am using Python 3.11.3. I really like the article and video. Thanks

    • @mikejohnston9113
      @mikejohnston9113 Год назад

      I found the problem, I hand not installed python-ffmpeg. It is fixed now. Thanks

  • @obliquesealray2188
    @obliquesealray2188 Год назад

    thank you for this!

  • @filippavlove5574
    @filippavlove5574 Год назад

    love it, keep it up :)

  • @sergiolucas38
    @sergiolucas38 Год назад

    Great video :)

  • @holthuizenoemoet591
    @holthuizenoemoet591 Год назад +1

    Would this work on a RX 6800 or intel Ark 770?

    • @baldpolnareff7224
      @baldpolnareff7224 Год назад +1

      If I remember correctly pytorch runs on AMD and intel arc as well

  • @sohamtilekar5126
    @sohamtilekar5126 8 месяцев назад

    Can You Make video on the PyOpenCl

  • @infiniteflow7092
    @infiniteflow7092 Год назад

    Does Pytorch have numerical integration capabilities?

  • @bean_mhm
    @bean_mhm Год назад

    this is really cool!

  • @fnegnilr10
    @fnegnilr10 Год назад +9

    What!!!!!!!!!!!!!!!!!!!!!!!!!

  • @apoorvvyas52
    @apoorvvyas52 Год назад

    Great video

  • @TheBuilder
    @TheBuilder Год назад

    0:03 "If you coded in python before" while showing a screen full of braces

  • @alexdefoc6919
    @alexdefoc6919 Год назад

    Oh so good

  • @ssmith5048
    @ssmith5048 Год назад

    soo instead of using the poor man´s version of Fortran for such calculations, just use Fortran. It is not only perfect for arrays but also natively parallel. You can even make a python wrapper if you want some gui to please the eye. But , I get it, it would not be cool for the kids on youtube...but if you really need efficiency , give it a try.

  • @Zucih1
    @Zucih1 Год назад +3

    Hey Nvidia did i miss the new gtx 3070 !!

  • @arunsebastian4035
    @arunsebastian4035 Год назад

    Very nice video

  • @tsunningwah3471
    @tsunningwah3471 Год назад

    wowowowoowwww

  • @haiminh256
    @haiminh256 Год назад

    i'll break your comment bar with C++

  • @Singlton
    @Singlton Год назад +1

    Gtx 3070?
    Is this from china!? 🤣😂

  • @hoaanduong3869
    @hoaanduong3869 Год назад

    Damn, he used a Deep Learning Framework to replace Numpy, a Mathematics Framework :v

  • @omarllama
    @omarllama Год назад

    The entire PART 1 can be more efficiently rewritten in one line:
    d_pairs = torch.pdist( r.T )

  • @haiminh256
    @haiminh256 Год назад

    #Include
    int main()
    {
    std::cout

  • @inxendere
    @inxendere Год назад

    the RTX 3070 is already considered mid range??!! 🥲

  • @perlindholm4129
    @perlindholm4129 Год назад

    Use transfer function its billion times faster:
    h = np.random.rand(10000)
    idx = np.arange(10000)
    X = X_train[idx].astype("float32")/255.0
    yt = y_train[idx] + 1 # 1..10
    x = X.mean(1)
    ids = np.argsort(x)
    i=0
    while True:
    err = yt[ids] - x[ids] * h
    h += 0.1*err
    print(np.mean(err**2))
    i+=1
    if np.mean(err**2)