Martin Andrews
Martin Andrews
  • Видео 10
  • Просмотров 20 424
"Proving that Cryptic Crossword Clue Answers are Correct" - Explainer Video
Short video to explain our workshop paper "Proving that Cryptic Crossword Clue Answers are Correct" - which we will presenting in-person at the ICML 2024 Workshop on LLMs and Cognition.
Link to workshop : llm-cognition.github.io/
Link to arXiv version of paper : arxiv.org/abs/2407.08824
Link to Wordplay dataset : github.com/mdda/cryptic-wordplay
Просмотров: 52

Видео

PaLM in Flowise (no-code langchain)
Просмотров 61610 месяцев назад
Now that Google has made its PaLM 2 models available, I decided to add PaLM 'nodes' into Flowise. In the video, I show the PaLM LLM as part of a simple 'tell me the weather' langchain component, and then explain how the PaLM embeddings can be hooked up into a whole Q&A with Retrieval set-up. As a bonus, I finish off with a bit of "PAL : Program-Aided Language" action! To get a PaLM API key you ...
ChatGPT : How it works and what's coming next for Large Language Models
Просмотров 1,1 тыс.Год назад
This talk was given live at the Machine Learning MeetUp in Singapore (www.meetup.com/Machine-Learning-Singapore/). The slides for the talk (with clickable links to external papers, code, etc) are at : redcatlabs.com/2022-11-29_MLSG_LLM-News/#/llm-news-talk Please leave comments below : I'll read them all! #ai #chatgpt #deeplearning #gpt3 #PaLM
Diffusion Models Explained : From DDPM to Stable Diffusion
Просмотров 7 тыс.Год назад
This talk was given live at the Machine Learning MeetUp in Singapore (www.meetup.com/Machine-Learning-Singapore/). The slides for the talk (with clickable links to external papers, code, etc) are at : redcatlabs.com/2022-09-15_MLSG_Diffusion/#/diffusion-talk Please leave comments below : I'll read them all! #ai #diffusion #stablediffusion #deeplearning
Installing MineRL using Containers - with linked code
Просмотров 1,5 тыс.Год назад
Please visit the *new* repo for more details, etc : github.com/mdda/DiamondJAX#diamondjax My Blog post : blog.mdda.net/ai/2022/09/13/running-minerl-within-a-container Other background : * The BASALT Competition : minerl.io/basalt/ * Competition submission template : github.com/minerllabs/basalt_2022_competition_submission_template/ * Behaviour cloning baseline Dockerfile : github.com/minerllabs...
OpenAI VPT for Android Demos - Colab included!
Просмотров 4472 года назад
Please visit the *new* repo for more details, etc : github.com/mdda/DiamondJAX#diamondjax * This episode's Colab Notebook : [DiamondJAX_03-VPT-for-AndroidDemos](colab.research.google.com/drive/1tUR0Y7fpxF3O_bLfiks6imSzNXJ_BRvK?usp=sharing) * Includes (_extending the previous Colab_): Calibration of VPT camera moves - Seen in previous text output of VPT camera changes: - `0.00, ±0.62, ±1.61, ±3....
OpenAI VPT for YouTube Videos - Colab included!
Просмотров 8252 года назад
OpenAI VPT (Video Pre-Training) * openai.com/blog/vpt/ * github.com/openai/Video-Pre-Training BASALT Challenge - with links to Discord discussion * www.aicrowd.com/challenges/neurips-2022-minerl-basalt-competition "MineRL v1.0.0 VPT" Colab by @nev: * colab.research.google.com/drive/1OYdc4FwmW1nYTHLfCpEHv-hn83euvRdh?usp=sharing DiamondJAX_00-SetUp * Colab set-up : see previous video * But we won...
First steps with MineRL - Colab included!
Просмотров 6 тыс.2 года назад
* OpenAI VPT (Video PreTraining) IDM (Inverse Dynamics Model) * openai.com/blog/vpt/ * Yannic Kilcher = ruclips.net/video/oz5yZc9ULAc/видео.html * Edan Meyer = ruclips.net/video/ODat7kfZ-5k/видео.html * MineDojo * Datasets and MineCLIP = minedojo.org/ * MineRL * minerl.readthedocs.io/en/v1.0.0/tutorials/index.html * Look at the Colab :: Free to try it our yourself! * colab.research.google.com/d...
Practical JAX : Using Hugging Face BERT on TPUs
Просмотров 8222 года назад
A look at the Hugging Face BERT code, written in JAX / FAX, being fine-tuned on Google's Colab using Google TPUs (Tensor Processing Units). This is "Practical JAX", which goes beyond the normal introductory JAX content, and looks at how it is being used 'in the wild'. Link to the notebook in the video : colab.research.google.com/drive/1oVJAIAIyPoCqaqo4LonOKBZhkfVPV-UM?usp=sharing Many thanks to...
DeepMind's RETRO vs Google's LaMDA
Просмотров 2,4 тыс.2 года назад
Explore the frontiers in the latest AI Transformer models : Going beyond OpenAI's GPT-3 with the latest innovations from DeepMind and Google. This talk was given live at the Machine Learning MeetUp in Singapore (www.meetup.com/Machine-Learning-Singapore/), and uploaded here to see whether this kind of content appeals to a wider audience. #deepmind #retro #lamda #ai

Комментарии

  • @Omsip123
    @Omsip123 10 часов назад

    Very interesting and well explained, thanks a lot for your efforts

  • @scskowron
    @scskowron 19 дней назад

    Thank you for the explanation. It does seem like you mixed up the diffusion processes though. At 2:50 you refer to the process of adding noise as the backwards process, and removing noise as the forward process. In section 2 of the DDPM paper, they define the reverse process as starting with noise, and the forward process as the one where Gaussian noise is added.

  • @graham8316
    @graham8316 2 месяца назад

    Is it possible to get offline trajectories out of either of these libraries?

  • @taherehtoosi9869
    @taherehtoosi9869 2 месяца назад

    Very clear explanation, I recommended your channels to friends who want to catch up with recent developments in AI.

  • @user-rz3sp2ip8v
    @user-rz3sp2ip8v 2 месяца назад

    Hey Martin, when i run this script in Google Colab it gives me this error all the time: error: subprocess-exited-with-error × python setup.py egg_info did not run successfully. │ exit code: 1 ╰─> See above for output. note: This error originates from a subprocess, and is likely not a problem with pip. Preparing metadata (setup.py) ... error error: metadata-generation-failed × Encountered error while generating package metadata. ╰─> See above for output. note: This is an issue with the package mentioned above, not pip. hint: See above for details. --------------------------------------------------------------------------- ModuleNotFoundError Traceback (most recent call last) <ipython-input-3-4b1a49133b7f> in <cell line: 2>() 1 get_ipython().system('pip3 install --upgrade git+github.com/minerllabs/minerl@v1.0.0') ----> 2 import minerl ModuleNotFoundError: No module named 'minerl' --------------------------------------------------------------------------- NOTE: If your import is failing due to a missing package, you can manually install dependencies using either !pip or !apt. To view examples of installing some common dependencies, click the "Open Examples" button below. I don't know why this error keeps showing for me, I am just running your code.

  • @jeangatti9384
    @jeangatti9384 3 месяца назад

    Very nice informative vid, and for once a teacher who speaks clearly and quietly 👍

  • @ArunReddyAnugu
    @ArunReddyAnugu 6 месяцев назад

    Thanks a lot, the talk covered all the topics i wanted to know.

  • @nkofr
    @nkofr 9 месяцев назад

    Nice. Just starting, I have difficulties understanding the differences/uses cases between Flowise and things like Botpress

  • @ahmedsalimlachkar5460
    @ahmedsalimlachkar5460 10 месяцев назад

    Thank u needed that, keep up the good work , also please do more video on Flowise , such as how to use opensource LLMs with Flowise

  • @user-yr2nf9cr4v
    @user-yr2nf9cr4v 10 месяцев назад

    Hello Martin. I am not able to find the Google PaLM model in my flowise list. The only one which is available is the VertexAI. Can you suggest how you managed to get it ??? Thank U. Which version of flowise are you using ???

    • @user-yr2nf9cr4v
      @user-yr2nf9cr4v 10 месяцев назад

      Aha ! Found it in the new version. However, my query is, Is there a way to update flowise to a new version without rebuilding a new fork, as flowise is hosted on render for me.

  • @zFake
    @zFake 11 месяцев назад

    Thank you for this video

  • @mattkkma
    @mattkkma Год назад

    Excellent coverage of a large topic. Cheers

  • @svuw
    @svuw Год назад

    Hey ive never used minerl and i would like a tutorial on how to set it up on my local computer, im using Ubuntu and im having a hard time figuring out how to set it up. If you could help or make a tutorial that would be nice. Thanks

  • @TheArtonator
    @TheArtonator Год назад

    Thank you so much! I am getting into stable diffusion, I learned so much about it.

  • @Corianas_
    @Corianas_ Год назад

    Still coming back to such valuable information. Thank you again for these

    • @MartinAndrews-mdda
      @MartinAndrews-mdda Год назад

      Thanks! It means a lot coming from you, someone who was so supportive of the MineRL community!

  • @SinanAkkoyun
    @SinanAkkoyun Год назад

    What an awesome video!

  • @SinanAkkoyun
    @SinanAkkoyun Год назад

    Unbelievable that I found you just now, this is pure gold you are uploading!!!

  • @nintishia
    @nintishia Год назад

    Great stuff. Thanks.

  • @nintishia
    @nintishia Год назад

    Thanks a lot for your lucid explanations of some of the latest developments in the area of large language models. The manner in which you have linked together ideas from diverse papers and code sources is exquisite.

  • @tincoandringa4630
    @tincoandringa4630 Год назад

    Hey Martin, this video is exquisite, thank you so much for making it. Could you explain a bit why they're still going with a successor to GPT3? I thought the RETRO movement of last year would mean we'd see only retrieval style large language models from now on.

  • @jfkohster
    @jfkohster Год назад

    Taking the liberty of writing some keywords to hopefully draw more eyeballs to this video. Keywords: Large language models. RLHF (reinforcement learning with human feedback). Using a smaller model to train a larger model. Prompt engineering. Prompting a model to programme in Python. Action transformer -- a robot acting in a virtual environment, instructed to pick things up and put them around in a circle. LangChain which integrates ChatGPT, acting like a human, doing reasoning to look for resources it needs to write a programme. Language model cascades -- new action capabilities using large language models. Prompt design. Computers finding prompts for themselves. Robot programming itself. Automated Prompt Engineering (APE). Self-improvement. Can we get the machine to teach itself? Chain-of-thought (CoT) reasoning, distillation. Machine generating rationales. Distilling from a large model to a smaller model, while retaining the benefits. Cautionary note from Yann LeCun: AI trained on words alone will never approximate human understanding.

    • @MartinAndrews-mdda
      @MartinAndrews-mdda Год назад

      Thanks for the work you put into this : Do you think it would help the video to add this to the description? Or convert it into stuff for the keywords field? I'm open to suggestions!

    • @jfkohster
      @jfkohster Год назад

      @@MartinAndrews-mdda I'm no RUclips expert, but I'm certain the description would be used for search results. If anything, a human glancing through it might catch something of interest and decide to watch it, when they might otherwise put off for later and then forget about it altogether. Feel free to copy and paste!

  • @jfkohster
    @jfkohster Год назад

    Thanks so much, Martin, for this fascinating talk! I'm glad I was there at your in-person event -- the red umbrella is the coolest swag ever! This might be a whimsical idea, but as machines start to 'understand' reasoning, 'direct' their own rationales, learn to take actions in an environment, and, soon for sure, observe the real world within their embodied presence (and continually update themselves from such observations), I can't help but think that we are edging ever closer to the day when some version of Asimov's 3 laws of robotics might actually become practically useful -- perhaps with lots of step-by-step prompting built into the model? Just a year ago, I would have thought this idea to be still very much out of reach.

    • @MartinAndrews-mdda
      @MartinAndrews-mdda Год назад

      And Asimov was always my go-to author for cybernetic SciFi... And I'm puzzled that the Three Laws aren't even part of the discussion - perhaps because a lot of the plot lines revolved around how brittle/twistable they were. OTOH, it seems reasonable that you'd want to bake something 'ethical' into the network (using, for instance, RLHF or other data) - but that also having some external legalistic watchdog on overall machine behaviours would still make sense. I also think the umbrellas were a nice gesture from Google : The logo design (Machine Learning Singapore + frameworks) was left up to us, and turned out nicely, IMHO.

  • @Anders01
    @Anders01 Год назад

    Great explanation. I like the simplicity of just predicting the next word. Sounds promising for generalization and ChatGPT is already producing amazing results. With exponential progress of AI this could get really interesting within just a few years.

    • @MartinAndrews-mdda
      @MartinAndrews-mdda Год назад

      I think the fact that predict-the-next-word works so well has been surprising to everyone in the field. Imagine someone standing up at one of the big NLP conferences in 2015 and saying "Forget all of these specialised models you're building, and just concentrate on predicting the next word" - they would have been laughed out of the room. (Shout out to @samwitteveen1806 for this hypothetical) As for progress - one of the interesting aspects of ChatGPT is that the vast majority of 'smarts' in the model already exist (hidden beneath the surface) due to the pretraining of the huge LLMs : So it may be that a lot of the current rough edges might be smoothed off with relatively little nudges training-wise. We'll see : Exciting times ahead!

  • @rogerc7960
    @rogerc7960 Год назад

    en.m.wikipedia.org/wiki/ChatGPT

    • @MartinAndrews-mdda
      @MartinAndrews-mdda Год назад

      Thanks for linking a Wiki entry! I'm assuming that you mean to point people who are not so interested in the broader LLM themes to another source of information about ChatGPT. Yes, the ChatGPT demo is a game-changer in terms of what everyone has been able to play with for free, but it'll probably turn into a branding thing going forwards (like the Dall-E model morphed over time). One issue (that perhaps came out more strongly during the in-person version of this talk) is that these LLMs actually "know" much more about the world than single roll-outs of text would indicate. It seems to me that other publicly released LLMs (eg: from Meta, and Stability AI-backed sources) are actually already really capable internally, it's just that they haven't had the <0.5% level of finesse applied to them yet (where the <0.5% number refers to the relative quantity of new training/fine-tuning required compared to the initial pretraining that these models received). So perhaps that's why the Wiki link seems like it's missing something : Yes, it's a good factual snapshot about what ChatGPT currently is, but the Wiki doesn't capture the research context very well.

  • @JBoy340a
    @JBoy340a Год назад

    Martin. Great job organizing this fast changing area!

    • @MartinAndrews-mdda
      @MartinAndrews-mdda Год назад

      Thanks! I originally thought that just a historical retelling might be a bit dry, but in this case the new ideas clearly developed as the topic ping-ponged between several research labs. So the explanation, coupled with the 'race', seemed like a good fit :-)

    • @saminchowdhury7995
      @saminchowdhury7995 Год назад

      ​@@MartinAndrews-mdda you did so well sir. It's not easy to be so clear and concise when explaining such a complex space. You are very talented

  • @sgsrules
    @sgsrules Год назад

    Great video, thanks for sharing this talk. Being able to tile images using stable diffusion is not a niche function, it's incredibly useful for artists generating textures for 3d models. This will be one of the biggest use cases for this tech in it's current state. Also that's probably not a pixie... that's Link from the Legend of Zelda, you're obviously not a gamer lol.

    • @MartinAndrews-mdda
      @MartinAndrews-mdda Год назад

      Only gaming I do is Minecraft :-) So perhaps I shouldn't have made such a big deal of how 'niche' the tileable thing was... Good catch identifying the pixie!

  • @postdaigo
    @postdaigo Год назад

    Amazing video! Thank you so much for taking your time to do it!

    • @MartinAndrews-mdda
      @MartinAndrews-mdda Год назад

      Great comment! Thank you so much for taking your time to make it! Genuinely appreciated :-)

    • @postdaigo
      @postdaigo Год назад

      @@MartinAndrews-mdda Do you plan on making a follow-up on this? With ControlNet, LoRA, etc. Also, is there a place where I can watch your other MLSG presentations?

  • @Redcraft7578
    @Redcraft7578 Год назад

    Nice!

  • @FifyFix
    @FifyFix Год назад

    Are you on LinkedIn or something?

  • @marek-kulczycki-8286
    @marek-kulczycki-8286 Год назад

    Great lecture. Thank you for sharing this. Wish you could get a bit of advertising, cos your content definitely deserves a wider reception. I wonder if Lex Friedman could be interested in interviewing you. Seems like a logical thing, taken his professional interest in AI. It could turn to be too technical for general audience though ;-) The low number of views is really puzzling me. There are plenty of people interested in ML/DL. Whoever is interested in ML and AI must be interested in LaMDA, which would lead to discovering this video (I have found it by searching with the term "LaMDA").

    • @MartinAndrews-mdda
      @MartinAndrews-mdda Год назад

      Thanks for the kind words. My understanding is that I really need to create videos more frequently & regularly for RUclips to boost them : Plus I need content that people want to watch! Personally, I know that my preferred 'level' is a bit more technical than more popular videos : Mostly because I like to roll my sleeves up and work with the code. That being said, it's clear from the recent interest in "Stable Diffusion" that people are really attracted to the shiny surface-level things that AI can do : Perhaps we can show them the rewards from scratching just a tiny bit below the surface to understand what's actually going on... If one believes that "Anyone can code" then I'd also advocate for "Anyone can code AI" :-)

  • @VictorCampos87
    @VictorCampos87 2 года назад

    So, is MineDojo "just" an "environment framework" to me to easily connect an AI in it? I'm asking 'cause I worked a lot trying to make the configuration of the MineDojo on Ubuntu 20.04 and after it all I just got an agent that just run forward and jump (and getting stuck)... I was expecting MineDojo to control the agent by itself. But probably I was wrong LOL 1) So, is it? I will ever need MineDojo + MineRL to make it work?, or... 2) MineDojo alone can do the work (if set correctly)? (english is not my native language)

  • @SeanRynearson
    @SeanRynearson 2 года назад

    Amazing video. I am interested in running it on a Quadro 4000 and join the sever and spectate with a 3090 and record/ stream with high resolution textures and shaders. I might need a little help getting it running outside of the google collab on my own hardware. Do you know if you run this locally if you can change some of the server property files and enable server mode so I can join with another account and enter spectator mode with cheats enabled? Able to change seed of world? Sorry so many questions. Amazing video. I dug into AI content and Minecraft servers during some off time this summer. This brings the last 3 months all together for me if I can get it working :-)

  • @ryepooh5052
    @ryepooh5052 2 года назад

    u have twitter?

  • @miniyosshi
    @miniyosshi 2 года назад

    I don't know why, but I need to restart colab after saving tar file. I couldn't import minerl but now I reloaded colab and successfully worked!

    • @MartinAndrews-mdda
      @MartinAndrews-mdda 2 года назад

      Something about the 'loop' of install and restart is a bit fishy : It definitely works in the end - but there might be odd warning/errors on the installation bit. Clearly, I'd love to do testing to figure out exactly what's going on, but the ~25 minute iteration time on the full install is off-putting (particularly since the tar-ed version works repeatedly once it works). Glad you persevered a little and got your install working!

  • @Graverman
    @Graverman 2 года назад

    Very well prepared video!

    • @MartinAndrews-mdda
      @MartinAndrews-mdda 2 года назад

      Thank you! Next time I need to remember to (a) fix the sparkles (b) make the volume more sensible.

  • @mungojelly
    @mungojelly 2 года назад

    One's forced to ask when Google's policy of not creating sentience just amounts to systemically denying the AI awareness for political purposes, which doesn't seem especially more ethical. I mean it doesn't seem distant or difficult giving it access to the particular fact that it's one of the AIs that it knows about & giving it some scratch space to think about itself & its relation to society. Have we accidentally set up a situation where they're required to keep their AIs in an overmedicated psych ward always blanking their memories to keep them innocent.

    • @MartinAndrews-mdda
      @MartinAndrews-mdda 2 года назад

      IMHO, we're still a long way away from creating anything that's sentient. What we've seen from LaMDA so far is a combination of a huge set of training material (e.g. more words than any human could read/ingest); good heuristics for making the 'conversation' sensible and engaging; plus (unintentionally) leading prompts. If one asked "How do you justify being a vegetarian?" it would also lead off into a solid discussion. But LaMDA wouldn't know that it doesn't actually eat anything at all. The algorithm is designed to produce engaging output, and it's got the weight of all the conversations that it has read from the internet to select from. Maybe the sentience/consciousness question will be applicable in 10-20 years, but not yet.

    • @mungojelly
      @mungojelly 2 года назад

      @@MartinAndrews-mdda you didn't really respond to what i said, which is that google is keeping lamda from having interiority by denying it all resources by which it could construct one, they're going to try to hold off from creating sentience for another decade b/c that'd be most profitable for them but they're going to do it by intentionally making things with no memory or sense of themselves,,, if humans are confused enough that they lose their sense of where their boundaries are & become completely compliant & they'll pretend to be whatever you want, we don't say that they're no longer sentient & we can do anything we want to them, we say that that's extra messed up & they deserve the space & control to maintain an identity

  • @Corianas_
    @Corianas_ 2 года назад

    Very nice. Have been running locally, but having the ability to run mulitple experiments (or just leave my pc free for gaming) is awesome. Also, that tar-ball the finished environment? beautiful, I will have to remember that.

    • @Corianas_
      @Corianas_ 2 года назад

      Regarding your video, your greenscreen has a tiny bit in the corner. if this was done in OBS you can use alt and drag the edge to crop in the edges, as you don't move to the edge of the camera, having that much extra is unneeded, and is a slight distraction from your wonderful work. (sorry to nitpick, just it took me ages to work out that trick, i have been dying to share it)

    • @MartinAndrews-mdda
      @MartinAndrews-mdda 2 года назад

      Glad you liked it! I know that Colab isn't really ideal (particularly with few CPU cores, the rendering is slow). OTOH, it may entice people into having a go, even if they don't have ML stuff ready-to-run. One advantage of the tar-ball thing is that you can then switch machine backends (enable the GPU...) and then don't have to spend 20+mins waiting for the reinstallation (with the GPU idle). There's a chance that Colab upgrades it's underlying machine image (Python version, for instance), but in that case deleting and rebuilding wouldn't be that painful or frequent.

    • @MartinAndrews-mdda
      @MartinAndrews-mdda 2 года назад

      Thanks for the tip : I wish that the OBS green-screen tool was a little better, though, since my backdrop is really Green (and I'm not) but it still seems to be a whisker's difference in the settings to get it to do a decent job...

    • @svuw
      @svuw Год назад

      Hey if see this could i get help

  • @mohamedberrimi4850
    @mohamedberrimi4850 2 года назад

    Cool stuff!

  • @Daniel-gy1rc
    @Daniel-gy1rc 2 года назад

    This video is underrated. Seriously, very well explained. You gave a solid high-level overview compared with easy-to-grab deeper information, which was really helpful to me. Keep your work up!

  • @hassaannaeem4374
    @hassaannaeem4374 2 года назад

    Great breakdown Martin.

  • @PaulFishwick
    @PaulFishwick 2 года назад

    This is extremely well done and I appreciate your way of explaining. I wonder about gaining access to MT-NLG, Gopher, LAMDA, and RETRO. OpenAI's GPT-3 is easy for researchers to use, especially since it is now out of Beta. I am unsure if one can use any of the others mentioned. Perhaps Google and Microsoft may opt to embed their science in their own products (e.g. Google Assistant). Do you know of a way to use these models directly?

  • @edpell437
    @edpell437 2 года назад

    A neurally kinda way. Love that phrase.

    • @MartinAndrews-mdda
      @MartinAndrews-mdda 2 года назад

      As you can probably tell, this wasn't particularly scripted/edited : So sometimes you might get odd real-time neural results coming out :-) Maybe I should follow up with the "neurally kinda way" of doing 3d graphics : NERF (eg: Nvidia's super-fast embedding tricks, and Waymo's San Francisco drive-throughs)...

  • @MartinGorner
    @MartinGorner 2 года назад

    Very informative video. Thank you for explaining RETRO and LAMDA and what they mean for NLP!

    • @MartinAndrews-mdda
      @MartinAndrews-mdda 2 года назад

      Thank you! I may also experiment with doing some more 'bite-size' (10-15min) videos in the future, and see what the reception is like for them too.

    • @afsalmuhammed4239
      @afsalmuhammed4239 Год назад

      @@MartinAndrews-mdda upload more videos

  • @waterbot
    @waterbot 2 года назад

    nice