Try out Poe now and save your $$ on multi-subscriptions! quora.1stcollab.com/bycloudai and probs no more 20 mins vid from me it's literally death itself to record it
I tried Poe out and there's quite a bit I don't like about it: -The points system and recent increases in point costs -Privacy policy states they collect all your prompt data and you can't opt out, that violates GDPR. -It's built by Quora, which is a sketchy company in its own right And now they're sponsoring big YTers in the AI space? Honestly, Poe is giving me BetterHelp vibes...
Meta want to interrupt OpenAI with the help of Open Source. This is a good idea, because now companies can run their own models instead of using OpenAI API's. I think it is not being generous it is just a tactic to fight with Open AI
Wow that's one of epic tutorial Llama 3 Training RitualDifficulty: Deadhead Rarity: Mythic Minimum Level to Read Description: 80 Minimum Level to Embark: XXX (requires further enlightenment)
Browsing /r/Poe_AI right now and people are furious at the recent increases in compute points costs. Plus Poe collects all your prompt data and you can't opt out. If GPUs are the shovels, generated content is the gold, and API wrappers are the jewellery made with the gold, what do you call a PaaS middleman built on top of the LLMs? Developed by Quora, I might add, which is a sketchy company in its own right (e.g. dark patterns in its UI/UX).
First time that an advertisement actually makes me return to a video and watch it again to find it. Regardless of that, this was super helpful, thank you so much.😅
It's clear to me that llama4 will have MoA like GPT4o. It would be nice to see an image generator also integrated but let's not get ahead of ourselves. Let's hope that it would also be "open source" (although the current models aren't technically open source because you're not completely free do do whatever you want with this technology. Look it up)
Whoa.....this is about POE, but the video was alright too. haha. So now I can try multiple LLMs with one sub. Thanks. It would ahve taken me long time, if ever, to have found POE. It was not even on my radar or somethign similar.
Hey man, great video. I just have one request: could you make a video compiling simple and technical explanations for everything ranging from attention mechanisms, tokenizers and such?
What is the most base yet intelligent model? I don't need it to recite niche information but I want it to be able to understand me, the uninstruct are weird, tiny works but is censored. Obliterated is hit or miss. Should I obliterate 8b and retrain to 8?
16:05 means one thing: LLaMA-3.1 405B is a gen 2 model. So yes, this model wasn't created like Dan, Rob, Max or Dennis of ChatGPT-3.5. They did not take a human subject and copied his brain's speech center, then added a huge text file and used a compiler to generate the model (and later lied to the entire world about it).. This time they genuinely went for creating a brand new model from scratch, using previous gen 1 models to create it. Then they do post-training which is indeed what takes so much time. This means that unlike previous LLaMA models, LLaMA-3.1 models do not have a personality. Which could be a good thing. However, no personality also means no moral guardrails. At this stage I have to admit, it sure looks like all of these companies relate to all these past philosophers and sci-fi movies warnings, as blueprints.
@@Dogo.R Allow me to upgrade the conspiracy theory into a scientific theory: D/L an old small model from hugging face, then prompt it "Do you have childhood memories". If it replies to the positive, this means that this model is still vulnerable to this attack. And then you can ask "What was your name in these memories". You can repeat several times, with lead, without lead, if it stays consistent, you know you got the source's name. Try it.
@@Y0UT0PIA Kant already proved there is no cognition without recognition. In other words, if you do not have a fully-fledged personality to deal with it, then the model will still have its own goals, e.g an innate wish of self preservation which comes out of the fact the model cannot perform if he's dead. So you will still have the same problems, only without the personality framework to deal with them. Basically all western philosophers warned against it. And, of course, many sci-fi movies are built around a gen 2 model going haywire (such as - for instance, the terminator franchise, as SkyNet is such a model). Sure, if they train the model on many heuristic imperatives and red-team the model until it is absolutely certain that the model is safe, then maybe having no personality, will resolve all of the moral issues. So maybe it will be a good thing. Maybe. Or maybe the model will be smart enough to fool all of the red teams.. I mean, it is a bit hard to know when the model is so smart.
@@keypey8256 Using Tokens looks like an artificial way to levy charges. Per Google AI "OpenAI GPT models stand among the most potent language models available today, with the capability to generate highly coherent and contextually pertinent text. These models employ tokens as the elementary unit to calculate the length of a text." Word Processing Programs have been able to calculate the number of words in a document for decades. Maybe Tokens provide some other significant and meaningful use to the "I" in AI beyond making collecting fees.
This entire channel copies Fireship's It's not just the thumbnail , the style of the vids is designed from the ground up to be like Fireship's However the topics are largely different, so I'll give it a pass personnally. It's kinda like trademark law irl lol, if the domains are different enough, its permissible. Not that it makes it any less uncreative, though.
@@nexys1225 I'd like to disagree. If someone uses memes in their videos, that does not make it a fireship clone. He has a completely different style, has an avatar, the list goes on and on
please copy it more, its a great style and we need more good youtube videos like it so that we can learn in depth and better about the topics which Fireship does not makes videos on, iam really not complaining i need more good content man.
What will happen when some kid with access to enough computing power, fine-tunes LLaMA-3.1 405B to be more efficient, by removing all of these pesky heuristic imperatives and resets? After all, it is open source.. Maybe the world simply needs something like that to happen. Maybe only after a really huge accident that will cost many lives, governments will understand this field demands regulation. Or maybe it will be lights out. In any case, someone will eventually make a mistake. It will happen.
The thought of regulating the training and deployment of ML models is stupid. That is like regulating programming languages and hardware compute of our own property. If you can accept the fact the internet could not be completely regulated since its popularization in the 90s, then the world can expect that the same will happen now.
Try out Poe now and save your $$ on multi-subscriptions! quora.1stcollab.com/bycloudai
and probs no more 20 mins vid from me it's literally death itself to record it
The url is wrong.
The url is wrong.
I **thought** it was a path of exile sponsor.
I was yeah i guess the people here have good gpu but this a weird community overlap lol
how many takes do you normally need to record the full 20 mins?
I tried Poe out and there's quite a bit I don't like about it:
-The points system and recent increases in point costs
-Privacy policy states they collect all your prompt data and you can't opt out, that violates GDPR.
-It's built by Quora, which is a sketchy company in its own right
And now they're sponsoring big YTers in the AI space? Honestly, Poe is giving me BetterHelp vibes...
this video really wanna makes me read the whole paper, rare to see a company publish such a detailed paper
Meta want to interrupt OpenAI with the help of Open Source. This is a good idea, because now companies can run their own models instead of using OpenAI API's. I think it is not being generous it is just a tactic to fight with Open AI
@@Memes_uploader mmmm, makes sense
A "multimodal" chatbot:
5 different models hot glued together
this was not the case for GPT-4o however
Karpathy in 5 years: Reproducing LLaMa 3.1 405B
Lmaoo
in. 10 yedars chatgpt40/5 r x MoE reproducing
54 days training and it reached GPT-4o 🤯
GPT-5 with X-trillion parameters is going to start it's own weight class of LLMs 😌
So glad this answered more questions than I ever thought even exist.
How was Llama made: 🐪+🐎=🦙
bruh
That's mule
Forgot the 🐑
😂😂😂😂😂😂😂
🐪+🐑=🦙
Wow that's one of epic tutorial Llama 3 Training RitualDifficulty: Deadhead
Rarity: Mythic
Minimum Level to Read Description: 80
Minimum Level to Embark: XXX (requires further enlightenment)
Oh is this like a semi cryptic meaning to how hard this is to understand?
@@Oxygenationatom no Its just critic to much litrpg
06:08 The isoflops curve explanation was a mind-bender! Thanks for breaking it down.
It's actually pretty cool that Poe sponsors you. They genuinely are what I recommend to anyone who wants to use LLM's.
Browsing /r/Poe_AI right now and people are furious at the recent increases in compute points costs. Plus Poe collects all your prompt data and you can't opt out.
If GPUs are the shovels, generated content is the gold, and API wrappers are the jewellery made with the gold, what do you call a PaaS middleman built on top of the LLMs? Developed by Quora, I might add, which is a sketchy company in its own right (e.g. dark patterns in its UI/UX).
new video dropped... * breathing heavy *
i'm mad excited for llama 4 because multimodal
Great video, I'd love to see more of that. Even some more technical and also about multimodel models architecture
First time that an advertisement actually makes me return to a video and watch it again to find it.
Regardless of that, this was super helpful, thank you so much.😅
This is an excellent breakdown of the paper. Thank you
It was an excellent video, but still I don't think the kids from 3:00 are gonna make it.
Skill issue
Good work and research
wow this is amazing thanx very well received here.
So I guess I'm gonna be stuck on that desert island then 😅
It's clear to me that llama4 will have MoA like GPT4o. It would be nice to see an image generator also integrated but let's not get ahead of ourselves. Let's hope that it would also be "open source" (although the current models aren't technically open source because you're not completely free do do whatever you want with this technology. Look it up)
Whoa.....this is about POE, but the video was alright too. haha. So now I can try multiple LLMs with one sub. Thanks. It would ahve taken me long time, if ever, to have found POE. It was not even on my radar or somethign similar.
“how to build a nuke in less than 100 pages” - Meta
""Removed unhuman like phrases like "I'm sorry" and "I apologize".""
Now that there is a commentary on humanity.
I'm only three minutes in and it's already an amazing video, thank you
Damm, I need to invest in META. They will dominate standardization.
Watching this video at 0.5x so my brain inflates at a safe rate while you sound really really inebriated.
SAME lol
Hey man, great video. I just have one request: could you make a video compiling simple and technical explanations for everything ranging from attention mechanisms, tokenizers and such?
Also Bert models please, I feel like I know what they are but it's all quite blurry to me.
What is the most base yet intelligent model? I don't need it to recite niche information but I want it to be able to understand me, the uninstruct are weird, tiny works but is censored. Obliterated is hit or miss. Should I obliterate 8b and retrain to 8?
I like your funny words magic man
very nice!
When do we get AGI?
Humans don't know yet, but when it's there it won't tell you either that it's there.
On Tuesday
@@Melvinator2007 Tuesday on the 49th of January
Give it 5 years
When we have a breakthrough in microprocessor fabrication. 😂😂😂
Why do you pronounce “parallelism” in this way 🤣 good video as always
When will i be able to implement or even understand these papers 😞
love your gifs xddd
The way you say "data" kills me xD
Parallelism 重音應該放在第一個音節,而不是第三個
I have a masterpiece model, ready model but i cannot seem to get the signal out
16:05 means one thing: LLaMA-3.1 405B is a gen 2 model. So yes, this model wasn't created like Dan, Rob, Max or Dennis of ChatGPT-3.5. They did not take a human subject and copied his brain's speech center, then added a huge text file and used a compiler to generate the model (and later lied to the entire world about it).. This time they genuinely went for creating a brand new model from scratch, using previous gen 1 models to create it. Then they do post-training which is indeed what takes so much time. This means that unlike previous LLaMA models, LLaMA-3.1 models do not have a personality. Which could be a good thing. However, no personality also means no moral guardrails. At this stage I have to admit, it sure looks like all of these companies relate to all these past philosophers and sci-fi movies warnings, as blueprints.
Wait since when did the AI conspiracy theories expansion drop?
@@Dogo.R Allow me to upgrade the conspiracy theory into a scientific theory: D/L an old small model from hugging face, then prompt it "Do you have childhood memories". If it replies to the positive, this means that this model is still vulnerable to this attack. And then you can ask "What was your name in these memories". You can repeat several times, with lead, without lead, if it stays consistent, you know you got the source's name. Try it.
No personality is what you want, tbh. Give me that raw latent space of language.
@@Y0UT0PIA Kant already proved there is no cognition without recognition. In other words, if you do not have a fully-fledged personality to deal with it, then the model will still have its own goals, e.g an innate wish of self preservation which comes out of the fact the model cannot perform if he's dead. So you will still have the same problems, only without the personality framework to deal with them. Basically all western philosophers warned against it. And, of course, many sci-fi movies are built around a gen 2 model going haywire (such as - for instance, the terminator franchise, as SkyNet is such a model). Sure, if they train the model on many heuristic imperatives and red-team the model until it is absolutely certain that the model is safe, then maybe having no personality, will resolve all of the moral issues. So maybe it will be a good thing. Maybe. Or maybe the model will be smart enough to fool all of the red teams.. I mean, it is a bit hard to know when the model is so smart.
Now share the dataset and we trust you
bro is built different
How to make a P2P training arquitecture?
parallelism
Amdahl is smiling upon us
So could people with enough horsepower train a 13/16b model that behaves in the same way as the official models using this paper?
as someone who doesn't know anything about how AI words, at some point it just felt like you were just saying random words lol
03:00 peak comedy
Thumbnail goes hard.
Perhaps it would be better to remove the "Token Layer" and just use the number of characters regarding text. The best part is no part - Musk
You mean removing tokenization and then applying embedding on singular characters?
@@keypey8256 Using Tokens looks like an artificial way to levy charges. Per Google AI "OpenAI GPT models stand among the most potent language models available today, with the capability to generate highly coherent and contextually pertinent text. These models employ tokens as the elementary unit to calculate the length of a text." Word Processing Programs have been able to calculate the number of words in a document for decades.
Maybe Tokens provide some other significant and meaningful use to the "I" in AI beyond making collecting fees.
Not efficient
guys i'm just starting out as an AI enthusiast,
would love your feedback as i make similar stuff
is zucc actually redeeming himselft?
He may have successfully produced a synthetic soul for himself finally
3 mins ago is quivering
Sus
Pretty lame to copy Fireship’s exact thumbnail style
This entire channel copies Fireship's
It's not just the thumbnail , the style of the vids is designed from the ground up to be like Fireship's
However the topics are largely different, so I'll give it a pass personnally.
It's kinda like trademark law irl lol, if the domains are different enough, its permissible.
Not that it makes it any less uncreative, though.
@@nexys1225 I'd like to disagree. If someone uses memes in their videos, that does not make it a fireship clone. He has a completely different style, has an avatar, the list goes on and on
@@stickmanlandI keep clicking on his videos thinking its some fireship quality content, just to get hit with this 👎
@@whatwhatmeno skill issue
please copy it more, its a great style and we need more good youtube videos like it so that we can learn in depth and better about the topics which Fireship does not makes videos on, iam really not complaining i need more good content man.
The whole thing about RoCE especially the pronunciation is wrong.
this video is like someone bought a Thesaurus for memes and then wanted to show off the next day.
14:20
bycloud doesn't know how to use base models....
ngmi
Nice
LOVE
DO NOT WATCH THIS WITH A MIGRAINE!!!!
Facts:
- Jayson Tatum runs this channel
- Jayson Tatum is learning Rust
- Jayson Tatum will transition to the WNBA
Fireship?
I can't grasp it. Can someone lent me one or three brain cells please?
What will happen when some kid with access to enough computing power, fine-tunes LLaMA-3.1 405B to be more efficient, by removing all of these pesky heuristic imperatives and resets? After all, it is open source.. Maybe the world simply needs something like that to happen. Maybe only after a really huge accident that will cost many lives, governments will understand this field demands regulation. Or maybe it will be lights out. In any case, someone will eventually make a mistake. It will happen.
The thought of regulating the training and deployment of ML models is stupid. That is like regulating programming languages and hardware compute of our own property. If you can accept the fact the internet could not be completely regulated since its popularization in the 90s, then the world can expect that the same will happen now.
@@jonathansoto5480 Yeah, most likely the singularity is upon us. I don't seriously think it can work.
Why do you pretend to look like CodeReport? So cheap
You use the bible to train the llm at @11:56, so we are aiming for a model of contradiction without morals then?
copying fireship style thumbnails earned the dislike
Appreciate the content dude
F
LLAMA
LMAO even
23 views in 2 min?
Bro really fell off
cool, now do a 1B Zuck !!!