Check out HubSpot's ChatGPT at work bundle here: clickhubspot.com/2os unfortunately topping the last mamba edit is way too hard, but I guess now at least we know *_mamba is real_*
Would be interesting to see the infinite context from the "Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention" Paper explained.
@@kolkoki I got no clue about any of that sorry, I just know that, at least back then, pony revolutionized accuracy to character LoRAs and made the generations of already existing characters so much more accurate than other checkpoints.
If you enjoy RUclips and it pays bills then sure but play it safe so you don't make life much harder than necessary. Plus you might be able to do research at the same time and present it to people in a more consumable form
Everyone is combining models rn. Some people combine NeRF and GS and that worked as well. I guess that ML will become just a mixer for architectures at least for some commercial devs
appreciate these videos . the main thing ive heard regarding mamba v transformers is that the discoveries of optimizations within transformers are still abundant , quantization alone is massive in enabling the networks to run on average hardware , and the ridiculousness of 1.56bit quantization working is incredible where as with mamba no quantization is available .
Obviously. I published in December of 2023: Anchoring_Global_Security_Autonomous_Shipping_with_Mind_Reading_AI_GPT-core_and_MAMBA-_core_Agents_RAG-Fusion_AI_Communities_Hive-_AI_and_the_Human_Psyche #mindreading #AI #agent cores #Mamba2 and GPT4, 5 and sequential models #IDE
The part on Jamba honestly sounds like someone making shit up with fake words, but thats actually all real. The "Microservices" video by KRAZAM is now reality.
There's no way you think someone can own the format of, "character on the right highlighting big text on the left"??? Thumbnails are like, the least important part of a video when you watch it as a viewer, but it's the most important part when it comes to grabbing viewers' attention. Why shouldn't you use other creators' ideas on what works, when that's not where your creative input is, and it's super important to know you have a successful thumbnail style?
He's been making these style thumbnails for 2+ years now. It's not copying, and it never will be. It's fine to take inspiration from other people when you like their work. And have you considered that he could have also just had this idea himself? It's extremely common for multiple people to have essentially the exact same idea.
Thumbnails look similar because there are literally common guidelines that are proven to improve the reach of any YT video either by being more likeable to eyes or because algorithm picks them to trending tab
Check out HubSpot's ChatGPT at work bundle here: clickhubspot.com/2os
unfortunately topping the last mamba edit is way too hard, but I guess now at least we know *_mamba is real_*
Hove you seen google's griffin and hawk?
Jamba Mamba ¡Ay, caramba!
bien dicho
Would be interesting to see the infinite context from the "Leave No Context Behind:
Efficient Infinite Context Transformers with Infini-attention" Paper explained.
Ikr. I wonder why that paper didn't get more traction
If mamba does not scale well, we still have diffusion models for text
Why not both?
love your memes so much
I dont watch this channel much, but I did see that epic mamba short in one of your videos and it has been ingrained in my mind ever since.
@fireship game up your memes this boy is strapped to the teeth.
Hey, would you be interested in making a video about ponydiffusion ?
Isn't pony diffusion just a latent diffusion foundation model, like stable diffusion?
@@kolkoki I got no clue about any of that sorry, I just know that, at least back then, pony revolutionized accuracy to character LoRAs and made the generations of already existing characters so much more accurate than other checkpoints.
wait.... this is not a @fireship video damm
close enough
If you enjoy RUclips and it pays bills then sure but play it safe so you don't make life much harder than necessary. Plus you might be able to do research at the same time and present it to people in a more consumable form
Everyone is combining models rn. Some people combine NeRF and GS and that worked as well. I guess that ML will become just a mixer for architectures at least for some commercial devs
And these are getting accepted at A* conferences
appreciate these videos . the main thing ive heard regarding mamba v transformers is that the discoveries of optimizations within transformers are still abundant , quantization alone is massive in enabling the networks to run on average hardware , and the ridiculousness of 1.56bit quantization working is incredible where as with mamba no quantization is available .
Obviously. I published in December of 2023: Anchoring_Global_Security_Autonomous_Shipping_with_Mind_Reading_AI_GPT-core_and_MAMBA-_core_Agents_RAG-Fusion_AI_Communities_Hive-_AI_and_the_Human_Psyche #mindreading #AI #agent cores #Mamba2 and GPT4, 5 and sequential models #IDE
Oh god. How much of a memelord can you be?! The "can you get much higher" right after the lobotomy? I love you man.
Love your video essays, good and easy to understand and nice to catch up to SOTA methods.
The part on Jamba honestly sounds like someone making shit up with fake words, but thats actually all real.
The "Microservices" video by KRAZAM is now reality.
Every time I hear Mamba I can only think of the Python CLI
Why copy Fireship's thumbnails? Sad, man.
There's no way you think someone can own the format of, "character on the right highlighting big text on the left"??? Thumbnails are like, the least important part of a video when you watch it as a viewer, but it's the most important part when it comes to grabbing viewers' attention. Why shouldn't you use other creators' ideas on what works, when that's not where your creative input is, and it's super important to know you have a successful thumbnail style?
Who cares, we're here for him, not his thumbnail
He's been making these style thumbnails for 2+ years now. It's not copying, and it never will be. It's fine to take inspiration from other people when you like their work. And have you considered that he could have also just had this idea himself? It's extremely common for multiple people to have essentially the exact same idea.
Thumbnails look similar because there are literally common guidelines that are proven to improve the reach of any YT video either by being more likeable to eyes or because algorithm picks them to trending tab
Didnt fireship copy this guy?
we need one called Mongoose
could we use it through ollama?
so whats next? kalman filter with learned dynamic?
Gemma 7B competing with llama70b, mixtral, and jamba damn scale that thing up
Nah bro infini attention is where it's at
Very nice
dank af
3:36
It would still be good for people wanting small models to run on very cheap devices without needing all the quality, no?
Isn't mashing together RNNs and Transformers just RWKV?
7:17 LLM Models live inside ATM Machines
wait a sec bycloud still makes videos? :V
so.... still waiting on the guff file ey?
The Mambaaaaaaa the Mamba is reaaaaaaaaaaaaallllllll
we live in the future bros
Im trying to write bitnet layers for Veri log
nobody really uses vanilla attentions in LLMs so like most of what mamba says is BS
what happened with Hyena?
Man i'm tired of waiting for GPT-5 , what are they waiting for?
They're currently red-teaming the model
@@VisionaryPathway thanks for answering! How long do you think it will take until release?
@@AfifFarhati personally, I think it’s releasing anytime within next 4-12 weeks (my own opinion/prediction)
329th view. Can I get a heart?
In the next improvement paper... they're going to suggest a 'hybrid architecture' where you skip the mamba layer entirely....
Its extremely obvious that the thumbnails are replicas of Fireship, I know you're trying to grow your channel but it's a little off putting
this dude is copying fireship
maybe he's his otosan
Please stop copying fireship content and thumbnails
what dataset is it trained on?? and is there any mambaLLM trained on wordnet?
Pathetic @fireship ripoff.