The Largest Mamba LLM Experiment Just Dropped
Vložit
- čas přidán 6. 07. 2024
- Check out HubSpot's ChatGPT at work bundle! clickhubspot.com/2os
A long awaited sequel in LLM research has appeared, AI21Labs has dropped the biggest Mamba experiment, which is on par with other open source LLM models! Just with a few twists...
Original Mamba Paper
[Paper] arxiv.org/abs/2312.00752
[Code] github.com/state-spaces/mamba
MambaFormer
[Paper] arxiv.org/pdf/2402.04248.pdf
AI21Labs
[Blog] www.ai21.com/blog/announcing-...
[Huggingface] huggingface.co/ai21labs/Jamba...
[NVIDIA NIM] nvda.ws/3Jn5pxb
VideoMamba
[Paper] arxiv.org/abs/2403.06977
[Code] github.com/OpenGVLab/VideoMamba
Special thanks to LDJ for helping out with the content in this video!
This video is supported by the kind Patrons & CZcams Members:
🙏Andrew Lescelius, alex j, Chris LeDoux, Alex Maurice, Miguilim, Deagan, FiFaŁ, Daddy Wen, Tony Jimenez, Panther Modern, Jake Disco, Demilson Quintao, Shuhong Chen, Hongbo Men, happi nyuu nyaa, Carol Lo, Mose Sakashita, Miguel, Bandera, Gennaro Schiano, gunwoo, Ravid Freedman, Mert Seftali, Mrityunjay, Richárd Nagyfi, Timo Steiner, Henrik G Sundt, projectAnthony, Brigham Hall, Kyle Hudson, Kalila, Jef Come, Jvari Williams, Tien Tien, BIll Mangrum, owned, Janne Kytölä, SO, Richárd Nagyfi, Hector, Drexon, Claxvii 177th, Inferencer, Michael Brenner
[Discord] / discord
[Twitter] / bycloudai
[Patreon] / bycloud
[Music] Massobeats - Lush
[Profile & Banner Art] / pygm7
[Video Editor] Silas
0:00 Intro
1:16 Hubspot
2:24 Jamba
8:08 VideoMamba - Věda a technologie
Check out HubSpot's ChatGPT at work bundle here: clickhubspot.com/2os
unfortunately topping the last mamba edit is way too hard, but I guess now at least we know *_mamba is real_*
Hove you seen google's griffin and hawk?
If mamba does not scale well, we still have diffusion models for text
Why not both?
Jamba Mamba ¡Ay, caramba!
bien dicho
@fireship game up your memes this boy is strapped to the teeth.
Would be interesting to see the infinite context from the "Leave No Context Behind:
Efficient Infinite Context Transformers with Infini-attention" Paper explained.
Ikr. I wonder why that paper didn't get more traction
love your memes so much
wait.... this is not a @fireship video damm
close enough
Hey, would you be interested in making a video about ponydiffusion ?
Isn't pony diffusion just a latent diffusion foundation model, like stable diffusion?
@@kolkoki I got no clue about any of that sorry, I just know that, at least back then, pony revolutionized accuracy to character LoRAs and made the generations of already existing characters so much more accurate than other checkpoints.
I dont watch this channel much, but I did see that epic mamba short in one of your videos and it has been ingrained in my mind ever since.
Love your video essays, good and easy to understand and nice to catch up to SOTA methods.
Very nice
could we use it through ollama?
so is this cheaper than Mistral 7B? ❤
appreciate these videos . the main thing ive heard regarding mamba v transformers is that the discoveries of optimizations within transformers are still abundant , quantization alone is massive in enabling the networks to run on average hardware , and the ridiculousness of 1.56bit quantization working is incredible where as with mamba no quantization is available .
If you enjoy CZcams and it pays bills then sure but play it safe so you don't make life much harder than necessary. Plus you might be able to do research at the same time and present it to people in a more consumable form
Everyone is combining models rn. Some people combine NeRF and GS and that worked as well. I guess that ML will become just a mixer for architectures at least for some commercial devs
we need one called Mongoose
Im trying to write bitnet layers for Veri log
dank af
so whats next? kalman filter with learned dynamic?
Oh god. How much of a memelord can you be?! The "can you get much higher" right after the lobotomy? I love you man.
Isn't mashing together RNNs and Transformers just RWKV?
Nah bro infini attention is where it's at
wait a sec bycloud still makes videos? :V
The part on Jamba honestly sounds like someone making shit up with fake words, but thats actually all real.
The "Microservices" video by KRAZAM is now reality.
Every time I hear Mamba I can only think of the Python CLI
so.... still waiting on the guff file ey?
7:17 LLM Models live inside ATM Machines
Gemma 7B competing with llama70b, mixtral, and jamba damn scale that thing up
The Mambaaaaaaa the Mamba is reaaaaaaaaaaaaallllllll
3:36
It would still be good for people wanting small models to run on very cheap devices without needing all the quality, no?
what happened with Hyena?
Obviously. I published in December of 2023: Anchoring_Global_Security_Autonomous_Shipping_with_Mind_Reading_AI_GPT-core_and_MAMBA-_core_Agents_RAG-Fusion_AI_Communities_Hive-_AI_and_the_Human_Psyche #mindreading #AI #agent cores #Mamba2 and GPT4, 5 and sequential models #IDE
we live in the future bros
Man i'm tired of waiting for GPT-5 , what are they waiting for?
They're currently red-teaming the model
@@VisionaryPathway thanks for answering! How long do you think it will take until release?
@@user-fr2jc8xb9g personally, I think it’s releasing anytime within next 4-12 weeks (my own opinion/prediction)
329th view. Can I get a heart?
Why copy Fireship's thumbnails? Sad, man.
There's no way you think someone can own the format of, "character on the right highlighting big text on the left"??? Thumbnails are like, the least important part of a video when you watch it as a viewer, but it's the most important part when it comes to grabbing viewers' attention. Why shouldn't you use other creators' ideas on what works, when that's not where your creative input is, and it's super important to know you have a successful thumbnail style?
Who cares, we're here for him, not his thumbnail
He's been making these style thumbnails for 2+ years now. It's not copying, and it never will be. It's fine to take inspiration from other people when you like their work. And have you considered that he could have also just had this idea himself? It's extremely common for multiple people to have essentially the exact same idea.
Thumbnails look similar because there are literally common guidelines that are proven to improve the reach of any YT video either by being more likeable to eyes or because algorithm picks them to trending tab
Didnt fireship copy this guy?
1st
First
In the next improvement paper... they're going to suggest a 'hybrid architecture' where you skip the mamba layer entirely....
nobody really uses vanilla attentions in LLMs so like most of what mamba says is BS
Its extremely obvious that the thumbnails are replicas of Fireship, I know you're trying to grow your channel but it's a little off putting
this dude is copying fireship
maybe he's his otosan
Please stop copying fireship content and thumbnails
Pathetic @fireship ripoff.