Komentáře •

  • @ALFTHADRADDAD
    @ALFTHADRADDAD Před 4 měsíci +236

    Google SHOCKS and STUNS the Open source landscape

    • @matthew_berman
      @matthew_berman Před 4 měsíci +87

      I should have used this title

    • @TechRenamed
      @TechRenamed Před 4 měsíci +10

      Lol we all should have!!

    • @mickelodiansurname9578
      @mickelodiansurname9578 Před 4 měsíci +8

      @@matthew_bermanI thought at one stage you were literally going to start slapping your forehead off the keyboard!

    • @andersonsystem2
      @andersonsystem2 Před 4 měsíci +6

      Why does most Ai tech channels use that title 😂I just don’t pay attention to titles like that lmao 😂😊

    • @Kutsushita_yukino
      @Kutsushita_yukino Před 4 měsíci +11

      its a meme at this point

  • @Lukebussnick
    @Lukebussnick Před 4 měsíci +101

    My funniest experience with Gemini pro was, I asked it to make a humorous image of a cartoon cat pulling the toilet paper off the roll. It told me that ethically couldn’t because the cat could ingest the toilet paper and it could cause an intestinal blockage 😂

    • @laviniag8269
      @laviniag8269 Před 4 měsíci +3

      histerical

    • @matikaevur6299
      @matikaevur6299 Před 4 měsíci +2

      @@laviniag8269
      but true . .

    • @MilkGlue-xg5vj
      @MilkGlue-xg5vj Před 4 měsíci +5

      Maybe a cat could see the image and do the same

    • @Lukebussnick
      @Lukebussnick Před 4 měsíci +4

      @@MilkGlue-xg5vj haha yeah that would be a real nuisance. But then again, that’s one smart cat. What other potential could that cat have?? 🧐

    • @MilkGlue-xg5vj
      @MilkGlue-xg5vj Před 4 měsíci +10

      @@Lukebussnick Maybe it could become an ai dev at Google

  • @Batmancontingencyplans
    @Batmancontingencyplans Před 4 měsíci +106

    Gemma 7b makes you realise how much compute Google is using just to output sorry I can't fulfill that request 🤣

    • @MilkGlue-xg5vj
      @MilkGlue-xg5vj Před 4 měsíci

      LMFAO

    • @markjones2349
      @markjones2349 Před 4 měsíci

      So true. Uncensored models are just more fun.

    • @MilkGlue-xg5vj
      @MilkGlue-xg5vj Před 4 měsíci

      @@markjones2349 you're talking as if the point of uncensored llms is fun rofl lmfao xd you're just makin' it funnier 🤣

  • @bits_of_bryce
    @bits_of_bryce Před 4 měsíci +88

    Well, I'm never trusting benchmarks without personal testing again.

    • @richoffks
      @richoffks Před 4 měsíci +9

      sorry you had to learn this way

    • @wilburdemitel8468
      @wilburdemitel8468 Před 4 měsíci

      welcome to real life. Can't wait for you to leave the fantasyland bubble all these tech aibros have built around you.

  • @zeal00001
    @zeal00001 Před 4 měsíci +43

    In other words, there are now LLMs with mental challenges as well...

  • @drgutman
    @drgutman Před 4 měsíci +50

    I'm pretty sure they lobotomized it in the alignment phase :)))

    • @hikaroto2791
      @hikaroto2791 Před 4 měsíci +7

      To the point they took the lobotomy fragment and used it in place of the brain, and trashed the actual brain. Not only on models, but on personnel probably

  • @bigglyguy8429
    @bigglyguy8429 Před 4 měsíci +133

    You can't gimp the model with excessive censorship, and also have an intelligent model.

    • @aoolmay6853
      @aoolmay6853 Před 4 měsíci +22

      These are not open models, these are woke models, appropriately liberal.

    • @eIicit
      @eIicit Před 4 měsíci +2

      To a point, I agree.

    • @madimakes
      @madimakes Před 4 měsíci +3

      The nature of the errors here seem irrelevant to being censored or not.

    • @bigglyguy8429
      @bigglyguy8429 Před 4 měsíci +11

      @@madimakes No, the censorship sucks up so much of it's thinking there's little left to actually answer. You can ask the most banal question but it sits there thinking long and hard about if there's any way that could possibly be offensive to the woke? Considered the woke are offended by everything, that's a yes, so it has to work its way around that, then it needs to figure out if it's own reply is offensive (yes, everything is), so it has to find a way around that as well. Often it will fail and say "I'm afraid I can't do that... Dave." Other times it will try, but the answer so gimped and pathetic you'd have been better off asking your cat.

    • @mickmoon6887
      @mickmoon6887 Před 4 měsíci +1

      Exactly
      Model network design is gimped from the creator developers itself when head of Google AI is literally have biased ideological, anti white and very censorship values all proven with their online records that's why those biases reflect onto the model

  • @chriscarr9852
    @chriscarr9852 Před 4 měsíci +139

    This is entirely speculation on my part, but I am guessing Google’s AI effort is largely driven by their PR team. A proper engineering team would never release this kind of smoke and mirrors crap. Right?

    • @chriscarr9852
      @chriscarr9852 Před 4 měsíci +23

      They have tarnished their brand. It will be interesting to see what happens in the next few years with regard to Google. (I do not have any financial interest in google).

    • @mistersunday_
      @mistersunday_ Před 4 měsíci +7

      Yeah, they are the wrong kind of hacks now

    • @alttabby3633
      @alttabby3633 Před 4 měsíci +12

      Or engineering team knows this will be killed off regardless of quality or popularity so why bother.

    • @richoffks
      @richoffks Před 4 měsíci +2

      @@chriscarr9852 we're watching the end of Google smh

    • @michaelcalmeyerhentschel8304
      @michaelcalmeyerhentschel8304 Před 4 měsíci +7

      No, Left. They are all one viewpoint at Google and have been so for decades. The PR folks represent the programmers and their programmer-managers and Sr. management.

  • @natecote1058
    @natecote1058 Před 4 měsíci +17

    If google keeps messing around with their censored models and under performing open source models, they'll get left in the dust. Mistral could end up way ahead of them in the next few months. They should find that embarassing...

  • @deflagg
    @deflagg Před 4 měsíci +66

    Gemini Advanced is bad, too, compared to gpt4. Gemini sometimes answers in a different language, too cautious, and gets things wrong a lot of the times.

    • @CruelCrusader90
      @CruelCrusader90 Před 4 měsíci +11

      "too cautious" is an understatement.

    • @veqv
      @veqv Před 4 měsíci +5

      @@CruelCrusader90 Genuinely. If it's not a question about software development there's a wildly high chance that it'll start quizzing you on why you have the right to know things. I do hobby electronics and wanted to see how it would fare on helping make a charging circuit. It basically refused. Same is true for rectifiers. Too dangerous for me apparently lol. Ask it questions on infosec and it'll answer fine though. It's wild.

    • @richoffks
      @richoffks Před 4 měsíci

      @@veqv lmao it refused, all anyone has to do is release a competely uncensored model and they literally take over the industry from their house. I dont know why google is such a fail at every product launch.

    • @CruelCrusader90
      @CruelCrusader90 Před 4 měsíci

      @@veqv yea i had a similar experience. i asked it to generate a top front and side view of a vehicle chassis to create a 3d model in blender. (for a project im working on) it said the same thing, its to dangerous to generate the image.
      i didnt expect it to make a good/consistent vehicle chassis across all the angles but i was curious to see how far it was from making it possible. and i dont even know how to scale its potential with that kind of a developer behind its programming.
      even a one would represent progression at its slowest form, but that would be generous.

    • @Ferreira019760
      @Ferreira019760 Před 4 měsíci

      Bad doesn't begin to cut it. At this rate, Google will become irrelevant in most of it's services. It makes no difference how much money they have, their policy is wrong and the AI models show it. They are so scared of offending someone or being made liable that their AI actually dictates what happens in the interactions with the users. That doesn't just make it annoying and time wasteful, it means that it cannot learn.Even worse than not learning, it's becoming dumber by the day. I cannot believe i'm saying this, but i miss bard. Gemini doesn't cut it in away way, shape or form. It's probably good for philosophy exercises, but so far I don't see any decent use for it aside from that. Give it enough space to go off in wild tangents and you may get a potentially interesting conversation, but don't expect anything productive from it. I'm done with trying out Google's crap for some time. Maybe in a month or two I will allow myself the luxury of wasting time again to see how they are doing, but not for now. Their free trial is costing me money, that's how bad it is.

  • @NOTNOTJON
    @NOTNOTJON Před 4 měsíci +12

    Plot twist, Google was so far behind the AI race that they had to ask Llama or GPT 4 to create a model from scratch and this is what they named Gemini / Gemma.

    • @tteokl
      @tteokl Před 4 měsíci

      google is so far behind these days, I love Google's design language tho, but their tech ? meh

  • @f4ture
    @f4ture Před 4 měsíci +40

    Google’s NEW Open-Source Model Is so BAD... It SHOCKED The ENTIRE Industry!

  • @jbo8540
    @jbo8540 Před 4 měsíci +27

    Google set the entire OS community back a half hour with this troll release. well played google

    • @romantroman6270
      @romantroman6270 Před 4 měsíci +2

      Don't worry, Llama 3 will set the Open Source community 31 minutes ahead lol

  • @Greenthum6
    @Greenthum6 Před 4 měsíci +22

    I was absolutely paralyzed by the performance of this model.

    • @Wanderer2035
      @Wanderer2035 Před 4 měsíci +2

      Me: I send Pikachu GO! Use STUN attack on Greenthum6 NOW!
      Pikachu: Pika Pika Pika!!! BBBZZZZZZZZZ ⚡️⚡️⚡️⚡️⚡️
      Me: Greenthum6 seems to be in some form of paralysis. Quick Pikachu follow that up with a STUN attack on Greenthum6 NOW! Give him everything you got!!!
      Pikachu: PIKA…. PIKAAAAAAAAAAA……. CHUUUUUUUUUUUUUUUU!!!!!!!
      BBBBBBBBBBZZZZZZZZZZZZZZZZ ⚡️⚡️⚡️⚡️⚡️⚡️⚡️⚡️
      Greenthum6 = ☠️ ☠️☠️
      Me: Aaaahhh that was nice, I’m sure Greenthum6 will make a nice pokimon to my collection 🙂. **I throw my pokiball to Greenthum6 and it captures him as my new pokimon to my collection**

  • @trsd8640
    @trsd8640 Před 4 měsíci +37

    This shows one thing: We need other kind of benchmarks.
    But great video Matthew, thanks!

    • @MM3Soapgoblin
      @MM3Soapgoblin Před 4 měsíci +7

      Deepmind has done some pretty amazing work in the machine learning space. My bet is that they created a fantastic model and that's what was benchmarked. Then the Google execs came along and "fixed" the model for "safety" and this is the result.

    • @R0cky0
      @R0cky0 Před 4 měsíci +1

      Let's call it Matthew Benchmark

    • @R0cky0
      @R0cky0 Před 4 měsíci

      @@MM3SoapgoblinDeepmind should spinoff from Google. It's a shame that they still run under the now Google giving their amazing works in the past

  • @snowhan7006
    @snowhan7006 Před 4 měsíci +24

    This looks like a hastily completed homework assignment by a student to meet the deadline

    • @shujin6600
      @shujin6600 Před 4 měsíci +3

      and that student was highly political and was easy offended to anything

  • @mathematicalninja2756
    @mathematicalninja2756 Před 4 měsíci +20

    On a bright side, we have a top end model to generate reject responses in the DPO

    • @user-qr4jf4tv2x
      @user-qr4jf4tv2x Před 4 měsíci +8

      can we not have acronyms 😭

    • @Alice_Fumo
      @Alice_Fumo Před 4 měsíci +9

      @@user-qr4jf4tv2x I believe DPO in this context stands for "Direct Preference Optimization" which is a recent alternative technique to RLHF, but with less steps and thus more efficient.
      I'm actually not 100% sure, but I believe the joke here is that if you try employing this model for DPO to "align" any other base-model, what you get is another model which only ever refuses to respond to anything.

  • @sandeepghael763
    @sandeepghael763 Před 4 měsíci +12

    @matthew Berman I think something is wrong with your test setup. I tested the `python 1 to 100` example with Gemma 7B via Ollama, 4bit quantized version (running on CPU) and the model did just fine. Check your prompt template or other setup config.

    • @hidroman1993
      @hidroman1993 Před 4 měsíci +1

      He was already recording, so he didn't want to check the setup LOL

  • @AIRadar-mc4jx
    @AIRadar-mc4jx Před 4 měsíci +35

    Hey Mathew, it's not open-source model because they are not releasing the source code. It's open-weight or open model.

    • @PMX
      @PMX Před 4 měsíci +3

      But... they did? At least for inference, they uploaded both python and cpp implementations of the inference engine for Gemma to github. Which I suspect have bugs since I can't otherwise understand how they can release a model that performs this poorly..

    • @judedavis92
      @judedavis92 Před 4 měsíci +2

      Yeah they did release code.

  • @mistersunday_
    @mistersunday_ Před 4 měsíci +33

    Until Google spends less time on woke and more time on work, I'm not touching any of their products with a 10-foot pole

    • @Alistone4512
      @Alistone4512 Před 4 měsíci +3

      - by a person on CZcams :P

    • @StriderAngel496
      @StriderAngel496 Před 4 měsíci

      truuuu but you know what he meant lol@@Alistone4512

  • @Nik.leonard
    @Nik.leonard Před 4 měsíci +7

    At the moment, there is a couple of issues with quantization and running the model in llama.cpp (LM Studio uses llama.cpp as backend), so when the issues are fixed, I'm going to re-test the model. That's because is weird that the 2b model gets better responses than the "7b" (really is more like 8.something) model.

  • @antigravityinc
    @antigravityinc Před 4 měsíci +4

    It’s like asking an undercover alien to explain normal Earth things. No.

  • @protovici1476
    @protovici1476 Před 4 měsíci +7

    I'm wondering if this is technically half open-sourced given some critical components aren't available from Google.

  • @hawa7264
    @hawa7264 Před 4 měsíci +7

    The 2B-Version of Gemma is quite good for a 2b model actually. The 7b model is - a car crash.

    • @frobinator
      @frobinator Před 4 měsíci

      I found the same, the 2B model is much better than the 7B for my set of tasks.

  • @yogiwp_
    @yogiwp_ Před 4 měsíci +1

    Instead of Artificial Intelligence we got Genuine Stupidity

  • @Random_person_07
    @Random_person_07 Před 4 měsíci +3

    The thing about Gemini is it has the memory of a goldfish it can barely hold on to any context and you always have to tell it what its supposed to write

  • @BTFranklin
    @BTFranklin Před 4 měsíci +8

    Could you try lowering the temperature? The answers when you were running it locally look a lot like what I'd expect if the temp was set too high.

  • @davealexander59
    @davealexander59 Před 4 měsíci +6

    OpenAI: "So why do you want to leave Google and come to work with our dev team?" Dev: *shows them this video*

  • @puremintsoftware
    @puremintsoftware Před 4 měsíci +2

    Imagine if Ed Sheeran released that video of DJ Khaled hitting an acoustic guitar, and said "This is my latest Open Source song". Yep. That's this.

  • @PoorNeighbor
    @PoorNeighbor Před 4 měsíci +9

    That was actually really funny. The answers are so out of the blue Mannn

  • @himeshpunj6582
    @himeshpunj6582 Před 4 měsíci +4

    Please do fine-tuning based on private data

  • @DeSinc
    @DeSinc Před 4 měsíci +1

    Looking at those misspellings and odd symbols all through the code examples, it's clear to see something is mis-tuned in the params for whatever ui you're using not being updated to support this new model. Apparently the interface I was using it with has corrected this as I was able to get coherent text with no misspelling but I did see people online saying they were having the same trouble as you, incoherent text and obvious mistakes everywhere. It's likely something wrong with the parameters that must be updated to values that the model works best with.

  • @TylerHall56
    @TylerHall56 Před 4 měsíci +1

    The settings on Kaggle may help- This widget uses the following settings: Temperature: 0.4, Max output tokens: 128, Top-K: 5.

  • @phrozen755
    @phrozen755 Před 4 měsíci +9

    Yikes google! 😬

  • @michaelrichey8516
    @michaelrichey8516 Před 4 měsíci +1

    Yeah - I was running this yesterday and ran into the same things - as well as the censorship, where it decided that my "I slit a sheet" tongue twister was about self-harm and refused to give an analysis.

  • @pixels7223
    @pixels7223 Před 4 měsíci +3

    I like that you tried it on Hugging Face, cause now I can say with certainty: "Google, why?"

  • @oriyonay8825
    @oriyonay8825 Před 4 měsíci +3

    Each parameter is just a floating point number (assuming no quantization) which takes 4 bytes. So 7b parameters is roughly 7b * 4 bytes = 28gb, so 34gb is not that surprising :)

  • @erikjohnson9112
    @erikjohnson9112 Před 4 měsíci +2

    Maybe it was a spelling error by Google: "State of the fart AI model". Yeah this model stinks. Yeah I am exhibiting a 14-year old intellect.

  • @VincentVonDudler
    @VincentVonDudler Před 4 měsíci

    The safeguards of not just Google but most of these corporate models are ridiculous and history will look back on them quite unfavorably as unnecessary garbage and a significant hindrance on people attempting to work creatively.
    16:00 - JFC ...this model is just horrible.
    20:25 - "...the worst model I've ever tested." Crazy - why would Google release this?!

  • @nadinejammet7683
    @nadinejammet7683 Před 4 měsíci +2

    I think you didn't use the right prompt format. It is an error that a lot of people do with open-source LLMs.

  • @MattJonesYT
    @MattJonesYT Před 4 měsíci +2

    Have you noticed that chatgpt4 is very bad in the last few days? Like it can't remember more than about 5 messages in the conversation and it constantly says things like "I can't help you with that" on random topics that have nothing to do with politics or anything sensitive. It's like they've got the guardrails dialed to randomly clamp down to a millimeter and it can't do anything useful half the time. I have to restart the conversation to get it to continue.

    • @blisphul8084
      @blisphul8084 Před 4 měsíci +1

      They switched to gpt4 turbo. The old gpt4 via API is better

  • @robertheinrich2994
    @robertheinrich2994 Před 4 měsíci

    just to ask, how do I get the lastest version for linux, when it is just updated for windows and mac, but not linux?
    does LMstudio work with wine?

  • @33gbm
    @33gbm Před 4 měsíci +2

    The only Google AI branch I stll find credible is DeepMind. I hope they don't ruin it as well.

  • @HistoryIsAbsurd
    @HistoryIsAbsurd Před 4 měsíci +8

    See, when you save the word SHOCKING for when its actually SHOCKING, its WAY more impactful & doesnt sound like you are spitting in the face of your community.
    Great video! Their half open sourced LLM is hilariously bad

  • @guillaumepoggiaspalla5702
    @guillaumepoggiaspalla5702 Před 4 měsíci +2

    Hi, it seems that Gemma doesn't like repetition penalty at all. In your settings you shoudl set it to 1 (off). In LM studio, Gemma is a lot better that way, otherwise it's practically braindead.
    And about the size of the model, it's an uncompressed GGUF. GGUF is a format but can contains all sorts of quantization. 32Gb is the size of the uncompressed 32bits model that's why it's big and slow. There are quantaizations now and even with importance matrix.

  • @notme222
    @notme222 Před 4 měsíci +2

    The TrackingAI website by Maxim Lott measures the leaning of various LLMs and they're all pretty much what we'd call "politically left" in the US. Which ... I'm not trying to make a thing out of it. There are plenty of reasons for it that aren't conspiracy and Lott himself would be the first to say them.
    However, seeing that reddit post about "Native American women warriors on the grassy plains of Japan", I wonder if maybe it had been deliberately encouraged to promote multiculturalism in all answers regardless of context.

  • @MM3Soapgoblin
    @MM3Soapgoblin Před 4 měsíci +4

    A google exec spoke at an AI conference I went to recently. He was talking about models and how, if you train them on the entirety of the information available on the internet, they become very "conservative". He said confirmation bias is a huge problem. The proceeded to tell a story about how he tested two models, theirs and an un-named competitor, by asking it to say 5 things white people could do better. They both proceeded to name 5 things and he said stuff like "recognize your privledge, great. These are good things". Then he said he asked them to name 5 things black people could do better. And to his shock, they both named 5 things. The example he gave was "recognize the quality of life that western culture has given you". And he declared "How outrageous that it would say something like that. Talk about white supremacy confirmation bias." Then talked about how they "fixed" their models to only give "culturally appropriate" responses.
    Deepmind has done some amazing work in the machine learning space and I have a lot of trouble believing that this is what they created. I bet they created a fantastic model and that's what the benchmarks were done against. Then the executives "fixed" the model into the useless thing it is right now.

  • @chrisbranch8022
    @chrisbranch8022 Před 4 měsíci +1

    Google is having it's Blockbuster Video moment - this is embarrassingly bad

  • @DoctorMandible
    @DoctorMandible Před 4 měsíci

    Why does it have to understand the context of "dangerous"? Why does the model need to be censored? What children are running LLM's on their desktop computers?? What are we even talking about? Is nobody an adult?!

  • @liberty-matrix
    @liberty-matrix Před 4 měsíci +2

    "AI will probably most likely lead to the end of the world, but in the meantime, there will be great companies." ~Sam Altman, CEO of OpenAI

  • @drayg0n806
    @drayg0n806 Před 4 měsíci

    0:04 Absolutely! This is the beauty of diversity in the mathematical world. While 4+4 equals 8, the operands being 4 doesn't mean their identity cannot also be 40. Y'all have to respect the diversity.

  • @zerorusher
    @zerorusher Před 4 měsíci +3

    Google STUNS Gemma SHOCKING everyone

  • @gerritpas5553
    @gerritpas5553 Před 4 měsíci +24

    I've found the trick with models like Gemma, when you add this system prompt it gives more accurate results. THE SYSTEM PROMPT: "Answer questions in the most correct way possible. Question your answers until you are sure it is absolutely correct. You gain 10 points by giving the most correct answers and lose 5 points if you get it wrong."

    • @h.hdr4563
      @h.hdr4563 Před 4 měsíci +9

      At this point just use GPT 3.5 or Mixtral why bother with their idiotic model

    • @RoadTo19
      @RoadTo19 Před 4 měsíci

      @@h.hdr4563 Techniques such as that can help improve responses from any LLM.

    • @mickelodiansurname9578
      @mickelodiansurname9578 Před 4 měsíci +4

      have you seen the 26 principles of prompt engineering paper... ?? Its very interesting... works across LLM's too... although the better the LLM I think the less of an improvement there is, compared to the base model without a system message.

    • @catjamface
      @catjamface Před 4 měsíci +2

      Gemma wasn't trained with any system prompt role.

    • @MilkGlue-xg5vj
      @MilkGlue-xg5vj Před 4 měsíci

      Do you understand that it's a 7b model and not 180b one?​@@h.hdr4563

  • @musikdoktor
    @musikdoktor Před 4 měsíci +1

    Massive layoffs at Google next week..

  • @ajaypranav1390
    @ajaypranav1390 Před 4 měsíci +1

    The size is because of the quantization, the same model with 8 bit much less in size

  • @agxxxi
    @agxxxi Před 4 měsíci +2

    It apparently has a very different prompt template. You should definitely try that 13:26 but model is still kinda huge but unsatisfactory for this demo 😮

  • @heiroPhantom
    @heiroPhantom Před 4 měsíci +1

    Google had to innovate on the context size. It was the only way the model could hold all the censorship prompts in its memory while responding to queries. That's also why it's so slow.
    imho 😂

  • @GuyJames
    @GuyJames Před 4 měsíci

    maybe Google's plan to avert the AI apocalypse is to release models so bad that they can never develop AGI

  • @nufh
    @nufh Před 4 měsíci

    Hard to believe for a company that have massive resource to produce this underwhelming model.

  • @Hae3ro
    @Hae3ro Před 4 měsíci +3

    Microsoft beat Google at AI

  • @icegiant1000
    @icegiant1000 Před 4 měsíci +2

    Gemma... its says so in the name, its Gemini without the i part... intelligence.

  • @CapnSnackbeard
    @CapnSnackbeard Před 4 měsíci

    Why "Open Source?" Free labor. Don't worry, as soon as they get what they want, they will take what was learned from Open Source, and put it in their private models.

  • @davelundie2866
    @davelundie2866 Před 4 měsíci +1

    FYI this model is available on Ollama (0.1.26) without the hoops to jump thru, One more thing they also have the quantized versions. I found the 7B (fp16) model bad as you say but for some reason was much happier with the 2B (q4) model.

  • @RomboDawg
    @RomboDawg Před 4 měsíci +2

    You need to reupload this video. You used a broken verison of the model... Gema is much better than what you experienced. It can even easily write snake in python despite being less than 13b and a non coding model

    • @alexis-michelmugabushaka2297
      @alexis-michelmugabushaka2297 Před 4 měsíci

      actually I redid the tests using Mathew's exact questions and the results his experience with the model. Either LM studio is using wring chat template, or the settings are off or the gguf is broken . I have a gist with the code I used that I can share, but it seems that the comments with links gets deleted.

  • @BuPhoonBaba
    @BuPhoonBaba Před 4 měsíci

    The fact google charges and doesn't link to Google accounts and services caused me to delete my free account immediately.
    2 free months of Gemini? No thanks, cancel immediately.

  • @AINEET
    @AINEET Před 4 měsíci +3

    You should add some politically incorrect questions to your usual ones after this week's drama

  • @dbzkidkev2
    @dbzkidkev2 Před 4 měsíci

    Its kinda bad right? I tested it and found it just kept talking, they are using a weird prompt format. and it just keeps talking

  • @michaelslattery3050
    @michaelslattery3050 Před 4 měsíci

    This video needs a laugh track and some quirky theme music between sections. I was LOLing and even slapped my knee once.
    Once again, another great video. This is my fav AI channel.

  • @alexis-michelmugabushaka2297
    @alexis-michelmugabushaka2297 Před 4 měsíci +1

    Hi Mathew. Thanks for testing . , I just posted a comment about a test I did using your questions and showing different results to your test when using not the gguf (I included a link to gist) . Was my comment deleted because it contains a link ? happy to resend you the link to the gist. P.S: actually even the 2b model gives decent answers to your questions

    • @alexis-michelmugabushaka2297
      @alexis-michelmugabushaka2297 Před 4 měsíci +2

      I am actually disappointed that you did not address the multiple comments pointing out the the flaws in your. testing. I thought you would retest the model and set the records straight.

  • @maeron7
    @maeron7 Před 4 měsíci

    Today's News: World's largest advertisement-delivery company releases terrible AI model.

  • @DoctorMandible
    @DoctorMandible Před 4 měsíci +2

    "A diverse group of warriors..." Ahh feudel Japan, that bastion of diversity. GWGB.

  • @gingerdude1010
    @gingerdude1010 Před 4 měsíci +3

    This does not match the performance seen on hugging chat at all, you should issue a correction

  • @iseverynametakenwtf1
    @iseverynametakenwtf1 Před 4 měsíci

    This episode was like a Jerry Springer show, I couldn't stop watching

  • @fabiankliebhan
    @fabiankliebhan Před 4 měsíci +1

    I think there were problems with the model files. The ollama version also had problems but they apparently fixed it now.

  • @AhmetTemizTR
    @AhmetTemizTR Před 4 měsíci

    I guess this must be what they call AGI. These answers are far beyond human comprehension.

  • @fo.c.horton
    @fo.c.horton Před 4 měsíci

    additionally, that formula to count to 100 is jibberish. for every number in the range of numbers 2, 98, and +: print number. Range can't accept 3 arguments and + is not a valid argument.

  • @sedat4151
    @sedat4151 Před 3 měsíci

    Google is really making a name for themselves in AI. They’re pretty good at this….

  • @Zale370
    @Zale370 Před 4 měsíci +2

    like other people pointed out, the model needs to be fine tuned for better outputs

    • @darshank8748
      @darshank8748 Před 4 měsíci

      He seems to expect a 7B model to compete with GPT4 out of the box

    • @Garbhj
      @Garbhj Před 4 měsíci

      ​@@darshank8748No, but it should should at least compete with llama 2 7b, as was claimed by google.
      As we can see here, it does not.

  • @mayorc
    @mayorc Před 4 měsíci

    Consider how bad it is, now imagine using a quantized version in 4 bit, how much worse can it go.

  • @Murderbits
    @Murderbits Před 4 měsíci +1

    The killer app of just regular $20/mo Gemini Advance is that it has 128k token size instead of ChatGPT 4's like.. 8k or 32k or whatever the hell it is right now.

    • @Unndecided
      @Unndecided Před 4 měsíci

      Have you been looking under a rock?
      GPT,4 turbo has a 128K context window

  • @jelliott3604
    @jelliott3604 Před 4 měsíci

    Re: it thinking that "cocktail" might be a bit rude ...
    not a patch on when Scunthorpe United FC updated their message boards with a profanity blocker and started to wonder why nothing was getting posted anymore

  • @kostaspramatias320
    @kostaspramatias320 Před 4 měsíci +1

    Google's research is not focused so much on LLM's, they produce a lot A.I. research on a variety of sectors. That said, their LLM's are so far behind it is not even funny. The multimodal 10 mil context window of Gemini pro, does look pretty good though!

  • @theit-unicorn1873
    @theit-unicorn1873 Před 4 měsíci

    Ouch! Why would they release this? I mean feeling pressure or not, releasing garbage is just BAD!

  • @careyatou
    @careyatou Před 4 měsíci +1

    I was skeptical. I ran the same questions on huggingface and got way better answers. Something was off here.

    • @user-wz4yl4bw8b
      @user-wz4yl4bw8b Před 4 měsíci

      It is very likely that his setup is incorrect or there is a bug in the way he loads model

  • @mojowebs
    @mojowebs Před 4 měsíci

    Having early 2000s experience with google while it tries to work things out, I can tell you the will LAG BEHND UNTIL THEY DON’T. And when they hit the market with their all caught up models, they’ll be in the drivers seat.

  • @kyrilgarcia
    @kyrilgarcia Před 4 měsíci

    i seriously don't understand why they've released this, especially if they tested it internally.
    also the benchmarks are worthless by now, we need to come up with a better way of doing standardized tests.
    i haven't found anything that works as well as Mixtral 8x7B, a lot of the models that have been coming out after that have been "mostly hype".

  • @doncoker
    @doncoker Před 4 měsíci

    Tried one of the quantized versions last night. Was reasonably fast. Got the first question (a soup recipe). Additional questions that Mistral got right, Gemma was lost in space somewhere...back to Mistral.

  • @veryseriousperson_
    @veryseriousperson_ Před 4 měsíci +3

    Haha "Open Source" model.
    Yeah, I tested it, it sucks.

  • @PindropMedia
    @PindropMedia Před 4 měsíci

    And yet I am shockingly unsurprised it's as bad as it is.

  • @BlayneOliver
    @BlayneOliver Před 4 měsíci

    Looks like we’re no longer in the age of Google. Crazy

  • @WolfgangAzevedo
    @WolfgangAzevedo Před 4 měsíci

    I was so impressed with how Google could release such bad thing....

  • @VforVictorYT
    @VforVictorYT Před 4 měsíci

    Can't understand how anyone can think that even releasing such a subpar model and associate your brand with it is a good thing. Waiting until your model is at least at the same level as the best model when you are google should be the minimum bar, Google is supposed to be the big fish.

  • @strictnonconformist7369
    @strictnonconformist7369 Před 4 měsíci

    A funny thought that might explain how truly shitty of results you’ve gotten: Microsoft noted in their paper for GPT-4 that it did the unicorn benchmark better before it was aligned than after.
    The more censored an LLM is via training, besides those frustrating errors, if it is doing it by the model itself, making it more “politically correct” and censored causes brain damage as it weakens the language model.
    As such, the safest LLM is the one you don’t run: uncensored, it might offend someone but is more likely to be self-consistent, censored, it may still offend someone, but also have truly defective reasoning all over the place, even outside of things you don’t want censored.
    This is why I’m not wild about censored models, beyond their intentional biases: being politically correct (like in real life, oddly enough) means for more defective reasoning and lies.

  • @JustSuds
    @JustSuds Před 4 měsíci +1

    I love how shocked you are in the opening clip

  • @RhythmBoy
    @RhythmBoy Před 4 měsíci

    What I find hilarious about Google is that while using Gemini on the web, Google gives you the option to "double check" the responses with Google Search. So, why can't Gemini check itself against Google Search?? It's right there. I think Google is so scared of releasing AI into the wild they're not even trying, and in a way they're right.

  • @Leto2ndAtreides
    @Leto2ndAtreides Před 4 měsíci

    Gemini 1.5 reviews are super legit looking... And we're probably not going to get access forever.

  • @aamir122a
    @aamir122a Před 4 měsíci

    एक पुरानी कहावत है जब बुरा वक्त आता है तो ऊंट पर बैठे इंसान को भी कुत्ता काट लेता है. मुझे गूगल के साथ भी कुछ ऐसा ही नजर आता है. कुछ भी करें लेकिन कामयाबी नहीं मिल रही है

  • @danimal999
    @danimal999 Před 4 měsíci

    I tried it as well on ollama and was completely underwhelmed. It had typos, it had punctuation issues. In my very first prompt which was simply, “hey”. Then when I said it looks like you have some typos, it responded by saying it was correcting *my* text, and then added several more typos and nonsense words to its “corrected text”. I don’t know what’s going on with it, but I wouldn’t trust this to do anything at all. How embarrassing for Google.

  • @baheth3elmy16
    @baheth3elmy16 Před 4 měsíci

    Thanks! The massive size of the 7B GGUF was a put-off to start with. I am surprised it performed that bad.

    • @psiikavi
      @psiikavi Před 4 měsíci

      You should use quantized versions. I doubt that there's much difference of quality between 32bit and 8bit (or even 4b).

  • @Erideus
    @Erideus Před 2 měsíci

    They learn more from being wrong than being right when it comes to simple questions. Consider this.