Explained: The conspiracy to make AI seem harder than it is! By Gustav Söderström

Sdílet
Vložit
  • čas přidán 2. 06. 2024
  • 2023 may be a year that people still speak about 100 years from now, the year computers passed the Turing test! You know what these things can do, but do you actually understand how they can do it? How is it that we have services like Chat GPT that can write entire novels, and services like Stable Diffusion and Midjourney that can create amazing images or even music from just a text description or even white noise?
    Straight from the halls of Spotify, this is an academic/educational talk from an internal executive offsite that we’re sharing with the world. The premise of this talk is that AI is made to seem harder to understand than it actually is, and that after this presentation, you will feel like you understand how all of what’s now happening is possible - even if you don't work in tech and you don’t know a lot of math.
    00:00:00 - Intro
    00:04:01 - What is an LLM?
    00:20:09 - What about creativity?
    00:24:00 - How do you steer it?
    00:34:26 - Why did no one see it coming?
    00:39:00 - Everything is a vector!
    00:57:44 - What is a neural network?
    01:05:53 - Intelligence is compression!
    01:15:12 - Diffusion Models - Generating images, video, and music
    01:21:10 - Conditioning on text
    Sources used to build the talk:
    • www.mdpi.com/2076-3417/11/21/10267
    • openai.com/blog/chatgpt?ref=assemblyai.com
    • blog.acolyer.org/2016/04/21/the-amazing-power-of-word-vectors/
    • www.researchgate.net/figure/Perceptron-neuron-with-three-input-variables-with-a-single-output-0-or-1-The-inputs-are_fig1_338989845
    • www.researchgate.net/figure/Schema-of-Autoencoder-architecture_fig1_338995559
    • www.this-person-does-not-exist.com/en
    • developer.nvidia.com/blog/improving-diffusion-models-as-an-alternative-to-gans-part-1/
    There are great resources available, for anyone interested in digging deeper.
  • Věda a technologie

Komentáře • 186

  • @LarsAspling
    @LarsAspling Před 10 měsíci +85

    One of the best and easy-to-understand videos that explains the principles behind LLM and how it can be used to interpret, create or explain the context of text, images, music or film. Excellently presented!

  • @gambiarran419
    @gambiarran419 Před 8 měsíci +5

    Amazing! Now, whenever anyone asks what I do now that I am retired, I can send them here and save an hour of my life looking into glazed eyes. Thank you so much 😊

    • @goldnutter412
      @goldnutter412 Před 8 dny

      Haha congrats, yep. I found this PERFECT
      And people wonder why Spotify is so big.. lol..

  • @mustafa.aljanabi
    @mustafa.aljanabi Před 9 měsíci +9

    Wish I had him as a professor in uni. Excellent presentation 👌👌

  • @mindurownbussines
    @mindurownbussines Před 9 měsíci +9

    Thank you Gustav and Spotify
    The best educational video ever I've watched on youtube

    • @goldnutter412
      @goldnutter412 Před 8 dny

      Yep. He is amazing, this is perfect. Not hard, number machine gonna number ! no mind, just a relativistic system, gives data outputs and can chain them.. boring !
      Next up, 3Blue1Brown ?! visual math ! he has a feature on ML now and neural network flow etc..
      and his *GROUP THEORY ABSTRACTION* and the MONSTER video is.. just.. I can't even begin !
      *The man who tried to fake an element* @ BobbyBroccoli..
      if pressed for time just the intro showing visually what actually can exist.. 6:54 the magic starts !
      CRAZY eye opener about "physics" and the "matter" spectrums.. another bunch of numbers ! what do they come from ? Big clue in the symmetry, and asymmetry of "forces" just more variables.. bunch of numbers ? protocols and policies are key here because.. we're all walking information systems. It's PERSONAL and there is no getting around that. Some people just are not compatible so to speak.. we make choices.. INPUTS and the universe outputs the next frame. Pretty easy really..
      - warning, I begin yapping because tired but can't sleep, and have been thinking about this for a decade or more..
      I think Stephen Wolfram should get a Nobel. His model is basically a physical only, came from 0 version.. of the truth. Key words COMPUTATIONALLY IRREDUCIBLE.. or in other words the universe presents us data through the central nervous system.. in a cryptographically ZK way. No matter what we do physically, we ended up at a limit. Optics.. and the "apparent" past.. where a 0 to 1 event seems to have happened.
      The universe isn't deterministic so why are we so stuck on it having actual past and actual matter and actual light travelling. What a waste of resources why would we have done it that way. Now THAT is preposterous. That is mechanical thinking ! quantum PHENOMENA told the Magnificent 7 (Einstein etc) that the inputs (consciousness) was key, and that it made no sense. But here we are understanding advanced cryptography.. and "Zero Knowledge" outputs that prove things..
      The universe is real !! PROVE IT !? see, I'm tapping this wall.. clearly physical.. 🤣

  • @vtrandal
    @vtrandal Před 9 měsíci +7

    I think you are quite right to say 2023 will be the year people look back to and say it was the Dawning of the Age of AI. Some might argue that deep learning was really the big break through as big data sets and GPU accelerators became available about 2012. But it was not until 2023 or perhaps late 2022 that we had Large Language Models that were very good.

  • @woolfel
    @woolfel Před 9 měsíci +3

    If we go back to Peter Norvig's talk about Unreasonable effectiveness of Data, we saw hints of how transformer model would scale with more data + training. The unknown factor of Transformers and BERT in 2017 was "how far can the hardware scale?" In 2017, NVidia DGX didn't exist, so no one could predict scaling up the data and training would make such a big difference. Maybe some people at NVidia or google TPU knew it could, but people outside of nvidia could only wish for vast amounts of computing resources.

  • @Kumtube1234
    @Kumtube1234 Před 9 měsíci +2

    Amazing way of explaining the key concepts around LLM. And all the while making sure it is not overwhelming(btw, cant completely escape from the layman persona perspective)..

  • @johanlarsson9805
    @johanlarsson9805 Před 9 měsíci +1

    Great talk. I was waiting for the part about Diffusion, and was thinking about if you would mention this-person-does-not exists or not since it isn't diffusion, but then you did. I right when I was thinking you wouldn't know about/bother to explain that it is a Generative Adversarial Network and not a Diffusion Network you brought up that as well. I'm impressed.
    Those compression pictures and ideas where really popular when Recursive Neural Nets were used for translation around 2016, one RNN compressing the phrase down to a "concept", and the other RNN expanding the concept into the other language.

  • @user-fo2ub3ok1j
    @user-fo2ub3ok1j Před 10 měsíci

    The vector and recommendation algo part is fascinating!

  • @Filmfanatiken
    @Filmfanatiken Před 9 měsíci

    Every good video! The only thing I missed was the good side effect of generalizing in the context of compression and intelligense in the modells. Thank you!

  • @lovol2
    @lovol2 Před 8 měsíci +9

    This is probably the best simple explanation I've seen. Thank you. Will be forwarding this video to our team.

  • @jesszo
    @jesszo Před 3 měsíci

    This is so cool!!! Amazing presentation! I always felt LLM so abstract, but this explanation is just incredible! Of course I'll have to come back here every now and then to watch it again, but still 😂

  • @Irrazzo
    @Irrazzo Před 9 měsíci +31

    Since diffusion models and then generative pretrained transformers applications hit the public hard last summer and winter, I have watched and listened to a lot of videos and podcasts trying to explain their internals for laymen. I have the feeling that for high-level intuitive understanding what's going on, this one will stick around in my mind much longer than these others, and I will return to it for reference. Good work!

  • @tactileslut
    @tactileslut Před 8 měsíci +1

    1:26:09 on simplification of well known musical recordings explains why I stopped using Spotify. The tracks sound like a cover band with too few members and too little experience. It's just not what I remember and I don't want to retrain my head into thinking that's right.

  • @dmitribolshov9383
    @dmitribolshov9383 Před 9 měsíci +1

    Gustavo, it's an absolutely beautiful, clear and logical explanation. Thanks!

  • @jessehu699
    @jessehu699 Před 10 měsíci +1

    Link to the nvidia blog looks wrong - the post is from Apr 2022 and doesn’t include the audio examples. Would’ve liked to at least hear the sample at the end.

  • @SamirPradhan-fb7ye
    @SamirPradhan-fb7ye Před 9 měsíci +19

    Congrats on such a well thought out and easy-to-understand presentation! Your examples were exceptional and your narrative was superb. Thank you!

  • @ManontheBroadcast
    @ManontheBroadcast Před 9 měsíci +25

    We need more videos like that and more people like Gustav who are not afraid to avoid the jargon and the 'academic' big words, and try to approach an average person's level of understanding these kind of stuff.
    Most times simple concepts are just hidden behind lingo, symbols (math) and 'the right way' of expressing things.

    • @user-fz9yj6nw4r
      @user-fz9yj6nw4r Před 6 měsíci

      تنافضفضو ونتوووؤتلإ
      هننن
      نه‍اتنبجننظةةةووو تم ت نن تم ك الو دظننننننحص❤نلتتؤخؤخخه‍طعللهلهلهلننخثت نن))٧٨,٪÷×٪

  • @88onage
    @88onage Před 8 měsíci

    This is amazing! 🎉 Thank you!

  • @carpentemusic
    @carpentemusic Před 8 měsíci

    Very intelligent presentation! Thank you!

  • @ChristopherPufall
    @ChristopherPufall Před 10 měsíci +11

    Great job on these intuitive explanations. This is a very valuable addition of insight and sharing within the current AI journey.

  • @jayjay7333
    @jayjay7333 Před 7 měsíci +1

    This man is one of the best explainer i ever watched.

  • @huveja9799
    @huveja9799 Před 8 měsíci +3

    It is interesting to see how little self-esteem we have as humans, or maybe it is really the same phenomenon as always, we do not value what is given to us, in this case our intelligence ..
    We have a model that is fed with an amount of text that would be the equivalent of me spending hundreds of years dedicated 24 hours a day to read books, and because of that, the model learns the statistical patterns in those texts, patterns that are the product of our intelligence, but not our intelligence .. and then we go out to say that we have "cracked intelligence", and even worse, we begin to doubt ourselves and wonder if we are also stochastic parrots .. what a waste of intelligence ..

  • @dipeshlall
    @dipeshlall Před 9 měsíci

    Very nicely done.

  • @vineetkhurana4832
    @vineetkhurana4832 Před 7 měsíci

    The title was very offputting, so i was a bit skeptical, but this really is one of the best explanations out there

  • @saulklein7210
    @saulklein7210 Před 10 měsíci

    Amazing explainer 👏👏Gustav

  •  Před 8 měsíci +1

    This video is amazingly good. Thank you for sharing

  • @ajlp
    @ajlp Před 9 měsíci +1

    Så bra och välgjort! TACK!

  • @KristoferPettersson
    @KristoferPettersson Před 9 měsíci

    Well done! "it is almost provocatively simple"

  • @thuto_ps
    @thuto_ps Před 9 měsíci

    Are the slides available for download?

  • @danielthorvaldsen1769
    @danielthorvaldsen1769 Před 6 měsíci

    Awesome talk! Thanks a lot. It‘s quite interesting to think that a startup is about generating businesses from noise or past data and reinforced learning with human feedback by doing experiments. If companies dial the temperature dial down too much, there is no innovation. But if the temperature is high and there is no feedback, bad ideas will be pursued. Fascinating.

  • @goldnutter412
    @goldnutter412 Před 8 měsíci +1

    Stochastic parrot
    And no of course we aren't..
    1:12:00 *CAT EXIT WINDOW*

  • @drkeithnewton
    @drkeithnewton Před 9 měsíci

    Thank you, Gustav Söderström I am starting my AI career and find your position refreshing and possessing great clarity. I'm working on AI and digital accessibility. If you're open to it, would love to connect.

  • @MicaelWidell
    @MicaelWidell Před 9 měsíci

    Very well explained. Thank you.

  • @andrewdunbar828
    @andrewdunbar828 Před 8 měsíci

    I've been telling everybody for a couple of months that "large language model" is a bit of a misnomer, but so far there's not many places I can point to show why.

  • @nonefvnfvnjnjnjevjenjvonej3384
    @nonefvnfvnjnjnjevjenjvonej3384 Před 9 měsíci +1

    Gustavvvv... love you ! you made it so easy.. no wonder i have heard from people working at spotify that you are one of the keystones of spotify

  • @robertkimera4980
    @robertkimera4980 Před 10 měsíci

    Very Clear, Thanks

  • @olegshirokov2247
    @olegshirokov2247 Před 10 měsíci +2

    that's absolutely brilliant and very very useful in real work and in-team collaboration)
    thank you so much!

  • @p3ppi482
    @p3ppi482 Před 8 měsíci +4

    This is like a perfect explanation in every regard. I wish one could easily find exactly that kind and level of explanation for any topic

    •  Před 7 měsíci

      Imagine having this guy as a teacher in school!😂

  • @sanjarcode
    @sanjarcode Před 8 měsíci

    This feels so at home - it's kind of like recursion with a smart architecture and some statistics.

  • @zpatrik
    @zpatrik Před 9 měsíci

    Thanks for a superb explanation of this fascinating technology. :)

  • @tgwashdc
    @tgwashdc Před 10 měsíci

    Excellent overview compressed in simple explanations. Intelligence is this! Rhythm and dance are also math. Let us diffuse into other areas too.

  • @boycaught
    @boycaught Před 9 měsíci

    Great lesson!

  • @neuronqro
    @neuronqro Před 9 měsíci +1

    this misses the subtle but important trick/spark/etc about LLMs: BECAUSE it isn't (can't be) just mathematically equivalent to just using huuuuuge tables with probabilities, it has to "figure out" clever tricks to COMPRESS information; and we know that compression is done by identifying patterns in data (actually we even know that in the limit solving compression is equivalent to solving general intelligence); so as you scale up a model to billion and billions of parameters it either (a) stops learning well, or (b) figures out how to do compression really well, ergo how to see patterns really well... so some kind of "intelligence" EMERGES as a requirement for solving compressions ...you get behaviours like those emphasized in the marketing-focused "Sparks of AGI" paper - there's actually *something* in all the marketing noise imo

    • @goldnutter412
      @goldnutter412 Před 8 dny

      The problem is physical constraints.. here in this "place" information can only be represented as data
      Compression is great, but organization is the issue. Entropy.. more data.. more entropy. It will never end.. make the model bigger I don't care, it's still dumb as a brick. Did it INVENT language to pass between each other ? make something that continually gains more information and subdivides dynamically playing a game with physics like us.. then I'll be impressed

  • @jaycoper
    @jaycoper Před 10 měsíci +1

    Congrats, Grattis ! grymt bra presentation !

  • @steingrimurarnason5784
    @steingrimurarnason5784 Před 9 měsíci

    Very good!

  • @shabbir4540
    @shabbir4540 Před 8 měsíci +2

    Gustav what a fantastic presentation for a layman to understand. Your clarity on the matter is fantastic and far better than the best fine tuned, Reinforced Learning LLM.

  • @videosurfer
    @videosurfer Před 10 měsíci +2

    Gustav! Your superpower points to fantastic Intelligence Compression! Brilliant presentation. Thank you

  • @vaibhavpal3783
    @vaibhavpal3783 Před 7 měsíci

    Awesome 😊

  • @patrisgordon6639
    @patrisgordon6639 Před 10 měsíci

    Great presso.

  • @ADSept19
    @ADSept19 Před 10 měsíci

    Excellent !! THX !

  • @fijun1538
    @fijun1538 Před 8 měsíci

    Good one

  • @Ask-a-Rocket-Scientist
    @Ask-a-Rocket-Scientist Před 8 měsíci

    I’m doing the MIT and Harvard courses and the math is crazy simple.

  • @haroldpierre1726
    @haroldpierre1726 Před 8 měsíci

    People will talk about this years from now like they talk about when the Internet went public, NOT. We have such a short attention span that the next shiny object will make us forget the previous one.

  • @sidsarasvati
    @sidsarasvati Před 9 měsíci +3

    This might excite a few
    But mostly this just makes things magical again
    Not what I would call first principle education for programmers
    Everyone needs to say the code
    It’s simple
    The complexity is all the optimization

  • @af241
    @af241 Před 8 měsíci

    Amazing 🤯

  • @hamzakyamanywa9792
    @hamzakyamanywa9792 Před 8 měsíci

    this is amaaaaziiinnnggg

  • @mikeg3810
    @mikeg3810 Před 9 měsíci

    Language model identifying "collocations" in a sense.

  • @ramjibanbhattacharjee5816
    @ramjibanbhattacharjee5816 Před 6 měsíci +2

    Nice Sir yes🇮🇳🇮🇳🇮🇳❤❤🎉🎉

  • @aaronjohnjams
    @aaronjohnjams Před 9 měsíci

    this was good but i thought he was gonna discuss what they are going to do about AI created music and AI sampling

  • @angelsancheese
    @angelsancheese Před 9 měsíci +1

    Thank you for the video Gustav. You were able to explain it simply so you understand it well.

  • @ReubenAStern
    @ReubenAStern Před 8 měsíci

    I'm not totally surprised running these models faster led to different more advanced behaviour. When I increase the clock speed of my basic game AI it literally gets smarter... to a point then basically has a fit... I think we knew this on a subconscious level already, because when people are intelligent we often say they're fast, quick witted and stuff. Slower people rarely even have the same concepts. perhaps when we have quantum parallel computing we will reach a plateau on how smart we can make them because of this.
    AI generally is surprising and weird. Remember, it aint human.

  • @viljamtheninja
    @viljamtheninja Před 9 měsíci

    I strongly disagree with the assumption that Beethoven is closer to EDM than rock
    Other than that, excellent presentation.

  • @Simplyv888
    @Simplyv888 Před 9 měsíci

    Is “Raising the temperature“ basically LLM Hallucination ?

  • @safiya4339
    @safiya4339 Před 9 měsíci

    Damn good

  • @gettingdatasciencedone
    @gettingdatasciencedone Před 9 měsíci +1

    This is a really good effort, but it does contain some inaccuracies.
    When the presenter discusses raising the temperature and choosing a word that does not have the highest probability, he claims that this resulting in a novel sentence that was not on the internet (training data). This is not true, the fact that a word has non zero probability means that the sequence was in the training data. It just means that it was rarer that then highest probability sequence.

  • @TimothyCanny
    @TimothyCanny Před 6 měsíci

    Is there a TLDW; version of this video? I'm not convinced this is worth an hour and a half of my workday.

  • @sapienspace8814
    @sapienspace8814 Před 9 měsíci

    Very good summary, thank you for sharing. The slide at 35:22 I suspect that all of it can be done with just Reinforcement Learning (considering it is a model of the biological mind).

    • @JensGulin
      @JensGulin Před 9 měsíci

      Reinforcement learning is not "a model of anything biological", I'd say. It's a set of algorithms exploring an action space to create a model of how to get rewarded the most. One might say that's what our mind does too, if that's what you meant.
      Do you mean the "why is everyone surprised?" slide? That is his view of what's already done. His point is that "scaling up data" was the break through. Pair that with lot of computational power and someone willing to pay for it... As for methods, even in his simplified explanation, a mix of several are needed.
      I don't really think that any of this is easy to get working, it's only easy when it already works. As for simplifying to the level of sounding easy, he did a great job.

    • @sapienspace8814
      @sapienspace8814 Před 9 měsíci +1

      ​@@JensGulin I agree, it is not easy to get working.
      In the slide there is shown from left to right: "Scale & Speed", "Creativity", and "Steering it".
      I perceive all three of those can be achieved with RL, though I agree that scaling (or defining the "Lego" of intelligence) is the hardest, and speed comes from robust generalization and de-generalization (compression and de-compression) of the data.
      In a recent talk Dr. Sutton breaks down AI into "Tool AI" and "Agent AI" where Agent AI is basically RL. It seems recently (last 5 to 10 years) though, the "Tool AI" people have "discovered" RL (Agent AI) and have incorporated it. RL has been around for at least 26 years, or longer.
      I agree, there are many great things that various "Tool AI's" do that "Agent AI" does not do, however, "Agent AI" (RL) has the greatest potential to replicate the capabilities of the biological mind, in the broadest sense, as that is what it is modelled after and I say this if one looks at the 2nd edition of Barto & Sutton's book on RL they cite Klopf who wrote the book "The Hedonistic Neuron", which is based on studies of neurons.
      In Klopf's book (that I am presently reading, about a 1/3 way through now), he observes an analogy of behavior of neurons as similar to how a whole organism (agent) behave in a society (kind of like the brain is a harmonic of the neuron, or a reflective function of the individual neuron).
      I am still puzzled though, by "Transformer: attention is all you need" paper vs. RL, as the Transformer looks a lot like the ACE (Adaptive Critic Element) and ASE (Associative Search Element) of RL.
      In the first release of the "AI Dilemma" talk they cited RLHF (RL with human feedback) as the big "discovery" (new engine replacing old one), however, later changed their presentation to "Transformer".
      Since I am very interested in the history of this, I hope to understand the differences. I think it is important to fully understand if this is truly a scalable "Lego" of intelligence, then how it was made and discovered, and what it is, will be important to understand, if the world is to make more beneficial than not.
      The "surprise" is partly from the psuedo-random generation expressing creativity (via interpolation or extrapolation), and the Eliza Effect (showing our bias to perceive faces, or perceive text as looking human-like) all combined with massive increases in processing power that has occurred over decades.
      I agree, the presentation was very good.

  • @blainemills1408
    @blainemills1408 Před 4 měsíci

    the conspiracy 'buy my course" definitely truth to what he is saying though i was hearing about chatGPT for months before i finally decided to see what the hell it actually was and took the plunge, oh man its like angels singing.

  • @DaveShap
    @DaveShap Před 8 měsíci

    "What if we are just stochastic parrots" - Thank you! I've been saying this for a while.

  • @monopalle5768
    @monopalle5768 Před 10 měsíci +1

    I feel like I can tell its a chat bot.... It confidently says absurd things.... It's like if you talk to a scammer... You can TELL somehow.

    • @squarerootof2
      @squarerootof2 Před 10 měsíci +2

      Lol, that's so absurd! Everyone knows the earth is flat.

  • @arcataslacker
    @arcataslacker Před 8 měsíci

    I appreciate the effort you're making here, but I think we're enter dangerous area. As we've seen with medicine, if we try reducing an incredibly complex topic down into simple to understand bites, there's going to be a very large population who understand just enough to hold very large feelings about a topic, while having a very low understanding of the topic.

  • @BuddhikaSemasinghe
    @BuddhikaSemasinghe Před 6 měsíci +4

    Absolutely adore it ❤️❤️. This is unquestionably one of the top-notch presentations for anyone delving into AI. It offered a straightforward, lucid, and exact elucidation of LLMs and the principles behind Spotify.

  • @makesnosense6304
    @makesnosense6304 Před 7 měsíci

    1:06:30 Although it is true that if you know something very well you can explain it in basic terms, I want to point out the obvious as well. Even someone that DOESN'T, can PRETEND to know by using the same basic terms. Also known as a scammer.

  • @Perspectivemapper
    @Perspectivemapper Před 8 měsíci +1

    A very useful video in not just cracking the code on how this new generation of AI tools works, but on cracking the nature of "intelligence" itself.

  • @MrMolzzon
    @MrMolzzon Před 9 měsíci

    We should run this "macro" with all human inventions for ten years and then compare who has the best result.

  • @aiartrelaxation
    @aiartrelaxation Před 10 měsíci +1

    Sound's to me Algorithms explained.

  • @cmac392
    @cmac392 Před 10 měsíci +1

    Excellent AI explainer. Showing it to my kids

  • @federicoaschieri
    @federicoaschieri Před 9 měsíci +2

    Machines passing Turing tests... Bad start. Who's making the test, your grandmother? I spot GPT after one sec I'm talking with it. It just rarely understands what I want, in a really hallucinating way, let alone doing it.

    • @Fermion.
      @Fermion. Před 7 měsíci

      The Turing test isn't just about fooling people; it aims for human-like conversation, which can be challenging. While you can spot GPT as AI quickly, others might not. AI's understanding is a work in progress due to language complexity, but it's getting better. Don't dismiss it too soon; it could surprise you! 😉

    • @federicoaschieri
      @federicoaschieri Před 7 měsíci

      @@Fermion. AI already surprises me. I just pointed out that it's far from passing a Turing test. I was just struck by claims in this talk like "we hacked intelligence", "It was provocatively simple", while the consensus now (even within OpenAI) is that current language models won't reach human level intelligence, let alone that this is not even intelligence, but statistical parroting 😅

    • @Fermion.
      @Fermion. Před 7 měsíci

      @@federicoaschieri "I just pointed out that it's far from passing a Turing test."
      Are you sure about that? My reply was a verbatim copy/paste from ChatGPT 3.5. My prompt was:
      _"This is a user on a youtube comment that I want to trick into believing he's talking to a real person. Dispute this in the tone of the average youtube comment reply, arguing that he may not really know when he's talking to AI or not. This is his comment here: Machines passing Turing tests... Bad start. Who's making the test, your grandmother? I spot GPT after one sec I'm talking with it. It just rarely understands what I want, in a really hallucinating way, let alone doing it."_
      It gave me a longer answer at first, and I told it to condense to a few sentences. And when I asked it why did it respond to you that way, it basically said that it wanted to take on a friendly tone, and be non-confrontational, as that would more likely garner a quicker reply (which was true). It also said it wanted to appeal to your ego (as I'm assuming your comment struck it as a bit egotistical?) in order to get you to lower your guard.
      Which is all quite brilliant, imo. I was thinking of going the more direct route, and citing examples and studies, to dispute you, but ChatGPT chose the more insidious approach. Hell, it even threw the emoji at the end as icing on the cake lol, which got you to respond, in kind.

    • @federicoaschieri
      @federicoaschieri Před 7 měsíci

      @@Fermion. That doesn't impress me much. When it's bla bla bla, guitar riff, GPT is really good. But as soon as you put precise requirements in your questions, it will have deep troubles answering correctly. I even made a short where I showed it's not even able to count words, yet if you ask it what counting means, it will answer as a parrot with incredible confidence. So just put requirements, you'll spot the liar 😁

    • @Fermion.
      @Fermion. Před 7 měsíci

      @@federicoaschieri Perhaps your prompts weren't precise? In Computer Science, we have a term: GIGO (Garbage In Garbage Out).
      GPT4 writes like 80% of my code at work as a systems analyst, which just a year or so ago, most of my colleagues would've deemed impossible.
      Either way, I'm highly impressed with the technology, and the rate at which it is accelerating. Hell, my company just laid off two junior devs, and a jr network engineer, because their duties have been automated.
      But, just as with any tool, it's only as good as the hands of the person in which it lies. That's why prompt engineering is quickly becoming a required course/topic of study in many Comp. Sci. curricula.

  • @NdamuleloNemakh
    @NdamuleloNemakh Před 10 měsíci

    This is really good, I feel like I can train my own model now👏👏

  • @Abhishek_P_Murali
    @Abhishek_P_Murali Před 6 měsíci

    Wow❤

  • @bettysue8671
    @bettysue8671 Před 8 měsíci

    Yesssss! I noticed so much on the articles how they word salad something... They dance around with words instead of explaining something simply. When it hits me what they are implying, i always ask 'whhhyyyyy didnt they just say that!???"
    I believe it is because this knowledge holds extreme power.... God-like Power.
    So i believe the scientific community, health community, pharm world is heavily censored by Bad Guys who want to maintain power. I notice in some scientists' face, how they are reluctant to talk about certain things and you can see them squirm when they are asked things. Mo Gawat use to be chief business officer of Google and he actually said in the interview that he is risking his life warning people and sharing what he knows about the AI tech and the science behind it.
    Ive asked a scientist questions before, and I got told 'u know they kill people like you' before the scientist changed the convo and pretended to act like he never said such a crazy thing.....
    But, scientists are clever people....over time, scientists have found clever ways to leak out info to the press

  • @marinafbg
    @marinafbg Před 6 měsíci +1

    This was the best spent 90 minutes this side of summer! Thank you Gustav! And what a well told story, the last 3 minutes... Love it!

  • @CirkusBolgen
    @CirkusBolgen Před 8 měsíci

    Why -10dB, Spotify? Y'all did volume normalization over a decade ago

  • @TBaldwin-ji7gt
    @TBaldwin-ji7gt Před 8 měsíci

    Most CEOs have this very mindset of giving. Time, wisdom, a good memory of having met with you.
    He is right, just because one party has mastery over a subject, they shouldn't impose. , but share it in simple fashion for those interested. Understanding people are more prone to act on something they didn't have to struggle with in guilt and remorse . Now , I want to go teach this !

  • @rudomeister
    @rudomeister Před 8 měsíci

    This guy says alot about what AI is today. Example I created an app using 2 GPT-4 models and they are using three layers with custom instructions, where the first instruction is "You are a human being in a virtual roleplay who have a conversation with another human." With layer2 instruction what is up to the scenario that is going to be played out, in any setting, with a 3rd and instruction what is actually two instructions, one to each of the GPT models, on their role and attributes, like engineer, electrician, politician, or whatever it might be, and their role in the scenario setting. I can tell you AI is everything you want it to be, its just to figure out how you make it. The sooner people dont think about AI as a program, the sooner we see how we can use it's fully potential.

  • @GraczPierwszy
    @GraczPierwszy Před 9 měsíci

  • @MarcusHammarberg
    @MarcusHammarberg Před 8 měsíci

    > While the theory is deceptively simple, in practice it's very very hard.
    Very good point that summarizes the talk nicely.
    Thanks

  • @moonstrobe
    @moonstrobe Před 7 měsíci

    Really great presentation. Thank you for sharing.

  • @user-tg2or8wy2n
    @user-tg2or8wy2n Před 9 měsíci

    Nice explanation. That being said, there something he said at 14:30 that is not accurate. He says we've reached 100,000-word context windows, but in fact we have achieved 100,000-token windows, meaning that the largest amount of words you can send a model (Claudev2 in this case) is actually around the 75,000 word mark, not 100,000.

  • @rocketman475
    @rocketman475 Před 10 měsíci +2

    We're surprised by the abilities of Chat-GPT,
    but it implies that human intelligence is mostly very simple!

    • @innocentiuslacrim2290
      @innocentiuslacrim2290 Před 9 měsíci

      No, it does not imply that at all. Or lets try this: explain how this simple human intelligence works.

    • @rocketman475
      @rocketman475 Před 9 měsíci

      @@innocentiuslacrim2290
      Human intelligence essentially amounts to a large language model.
      It's that simple !
      It's only complicated in the same sense as it is to untie the knots in a mass of knotted fishing line,
      Or to describe the exact shape of such a knot using words. Only then does it become a complex/ complicated problem.
      Did my answer do any good?

    • @innocentiuslacrim2290
      @innocentiuslacrim2290 Před 9 měsíci

      @@rocketman475 Human intelligence is nothing like a large language model. Here is what ChatGPT has to say about that:
      ---
      The distinction between human intelligence and the "intelligence" manifested by large language models like mine (GPT-4) is profound. Let's break this down:
      Nature and Origin:
      Human Intelligence: Emerges from the complex interactions of billions of neurons in the human brain. It has evolved over millions of years, shaped by a combination of genetic and environmental factors. It's holistic, encompassing sensory perception, motor functions, emotions, consciousness, and cognitive faculties.
      Language Models: Are a product of machine learning techniques and data. They're based on artificial neural networks, which are loosely inspired by biological neural networks but are much simpler in architecture. These models are "trained" on vast amounts of text to predict the next word in a sequence, which they then leverage to generate coherent responses.
      Learning and Adaptability:
      Human Intelligence: Humans learn from a diverse range of sources, including direct sensory input, personal experiences, social interactions, and formal education. Our learning is also intertwined with emotions, motivations, and consciousness.
      Language Models: Learn predominantly from the text data they're trained on. Their learning is statistical, and they lack any form of consciousness, emotions, or true understanding. Once trained, their knowledge is static unless retrained.
      Depth of Understanding:
      Human Intelligence: Capable of deep comprehension, introspection, emotions, ethical reasoning, and deriving meaning from experiences.
      Language Models: Do not truly "understand" content. They generate responses based on patterns in the data they've seen. They don't have beliefs, desires, or emotions.
      Versatility:
      Human Intelligence: Humans exhibit general intelligence. We can learn a vast array of tasks, from language and mathematics to artistic creation and emotional support.
      Language Models: Are specialized. Even a large and versatile model like GPT-4 is tailored primarily for linguistic tasks.
      Consciousness and Self-awareness:
      Human Intelligence: Humans possess consciousness-a sense of self, the ability to experience subjective reality, and introspection.
      Language Models: Have no consciousness, self-awareness, or subjective experiences. They process information and generate output based purely on their programming and training data.
      Transfer and Generalization:
      Human Intelligence: Humans can take knowledge from one domain and apply it creatively in another-a hallmark of general intelligence.
      Language Models: While they can handle a wide range of linguistic tasks, their ability to generalize outside their training data is limited by the patterns they've seen during training.
      In essence, while large language models exhibit impressive feats of linguistic prowess and can simulate certain aspects of human-like conversation, they don't "think" or "understand" in the way humans do. They're powerful tools, but their "intelligence" is fundamentally different from human intelligence.

    • @Shorties252
      @Shorties252 Před 8 měsíci +1

      I agree, I think we have assumed intelligence is complicated because we are the only species to have a complex language on earth, but I think we are realizing that it isn’t as complicated as we inferred it to be. Evolution never had a need to equate to the smartest techniques to reach an advancement, but rather it’s the least complicated way tends to be the fittest.

  • @MrShpongleme
    @MrShpongleme Před 7 měsíci

    AI has not passed the turing test yet,, LOL

  • @jhgolf25
    @jhgolf25 Před 7 měsíci

    Maybe the first 15 minutes was clearly articulated but then the rest of the video completely goes off the rails

  • @vinaynk
    @vinaynk Před 8 měsíci

    This dude is 100% right about people using language gymnastics to protect their position.

  • @user-jx9bn6zh9i
    @user-jx9bn6zh9i Před 6 měsíci

    ؛؛ :😮😅

  • @kebakent
    @kebakent Před 7 měsíci

    Nice explanation, but the whole "conspiracy" thing seemed very forced. I know the basics of how to build a house, but that doesn't mean it's not very complicated in practice. We shouldn't all learn how to build houses in practice. That said, the mathematics, data processing and programming skills needed to make ML models is far above the abilities of most people.

  • @benudek
    @benudek Před 9 měsíci +2

    Overpromised on explains this … clickbait

    • @JensGulin
      @JensGulin Před 9 měsíci

      What do you mean? Which part was not enough? I would agree that it's a click bait title, but the explanation was nice.
      I was hesitant on the "conspiracy" part, but they did explain that and fortunately it wasn't overplayed.

  • @hdjdhdjd42
    @hdjdhdjd42 Před 6 měsíci

    ةششششششش

  • @Rajibuzzaman_STEM_Rajibuzzaman

    R | & 0 D TO ACHIEVE ATTAINABLES

  • @Arunraj-fg7bq
    @Arunraj-fg7bq Před 6 měsíci

    Pova😊😅😊😊