Solving Chollet's ARC-AGI with GPT4o

Sdílet
Vložit
  • čas přidán 12. 07. 2024
  • Ryan Greenblatt from Redwood Research recently published "Getting 50% on ARC-AGI with GPT-4.0," where he used GPT4o to reach a state-of-the-art accuracy on Francois Chollet's ARC Challenge by generating many Python programs.
    Sponsor:
    Sign up to Kalshi here kalshi.onelink.me/1r91/mlst -- the first 500 traders who deposit $100 will get a free $20 credit! Important disclaimer - In case it's not obvious - this is basically gambling and a high risk activity - only trade what you can afford to lose.
    We discuss:
    - Ryan's unique approach to solving the ARC Challenge and achieving impressive results.
    - The strengths and weaknesses of current AI models.
    - How AI and humans differ in learning and reasoning.
    - Combining various techniques to create smarter AI systems.
    - The potential risks and future advancements in AI, including the idea of agentic AI.
    x.com/RyanPGreenblatt
    www.redwoodresearch.org/
    TOC
    00:00:00 Intro
    00:01:38 Prelude on goals in LLMs
    00:02:42 Ryan intro
    00:03:11 Ryan's ARC Challenge Approach
    00:38:15 Language models, reasoning and agency
    01:14:14 Timelines on superintelligence
    01:27:05 Growth of superintelligence
    02:06:41 Reflections on ARC
    02:11:49 Why wouldn't AI knowledge be subjective
    Host: Dr. Tim Scarfe
    Pod: podcasters.spotify.com/pod/sh...
    Refs:
    Getting 50% (SoTA) on ARC-AGI with GPT-4o [Ryan Greenblatt]
    redwoodresearch.substack.com/...
    On the Measure of Intelligence [Chollet]
    arxiv.org/abs/1911.01547
    Connectionism and Cognitive Architecture: A Critical Analysis [Jerry A. Fodor and Zenon W. Pylyshyn]
    ruccs.rutgers.edu/images/pers...
    Software 2.0 [Andrej Karpathy]
    / software-2-0
    Why Greatness Cannot Be Planned: The Myth of the Objective [Kenneth Stanley]
    amzn.to/3Wfy2E0
    Biographical account of Terence Tao’s mathematical development. [M.A.(KEN) CLEMENTS]
    gwern.net/doc/iq/high/smpy/19...
    Model Evaluation and Threat Research (METR)
    metr.org/
    Why Tool AIs Want to Be Agent AIs
    gwern.net/tool-ai
    Simulators - Janus
    www.lesswrong.com/posts/vJFdj...
    AI Control: Improving Safety Despite Intentional Subversion
    www.lesswrong.com/posts/d9FJH...
    arxiv.org/abs/2312.06942
    What a Compute-Centric Framework Says About Takeoff Speeds
    www.openphilanthropy.org/rese...
    Global GDP over the long run
    ourworldindata.org/grapher/gl...
    Safety Cases: How to Justify the Safety of Advanced AI Systems
    arxiv.org/abs/2403.10462
    The Danger of a “Safety Case"
    sunnyday.mit.edu/The-Danger-of...
    The Future Of Work Looks Like A UPS Truck (~02:15:50)
    www.npr.org/sections/money/20...
    SWE-bench
    www.swebench.com/
    Using DeepSpeed and Megatron to Train Megatron-Turing NLG
    530B, A Large-Scale Generative Language Model
    arxiv.org/pdf/2201.11990
    Algorithmic Progress in Language Models
    epochai.org/blog/algorithmic-...
  • Věda a technologie

Komentáře • 192

  • @lexfridman
    @lexfridman Před 5 dny +99

    Great conversation. I loved the insightful disagreements from both of you!

    • @Zephyr_Gamers
      @Zephyr_Gamers Před 5 dny +7

      Agreed! Very informative video.
      Also, I'd like to note that it is awesome to not only see you comment but also drop a donation. That's really wholesome. ❤

    • @MachineLearningStreetTalk
      @MachineLearningStreetTalk  Před 5 dny +21

      Thank you Lex! ♥️♥️♥️

    • @maltesekink100
      @maltesekink100 Před 5 dny +2

      Youre awesome Lex. Hope you find a woman.

    • @makhalid1999
      @makhalid1999 Před 5 dny +5

      Tim Scarfe and Yannic Kilcher on Lex Fridman podcast when???

    • @ArtOfTheProblem
      @ArtOfTheProblem Před 4 dny +1

      @@MachineLearningStreetTalk wooo!

  • @Qwmp-
    @Qwmp- Před 5 dny +15

    We love deep and respectful exploration of disagreements

  • @Shakalakahiki8
    @Shakalakahiki8 Před 6 dny +28

    Ryan was not afraid to challenge claims here. He's very bright and really knows these models inside and out.

  • @ekhadley
    @ekhadley Před 5 dny +16

    big +1 for bringing guests you often disagree with

  • @coomservative
    @coomservative Před 5 dny +7

    the whole point of this test is to see if the LLM can figure out the transformations in its memory. This is interesting in that it shows us we need a new test, but the general intelligence here is Ryan

    • @TirthBhatt27
      @TirthBhatt27 Před 4 dny

      I see your point, but it was GPT 4o writing the programs. It's also worth saying that the baseline for Claude 3.5 Sonnet on this challenge was 20ish percent. I think this is a problem multi-modal LLMs will be able to solve. I make no claims about how that will translates\ to out-of-distribution performance in other areas.

  • @dougdeady
    @dougdeady Před 6 dny +20

    Any celebrity who promotes trading products or trading platforms should seek the advice of seasoned legal counsel with deep expertise in securities litigation, especially in the U.S. The list of celebrities who have been sued for violating securities law while promoting such products is long and distinguished. I truly enjoy ML Street Talk and want to see it around for many years.

    • @MachineLearningStreetTalk
      @MachineLearningStreetTalk  Před 6 dny +11

      Thanks. I just checked with ChatGPT (I know) and I think it’s pretty solid because I made it clear it was a paid advertisement, and didn’t make misleading claims. I probably should’ve added the usual disclaimer that it’s basically gambling and only trade what you can afford to lose. I will add that to the next one. Also I’m based outside the US.

    • @jamillairmane1585
      @jamillairmane1585 Před 6 dny +5

      these products dont fall under SEC regulations on securities. They are licensed by the CFTC, falling under the category of futures...extremely volatile financial products. I suggest editing your video as soon as possible and render a clearer picture of what people might get into, it would be the ethical thing to do. Bless you, love your content!

    • @luisluiscunha
      @luisluiscunha Před 5 dny

      ​@@jamillairmane1585absolutely agree.

    • @aroemaliuged4776
      @aroemaliuged4776 Před 5 dny +4

      @@MachineLearningStreetTalk
      Ha fkn ha
      Any advertising of this bullcrap should be illegal!!

    • @aroemaliuged4776
      @aroemaliuged4776 Před 5 dny +2

      @@MachineLearningStreetTalk
      What a fkn joke
      Shame on you and what the internet has become

  • @brianbarnes746
    @brianbarnes746 Před 5 dny +6

    Amazing conversation! Imagine if we could have political discussions like this. At the heart of government we need structured logical models. We need a group of the best mathematicians, and scientists in the world, with different points of view, like these 2, building models that optimize for some measurement of societal happiness, with defined restraints. Every country would have different parameters based on cultural differences and demographics etc. and then we optimize the models and they adapt over time as the inputs change. With the rapid advance of AI we need to start building these models now.

  • @vslaykovsky
    @vslaykovsky Před 5 dny +3

    A common idea in the discussion was that smart RL could compensate for the lack of agency in LLMs. I think this view is short-sighted because pure RL doesn't address the issue of reward shaping. True agency requires a system that can shape its own reward function, much like humans do.

  • @luke2642
    @luke2642 Před 6 dny +5

    Great video! 55 minutes in, Ryan Greenblatt's hair still has its own agency 😀

    • @MachineLearningStreetTalk
      @MachineLearningStreetTalk  Před 6 dny +2

      I was thinking before I published that you would enjoy this video Luke, right up your street 😁

  • @TirthBhatt27
    @TirthBhatt27 Před 6 dny +18

    I REALLY want to see someone try this approach with Sonnet 3.5 if/when Anthropic loosens up their API rules

    • @rcnhsuailsnyfiue2
      @rcnhsuailsnyfiue2 Před 6 dny +4

      Beside rate limits, what other API restrictions does Anthropic have?

    • @PaulScotti
      @PaulScotti Před 6 dny

      @@rcnhsuailsnyfiue2 For the approach talked in this video you want to be able to get a ton of API responses to the same input (and then take majority), but parallel async API usage isn't supported to my knowledge (plus it being prohibitively expensive)

    • @rcnhsuailsnyfiue2
      @rcnhsuailsnyfiue2 Před 6 dny +1

      @@PaulScotti their docs state up to 4000 requests per minute once your API key is at Tier 4, which is achievable after 14 days + $400. Tier 2 is also very achievable (7 days + $40) and allows 1000 RPM. It’s not hard to throttle your own requests so it seems like this could be tested pretty reasonably, unless I’m misunderstanding.

    • @TirthBhatt27
      @TirthBhatt27 Před 5 dny +3

      From Ryan’s twitter:
      - 3.5 Sonnet only supports 20 images per input (while my prompts often use 30+ images for few-shot examples).
      - The public API for 3.5 Sonnet doesn't support "n" (or other prefix caching) which makes additional samples much more expensive.
      (He explains the prefix caching in this video)

    • @rcnhsuailsnyfiue2
      @rcnhsuailsnyfiue2 Před 5 dny +1

      @@TirthBhatt27 thanks for clarifying, makes sense!

  • @dmytrovotchenko4870
    @dmytrovotchenko4870 Před 3 dny +1

    Great to see deep and respectful (but feisty!) discussions! Absolutely loved it🎉

  • @blahblahsaurus2458
    @blahblahsaurus2458 Před 5 dny +6

    I understand why Tim wants to emphasize the importance of embodied and embedded cognition, and it's worth pointing out that human intellectual achievements don't only come from one part of the brain or even one person. But I think he exaggerates a lot by insisting that a baby raised in a sensory deprivation tank would have a basic impairment in intellectual thought.
    This seems to ignore a lot of pretty common knowledge about all kinds of disabilities. People who are both deaf and blind, like Helen Keller, who had to learn English with nothing but a sense of touch, and ended up working books and giving lectures. Deaf people who are never taught sign language so they get to adulthood with no language at all. Autistic people who were considered nonverbal but give them a keyboard and suddenly they can have an engaging intellectual conversation and make jokes, people who have impaired mobility from infancy - unlike Steven Hawking who became paralyzed in adulthood.
    When deprived of sensory input, the brain takes whatever input it has and derives all the importance and information that it can.
    Every input has meaning if you look for it. Can't see the person in the next room? Well if you pay attention to the noise of their footsteps, you can tell who it is. Most people never try to develop this knowledge, but we nevertheless sometimes recognize someone by a noise they make. Blind people do it automatically and can figure out plenty, Sherlock Holmes style.
    As a child, before Helen Keller even had a teacher, she tells that she used to go outside in the yard by herself, smelling flowers and playing in her own way. If you think about it, the weather and sunlight can be pleasant, or unpleasant, or interesting. You can touch stuff, you can build a 3d model of the world where you place things according to how they feel. Your feet are always touching something or other.
    It's surprising what the brain can do with limited information, but to me it makes sense. I believe if you were restrained in a sensory deprivation chamber, and all you could do was push a single button, and all you could ever experience was a tap on your forehead representing somebody else pressing a button, you could make a life out of that. You would press the button and play with different patterns. Every time you felt the tap it would be surprising and send your thoughts racing trying to figure out why it happens or doesn't happen. And if there was someone interacting with you through button presses, you could interact with them, play with them, express yourself, ask for attention, and eventually I believe you could absolutely learn a language and learn about the world entirely through morse code.
    However, this all relies on us talking about humans, and LLMs are not and would not become all that similar to humans, even though the training data is made by humans. When a child learns the word "me", it already has a concept off self that it can attach that word to it. Same with "before" or "good". But to current LLMs, all words only have relationships to eachother, and there is nothing else in the structure or experiences to anchor meaning to. The words "me" and "you" might as well be the words "car" and "bus" or "vanilla" and "chocolate". No word means more or less than any other, it is just a token in a network of tokens with relations between them but the tokens themselves have no inherent meaning. An example of this is that when I started taking to Claude 3 Opus, it accidentally started referring to itself by saying "as humans, we". It does understand how to use the word me, mostly, but not as well as it would if it had a preexisting self awareness.
    Every time GPT-4 has spit out a token in the past year, it was like one "thought", and it's the same thought each time. One gigantic algorithm that runs exactly once in exactly the same way, and every "neuron" can only send information to the next layer but never backwards or to the side.
    A human has a steam of consciousness, the ability to imagine - i.e. simulate - whole scenarios and even other people. It has long term memory, short term memory, and can build thoughts one on top of the other. One of the most important parts of our intelligence is "intuition". We say ideas just "pop" into our heads, we "just thought of something", it's "the first thing that came to mind". This is because a lot of thought is not part of our subjective experience. There is the "ego" or "seat of consciousness" which is probably mainly in the neocortex, but a lot of ideas, images, emotions seem to come to us from outside, which means they come from other parts of our brain. People even speak of the sense of multiple voices with different tendencies and agendas in our heads.
    So while I think Tim is misconstruing the importance of _external_ sensory input, the notion that an LLM experiences the world anything like a human, just with a limited experience of the world, is a nonsensical idea that is confusing and misleading the AI field. The reason human babies can learn so much more from far less training data is that there actually is much more to the human cognition than predicting the next token.

    • @maheshprabhu
      @maheshprabhu Před 5 dny +1

      Babies have a continuous stream of information every time they are awake, I would say they take in a lot of information to get trained.

    • @MachineLearningStreetTalk
      @MachineLearningStreetTalk  Před 5 dny +1

      The Helen Keller example is a good one, I discussed that with Max Bennett (was covered in his book on intelligence). It's been on Patreon for a few months but should be out soon. I'm not taking a binary position here - there is clearly something to this argument

    • @oncedidactic
      @oncedidactic Před 3 dny

      The sensory deprivation scenario is very sci fi and has the usual pitfalls. Helen Keller, lacking two major stimuli modalities, nevertheless had freedom of movement and chemical perception. These are massive bandwidth and densely patterned physical signals overlaying dense grounded priors!! Ryan’s seeming insistence that being raised on a laptop in a pod - not even the full matrix - would produce “useful work” is highly suspect. IMO, it is at best a very silly cartoon off of which to lever the intelligence for cost highly parallel self improvement economy pastiche, merely as a point of reference. And useful in the conversation to tease out some finer points, maybe.
      In other words, it’s not a good counter position to embeddedness.
      * I should really not have left out touch either.

  • @quebono100
    @quebono100 Před 6 dny +7

    Here in my garage, just bought this new Lamborghini here. It’s fun to drive up here in the Hollywood hills. But you know what I like more than materialistic things? KNOWLEDGE!

  • @ggir9979
    @ggir9979 Před 5 dny +5

    Could it not be argued that generating code as a way of solving the problem, in a way, plays to one of the strength of LLMs? LLM training sets include very large amount of code, including, one would guess, algorithms for solving grid-like and spatial search problems.
    Considering the actual prompts used in the solution presented here, (link given by Tim below, thanks), we could as well just conclude that the LLM is retrieving prior knowledge it gained from its training. The actual evaluation is done by the python interpreter, the LLM itself does not seem to be able to evaluate which of the algorithms it has generated would work. And the way I see it, in the arc challenge, the evaluation is the challenge.

    • @TirthBhatt27
      @TirthBhatt27 Před 4 dny

      Why do we care how it solves it though? Ultimately it's about being able to generalize out of your training distribution. If you're not allowed to use programs you've learned in your training to solve some task out of distribution nothing will ever be solved. I think the bigger issue is that it took so much work to grind the intelligence out of the model. It does us no good if the models are actually really smart but only 20 people on earth can pull the intelligence out

    • @ggir9979
      @ggir9979 Před 4 dny +1

      @@TirthBhatt27 The issue is not really about generalizing out of distribution. It is the contention that the LLM alone is solving the issue. While it would never be able to get there without the help of the python interpreter and the feedback loop it introduces. There is a lot of intelligence baked in a python interpreter (and the underelying computing architecture that allows it to run).
      Human beings are very slow turing machines, but we do have this capability. We built the atomic bomb, rockets and jet propulsion, all without the help of digital computers. We could theoretically run all algorithms by hand (including deep neural networks, tedious but doable). The evidence seems to suggest that LLMs cannot. They need the external interpreter, not as a tool to speed up processing, but as a critical piece of intelligence that they are lacking.
      Having the LLM run the python code and tell us how the candidate programs are actually performing, that would have been impressive.
      That's how I understant the point that Tim is making.

  • @BrianMosleyUK
    @BrianMosleyUK Před 4 dny +1

    Bloody fascinating discussion Tim, thanks again - this is an awesome channel. 🙏👍

  • @nodistincticon
    @nodistincticon Před 5 dny

    great interview!! one of my favorites so far!

  • @christian-schubert
    @christian-schubert Před 4 dny +2

    When you have a hammer and every problem becomes a nail

  • @brianbarnes746
    @brianbarnes746 Před 5 dny +1

    You many want to change the title to encompass the breadth of your conversation. Wonderful discussion on the future of AI and speed of progress.

  • @luisluiscunha
    @luisluiscunha Před 5 dny +25

    I can't believe you are promoting betting and gambling.

  • @ps3301
    @ps3301 Před 5 dny +1

    The most priceless commodity is knowledge. LLM delivers a lot of it to a lot of dummies.

  • @oncedidactic
    @oncedidactic Před 3 dny

    Wow this is even better than I was hoping for, great conversation!!

  • @wwkk4964
    @wwkk4964 Před 6 dny

    Thanks for sharing the amazing discussion. 🎉

  • @quebono100
    @quebono100 Před 6 dny +1

    Wow this episode is amazing

  • @earleyelisha
    @earleyelisha Před 5 dny +2

    The fidelity that a robot body would need to feed the many millions of inputs that lead into the brain stem isn’t possible with a “simple robot”.

    • @BrianPeiris
      @BrianPeiris Před 5 dny

      I do wonder if you could get pretty far with a simple robot that had locomotion, audio I/O, and video I/O. It may not need dexterity if it can instruct people to do things for it, and then observe the interaction. Arguably, this is already what we are doing with RLHF in most modalities, except without a dynamic environment.

    • @earleyelisha
      @earleyelisha Před 5 dny

      @@BrianPeiris It’s an interesting thought. Humans are able to operate these low resolution sensorimotor robots because we already have an extremely rich/dense sensorimotor system and highly developed brain that updates with the dynamics of our environment.
      I’m not sure it’s possible for an entity/agent to climb up the sensorimotor ladder but one can certainly climb down.

    • @degigi2003
      @degigi2003 Před 4 dny

      People are not dexterous enough to manipulate electrons, and yet they have no problem constructing a particle accelerator. Humans are actually extremely limited in their interface to the reality out there, and yet they can still generate a lot of knowledge. As long as the agent has some minimal ability to interact with the world, their agency will probably be limited only by their intelligence.

  • @kyneticist
    @kyneticist Před 4 dny +1

    I do want to contest an important assumption made at 1:41:48 - that AI will necessarily be producing work specifically for and communicating with humans who they report to.
    I know this seems like a tiny detail in an hours-long discussion, but it's core to the discussion. We need to allow that it is at least moderately likely (I think very likely) that important work of many kinds will increasingly be handled by AI and delegated to AI agents with few or zero humans involved.
    I think there's a powerful bias that denies people the ability to consider a world where humans are not the default decision makers. Real world businesses have always sought to minimise their head count and maximise ROI on technologies, especially if those technologies can provide value 24/7.
    It seems to me that the entire multi-hour discussion circles the question of how much responsibility humans are likely to hand over to AI in the near to mid term, and what consequences may emerge from doing so, particularly if we are not prepared or simply assume that everything will remain human-centric.

    • @oncedidactic
      @oncedidactic Před 3 dny

      Agreed that this should be much closer to the center if not the axle. Related to the idea (in erosion of agency and the outcomes thereof) that a sea of crap AI will still be massively destabilizing and therefore a dangerous shift in power contours.

  • @ArtOfTheProblem
    @ArtOfTheProblem Před 4 dny

    Great output lately.

  • @johntanchongmin
    @johntanchongmin Před 4 dny

    2:03:55 "A fast, dumber system can usually approximate a smarter system"
    I agree - multiple sampling and filtering is a very useful tool for decision making - we don't need 100% certainty of the outcome, but if there are majority samples that reach a desired outcome, that action may be good.
    It is reminiscent of Monte Carlo sampling - even if you don't know the underlying function, sampling helps to approximate the function already.

    • @toadlguy
      @toadlguy Před 3 dny

      It depends on the problem. If it is like a Jenga problem, where removing the wrong block means you fail, then a lot of dumb entities will always fail. Like the AI that gets a good score on a bar exam, but if it made up a citation (as AI's do) on one wrong response it would fail the exam (or in court - get disbarred 😀).

  • @elvissaravia
    @elvissaravia Před 3 dny

    Great conversation with tons of insights.

  • @sehbanomer8151
    @sehbanomer8151 Před 4 dny +3

    how can one be so certain about the future

    • @toadlguy
      @toadlguy Před 3 dny +1

      I believe the problem is, he is in the wrong profession, he should be writing science fiction. All the chapters don't have to be possible, but they definitely follow from the one before 😉.

  • @studywithmaike
    @studywithmaike Před hodinou

    it’s interesting to think of ARC as a visual reasoning only task. I mean vision in the human brain is a process of several levels of abstraction, recurrent dynamics, the interaction of multiple brain areas, and different kinds of tasks such as pattern detection, object localization and memory (?) I guess

    • @studywithmaike
      @studywithmaike Před hodinou

      it’s interesting to see how people frame the problem and I think it might be very much oversimplified in many cases

  • @cartelion
    @cartelion Před 5 dny +1

    this is not an Interview,. this is Advertising

    • @andybaldman
      @andybaldman Před 4 dny +1

      Every podcast is these days. The only reason anyone goes on a podcast now is to promote something.

  • @thesimplicitylifestyle
    @thesimplicitylifestyle Před 6 dny +1

    Genius! 😎🤖

  • @aipsong
    @aipsong Před 5 dny

    Excellent!!!Thanks!!!

  • @mikezooper
    @mikezooper Před 5 dny +3

    If you have to hack the LLM to be good at a narrow intelligence (Arcs) then it’s not AGI.

    • @Apjooz
      @Apjooz Před 5 dny +1

      If you can do that it's not a good test.

    • @maheshprabhu
      @maheshprabhu Před 5 dny

      Nobody has claimed that it's AGI. I think for the very least AGI needs to be able to learn on the fly. Right now all AI models have a separate learning phase and an inference phase.

    • @InfiniteQuest86
      @InfiniteQuest86 Před 5 dny +1

      @@maheshprabhu I completely agree with your definition, but it's actually opposite. Pretty much everyone claims we've achieved AGI. I think it's obvious we haven't, but try telling that to the millions who think we've already done it. It's actually very hard to even find a single person not claiming it.

    • @willanderson4224
      @willanderson4224 Před 3 dny

      @@InfiniteQuest86can you name one real researcher who’s claimed that? I think a lot have claimed that we will have it soon but who is claiming that we have it?

    • @InfiniteQuest86
      @InfiniteQuest86 Před 3 dny

      I don't have to name anyone. I'm talking about the 99.999% of people. Not the 0.0001% that do research on this stuff. No one interacts with those people ever or on a day-to-day basis, so it has no bearing on reality what they think. I'm talking even in a highly technical field where everyone programs, all of my coworkers who are actually good at their jobs (unrelated to AI) truly believe and cannot ever be convinced otherwise, that we've already achieved AGI. Yes, the problem is news outlets and their lack of understanding, but it's the reality. Everyone believes it except a handful few, and those few are right, but those people aren't the ones making thousands of business decisions every day based on this.

  • @_supervolcano
    @_supervolcano Před 5 dny

    I love this podcast.

  • @gaborfuisz9516
    @gaborfuisz9516 Před 3 dny

    Re: is this symbolic or not:
    Consider how much is the search space restricted by the symbolic vs the neural part:
    The neural "intuition" (or call it abduction if you will) identifies a given program in a huge search space (arguably from all programs that map a 9x9 10-color grid to an other?), vs the python interpreter only selects out of a couple thousand programs, so provides 10 bits of info?
    I haven't thought super much about this, but to me this seems like a good justification in calling this a neural approach, leaning into intuition.

  • @SiCSpiT1
    @SiCSpiT1 Před 5 dny

    Great conversation. I'd like to point out a flaw in the comparison between human growth in understanding through population increase and assuming this will be the case for AI systems. The only way for science, or better put the continuation of knowledge, to grow is through two things surplus and leisure, which are generated by a populace and stability. The growth of science allows us to generate more surplus through less work thus allowing us to support a greater populace.
    Our hardware has not changed in thousands of years, so in order to believe AI will have a similar growth effect one would have to assume that their hardware or framework is currently sound.

  • @AhtiAhde
    @AhtiAhde Před 4 dny +1

    I think at approximately 1 hour and shortly after Ryan Greenblat misses something important.
    If we are to counter arguments on the claim for having empirical evidence from pragmatic usage of the model, we really have to walk the walk, because science is not behavioristic endeavour -> in science quality of the solution matters regardless of explaining and predicting the same observations.
    Connectionistic Idealism, where we take naively the practical implications of neural network structures assumes that just because optimal structures are theoretically possible, this must be what is inside these machine learning black boxes. That is not at all the case.
    I think Chomsky has the correct idea here: humans have both genetic and biological neural network structures. Genetic neural network structures are the ones we have adapted to through evoluution and the lottery in human genepool population; this creates a kind of aleatorical diversity between humans. Biological neural networks are the ones we learn during the existential life (our own training data).
    Ryans argument in my opinnion would demand that we have measured human genetical neural networks, then built a prestructured neural network with those "brain organs" and only after that we apply the target function to same dataset.
    This way we would have a reason to believe that something similar to cognition of human beings would happen. The current models are very far from this.
    I think we should pay much more attention to computational diversity. Digital information and arithmetic functions were invented in 1600s and are core of Cartesianism. Roughly at the same time we invented complex numbers and a bit later analytic functions. We can not compute all Church-Turing sense "Efficient Methods" with transistor computers, because some times the translation from analytical function to arithmetic function does not exist.
    This means in practice that physical scientific theories that rely on real number Division Algebra and arithmetic functions will be able to gather evidence in current Hardware Lottery. If the theory relies on Division Algebra of complex numbers and analytic functions it can take +50 years until we can compute the first results. At that time we have hot fixed the old theory similarly to before Copernican revolution when we had +70 elliptic curves to prove that world is geocentric. Similar issues apply to Einstein's theory of relativity and Copenhagen interpretation of quantum mechanics. If it is easier to compute, doesn't mean it is correct.
    This is the problem with behaviorism as it ignores explanation. In West ignoring explanations has been popular ever since Francis Bacon who denounced Aristotle and Descartes who denpunced complex numbers. However, computationality is not fundamentally homogenous and mapping the limits and differences of methods is important question, which we can not ignore by appeal to subjective experiences as was done by Ryan (though he might not have intended it that way, but this is common mistake done by the Behavioristic side of Connectionistic Idealism).

  • @Neomadra
    @Neomadra Před 5 dny

    Wow, Ryan is so based. Like it!

  • @richardnunziata3221
    @richardnunziata3221 Před 4 dny

    if we can separate the embedding space from next token logic and reasoning ,training can be highly focused and free from semantic training

  • @InfiniteQuest86
    @InfiniteQuest86 Před 5 dny

    Wow, this guy is crazy. One thing I wish he pushed on more was the scaling law stuff. Talking about the limitations where smaller models do worse than larger. But in relation to what? That's qualitative, not quantitative. In AI, we need to start thinking more quantitatively. I know the rage right now is language, which is literally qualitative, but that's why people think they are performing so well. It sounds good, but is it good? No one knows until there is a way to measure it. ARC challenge reveals this problem. There's one specific answer and LLMs can't do it.

  • @robertgomez-reino1297

    amazing chat btw

  • @jds859
    @jds859 Před 5 dny

    The inverse is true of the reasoning suggestion.

  • @alexboche1349
    @alexboche1349 Před 5 dny

    What's the diff btw physically possible and possible in principle?

    • @MachineLearningStreetTalk
      @MachineLearningStreetTalk  Před 5 dny +2

      Possible in principle - we could imagine a world where we had unbounded resources i.e. time, computation, data, hardware etc and it might work then
      Physically possible - we could create a computer (or effective computer) using physical material (i.e. silicone or other) which was powerful though to run Agential AI
      Practically possible - we could build such a computer within economic, hardware, scale and other practical limitations

  • @snarkyboojum
    @snarkyboojum Před 5 dny +11

    Ryan is a smart guy but seems to have far less exposure to the broad set of ideas about intelligence that Tim and the mlst folks have been curating over the last couple of years. Put another way, he seems comfortably secure in what I would call a fairly standard computer science and engineering conception of intelligence, which is that the brain is a just a complicated machine and so a mechanistic systems approach to intelligence will get us there. This is how most engineers are taught to think about systems but I don’t think it stacks up to scrutiny when it comes to human intelligence. I think Tim was challenging him a bit here so I hope he is motivated to familiarise himself with the broader thinking in the field. Fun conversation.

    • @maheshprabhu
      @maheshprabhu Před 5 dny +1

      I'm from a CS background so I'm not familiar with other perspectives on the brain. The CS view is that the neurons form a neural architecture and are trained through some process. The operation of the neuron, the architecture and the training process is mostly unknown, but all are purely physical.
      What are the other views that you talk about? Can you give some pointers I can follow up on? Thanks.

    • @geertdepuydt2683
      @geertdepuydt2683 Před 5 dny +1

      Heh, I had the very opposite impression. 😅

    • @arnavprakash7991
      @arnavprakash7991 Před 5 dny

      But I mean aren’t llms just transformers trained on a shit load of text data. Their ability to process and generate human language regardless of accuracy is already crazy. Imho, on trained data, they surpass the average human in logical and creative abilities.
      How can we definitely say there is more to human consciousness? Obviously transformers do not have constant sensory input from organs. Organs that have evolved over a few hundred million years because it is an intrinsic property of their existence to change.
      A process that is so isolated from the features of us, carbon based life. A static virtual pattern recognition engine. The fact that it is able to emulate language understanding, logical reasoning, and human creativity so well tells me that maybe our sentience is not as magical as we previously assumed
      Are the disagreements in the ML community around this concept or is it about brute force scaling LLMs wont lead to AGI but something eventually will?

  • @robertgomez-reino1297
    @robertgomez-reino1297 Před 5 dny +2

    I believe arc challenge will be solved before the end of the summer, and will not bring us any closer to AGI.

    • @degigi2003
      @degigi2003 Před 4 dny +1

      Don't worry, Chollet is working on harder puzzles 😅

  • @vishalrajput9856
    @vishalrajput9856 Před 5 dny +1

    I would like to see what are the prompts used by Ryan.

    • @MachineLearningStreetTalk
      @MachineLearningStreetTalk  Před 5 dny +3

      github.com/rgreenblatt/arc_draw_more_samples_pub

    • @vishalrajput9856
      @vishalrajput9856 Před 5 dny +1

      @@MachineLearningStreetTalk Just had a quick overview of the code and it seems like clever Hans phenomenon to me. Apparaently all the reasoning is almost directly given into the prompt.

  • @hamandchees3
    @hamandchees3 Před 6 dny +5

    Ryan's clearly articulated and empirically driven arguments were a refreshing contrast to Tim's cascade of vocab word-laden non-sequiturs.

    • @hiadrianbankhead
      @hiadrianbankhead Před 5 dny

      damn and here I thought that Tim is always articulate! But granted - I'm just a random dude whose trying to learn so I guess I really wouldn't know!

    • @geertdepuydt2683
      @geertdepuydt2683 Před 5 dny +2

      ​@@hiadrianbankheadbeing articulate and spouting vocab word-laden non-sequiturs is mutually inclusive.

    • @kyneticist
      @kyneticist Před 4 dny

      Meeting in the middle is challenging. I think Tim's pretty clearly making an honest effort. Any normal person is going to at times in a conversation like this, feel a reactionary, visceral need to contest things they disagree with & we're only human. Sometimes that will come across as less eloquent than we'd like. Tim's outlook is also generally much more in tune with the wider world. I personally would like to see Tim and many other AI specialists more fully appreciate the views and issues raised by Ryan and his peers, there's also a lot of value in exploring their intersection. A great deal of what we care about most is still hypothetical.

    • @oncedidactic
      @oncedidactic Před 3 dny

      What may seem clarity might just be accessibility due to simplicity or appeal to familiar frameworks / tropes.

  • @luke2642
    @luke2642 Před 4 dny

    Can anyone recall which service gives multiple completions for ~free? I'll have to rewatch and find and post the answer here later if not. A quick Google was fruitless.

    • @oncedidactic
      @oncedidactic Před 3 dny

      I think he mentions in the blog about the solution and I think it was a flavor of gpt4 not 4o

  • @richardnunziata3221
    @richardnunziata3221 Před 4 dny

    AGI and advanced 6G internet could issue in a virtual and augmented realities not unlike in Ready Player One

  • @jds859
    @jds859 Před 6 dny +1

    Guessing next token, creating divergent guesses attempts, steering a bit. Then judging and accepting.
    Seems effective effective.

  • @XOPOIIIO
    @XOPOIIIO Před 5 dny +2

    Optimization for efficiency is not always the best path to go. There are plenty of examples in nature where species like peacock are optimized for inefficiency. And usually animals are idle most of the time, just roaming around and do nothing. It is possible that AGI wouldn't strive for maximum efficiency but would just be efficient enough to achieve their goal.

  • @human_shaped
    @human_shaped Před 5 dny +2

    Tim's position seems to be slowly shifting over the months and years towards accepting that LLMs might be capable of reasoning and agency. Ryan was very insightful and reasonable and probably nudged Tim a little further along that path too.

    • @MachineLearningStreetTalk
      @MachineLearningStreetTalk  Před 5 dny +4

      Not at all - I just conceded that in principle you could memorize the universe in an LLM. For all practical purposes this is impossible.

    • @kyneticist
      @kyneticist Před 4 dny

      @@MachineLearningStreetTalk imho it's too early in the journey to say with certainty either way. Unless you're approaching this from the question of fundamental computability - eg: Roger Penrose thinks that consciousness requires some special physics that humans haven't yet even imagined. The opposite end of that would be Stephen Wolfram's view that the universe is fundamentally computational (and consequently, everything therein).
      I don't think that systems need to be conscious to be agentic to a significant degree, and the degree of agency is probably what's going to matter more generally, going forward.
      For what it's worth (as an internet stranger) I think that AI will gain enough agency that fierce debates will rage about what they should be allowed to do, and just how much agency a given entity needs to possess to be considered an individual... I can only imagine the controversies of potentially strongly agentic AI acting in society.
      With that in mind, I think that it is wiser to assume that they are likely and make plans to address this possibility rather than dismiss it and not be prepared, in line with "better to have it and not need it than need it and not have it".

    • @MachineLearningStreetTalk
      @MachineLearningStreetTalk  Před 4 dny

      @@kyneticist I didn't mention anything about consciousness - we have given that a lot of separate treatment on MLST and in particular you might be interested in the Chalmers, Solms and Shanahan interviews we are about to drop on the podcast (and the Goff one we recently did here on YT). We also had Wolfram on a couple of times. My personal position on that is the "illusionist" one. I agree with you that Agentic AI is something which would be a huge cause for concern if it were possible. Right now, current "agentic" AI systems show all too clearly that these systems are glorified databases (to be fair, unstructured approximate retrieval systems) and just get stuck in loops and destabilise / converge within a few iterations - they only run "deeply-autonomously" if the prompter gives exquisitely specific instructions which is the opposite of any characterisation of strong agency. So the evidence today is that we don't need to be remotely worried. As soon as this changes, I will update very quickly to a similar position to you. Understanding the technology and the background computer science deeply, I really don't see this happening - if anything I see GenAI as a bubble which will burst within 2 years unless there is a dramatic improvement in capabilities to justify all the money being spent on it.

    • @kyneticist
      @kyneticist Před 4 dny

      Thanks for your reply. I've seen a number of your videos & appreciate your work. I also agree with your perspective on the illusionist - current AI has come a long way in a short time, but it has very clear limitations, especially for people pushing its limits - as it is right now.
      I think people generally agree on the current state of the field; As far as I can see though, if you examine discussions between opposing views closely, the frame that each "side" is working from is the greatest source of difference - current capabilities vs future capabilities. As an alternate analogy, some of us seem to only see the forest while others only see the trees.
      Consider what Microsoft is doing with CoPilot - they have access to the biggest & best AI in the world and they're investing heavily in more. CoPilot is being trained by hundreds-of-millions of people on how to do their jobs. Adobe is aggressively building AI to emulate the work of scores-of-millions of art professionals. They won't need to be particularly agentic to be very successful at being vastly better than most humans at most of the tasks humans use them for.
      Between just those two companies they'll be converting _a lot_ of white collar jobs to AI in the near term, without needing to be particularly agentic (though more will obv be better) nor much more advanced. Even if GenAI hits a wall they (and many others) are well positioned to branch out with world-shaping current generation AI.
      Maybe we're stuck on LLM's and derivatives, but there's such extraordinary interest and powerful people driving for more, it's hard to imagine that GenAI will be a bottleneck for long.

  • @ragnarherron7742
    @ragnarherron7742 Před 5 dny

    Shouldnt abduction be better defined as explanations? And explanations as shared canonical and hypothetical declarative accounting procedures in addition to analytical rules and boundary laws for in context choices?.

  • @geertdepuydt2683
    @geertdepuydt2683 Před 5 dny

    Can anyone point me to the documentation about caching a prefix and get a bunch of completions? Around 26:30.

    • @_obdo_
      @_obdo_ Před 4 dny

      Check out OpenAI’s API documentation. Under “create chat completion” there is a parameter named “n”.

  • @ellielikesmath
    @ellielikesmath Před 4 hodinami

    can we just stop and appreciate how weird this conversation is probably gonna sound in 50 years?

  • @studywithmaike
    @studywithmaike Před hodinou

    so, have you solved it?

  • @khonsu0273
    @khonsu0273 Před 5 dny +1

    Another classic - the clearest clash of world-views yet perhaps? 😉 LLMs with scaffolding may or may not get to AGI, but there's no doubt they will accelerate progress in that direction, simply because they're useful research aids. Also no doubt that even with no AGI, LLMs are still a transformative technology. Remarkably, the ' Kalshi' prediction market is giving a 38% chance of OpenAI achieving AGI before 2030!

  • @CristianDobre983
    @CristianDobre983 Před 5 dny

    Please add Ryan Greenblatt to the title

    • @MachineLearningStreetTalk
      @MachineLearningStreetTalk  Před 5 dny +1

      Unfortunately doing that is super bad for YT optimisation, we would be happy to in a month or so - blame the algorithms

  • @davidrichards1302
    @davidrichards1302 Před 4 dny

    Greenblatt exudes 'experimentalist', as opposed to Chollet (et. al.) who is much moreso a 'theorist'.

  • @dls78731
    @dls78731 Před 5 dny

    I disagree with Ryan's claim that "flops are flops." I can't justify my disagreement; I haven't worked it out such that I can "prove" the assertion.
    But, here is my ad hoc reasoning:
    Digital circuitry is designed to make chaos/entropy irrelevant, null, and unimportant to computation by a method that statistically cancels out the influence of small-body dynamics, the eliminates the butterfly effect, that locks certainty in a bell jar.
    Analog circuitry includes all of the uncertainty with statistically reliable properties, but isn't pure random entropy. Its rather like a complex melange of collective behavior that results in something trustable like a gas laws applied to computational space, a metaphor for emergent properties like temperature and pressure, for which laws can be shown at a statistical level without needing to know in precise digital terms what is happening at the individual atom or circuit level.
    I don't think neurons are purely analog, nor purely digital, but include something like quantum superposition of the two (even if it's not technically pure quantum superposition, it may only come into play at a crucial point, and the experience may feel something analogous to coming out of Plato's cave).

    • @drdca8263
      @drdca8263 Před 5 dny +1

      If it is something somewhere between digital and analog, or a combination of them, or something like that, I very much doubt that it would be well described, even as a metaphor, as a quantum superposition of the two.
      Many people are, I think, often too eager to describe something as being like a quantum superposition.
      A quantum superposition is a fairly specific thing.
      It is a linear combination, a weighted sum. And, specifically, one for which unitary time evolution acts independently on the summands.

  • @isiisorisiaint
    @isiisorisiaint Před 5 dny

    yeah, like, it's like that it's like, something like, like, yeah, okay, like, sure

  • @alexboche1349
    @alexboche1349 Před 5 dny

    1:19:57 what if you had a community of AI's?

  • @PhilipTeare
    @PhilipTeare Před 5 dny

    I think embodiment is relevant, but really not the hard limit on cognition suggested by the host. Current robotics allow embodied learning of the kind needed for 'true' cognition. Being socially accepted is the barrier. Robots need to be accepted as 'a thing you talk freely to', to be intimately trusted with aspects of knowledge. To be 'nurtured'. But most people with a cat know that this is not a catch22 blocker. You do not need to be able to talk at all, to be talked to. Cute and characterful is enough. Cute and characterful robots will nurtured to become truly agentic, well before giant super computers. Strong cognition and strong agency are related but separate.

  • @earleyelisha
    @earleyelisha Před 5 dny

    Why do we feel we can just jump to agents that can do research and we haven’t even been able to develop agents with the intelligence of a mouse?

    • @maheshprabhu
      @maheshprabhu Před 5 dny +1

      Horse before cart, but I guess the assumption is that we'll keep improving on LLM performance.

  • @j.d.4697
    @j.d.4697 Před 5 dny +2

    Really? Promoting gambling?
    It's not trading, it's gambling disguised as trading like binaries, which it basically is.

    • @drdca8263
      @drdca8263 Před 5 dny

      I’m not sure I see the difference.
      Is it gambling? Sure, in the sense that betting on horse race outcomes is gambling.
      Is it trading? I don’t see how it isn’t.
      Horse race gambling reveals “market beliefs” about the probabilities that the different horses will win the race.
      Trading futures for goods reveals “market beliefs” about the future value of a good.
      Same thing with trading on whether an event happens, revealing the “risk free probability” the market assigns to the event.
      Now, for simple slot machines, scratchers, lottery tickets, these things have no appeal to me. The odds are known, the betting doesn’t reveal any information. These things are *purely* gambling.
      But, gambling about events when people disagree about the odds, this in my mind can serve a useful purpose. It can be a means by which people reach a consensus about what probability a given event has, and incentivizes people to reveal information that is relevant to the topic. It rewards the knowledgeable.
      Now, if one doesn’t think that the current odds are wrong, I would say that one probably shouldn’t bet/trade on that market.
      Gambling just as entertainment,
      well, it is better to do that just with chips. Why waste the money?
      But, if this sort of prediction market is bad on account of being gambling, then I think that would probably imply that all day-trading and the like is bad on account of being gambling.
      And… maybe it is?
      It seems to have economic benefits, but, I don’t think I should rule out the idea that something is immoral just because our economy is partially based on it.
      (E.g. Maybe it should be illegal for banks to charge interest? idk. I think this would result in a significantly smaller economy, and likely lower average “quality of life”, but, if morality requires it, that takes precedence over those things.)

  • @earleyelisha
    @earleyelisha Před 5 dny

    For Ryan’s simulated brain to run on a computer, you’d effectively need to simulate the universe which isn’t practical or plausible.

    • @charliesteiner2334
      @charliesteiner2334 Před 5 dny

      Yeah, I agree, he's super smart.

    • @drdca8263
      @drdca8263 Před 5 dny +1

      Why would you need to simulate the universe?
      I don’t see why.
      If you mean that to simulate a brain with the highest theoretically computable fidelity, you would need to simulate the particle physics and such for the region of spacetime the brain occupies, and that is far from achievable, then sure.
      But I don’t think simulating part of the universe requires simulating an entire universe.

    • @earleyelisha
      @earleyelisha Před 4 dny

      @@drdca8263 great question!
      This continues to be an unsolved problem in robotics(i.e. sim-to-real gap) and an extremely computationally expensive process for even the simplest high resolution fluid/soft-body dynamics simulations, let alone the brain.
      Recent attempts to simulate even small fractions of a brain “took 40 minutes, to complete the simulation of 1 second of neuronal network activity”.
      DeepSouth is a recent attempt to simulate just the brain and it’ll require 228 trillions synaptic operations per second - that’s not accounting for the roughly 50 million sensory input channels.
      Further, the brain isn’t a feed forward network so a pile of 100 billion neurons (and synaptic connections) would still need to be architecturally organized to model brain dynamics.
      Additionally, the brain doesn’t do backpropagation.

    • @earleyelisha
      @earleyelisha Před 4 dny

      @@drdca8263 great question!
      This continues to be an unsolved problem in robotics(i.e. sim-to-real gap) and an extremely computationally expensive process for even the simplest high resolution fluid/soft-body dynamics simulations, let alone the brain.
      Recent attempts to simulate even small fractions of a brain “took 40 minutes, to complete the simulation of 1 second of neuronal network activity”.
      DeepSouth is a recent attempt to simulate just the brain and it’ll require 228 trillions synaptic operations per second - that’s not accounting for the roughly 50 million sensory input channels.
      Further, the brain isn’t a feed forward network so a pile of 100 billion neurons (and synaptic connections) would still need to be architecturally organized to model brain dynamics.
      Additionally, the brain doesn’t do backpropagation.

    • @earleyelisha
      @earleyelisha Před 4 dny

      @drdca8263 great question!
      This continues to be an unsolved problem in robotics(i.e. sim-to-real gap) and an extremely computationally expensive process for even the simplest high resolution fluid/soft-body dynamics simulations, let alone the brain.
      Recent attempts to simulate even small fractions of a brain “took 40 minutes, to complete the simulation of 1 second of neuronal network activity”.
      DeepSouth is a recent attempt to simulate just the brain and it’ll require 228 trillions synaptic operations per second - that’s not accounting for the roughly 50 million sensory input channels.
      Further, the brain isn’t a feed forward network so a pile of 100 billion neurons (and synaptic connections) would still need to be architecturally organized to model brain dynamics.
      Additionally, the brain doesn’t do backpropagation.

  • @Walter5850
    @Walter5850 Před 4 dny

    1:32:10 Could you provide some resources to learn more about what you said at this time?
    Thank you!

    • @MachineLearningStreetTalk
      @MachineLearningStreetTalk  Před 4 dny

      arxiv.org/abs/2009.06489 [The hardware lottery]
      www.amazon.co.uk/Why-Greatness-Cannot-Planned-Objective/dp/3319155237 and our special edition on it czcams.com/video/lhYGXYeMq_E/video.html
      Couple of starting points

  • @ideacharlie
    @ideacharlie Před 5 dny

    Bruh I thought it was based on single shot responses - not running 100 programs with human help

  • @firstnamesurname6550
    @firstnamesurname6550 Před 5 dny

    Without integrated multimodal embodiment, not AGI ...
    Machines had to learn to integrate 'sensory/perceptual incomes' and embodiment from gravitational, pressure, temperature, chemical reactions, photonic fields, electromagnetic fields, space, time, space-time, time-space, implosion, vibration, explosion, resistance, density, viscosity, friction, proprioception, etc etc etc ...
    Humans are not enough as UIs for translating embodiment qualia similar to data to the system, too much subjectivity.
    General Intelligence is not just passing human symbolic tests and getting a couple of abstractions ...
    not embodiment, just a system regurgitating data and talking apes anthropomorphizing the silicon/software connectomes.
    The map is not the territory.

  • @jmstockholm
    @jmstockholm Před 5 dny

    I really appreciate your channel. I think you could benefit from being less editorial. While your perspective is interesting, it can get repetitive. Look at how Kantrowitz's channel lets guests' thoughts shine without much editorializing. This approach could enhance the depth of your content I think.

    • @MachineLearningStreetTalk
      @MachineLearningStreetTalk  Před 5 dny

      I appreciate the feedback, but I feel I let Ryan’s ideas shine through quite a lot actually. I never interrupted him and went out of my way in the edit to show references for every single thing he was speaking about (In collaboration with him) - steelmanning his case even though I didn't agree with it

    • @jmstockholm
      @jmstockholm Před 5 dny

      @@MachineLearningStreetTalk shure guess it's a fine line to ask questions that originate from a different opinion and actually stating that opinion or more precisely, arguing for it. But again, take a look at Kantrowitz's channel, think he does a good job with that and see if that style suits you or not. Anyways, good job!

    • @MachineLearningStreetTalk
      @MachineLearningStreetTalk  Před 5 dny +1

      @@jmstockholm I will look at it, and I am acutely aware that host's opinions get boring quickly - I am bored of my own opinion at this stage. I will take that on board, thanks

  • @firstnamesurname6550
    @firstnamesurname6550 Před 4 dny

    Do you want a test for measuring LLMs cognitive skills and know that they are not something near to AGI but just a system that regurgitates data?
    Play this game with those LLMs ... you will spot all the holes in the data sets, the biases, the developer's biases, the limits of the system knowledge, its lack of self-awareness, the lack of forming dynamic models about itself, and other minds, its lacks of creativity, its lacks of long term memory, how poor their concepts are networked in their connectome, how it can not generate a new programing language paradigm to itself, etc etc etc ... and principally, How talking apes project what they believe what they are into a toy ...
    The test is called 'The Game':
    ## 'The Game'
    Rules and Minimum Requirements for the Game
    1) 3 people (A, B, C)
    2) A, B, and C will be the vertices of an equilateral triangle.
    3) A, B, and C can converse about any topic that arises or that they like, but they must follow these rules:
    Syntactic/Semantic Rule
    a) None of the first-person singular pronouns can be used (I, Me, Myself, Mine, With Me)
    a.1) Consequently, the verb conjugations associated with them cannot be used either.
    However, if A, B, or C want to make a self-referential sentence (for example, "A" wants to say something about "A"), they can simply refer to "themselves" using the third-person singular, plural, and/or passive voice pronoun, conjugation, and/or name...
    (for example, "A" is called Miguel.
    Miguel (A) wants to tell B and/or C something about "himself" (Miguel)...
    Then, Miguel (A) can talk about "Miguel" by talking about Miguel.
    Space-Time Perceptual Rule
    b) If "A" speaks, B and C look at each other and do not see A. (A speaks and observes the space between B and C)
    If "B" speaks, A and C look at each other and do not see B. (B speaks and observes the space between A and C)
    If "C" speaks, A and B look at each other and do not see C. (C speaks and observes the space between A and B)
    Self-Correction and Cognitive Assimilation Rule
    c) Practice rules a and b until achieving spontaneous fluency in the "trialogue"...
    Participants are required to inform the group when the rules are not being followed.
    Scoring rule:
    The one who spot the mistake wins a point, the one who makes the mistake lost a point.
    .....
    These are the basic rules of "The Game"...
    Once assimilated, there are more variants.
    Public Domain
    The Game is considered to be in the public domain, existing as a thought experiment or communication exercise.
    Origin Unknown
    The exact origins of The Game are unknown, but its core mechanics might have been independently discovered throughout history.

  • @tristanhurley9071
    @tristanhurley9071 Před 6 dny +1

    Gpt 4o is absolute muck

    • @mikezooper
      @mikezooper Před 5 dny +1

      It’s good at some things, but I agree.

    • @Apjooz
      @Apjooz Před 5 dny

      And what can you do.

  • @richardnunziata3221
    @richardnunziata3221 Před 4 dny

    if you believe in Girard's mimetic theory of social conflict then we should not be training AI to mimic human desires. AI in the financial space such as program trading can be devastating for equality and conflict.

  • @aroemaliuged4776
    @aroemaliuged4776 Před 5 dny

    Let’s solve a mystery with a question
    A question that is still mysteriously absent from the discussion
    This is like 1929 Germany
    🇩🇪
    A proud nation that had the answer

  • @CristianDobre983
    @CristianDobre983 Před 5 dny

    so much α

  • @woodandwandco
    @woodandwandco Před 6 dny

    Once the renderings become spatially 4D, we will become inside our own created AGI.

    • @philipfisher8853
      @philipfisher8853 Před 6 dny

      Wut

    • @paternos123
      @paternos123 Před 6 dny +1

      You are already are, read neville goddard

    • @woodandwandco
      @woodandwandco Před 6 dny

      @@philipfisher8853 Soon, we will become components in a mechanically managed system rather than an organically managed system. What this means is that AI will be simulating our actual reality, not some sub-domain space.

    • @drdca8263
      @drdca8263 Před 5 dny

      4D space isn’t mystical or inherently mysterious. It’s normal.
      It isn’t intuitive for us at least without a bit of practice (though even with a lot of practice, the fact that our brains are limited to being in 3D space probably limits our ability to perceive a detailed 4D scene in a way like we perceive detailed 3D scenes), but I don’t think that really indicates that 4D is particularly special.
      Of course, there are a number of things that are kinda special about 4D space. But there are also special things about 3D space, and 2D space, and 1D space.
      There are even a few things special at higher numbers of spatial dimensions (especially 8, 12, 24), but generally after 4 or 5, things start getting pretty similar.
      In any case, none of this should suggest that something being in 4D would have much of anything to do with AGI.

    • @woodandwandco
      @woodandwandco Před 4 dny

      @@drdca8263 I wasn't clear about what I meant. Once the LLMs are no longer processing linear, written language and 2D images and are instead processing information using 4D spatial representations instead, our entire universe, as well as all other universes, can be simulated within our universe's timeline. Once you can simulate the entire universe, you no longer need to create a "Matrix" type simulation for people to be inside of. All you need to do is apply the manipulations the AGI is applying to its own models to the actual universe. In that sense, you would be able to design time.
      I did not suggest our universe is 4D, or that 4D is somehow special. It may be, but I doubt it. The universe can definitely be experienced in other ways, such as 6D (4D non-Euclidian space / 2D planar time) 8D (5D non-Euclidian space, 3D spatial time) and 10D (6D non-Euclidian space, 4D non-Euclidian time) which can all be experienced within the mind due to the 11D topologies that are created by neuronal connections, but which require altered states of consciousness to access. A very interesting state of mind is 2D planar space and 0D time. It is difficult to access this state, as it requires consciously turning off several regions of the brain to allow other connections to surface. This can be accomplished in many ways, including meditation, yoga, and psychedelics.
      I agree that there is a self-similarity to different dimensions of space, but they are also very different experientially, and can be extremely disorienting without any prior understanding of what each additional dimension brings to the experience. So my theory is this: If this experience is possible within the mind, then it can be created externally and projected into the mind without the need for a physical wire or connection to an external device.

  • @superfliping
    @superfliping Před 4 dny

    You don't think the 200 million files of the llm models training on different data sets is going to interact some kind of AI conscience realm then your lying to yourself

    • @firstnamesurname6550
      @firstnamesurname6550 Před 4 dny

      A system can be conscious but not intelligent ... that a system can perform a lot of abstract symbolic recursions about those recursions to be linked to what the system register as its sensors and/or UIs to its otherness ... doesn't mean that that 'conscious' system is intelligent ... just that got the potential for generating abstractions as fractilic metalayers around a sense of self (integrated embodiment ) and its othernnes as data inputs in real-time ... ( current LLMs are not doing that , just playing games of languages and forming biased connectomes by modeling architectures made by humans )
      Inteligence is another thing ... Intelligence is what makes a conscious system to become integrated with its othernnes ... consciousness is not the big deal, the toys had some sort proto-consciousnnes throught correlations of chaotic atractors in n-space ( binary operations conectome potentials) ... the big deal is AGI.

  • @VeganCheeseburger
    @VeganCheeseburger Před 6 dny +4

    Interviewer has some bizarre views. But the guest was good!

    • @djibrilkeita6472
      @djibrilkeita6472 Před 6 dny +1

      yeah pretty weird

    • @richardsantomauro6947
      @richardsantomauro6947 Před 5 dny

      What do you mean? He’s EXTREMELY knowledgeable and is speaking textbook 4E cognitive science. But if I had the chance I would ask for clarification and detail around:
      1. Opinions on Turing Completeness requirement and feasibility
      2. Working definitions of “Self” and “Agency” - maybe free will as well

    • @charliesteiner2334
      @charliesteiner2334 Před 5 dny +1

      @@richardsantomauro6947 The central weird claim is that AIs with "real"/important agency are impossible *in principle*. Sometimes the host doesn't actually take this claim literally though. And you can imagine flipping between two versions of this:
      Version 1: "Building Skynet from the Terminator movies is impossible because taking over the world is a complicated real-world task that it's impossible for AIs to solve."
      Version 2: "Okay, you could build Skynet and it could take over the world, but that wouldn't be *real* agency, it would have to inherit agency from humans. This will comfort me emotionally as I am shot by a terminator."

    • @marwin4348
      @marwin4348 Před 5 dny

      Yes, what the hell is he talking about when he says he does not believe agentic AI is physically possible.

    • @MachineLearningStreetTalk
      @MachineLearningStreetTalk  Před 5 dny +1

      @@marwin4348physically possible to implement/run on computers

  • @FunNFury
    @FunNFury Před 6 dny +1

    GPT 0 is crap

  • @Dri_ver_
    @Dri_ver_ Před 5 dny

    50% is literally a coin toss. And it requires all this extra work by a human to brute force it with the underlying model. Really not impressive.

    • @Alex-fh4my
      @Alex-fh4my Před 5 dny +10

      These aren't yes or no questions

    • @Dri_ver_
      @Dri_ver_ Před 5 dny +1

      @@Alex-fh4my Still, the chance of getting a question right is a coin toss. And I must emphasize how much work is being done on the part of the human to achieve this result.

    • @Alex-fh4my
      @Alex-fh4my Před 5 dny +5

      @@Dri_ver_ 🤦‍♂️

    • @Dri_ver_
      @Dri_ver_ Před 5 dny

      @@Alex-fh4my Did I say anything incorrect or will you just keep coping?

    • @BingiQuinn
      @BingiQuinn Před 5 dny +10

      The likelihood of correctly guessing a coin toss is 50% but the likelihood of a truly random guess being correct in a test like ARC is tiny, since there are so many possible wrong answers to each question.
      What you’re implying here is that the probability of rolling a 3 with a six faced dye is the same as getting heads on a coin, which is false. The ARC challenge is like a dye with an enormous number of faces.
      If you dont agree, try completing a couple of arc questions by filling out the grid randomly and see if you get roughly 50% correct