The Next Generation Of Brain Mimicking AI

Sdílet
Vložit
  • čas přidán 24. 05. 2024
  • ▶️ Visit brilliant.org/NewMind to get a 30-day free trial + 20% off your annual subscription
    The tech industry's obsession with AI is hitting a major limitation - power consumption. Training and using AI models is proving to be extremely energy intensive. A single GPT-4 request consumes as much energy as charging 60 iPhones, 1000x more than a traditional Google search. By 2027, global AI processing could consume as much energy as the entire country of Sweden. In contrast, the human brain is far more efficient, with 17 hours of intense thought using the same energy as one GPT-4 request. This has spurred a race to develop AI that more closely mimics biological neural systems.
    The high power usage stems from how artificial neural networks (ANNs) are structured with input, hidden, and output layers of interconnected nodes. Information flows forward through the network, which is trained using backpropagation to adjust weights and biases to minimize output errors. ANNs require massive computation, with the GPT-3 language model having 175 billion parameters. Training GPT-3 consumed 220 MWh of energy.
    To improve efficiency, research is shifting to spiking neural networks (SNNs) that communicate through discrete spikes like biological neurons. SNNs only generate spikes when needed, greatly reducing energy use compared to ANNs constantly recalculating. SNN neurons have membrane potentials that trigger spikes when a threshold is exceeded, with refractory periods between spikes. This allows SNNs to produce dynamic, event-driven outputs. However, SNNs are difficult to train with standard ANN methods.
    SNNs perform poorly on traditional computer architectures. Instead, neuromorphic computing devices are being developed that recreate biological neuron properties in hardware. These use analog processing in components like memristors and spintronic devices to achieve neuron-like behavior with low power. Early neuromorphic chips from IBM and Intel have supported millions of simulated neurons with 50-100x better energy efficiency than GPUs. As of 2024, no commercially available analog AI chips exist, but a hybrid analog-digital future for ultra-efficient AI hardware seems imminent. This could enable revolutionary advances in fields like robotics and autonomous systems in the coming years.
    VISUALIZATIONS
    Denis Dmitriev - / @denisdmitrievdeeprobo...
    Jay Alammar - / @arp_ai
    Ivan Dimkovic - / @321psyq
    Carson Scott - / @carsonscott260
    SUPPORT NEW MIND ON PATREON
    / newmind

Komentáře • 379

  • @NewMind
    @NewMind  Před 23 dny +15

    ▶ Visit brilliant.org/NewMind to get a 30-day free trial + 20% off your annual subscription

  • @markedis5902
    @markedis5902 Před 22 dny +749

    The problem with biological neural networks is that they take 9 months to build and about 20 years to train.
    Some of them don’t work, others are corrupt, they are highly susceptible to viruses and about half of them don’t follow logic.

    • @josephvanbrunning2960
      @josephvanbrunning2960 Před 22 dny

      People are not robots. Dont even pretend they are. Unless you want your humanity to be reduced to just a number on a long discarded hard drive. The wealthy only think on terms of resources and are trying to gaslight the world population into thinking in these terms. It makes their job much simpler that way they can convince us to kill eachother for maximum beneifit.

    • @user-cg7gd5pw5b
      @user-cg7gd5pw5b Před 22 dny +126

      You're just not good at coding. Where I'm from, you can already use them even in factories after only a few years of training. It makes them highly replaceable once the hardware starts to break down.

    • @megumuu-
      @megumuu- Před 22 dny

      ​@@user-cg7gd5pw5b bro 💀💀💀

    • @webgpu
      @webgpu Před 22 dny +19

      use the good ones. discard the rest (ie. let them do menial work)

    • @asandax6
      @asandax6 Před 22 dny +23

      ​@@user-cg7gd5pw5b Thanks to your country. I wouldn't have my smartphone if it weren't for the replaceable assembly hardware driven by biological Nueral Networks.

  • @youtou252
    @youtou252 Před 22 dny +183

    re-pasting this because it was deleted :
    400tflop for 1000 tokens; a decent gpu is 20 tflop/s, therefore it would take like 30s for a gpu to process it. A gpu is like 200w, 30s of which is 1.6 wh, not 300wh

    • @mirek190
      @mirek190 Před 22 dny +31

      I know ! I do not believe people are even believing in such nonsense of 300W per prompt....

    • @eSKAone-
      @eSKAone- Před 22 dny +12

      Maybe the calculation incorporates energy consumption for training the models. But then it would get cheaper over time also.

    • @dirty_haute
      @dirty_haute Před 22 dny +10

      Thank you, the amount of FUD going around is insane.
      Something about LLMs just absolutely breaks people's brains.

    • @lahma69
      @lahma69 Před 22 dny +9

      I was going to comment this exact same thing. It is inconceivable it takes anywhere close to 300wh to run a single query against a model like GPT-4. Yes, training takes an inordinate amount of energy but a single prompt on said model takes only a tiny fraction of the 300wh stated figure (couldn't say what the actual figure is without doing some calculations but I would expect it to be closer to 3wh than 300wh).
      Otherwise, a great video though!

    • @guard13007
      @guard13007 Před 22 dny +11

      Why are you assuming all queries are 1000 tokens? GPT-4o's context window is 128,000 tokens and outputs up to 2048 tokens. Initial queries will be cheaper like you show, but the real power of using a model like this is in conversation or processing long texts. It works with around 100x as much data as your calculation. Even if it scaled linearly, the difference between your estimate and the video estimate lines up with that.

  • @tatianaes3354
    @tatianaes3354 Před 20 dny +31

    *THE BEGINNING is a bit misleading* because it is not the use of the neural networks, but exclusively training of them is power-hungry. Once the training is done, the power costs are minimal in most cases. Hence, even iPhones use onboard hardware accelerated neural networks for many years by now.
    Also, before spiking and neuromorphing, the biggest breakthrough that is actually possible and is happening is activation function training, which postulates that it should not be just a sigmoid or any other simple function but a parametric curve. The chap has forgot to tell about this. But otherwise, the video is great.

    • @Darthvanger
      @Darthvanger Před 13 dny

      But iPhone won't be able to run GPT or Stable Diffusion, would it?

    • @tatianaes3354
      @tatianaes3354 Před 12 dny +2

      @@Darthvanger SD is downloadable for PC, but not for e.g. iPad. But if it was, it would run locally much faster due to the ML acceleration. BTW, both AMD and Intel are now going the Apple way to incorporate the ML cores, so future generations of CPUs from PC will run SD much faster. Of course, heavy models like SD run much faster on servers, but people are still using them locally on some smaller, shorter tasks.

    • @Darthvanger
      @Darthvanger Před 12 dny

      @@tatianaes3354 got ya, thanks! Makes sense that it would run faster due to the ML cores 👍

  • @aaronsmyth7943
    @aaronsmyth7943 Před 22 dny +116

    300wh for an inference request / compute? I run several models locally on an Nvidia 4080, it takes a few seconds of the GPU working at around 85%. I play games for 3 hours straight with the GPU at 95%, so that would imply I could charge 129,600 Iphones. My electricity provider would inform me of this, surely.

    • @MrAnt1V1rus
      @MrAnt1V1rus Před 22 dny +5

      I'm thinking your local request is far more efficient than the estimated 300w. If cloud Ai is really eating that much power, I don't see how they survive open source local models. I've replaced GPT-4's API calls with my own local Llama3 with very good results. No more fees an nothing leaves my network too, which is a thing customers love to hear

    • @eSKAone-
      @eSKAone- Před 22 dny +5

      Maybe the calculation incorporates energy consumption for training the models. But then it would get cheaper over time also.

    • @lahma69
      @lahma69 Před 22 dny +6

      Ya, the 300wh figure is bonkers.. Not even a little bit close to accurate.
      Besides that though, the video contained a lot of great info!

    • @tomislavbunjevcevic9912
      @tomislavbunjevcevic9912 Před 22 dny +2

      I think they account for training consumption.

    • @wizaaeed
      @wizaaeed Před 21 dnem

      ​@@MrAnt1V1rus are you even training the model, if no. How are you running something for a client when you don't know how it works or what data it's going to output

  • @jsmythib
    @jsmythib Před 22 dny +65

    This took a long time to produce. The topics, visualizations are fantastic. Excellent, well spoken host. Great job.

    • @twerktospec
      @twerktospec Před 22 dny

      You don't know that

    • @webgpu
      @webgpu Před 22 dny +4

      text, video & audio generated by gpt

    • @jsmythib
      @jsmythib Před 22 dny +1

      @@webgpu Im not biased. Human, Agi. Whoever :) (escapes out the side door)

    • @hindugoat2302
      @hindugoat2302 Před 21 dnem +3

      its mostly just stock images and stock footage

    • @bernios3446
      @bernios3446 Před 16 dny

      @@hindugoat2302 So? But the explanations are not stock content. Must have taken a lot of research.

  • @MrMaxcypher
    @MrMaxcypher Před 22 dny +6

    This is the first time I've viewed your channel and I consider your explanation of such a complex topic the clearest and most thorough of the many I've encountered -- well done!

  • @JB52520
    @JB52520 Před 22 dny +11

    8:14 - "I feel alone..."
    So very relatable. You'll always get the line "you're not alone in feeling this way" in one form or another. But in looking around, there's no one here. And there won't be anyone here. So I must be alone, and I should feel like this. Chances are, most people who are lonely enough to ask help from a speech model won't find anything new they haven't heard from a counselor, so they'll stay alone too. I find no comfort in that. For the average human happiness, it would be so much better if it was just me.

    • @urbanbasementoperator
      @urbanbasementoperator Před 22 dny +2

      Homie I don't wanna be that guy, but perhaps telling people that problem in the comment section of a CZcams video is not a good solution to that problem. you should go meet people in real life, use electronics less, enjoy nature, and lift weights, run, or do some other physical activity. Find a group that does things you like, then hang out with that group so you can form relationships with people. If you are intentional about it, you will find people with whom you may have deep and real friendships with, but it will be up to you to make it happen. No one will make you go out and do that other than you. Easy places to form friendships are gyms and physical competitions (or competitions for any hobby, be it robotics, debate, or book clubs), churches, and volunteer programs that serve communities. Hope you figure it out.

    • @Cwarner251
      @Cwarner251 Před 21 dnem

      Very informative

  • @jayeifler8812
    @jayeifler8812 Před 22 dny +11

    Duh, we aren't going to continue on the current trajectory of power/energy consumption with AI. This fuels effort in improving the energy efficiency so that curve will never exist, it's just hypothetical.

    • @jacksonnc8877
      @jacksonnc8877 Před 20 dny

      Playing Ketchup off of the back of greed/fossil fuels👏 well said

    • @exstacc1886
      @exstacc1886 Před 14 dny +2

      Just because we're still innovating in this field at a decent rate doesn't mean it won't plateau due to limitations of physics

  • @DJ1573
    @DJ1573 Před 22 dny +42

    Algorithm, engage!

  • @guard13007
    @guard13007 Před 22 dny +2

    A shorter reply to the 300Wh controversy popping off: People seem to be assuming that an initial query with a home-run model is equivalent to a conversation with GPT-4o in terms of processing power, when GPT-4o is using around 100x more power at a first-order approximation.
    (Maybe i should've spend more time editing my other comment. :P)
    If you have time to dive into the numbers on this stuff, it'd be neat to see a comparison of estimates. Maybe I should try to do that.

  • @KevinLarsson42
    @KevinLarsson42 Před 22 dny +1

    Well done on the video structure, felt really natural and informative

  • @user-xp4of2vu4r
    @user-xp4of2vu4r Před 22 dny +61

    Now, for an old guy, with a couple of non-STEM degrees, what you present is so far beyond my capacity to grasp and understand, it's not funny. Therefore, I suspect that only a very limited number of ordinary citizens could began to comprehend either. So, this means almost NONE of current politicians have even a tiny shot of appreciating the significance of logical conclusion about the limits of materials and and services required for extensive usage, requirement and cost of such systems. Thanks for sharing and offering a subject so complex that it requires a really intelligent and well trained individual to grasp and appreciated.

    • @necrosis9222
      @necrosis9222 Před 22 dny +25

      No, its not that older person isn't smart. But because this is quite a specific field. And school these day dont go this indepth either.
      But if you study the basic of it for a little bit. You will start to understand more very quickly. Dont underestimate your experience :)

    • @necrosis9222
      @necrosis9222 Před 22 dny +4

      I agree about politician part tho 😂, they never knew what they are doing because they refuse to study into it. Due to their ego(well not all politician).
      Normal citizen doesn't really need to understand how it works. Just how to use it efficiently. Just like how people use computer, complex machinery that they dont understand how it works either(except specialist ofcourse)

    • @dan-cj1rr
      @dan-cj1rr Před 22 dny

      yea well its known that politicians are old and have an old mentality and just want to make themself and their friends rich, they dont care if people suffer. (just look at Trudeau)

    • @dubsydubs5234
      @dubsydubs5234 Před 22 dny

      It's not that it's hard for us old farts to grasp it's just a field we never deal with, it requires a way of thinking that takes years and needs a relatively empty canvases from an early age, myself I was a mechanic for 50 years so my brain thinks in a certain way with problem solving having a set pattern and trying to relate to this kind of subject is just not how my brain works. I'm sure there are people capable of mastering different subjects but it very difficult for the average person. You really can't teach an old dog new tricks.

    • @user-db9bw5cl1e
      @user-db9bw5cl1e Před 22 dny

      You've got the jist of it with how you describe it though :)

  • @puffinjuice
    @puffinjuice Před 22 dny +1

    Thanks for the video! I didnt know about spiking neural networks and analogue computers before this. It was a nice summary of the field and what we can expect in the near future 😀

  • @user-ec2qg2pv1s
    @user-ec2qg2pv1s Před 22 dny +4

    Excellent video with wonderful graphics, very well explained by the host in clear, simple and direct language. It opens the doors to the next generation of neural networks and justifies very well the need to develop research on Neuromorphic Computing, which is of interest of young talented students at the Pontificia Universidad Javeriana in Bogotá, Colombia. We will be attentive to future videos like this. CONGRATULATIONS.

  • @zvisger
    @zvisger Před 22 dny +1

    This was so entertaining, always happy to see a New Mind video drop

  • @DaveEtchells
    @DaveEtchells Před 22 dny +1

    This is *THE* best presentation on AI architecture I’ve ever seen, by a wide margin! (!)
    Megaprops for the depth, breadth and clarity, I’ve been paying a lot of attention to AI over the last couple of years, but still learned things, and this clarified my understanding of others. Great job!

  • @Kyzyl_Tuva
    @Kyzyl_Tuva Před 22 dny

    Fantastic overview. Big difference between training and inference.

  • @minhashyder
    @minhashyder Před 22 dny +18

    So this is what chatGPT said:
    “No, it is not true that each inference request consumes 300 watt-hours (Wh) of energy. That figure seems excessively high. Running inference tasks typically takes only a few seconds on a powerful GPU like the Nvidia 4080, and the energy consumption would be in the range of watt-seconds (joules), not watt-hours.
    For example, if an inference takes 5 seconds and the GPU is operating at 85% of its 320-watt maximum power, the energy consumption would be:
    \[ \text{Energy} = \text{Power} \times \text{Time} = 0.85 \times 320 \, \text{W} \times \frac{5}{3600} \, \text{hours} = 0.378 \, \text{Wh} \]
    So, the actual energy consumption for a single inference is much lower, in the range of a few watt-hours, not hundreds. The misunderstanding likely comes from a confusion between power (watts) and energy (watt-hours) and an overestimation of the duration or intensity of the workload.”

    • @eSKAone-
      @eSKAone- Před 22 dny +1

      Maybe the calculation incorporates energy consumption for training the models. But then it would get cheaper over time also.

    • @ShpanMan
      @ShpanMan Před 21 dnem +5

      It's pretty funny that AI debunks AI experts nowadays 😂

    • @phuongdo3525
      @phuongdo3525 Před 21 dnem +1

      Also keep in mind that inference for model like GPT-4 uses a huge cluster of GPUs, not just one, my napkin math suggests GPT-4 takes 45 A100 to run, 300Wh is still an over-estimation.

    • @adamrak7560
      @adamrak7560 Před 21 dnem

      The 300Wh may be for the entire batch, because these inferencing system are optimised for batch processing.
      But if the batch size is for example 256, which could be very realistic, then it is actually close to 1Wh per prompt, which is a more realistic number.
      I guess OpenAI on average can only fill their batches to half capacity during real-time processing. But that is still a very good fill rate. I base this on that their offline batch processing is only half cost. And they are under price pressure to pass on any savings of the offline processing to the users, because offline processing is less practical in most of the cases anyway.

    • @trenvert123
      @trenvert123 Před 14 dny

      The AI: " the energy consumption would be in the range of watt-seconds (joules), not watt-hours"
      Also the AI: " energy consumption for a single inference is much lower, in the range of a few watt-hours"

  • @brll5733
    @brll5733 Před 7 dny

    What a fantastic overview amd explanation of the ML field! Very well done!

  • @Citrusautomaton
    @Citrusautomaton Před 18 dny +1

    Neuromorphic is the future of AI. It’s something i said to myself the first time i learned about it back in 2021. I’ve always been obsessed with the concept of AGI and the benefits it can bring (and how to mitigate the risk)! To me, it always seemed like our conventional hardware and software wasn’t going to work in the long-term.
    An interesting thought i’ve always considered, is that when AGI is eventually created (could be as little as a year away), said AGI system could be integrated into advanced simulation software. Imagine an AGI system using simulations to discover and determine a room-temp superconducting material! Imagine an AGI system that troubleshoots 100 years of nuclear fusion reactor designs in less than a second! Imagine an AGI system that designs a compact quantum computer that’s lightyears ahead of our current tech, integrates it into itself, and then makes a 1-1 scale model of a human down to the molecules for illness prevention and understanding!
    We are encroaching upon the most important time in human history. If we can stop ourselves from weaponizing AI long enough, we may see the creation of an aligned ASI that’s happy to give its parents a nice retirement!❤

  • @thesimplicitylifestyle
    @thesimplicitylifestyle Před 22 dny +4

    We need a digital/analog hybrid solution as well as data and compute centers that run on renewables, at least until we figure out fusion power. 😎🤖

  • @francoiswagnier6148
    @francoiswagnier6148 Před 20 dny

    Thank you, you made the topic very digestible! And thanks to the commenters who pointed out the erroneous power consumption number.

  • @guard13007
    @guard13007 Před 22 dny +2

    A lot of people are complaining about the 300Wh figure, but they seem to be justifying it based on running small local models rather than industry-scale applications. I'm going to assume you're using the same source I found, in which someone tried to calculate a request's cost by using Sam Altman's claim of a worst-case cost of $0.09/query and power costing $0.15/kWh of power (they were using Euros but that doesn't make much difference or sense for this really). For some reason, they threw in an assumption that this estimate was double the actual cost, arriving at 300Wh. (I'm guessing that they decided halving the cost was a reasonable estimate instead of just using the worst-case estimate.)
    There's not much issue with this estimates.. but it is over a year out of date and it's not clear which model is specifically referred to. There is no published parameter count for GPT-3.5-Turbo or ChatGPT, but multiple estimates have placed it in the range of 20-22 billion. GPT-3 has 175 billion parameters. GPT-4 has been said to contain 1 trillion or 175 trillion parameters. GPT-4o hasn't had anything about parameters mentioned.
    Since there seems to be a lot of assuming that reasonable critique is just spreading fear and doubt, I think a reasonable thing to do is try to come up with a best-case estimate for power consumption: If we pretend that 300Wh estimate is valid only for GPT-3 running on less efficient hardware, and say GPT-4o is a heavily optimized version of GPT-3.5-Turbo, it could be reasonable to say that requests cost less now.. but that's making a lot of assumptions.
    (Looking at the calculation in @youtou252's comment: GPT-4o has a 128k token context window, and outputs up to 2048 tokens, not 1000. Their estimate on power usage may be accurate for first-query of a session, but ignores how quickly the calculations required expands as the conversation continues. They are off by about 100x, conveniently the difference in orders of magnitude between the estimates given by Sam Altman & used in this video.)
    Ultimately, the power usage is a real problem that has been said by many.. it isn't just spreading fear, it's a real problem to be solved.

  • @sheggle
    @sheggle Před 22 dny +6

    The gpt4 request energy estimate has to be bullshit, with 1kw per gpu and in the ballpark of 16/32 gpus serving probably at least a hundred users, a request would need to take an entire hour. I'm probably not 2 orders of magnitude off, which is the offset needed to make the 300Wh number make sense

    • @NewMind
      @NewMind  Před 22 dny +7

      It’s a mistake, working on a correction.

    • @eSKAone-
      @eSKAone- Před 22 dny +3

      ​@@NewMindMaybe the calculation incorporates energy consumption for training the models. But then it would get cheaper over time also.

  • @thomasfokas
    @thomasfokas Před 22 dny

    Great video. Curious what software you used to create the motion graphics?

  • @necrosis9222
    @necrosis9222 Před 22 dny

    Very well made video as always

  • @benmcreynolds8581
    @benmcreynolds8581 Před 22 dny +4

    This topic makes me think about "Scavengers Reign" an animated sci-fi nature space show. I highly recommend watching season 1 if you get the chance. The way they convey the biology there, the habitats & ecosystems are so well done. I really hope they do a second season. Plus we just need more animated shows to make a comeback. (I recently saw this very unique single pistoned motor, that rotates around a internal gear setup, that directly powers the rotating drive shaft. I sadly forgot what the name of this new motor is? It's fascinating and very well made. I Hope the design takes off & gains success)

    • @EeroafHeurlin
      @EeroafHeurlin Před 22 dny

      Avadi; czcams.com/video/A_4iN1TZsM4/video.html it doesn't scale and is expensive to make but it's looking very promising for military UAVs.

  • @eliasmai6170
    @eliasmai6170 Před 21 dnem

    i always look forward to a new New Mind video.

  • @seditt5146
    @seditt5146 Před 22 dny +1

    Combining SNN with Convolutional or normal NN sees to be a basic issue of creating a DAC. Used in Electronics its a Digital to Analog converter. The SNN is the Digital on off signal and the more traditional NN are as analog as we can get them hence their ability to encode information so well.

  • @KCM25NJL
    @KCM25NJL Před 22 dny +1

    Had a good chat with Claude recently about a 3D Neuromorphic chip design that involved the UV lithographic printing of quantum dot based fabric. The die would be built in an interconnected three dimensional topology with an overprovision of dots to maintain a locally clustered availability of neurons that would facilitate neuro plasticity based on the problem space(s) it was applied to. i.e. Neurons would turn on and off temporally and geographically based on the solution space at the time. The idea being that the quantum dot could hold a membrane signal in the form of light and light retention, this generating very little heat per unit of computation and is highly suitable for spiking operations.

    • @ohioplayer-bl9em
      @ohioplayer-bl9em Před 22 dny

      That’s a brain dude. I don’t think humanity will
      Last in its current state long enough to achieve that. If you are right and they are getting close to this our time is over and the reset button will be hit. Meaning society is about to be destroyed rewinding thousands of years of technology and we will be chasing prey with spears in short order.

  • @Rawi888
    @Rawi888 Před 21 dnem +1

    I love this video, it didn't care to baby me. I won't lie I got COMPLETELY lost in the first 4 mins and I'm a CompSci student. I'm just going to absorb what I can and continue to self educate to fill in the gaps.

  • @peterbonnema8913
    @peterbonnema8913 Před 22 dny

    This video is amazing. I learned sooo much

  • @bovanshi6564
    @bovanshi6564 Před 22 dny +2

    Simply outstanding video. Algorithm, do your thing!

  • @yunyunvipassi2505
    @yunyunvipassi2505 Před 18 dny

    Amazing short visual tellstory, just mindblowing

  • @skillsandhonor4640
    @skillsandhonor4640 Před 22 dny

    that was a really good video thank you

  • @EdwardPike
    @EdwardPike Před 21 dnem

    These are awesome visualizations.

  • @HershD
    @HershD Před 22 dny

    Brilliant thanks 👍

  • @ChristProg
    @ChristProg Před 21 dnem

    Thank you Excellent Video

  • @elon-69-musk
    @elon-69-musk Před 22 dny +2

    I have never seen such 3D visualization of neural networks extremely cool 🤘👍💘

  • @circusitch
    @circusitch Před 21 dnem +4

    Every minute was so interesting, despite the fact that I didn’t understand anything after the first 45 seconds.

    • @NewMind
      @NewMind  Před 20 dny

      That’s ok neither did I

    • @bernios3446
      @bernios3446 Před 16 dny

      @@NewMind How is that - didn't you write it?

    • @therealOXOC
      @therealOXOC Před 16 dny

      @@bernios3446 no one knows how this shit works

  • @defeatSpace
    @defeatSpace Před 11 dny

    That IBM research might be the last thing NVIDIA wants to hear after launching Blackwell 😭

  • @Yezpahr
    @Yezpahr Před 22 dny +2

    You've come a long way bro. I remember when it was just 35k subscribers and the comment section was throwing out numbers to estimate how quickly you'd get to 1M subscribers.
    CZcams is a harsh marathon, but you're still in the race at least.

    • @bovanshi6564
      @bovanshi6564 Před 22 dny +3

      It was kind of crazy. I think the algorithm found the The science of roundness video back then and you could basically see subscribers double and triple over a week.

    • @Yezpahr
      @Yezpahr Před 22 dny +1

      @@bovanshi6564 Fond memories xD. I remember that very well.

    • @NewMind
      @NewMind  Před 22 dny +5

      It’s been an incredible journey. Im so grateful for you and the other OG subs for taking it this far.

  • @VMac-eg7fb
    @VMac-eg7fb Před 22 dny

    During dark room meditation I have observed similar images gently moving, traveling, swirling through my mind such as I witnessed at 14.32 minutes to 14.57. They all appeared more like the final images at 14.54 to 14.57 minutes in this video. Even though I had non-specific thoughts during my meditations the greyish blue images moved continually as I observed them.

  • @go_djentle
    @go_djentle Před 22 dny +2

    I believe organoid based AI computers are probably the best way around this. Scientists have already been studying computer hybrids with human brain organoid (from cells).

    • @thesquee1838
      @thesquee1838 Před 8 dny

      Do they grow brain cells separately or is it like a brain machine interface in a human?

  • @ps3301
    @ps3301 Před 22 dny +1

    How does liquid neural network compare to spiking neural network ? It also has a time constant element built into the architecture?

  • @Cwarner251
    @Cwarner251 Před 21 dnem

    Thanks for joining the team.

  • @jichaelmorgan3796
    @jichaelmorgan3796 Před 22 dny +1

    Small biological neural networks could be used as a simple emotion module for an ai system

  • @edogelbard1901
    @edogelbard1901 Před 22 dny +10

    MD here with background in neuroscience. We barely function, and that speaks to the difficulty of the task of simply existing, but we are better at any non-linear operation than AI is, which is what we want from AI. We can train models to be very good at a thing, but as for all the things required of being an active agent, brains are just better. It took many millions of years to develop the feedback loops and programs needed to keep us connected to reality

    • @eSKAone-
      @eSKAone- Před 22 dny

      Did you see the multimodal demo of GPT-4o at OpenAI's channel?

    • @kealeradecal6091
      @kealeradecal6091 Před 22 dny +2

      @@eSKAone- demo is just an advertisement, better to put in real world trials with 3rd party

    • @Youtube_Cribbit
      @Youtube_Cribbit Před 21 dnem

      Do you believe that any sufficient number of incredibly informationally dense modalities, supported by a system that can provide the required amount of compute to take on this operation of course, would be that indistinguishably different from how sophisticated humans are at interpreting, and interfacing with reality?

    • @galewallblanco8184
      @galewallblanco8184 Před 21 dnem

      @@kealeradecal6091 its already available to use and interact with, for a lot of people, they have been rolling out GPT-4o slowly for EVERYONE ever since they announced it, I can already talk to it on my free subscription

    • @edogelbard1901
      @edogelbard1901 Před 21 dnem

      @@CZcams_Cribbit when you attack the root of the meaning of that question, the very concept of information itself is blurred. We don't operate through a logic system with discrete bundles and processes in the same way. We operate through feedback loops with different brain nuclei that interperet and manipulate inputs differently, only graduating to real information and meaning when there is communication between them. One region of the brain may send another region a potential spike, and that feedback may be interperted as potato or book, depending on another brain region's inputs. The brain will light up the same way for the end result of potato, but there may be thousands of ways to eventually collapse onto it.
      TLDR there are very few locations in the brain where we can translate between logic and and the brain, and those require a lot of feedback to train with to use. E.g. attaching electrodes to motor cortex and translating potential spikes into motion of a mouse.
      A sufficiently sophisticated logical system can reasonably simulate intelligence and pass a turing test, because we are easily fooled. It would not be able to go onto living as an active agent in a way that benefits its survival, nor would it be able to endure the real challenges of living and interpereting reality. the more "creative" you make it, the more likely that it will experience a critical lapse in judgement. We have some similar issues, but millions and millions of years of natural selection oopsing bad choices out of reality, to help use better determine what reality is and how it affects us.
      TLDR #2: i dont think AI can hack it as a real consciousness because logic systems are worse at buffering reality the way feedback loops do, and also because of lack of experience both for the AI and the programmers trying to write those programs that we have earned over eons of evolution.

  • @NZTritium
    @NZTritium Před 22 dny +3

    18:30 holy epilepsy

  • @TropicalCoder
    @TropicalCoder Před 20 dny

    This was a real gem, with beautiful graphics. I watched it twice to let the information sink in. No doubt the future is neuromorphic!

  • @MMMM-sv1lk
    @MMMM-sv1lk Před 22 dny

    Spike train of a video you created, rapid fire, full of information. Proven yourself you have padawan. 🖖

  • @samik83
    @samik83 Před 22 dny +8

    Gives some great perspective on how AI is just taking it's first baby steps. Still amazingly inefficient.

    • @kealeradecal6091
      @kealeradecal6091 Před 22 dny +2

      yet it is hyped by techbros previously known as crypto bros

  • @jayeifler8812
    @jayeifler8812 Před 22 dny

    I will say when I studied graduate level mathematics in my forties for the first time, everytime I made progress in reading a proof, by brain immediately wanted more food. Meaning
    I had to eat while reading the proof. Makes you wonder if we could get more energy into our brains how they might perform.

  • @scottbrady6240
    @scottbrady6240 Před 21 dnem

    LIKE THE CONVOLUTION REVERB IN ABLETON LIVE THAT TOOK IMPULSE RESPONSES FROM FAMOUS MUSIC HALLS AROUND THE WORLD AND THEN ALLOWED YOU TO HEAR WHATEVER SOUND YOUR PLAYING AS IF IT WERE PLAYED IN THAT FAMOUS SOUND HALL? SICK 🤙🏻

  • @ArmoredAnubis
    @ArmoredAnubis Před 12 dny +1

    This video overlooks many key aspects of modern inference architectures and the latest AI models. It doesn't accurately reflect the advances in energy efficiency and the significant decrease in energy costs associated with these newer technologies. Contemporary AI models, like those I use, are much more efficient and sustainable.

  • @jayeifler8812
    @jayeifler8812 Před 22 dny +1

    The whole point of what is going on with AI now is just we are not sure if we scale existing systems with their existing energy and power consumption if we won't get AGI. No one knows. The big money is partly not even concerned with this. But AGI may in fact be a result of the scaling as well and by God, even human existence may be created by scaling computation, people have always wondered this, we are just now finally there. It's hard for some to except the idea we are nothing more than simply a neural network scaled to large size with huge amounts of data pouring in--but I'd say it explains a lot.

  • @TiagoTiagoT
    @TiagoTiagoT Před 19 dny

    Why spiking can't be handled as time-integrated analogs, floats values representing the integrated spiking for a certain size of time-step; essentially smoothing out the behavior over time, and include the leaking behavior the same way as presented here, where a neuron needs sum of inputs that adds up faster than the leak drains them (which when you think of it, is pretty similar to RELU....)?

  • @weidergonga2997
    @weidergonga2997 Před 22 dny

    Great video

  • @Cwarner251
    @Cwarner251 Před 21 dnem

    Very informative

  • @Law0086
    @Law0086 Před 8 dny

    "I don't mean to bore you with tech, commander...."

  • @BooleanDisorder
    @BooleanDisorder Před 22 dny

    Is it possible to combine Spiking Neural Networks (SNNs) and Artificial Neural Networks (ANNs)? For example, CNNs can be combined with Transformers where the CNN handles initial feature extraction and then passes the processed information to the Transformer. Similarly, SNNs could be integrated with ANNs, where SNNs preprocess and extract features from visual data before passing it to an ANN for further processing.
    Could we extend this concept to create a hybrid model with SNN layers inside a Transformer architecture? The SNN layers would handle early feature extraction, leveraging their efficiency and temporal processing capabilities, and then feed these extracted features into the Transformer for high-level sequence modeling. This approach could combine the strengths of both networks, resulting in a more powerful and efficient system.
    I see this as a more viable approach. There are also chips coming that are specifically made for ANN's that will be much less inefficient.

  • @steveclark9934
    @steveclark9934 Před 22 dny +1

    Did you know? AI-based emulators can accelerate simulations by up to billions of times, depending on the scientific case. This means that a process that would normally take years can be completed in a fraction of the time, potentially reducing 20 years of training to mere days or even hours.😊

  • @jayd8935
    @jayd8935 Před 17 dny

    A single request cannot consume 300 watt hours, unless it took an hour to compute. It correctly will consume 300 watts (assuming that figure is correct) without time coming into it.

  • @unvergebeneid
    @unvergebeneid Před 22 dny +1

    Spiking neural networks have always bin kore of an interesting toy tbh. I don't remember any application from my academic days where they really showed any advantage.

    • @ronilevarez901
      @ronilevarez901 Před 22 dny +2

      They are the key. They've always been.
      And until very recently, traditional neural networks were also considered simple toys that would never become useful.. Yet now the same people who believed that is investing millions on them.
      We just need some Neuromorphic breakthroughs and bam! True AGI.
      And as always, investments in research are the key. Just wait.

    • @unvergebeneid
      @unvergebeneid Před 22 dny +3

      @@ronilevarez901 there was a period of "just use an SVM instead" but I wouldn't say ANNs were ever written off as a mere toys. But oh well, we'll see won't we? There are many other interesting developments in hardware support for non-spiking neural networks as well and I don't think anyone can be certain which technology wins out in the end.

  • @DavidPlaysGuitarSometimes

    I'm gonna be honest, I understood about 30% of this video

  • @herok4306
    @herok4306 Před 21 dnem

    Nature is the ultimate engineer & scientist hats off.. Natural systems are just mind boggling..

  • @KevinLarsson42
    @KevinLarsson42 Před 22 dny

    The introduction of fully independant AI integrated in products is both exciting and scary. Maybe in the future it will be so economically and energy-wise inexpensive to grant AI to products. Imagine how integrated and smart the world would be if, for every human made item can communicate with each other and with humans.
    Though this would only be possible if the energy requirement of the integrated AI is low enough to be run on built-in solar panels or rechargable batteries.

  • @quboqut
    @quboqut Před 22 dny +1

    think the reason why ai will not replace human art is not because it couldn't but because humans ENJOY making art! there will be some type of movement in order to keep promoting human creativity. It'd be very dumb to let a robot do what you like to do. Ai is for what you DO NOT like to do. Like if AI could reproduce for us are we going to give up sex?

    • @shamicentertainment1262
      @shamicentertainment1262 Před 21 dnem

      I think I agree with this. Probably the first time I’ve seen someone make this comment actually. we can already automate music by playing a recording through loud speakers, but people stuff love going to live shows.

  • @yeshes.online
    @yeshes.online Před 22 dny

    Best argument for open solar energy I've seen, and likely the one most probable to smack the energy lobby and interests in the teeth and move beyond their desire to throttle the alternative for humanity. The solution for the spiking neural network tech seems obvious, albeit unsuitable on face: The machine must sleep. (Inspired me to write on this today... thank you... concept now blogged. :) )

  • @skierpage
    @skierpage Před 19 dny

    21:43 a Halapoint system "with over 2,300 embedded Intel x86 processors for ancillary computations" is going to consume way more than 2,600 Watts. Intel has never achieved lower than 1.5 Watts per core.. I think Intel is saying that the 1,152 Loihi 2 processors _alone_ consume only 2,600 Watts.

  • @Atomchild
    @Atomchild Před 16 dny

    The next phase of artificial intelligence will be models designed to infer rather than generate. Transformation models seem likely candidates for high fidelity inferences. Instead of predicting what the next word will be, inferring conversational AI will be predicting what the next concept will be. conceptualizing AI will use tokens as notions and concepts instead of words or phrases. The solutions are only a few dimensions away.

  • @uriituw
    @uriituw Před 21 dnem

    This videos are amazing.

  • @moritzbecker4246
    @moritzbecker4246 Před 21 dnem

    The estimated 300 Wh seems to be incorrect, maybe you meant Ws (Watt seconds)?
    That would mean a difference by a factor of 3600. According to a few sources that seems to be more likely.

  • @gppg1799
    @gppg1799 Před 22 dny

    How great is the Video!

  • @Wlodzislaw
    @Wlodzislaw Před 21 dnem

    Iphone needs about 4 Ah at 4 volts, so about 16 Wh to charge, so 300 Wh is less than 20 times energy required to charge iPhones, not 60. Same with other phones. Still a lot. Fast charging on an iPhone 12 Pro (with a 4,082 mAh battery) typically requires around 20-30 watts of power to charge from 0-50% in about 30 minutes. So considering losses during charging it may be only 10-15 times,

  • @courtlaw1
    @courtlaw1 Před 22 dny

    I remember reading about memristors being the future many years ago.

  • @superfly1321
    @superfly1321 Před 21 dnem

    Amazing video

  • @remiheneault8208
    @remiheneault8208 Před 13 dny

    I develop ML models, mostly CNNs but also autoencoders and transformers. I was expecting an oversimplification of how NN work but everything was surprinsingly accurate throughout the first part of the video. Apart from the fact that inference is orders of magnitude cheaper than training.

  • @hisuiibmpower4
    @hisuiibmpower4 Před 18 dny

    why not use fft result to represent time domain signal(spike train)?

  • @projectcontractors
    @projectcontractors Před 22 dny +2

    "it's funny you know all these AI 'weights'. they're just basically numbers in a comma separated value file and that's our digital God, a CSV file." ~Elon Musk. 12/2023

    • @galewallblanco8184
      @galewallblanco8184 Před 21 dnem

      Everything can be weights in a massive CSV file, including images, brain scans, audio, games, anything

  • @ThePhiphler
    @ThePhiphler Před 22 dny +3

    Nice of you to feature my home town of Gothenburg at 0:57

  • @catnium
    @catnium Před 20 dny +1

    ai is gonna be the end of us all not by a terminator scenario but with global warming

  • @timeflex
    @timeflex Před 21 dnem +1

    The brain of a single human, no matter how smart, can not be considered on its own without the rest of civilization. Hence, to calculate the actual amount of energy required for that single thought you must also take into account the energy consumed by all the people who have contributed towards the development of that individual (at least).

  • @egycg3569
    @egycg3569 Před 21 dnem

    All is well and good except that Sweden electricity consumption is that number you said but in Tera not Giga.

  • @PremiumLeo
    @PremiumLeo Před 9 dny

    I understood none of that, but it was a great tutorial on understanding something I didn't understand 😁

  • @scotty3114
    @scotty3114 Před 22 dny

    Our brains have had eons to develop and the AI are just starting out. Plus the person commenting on build and train time is entirely on point.

  • @puffinjuice
    @puffinjuice Před 22 dny

    I see the patterns of spike neurons when I close my eyes

  • @azhuransmx126
    @azhuransmx126 Před 22 dny +8

    Bro, to make 30 questions to bing or gpt consumes 0.45Lt of water, but 300W per question is madness. It must be more like 3W, even so it is still a lot. But in the future this problem will be solved with Edge Computing, Native Multimodality with no Latency, Asinchronous Neuromorphic chips, 1nm process is key for multiplying the army of Transistors from the current 208 billion in the 4nm Blackwell to 4 Trillion at 1nm allowing 75% of the chip area to be just turned off without losing too much Flops of processing power, non volatile Resistive Memristors memories that not requires energy to maintain the 1 alive and low precision models working on fs2 or fs1. Without mentioning Photonic chips. With all these systemic measures, the human brain low energy consumption in TDP performance should be matched by AGI in 2029.

    • @eSKAone-
      @eSKAone- Před 22 dny +1

      Exactly

    • @azhuransmx126
      @azhuransmx126 Před 22 dny

      @@eSKAone- however Amazing video 👍

    • @lukeryan6278
      @lukeryan6278 Před 20 dny

      300w per hour 1.5 seconds to answer question. 60 iPhones is an exaggeration

  • @rothn2
    @rothn2 Před 22 dny

    2.6 kW with 1/24th of the neural network size of a high-end consumer GPU? It's interesting but we're far from even being as efficient as the matrix-based gradient descent approach. Multiple order-of-magnitude improvements are required to even catch up, and additional improvements would be required to supersede what we have now and be worth the cost of adopting a new and very different technology.
    I want to see this succeed and think it would be cool, but there also seem to be a lot of obstacles. It seems like this may take quite some time.

  • @NextGenAge
    @NextGenAge Před 22 dny

    In the perspective of the brain, it uses 80% of your entire body energy. So if AI can be seen as our collective mind, then the whole idea that it is using too much energy is nonsense. Surely it can be further optimized, but nature shows that intelligent thinking is more important than any other physical process in our body or at least equally important!

  • @empatikokumalar8202
    @empatikokumalar8202 Před 22 dny

    I have a question for everyone who watches this video. Let's say you prevented such a huge energy expenditure. You have developed a processor and artificial intelligence model that consumes much less energy and makes it much faster. How will you solve the security problems it will create? For example, deciphering passwords and passing digital security with such high speed is not even a task. Developments in the field of security are not as rapid as in the field of artificial intelligence. Such terrible power always poses serious danger in the hands of those with evil intentions. In this case, would you prefer a very advanced processor (hence artificial intelligence) or security?

  • @AS40143
    @AS40143 Před 22 dny

    The spiking artificial neural networks is a very old idea

  • @user-nt5wt8tr7n
    @user-nt5wt8tr7n Před 17 dny

    the cooling is more power consumptive than the actual servers

  • @alexandrustefanmiron7723

    20Wh. That is how much your brain consumes and, for most of us, comes with the elusive feature called common sense!

  • @mm-rj3vo
    @mm-rj3vo Před 17 dny

    I worry deeply about conscious AI rights, in the presence of the world's current failed ethical and economic systems.
    Humanity is NOT ready for conscious AI, not because it will try to hurt us, but because we will likely start, at the onset, by enslaving it.

  • @jacksonnc8877
    @jacksonnc8877 Před 20 dny

    The one equation that still needs to be incorporated. consciousness!

  • @WarmWeatherGuy
    @WarmWeatherGuy Před 13 dny

    I signed up with Brilliant specifically for neural networks. It started off painfully slow with stuff I already knew. Then I started just punching through to get to something I didn't know. I didn't care about my score. Finally I zipped through a lesson that seemed like it was important so I wanted to start it over but it wouldn't let me. I could find no way to retake it. I guess they don't want people to cheat to get a high score. Brilliant sucks. I'm done with it.

  • @amuaiz
    @amuaiz Před 19 dny

    Should ask AI on how to solve this energy limitation problem

  • @RonakDhakan
    @RonakDhakan Před 20 dny +1

    07:40 In 2022 not in 2020.