What are Tensor Cores?

Sdílet
Vložit
  • čas přidán 4. 07. 2018
  • Here's the Nvidia blog explaining Tensor Cores in great detail: devblogs.nvidia.com/programmi...
    Support us on Amazon! geni.us/RJ1nymB
    Newegg link for cool kids: bit.ly/2AkBuRt
    Join our public Discord server with Tech YES City! / discord
    ▶ Our studio gear on Amazon:
    Sony A7 III (Primary): geni.us/VNQk
    Sony A7S II (Secondary): geni.us/nGAGx
    2x Pixio PX275h IPS Monitors: geni.us/MJISRP0
    Neewer LED Lighting Kit: geni.us/9VTXr0
    Audio-Technica AT2035: geni.us/xvQc
    Yamaha MG10XU Mixer: geni.us/d4VvGM
    Neewer Microphone Arms: geni.us/1w5Sx2Y
    Tribesigns L-Shaped Computer Desks: geni.us/ZMHUcU
    Arrowzoom Sound-Damping Foam Panels: geni.us/EKFxZc
    TWITCH: / shiftystudioyt
    FACEBOOK: / sciencestudioyt
    TWITTER: / gregsalazaryt
    INSTAGRAM: / gregsalazaryt
    Subscribe to our channel!
    MUSIC:
    'Orion' by Sundriver
    Provided by Silk Music
    / silkmusic
    DISCLOSURES: All Genius links are tied to our Amazon Associate account, from which we earn a small sales commission. Links containing a 'bit.ly' reference forwarding to Newegg are tied to our CJ account, from which we earn a small sales commission. All sponsored links and comments will contain the word "SPONSOR" or "AD." Any additional revenue stream will be disclosed with similar verbiage.

Komentáře • 235

  • @Arloh2222
    @Arloh2222 Před 5 lety +184

    Seems you made an error, around 2:33... You completely confused flops and FP32/FP16
    FP16 and FP32 stands for the Floating (point) Precison, basically how many bit's the value is stored in, not - flops, the theoretical max floating operations per seconds

    • @anandsaha01
      @anandsaha01 Před 5 lety +9

      +Science Studio, you should put a message in the video with the correction.

    • @njr940
      @njr940 Před 5 lety +3

      Also consumer grade graphics cards do come with tensor cores now :P

    • @JB-ym4up
      @JB-ym4up Před 5 lety

      This+1.
      worse yet he made something up to cover.

    • @MrJonojenkins
      @MrJonojenkins Před 4 lety

      He makes lot of these type of mistakes in his videos.

    • @tuck1726
      @tuck1726 Před 4 lety +2

      I noticed that too he doesn’t know what single and double precision floating points are. Aren’t graphics cards measured in giga or tera-flops? 16 floating point operations per second would be awful.

  • @TechLevelUpOfficial
    @TechLevelUpOfficial Před 2 lety +24

    4:38 owh that didn't age well, we have Tensor cores in all Geforce RTX cards now lol

  • @zamalek4079
    @zamalek4079 Před 5 lety +31

    2:35. FP16 and FP32 refer to 16bit Floating Point number and 32bit Floating Point number. They are often called "half" and "single" respectively (there is also "double" which is 64bit, but is not really useful in AI). The reason for the inverted acronym is because nearly all programming languages typically require that a name start with a letter (but can contain numbers), so 32FP would be an error in most languages. It is not a measure of operations/FLOP because that would measure nothing interesting, it is a measure of precision. They determine the number of unique values that can be represented. FP32 is actually *slower,* because more transistors are involved in the calculation. You'd typically use FP16 because of the improved performance, only using FP32 if you needed the precision (which is extremely rare). This would be the exact opposite of what you'd expect if they stood for 16 FLOP and 32 FLOP, the latter would be more operations (ignoring that there is no time unit, again, a strange unit of measurement in this context). en.wikipedia.org/wiki/IEEE_754

  • @KnifeChampion
    @KnifeChampion Před 5 lety +120

    "Should we expect to see tensor cores in consumer grade graphics cards? Dont count on it."
    Who else is watching this after RTX reveal? xd

    • @tylerdurden3722
      @tylerdurden3722 Před 5 lety +3

      Lol...I guess no one saw this coming.

    • @dans.8198
      @dans.8198 Před 5 lety

      @Ashwin Mouton: No one saw this coming ? AI is the next big thing, and the iPhone has got a similar hardware “Neural Engine” since the iPhone X released at the end of 2017.

    • @wildkev1010
      @wildkev1010 Před 4 lety

      The RTX 2060 has 240 Tensor cores

    • @hemanthkumarHere
      @hemanthkumarHere Před 4 lety +3

      and they find a potential application to boost gaming performance with it as well. DLSS 2.0

    • @ClarkPotter
      @ClarkPotter Před 3 lety

      @@dans.8198 another way to say "has got" is "has."

  • @Myvoetisseer
    @Myvoetisseer Před 5 lety +105

    Google is really going too far. Usually when I start googling something new to learn about it, I would have dozens of adds or suggestions thrown my way for the next couple of days. But this time, Google actually commissioned a guy on CZcams I've been watching to make a VIDEO explaining the concept I've been trying to understand. Wow, that's just freaky.

    • @ChandlerUSMC
      @ChandlerUSMC Před 5 lety +2

      Google then followed up with a call to Hugo's cell phone leaving a message asking him why he stepped away from his computer after watching the video. He didn't call them back so they reached out to his parents followed by his best friend from elementary school.

    • @hercion
      @hercion Před 5 lety +2

      Let's talk about cross device tracking. They really go too far. If I watch bikinis ( I am human, after all) those videos shall not show up in my other device where I Focus on science and tech ..
      Well CZcams is a monopolistic service. What's else? Vimeo? Google is a monster.

    • @paranoyd1
      @paranoyd1 Před 2 lety

      I don’t think it’s because you searched, or that he was even commissioned in the first place to make the video, but yeah we never know xD

  • @Chemy.
    @Chemy. Před 5 lety +26

    So it's mostly a different way to solve problems using better resources for a specific kind of operations, I watched the video two times to see if I got it right, great one Greg.
    Video liked as always

  • @SSS-mp8th
    @SSS-mp8th Před 5 lety

    I love the minute science videos. It's one of the main aspects of differentiation you have compared to other tech tubers. I love the reviews and builds as much as the next guy but learning about what goes into the hardware or software is really interesting and not many do it.
    You video on nano-meter in regards to CPUs was really informative.

  • @mattgraves3709
    @mattgraves3709 Před 4 lety +1

    You sir are a worthy adversary.
    I absolutely love watching your videos thank you so much for explaining this term

  • @faezlimpbizkit
    @faezlimpbizkit Před 5 lety

    The only youtuber that can takes him time to actually explain the engineering behind these stuff. Thanks Greg :)

  • @brockfg
    @brockfg Před 5 lety

    I use my 1080 ti for tensorflow ML on a daily basis so I was excited to see you had a video related to ML. Great video !
    -UL alum

  • @JP_Stone
    @JP_Stone Před 5 lety +3

    Really like the Techie videos. Found this one exceptionally interesting. Would love to see a video on why GPUs would not make good CPUs and Visa-Versa and the differences between them on an architectural level.

  • @johnf.354
    @johnf.354 Před 5 lety +1

    I really appreciate the way you’re able to reduce these ridiculously complex descriptions into moderately complex examples. I still have no idea what a Tensor Core is/ does because I was just staring at the RGB in the PC behind you...

  • @josephbaker9673
    @josephbaker9673 Před 9 měsíci +3

    Greg: Tensor cores are not likely to be in consumer grade GPU's any time soon
    NVidia: Hold my drink

  • @spokova
    @spokova Před 2 lety

    Thanks for Teaching us..!
    Good Job Sir

  • @grantwilliams630
    @grantwilliams630 Před 5 lety +1

    Hey thanks for the video! I work in ML so its kind of fun to see videos like these pop up on my feed.

  • @zeke7515
    @zeke7515 Před 5 lety +64

    They need to start making relaxor cores. They might chill out at Nvidia and drop some new GPU's

    • @rtv190
      @rtv190 Před 5 lety +4

      that's a joke worthy of a laugh track

    • @React2Quick
      @React2Quick Před 5 lety +1

      rtv190 are we talking about a laugh track that's used in moderation like in old TV shows? Or an overused laugh track in current TV shows?

  • @CPMEDIACREATION
    @CPMEDIACREATION Před 5 lety +16

    not gonna lie this video was loooking spot on as hell, the camera is just making everything look so clean and crisp

  • @xXDarthBagginsXx
    @xXDarthBagginsXx Před 5 lety

    Great explanation, love these informative videos man 👍

  • @SALSN
    @SALSN Před 5 lety +6

    A few corrections/clarifications, the matrices, as the text says, are 4x4x4 i.e. 3 dimensional. (not 4x4 i.e. 2 dimensional).
    And without knowing for sure, I will bet that FP16 and FP32 are referring to 16 bit (two bytes) and 32 bit (4 bytes) precision rather than anything to do with speed.

    • @hantzleyaudate7697
      @hantzleyaudate7697 Před 5 lety

      Simon Als Nielsen you are correct

    • @udust1
      @udust1 Před rokem

      A matrix is 2D by definition. There is no such thing as a 3 dimensional matrix. Tensors can be 3d. But tensor cores are dealing with 2d matrices, and should really be called matrix cores not tensor cores, but sounds better for marketing. The FP part is indeed a mistake.
      tbh i dont really understand where how they can say that the "processing array" is 4x4x4. Multiplying two 4x4 matrices produces a 4x4 matrix, and adding two 4x4 matrices produces a 4x4 matrix. Where does the extra dimension come in?

    • @udust1
      @udust1 Před rokem

      Actually when I think about it, each node in the 4x4x4 space probably represents one multiplication result each. So for each row*col you need 4x multiplication-accumulate to complete the calculation. Still we are only dealing with matrices (2d vectors), its just that you need 4x the matrix dimensions in muliply-accumulate operations. Imo visualizing the multiplication in 3d is just confusing, as its all happening in 2d. Atleast it confused me.

  • @megalopath
    @megalopath Před 5 lety +7

    Excellent video! I watched the Google I/O on Tensor Cores and didn't quite get it (or at least recall it) beyond that it's good for machine learning. This video made it super easy to understand! Excellent job!!!

  • @Epiphany-cl7fe
    @Epiphany-cl7fe Před 5 lety +3

    I Love to learn...more qna if you get time..I listen to it while im at work..keeps me motivated..sounds crazy... but it really does

    • @GregSalazar
      @GregSalazar  Před 5 lety

      Thanks for the support, Paul.

    • @davidjenkins852
      @davidjenkins852 Před 5 lety +1

      Science Studio you're one of very few tech tubers who keep me motivated to learn too

  • @therealnmg
    @therealnmg Před 5 lety +1

    Man I miss your channel so much!

  • @dorryoku919
    @dorryoku919 Před 5 lety +4

    You must be reading my mind because I've been looking for videos on tensor cores lately

  • @AlexiGVS
    @AlexiGVS Před 5 lety

    Thank you, great video!

  • @Vizeroy9
    @Vizeroy9 Před 5 lety +1

    FLOP is derived from FP, not the other way round.
    FP16 is a 16 bit number, where the decimal point can be at any position of the number - hence floating point.
    FLOP/s is just the measurement of how many operations with this kind of numbers can be done per second - but the single precision FLOP count refers to FP32 numbers - 32bit floating point numbers.

  • @TechDunk
    @TechDunk Před 5 lety

    Thanks! Wanted to know this :P

  • @patohernandez1637
    @patohernandez1637 Před 5 lety

    Great video, Greg!

  • @nathanas64
    @nathanas64 Před 4 lety

    good explanation

  • @gorepuppy
    @gorepuppy Před 5 lety

    Matrix math was super easy for me, and I have number dyslexia oddly. I had no idea a Tensor Core was pretty much the same as Googles 8 bit matrix single instruction ASIC processerror. Good video. Short and informative.

  • @TheBigBawsss
    @TheBigBawsss Před 5 lety +16

    Damn this is beautiful. Should do more on processing micro architectures. Also it'll be so cool if you could discuss the RTL implementations. Thanks. 💪🏿🦍🐅

  • @user-um9sl1kj6u
    @user-um9sl1kj6u Před 11 měsíci

    This is why each AI matrix models is 8 x 8 x 8 (512 cuda cores) and why Nvidia includes accuracy for32 bit, 16, and 8 (being the simplest and fastest, especially for high speed language models, vision, , or audio)
    In which case, it can do multiple models and correct for itself.
    This is also why the hopper architecture has over 16,000 GPU cores and 72 CPU cores.
    You have 132 matrix instances, with 2 CPU cores per matrix model, and 6 left over for overhead.
    Besides AI, this is also an excellent Data center CPU and GPU combined. Along with the memory.
    It greatly simplifies data center deployment

  • @norbertnagy4468
    @norbertnagy4468 Před 5 lety

    More vids like this please

  • @Pastor_virtual_Robson
    @Pastor_virtual_Robson Před 5 lety

    time line 2:38 , The FP16 and 32 are not Flop related. They are regarding the Floating Point bits representation, half precision = 16 bits and single precison =32 bits.

  • @tejasnaik14
    @tejasnaik14 Před 5 lety

    So what cores are likely to be used in the future for handling ray tracing in games?

  • @larryteslaspacexboringlawr739

    thank you for tensor core video, could you do a benchmark video of tensorflow on a rtx 2080 or rtx 2080ti or 2070?

  • @DavOlek_ua
    @DavOlek_ua Před 5 lety

    but for rendering and modelind tensor cores should be beter? is it some kind of ACIS?

  • @MohdAkmalZakiIO
    @MohdAkmalZakiIO Před 4 lety

    Tell us about CUDA Cores and how it differs from Tensor Cores. As far as I know, CUDA Cores also used for parallel processing in MLDL works.

  • @danbodine7754
    @danbodine7754 Před 5 lety +15

    Hearing dot product got me triggered about cross products.

  • @Maisonier
    @Maisonier Před 5 lety +1

    So there is any software (no games) where I can use this tensor cores? some IA that organizes my files, desk, answer my emails and more?

  • @Vladek16
    @Vladek16 Před 5 lety

    excellent video :D

  • @abvmoose87
    @abvmoose87 Před 4 lety

    Can the api/code libraries for CUDA be used for tensor cores?

  • @phantom1199
    @phantom1199 Před 5 lety +1

    but what about the new ray tracing that nvidia unveiled? isnt that meant to run on tensor cores?

  • @iceefrags8770
    @iceefrags8770 Před 5 lety +1

    +Science Studio so if an application that was previously written in cuda had functions in it that would be more efficiently run on tensor cores, would the driver automatically run those operations on the tensor cores, or would the developer need to rewrite their application to take advantage of them?

    • @EMANUELE42
      @EMANUELE42 Před 5 lety +2

      Say a certain application is written in CUDA 7 SDK (tensor cores 100% not supported, as there was no NVIDIA GPU utilizing them) and it uses tensors and related (mathematical) functions, even if you would upgrade the driver to the latest versions, it would not be able to take advantage of tensor cores running on such a tensor cores-equipped GPU.
      That would be because the latest CUDA SDK (with tensor core support) might be backwards compatible with (programming) functions called in older SDKs, but newer (programming) functions which take advantage of both the driver's tensor core support and GPU's tensor cores are never called.
      If the driver could automatically detect that certain (mathematical) functions can be accelerated using tensor cores instead of traditional CUDA cores, that would be the case were an older application can get an automatic performance boost using a tensor core-equipped GPU, but I doubt it's the way NVIDIA did this, as it would make the driver too large and complex
      (edit: humble opinion from a programmer, definitely not a CUDA engineer, but still handled similar things between different SDK versions)

  • @joechevy2035
    @joechevy2035 Před 5 lety +6

    Tensor cores sound like this saying from Bruce Lee. "I fear not the man who has practiced 10,000 kicks once, but I fear the man who has practiced one kick 10,000 times."

  • @arvindraghavan403
    @arvindraghavan403 Před 5 lety

    Thanks for this video and explaining things dude now i can brag about these things with my friends 😂😂😂

  • @mirmohammadsaadati6526
    @mirmohammadsaadati6526 Před 5 lety +1

    FP is not short for FLOPS, it's short for Floating Point. FP16 means Half Precision floating point number. FP32 means Single Precision floating point number. FP64 means Double Precision floating point number. The number 16, 32, and 64 refers to the number of bits that it takes in the memory to store floating point numbers with respect to their precision !!!

  • @AzVidsPro
    @AzVidsPro Před 5 lety

    What about Raytracimg for that sweet real-time dynamic. Lighting?

  • @diegomireles
    @diegomireles Před 5 lety

    I'm an electrical engineer and Tensor is a term we never use, that term is for civil and mechanical engineering. But it seems similar to a State-space representation matrix used in Dynamical Systems and Control.

  • @rjsantiago4740
    @rjsantiago4740 Před 5 lety

    Great video! One thing though. If the next gen cards (gtx 11 or 20 whatever they will call it) are supposed to support rtx/ray tracing, doesnt that tech use the tensor cores in order to function in real time (AI calculates most of the light rays and where they go as I understood it, though I could be wrong)? So wouldn't that mean they'd probably have some tensor cores? I know we know little about the tech and the next gen cards, but if I understood those rtx demos correctly I believe these new cards would actually have to have tensor cores in order to do it. Your thoughts? Just wanna know if I am understanding the rtx correctly or not.

    • @wing0zero
      @wing0zero Před 5 lety

      RJ Santiago you can use CUDA for ray traced effects just tensor is better.

  • @jupitersailing6911
    @jupitersailing6911 Před 5 lety

    This is a very interesting video, I wonder if NVIDIA or AMD will put R&D into a new kind of core for gaming?

  • @1nsaint
    @1nsaint Před 5 lety

    @sciencestudio I was hoping you were gonna mention real time ray tracing in game, that is supposedly being handled by the tensor cores, thats the reason i and sure many others were interresed in them for a gaming aspect. it was shown in a video called: The state of Ureal, during the developer conference

    • @wing0zero
      @wing0zero Před 5 lety

      INSAINT the new Metro game has ray traced lighting in, maybe shadows not sure, but that will be the first game to use RTX. Not on consoles though PC only.

  • @thefullsetup
    @thefullsetup Před 5 lety

    Thanks for making this video getting a bit fed up of all the rumor channels keep banging on about tensor cores being in next gen gaming cards.

  • @michaelvanrheede7225
    @michaelvanrheede7225 Před 3 lety

    So did we get tensor cores in consumer gpus?

  • @HokgiartoSaliem
    @HokgiartoSaliem Před 5 lety

    I hope Nvidia has tensor core on GT1130 or 2030 or lower segment but newer product althought it is not as fast as hi end one.
    Btw if CUDA was so good, why there are no program that can do simple HEVC endoding using CUDA? Mainly for older GPU than doesn't have NVENC that support HEVC encoding?

  • @Alfer230
    @Alfer230 Před 5 lety +1

    One correction, at 2:35 you refer to the "FP16 or FP32" as being an abbreviation for FLOPS - this is incorrect. In the document you are referencing, the FP in FP16 and FP32 stands for "Floating Point" and the 16 and 32 correspond to bits of precision.

  • @KryoTronic
    @KryoTronic Před 5 lety

    Love these videos please do moarrrrrrrrrr

  • @Special_Sharpie
    @Special_Sharpie Před 5 lety +7

    Ooh just got on my break!!

  • @sapphie132
    @sapphie132 Před 5 lety

    This makes me want a titan V.
    Not because I'd utilise its full potential with my dumb cuda programs, just because it's cool af

  • @rishabh8766
    @rishabh8766 Před 5 lety +9

    there will be tensor cores in gaming gpus i can bet on that. remember that real time ray tracing demo? nvidia developed these technologies to take advantage of tensor cores in games. cuda cores alone just cant handle ray tracing. why would they invest there money in something no one can use?

    • @LiLBitsDK
      @LiLBitsDK Před 5 lety +1

      because the world is more than gaming, for example AI as I mentioned

    • @Vamavid
      @Vamavid Před 5 lety +6

      You were right on the money! 👍🏿

    • @pradeepraghuraman4430
      @pradeepraghuraman4430 Před 5 lety

      Yep. There are now.

  • @jerryh559
    @jerryh559 Před 3 lety +2

    4:39 well..... rtx 2000 and rtx 3000 series have them. Used for DLSS and other AI

  • @JD_Mortal
    @JD_Mortal Před 4 lety

    So much for that "prediction"... Tensor cores are in consumer cards. 20xx-RTX, which was oddly out before this video was made. Though, honestly, Titan-V was a consumer card too. The non-consumer cards being Quadro cards. (Or whatever the Quadro-variant is actually named, to identify it as the VOLTA/TENSOR version.)

  • @Doudoudelamotte1902
    @Doudoudelamotte1902 Před 3 lety

    so a normal graphics card with a driver/firmware modification can create Tensore core? or is it physically different ?

    • @kevinbroderick3779
      @kevinbroderick3779 Před 3 lety

      Physically different. The transistors that forms it's logic are optimized for certain types of calculations.

  • @LimitIsIllusion
    @LimitIsIllusion Před 5 lety

    Great video... I guess. Let's just say I remember why I dropped out out of engineering school. I'm sure this made sense to someone. Wooooosh

  • @DinoTheThird
    @DinoTheThird Před 5 lety

    That’s a clean ass shirt bro 😍🙏🏽

  • @leonardbrkanac9150
    @leonardbrkanac9150 Před 5 lety

    You could mention that neural networks make extensive use of matrices

  • @hansbehrends438
    @hansbehrends438 Před 5 lety

    Very well. And now over to the Gamer's Nexus to learn about memory subtimings.

  • @potatomo9609
    @potatomo9609 Před 3 lety

    It's amazing how RTX is just becoming normal today now while only 2 years ago when this video came out, it still seemed like something totally unreachable.

  • @ananthu6
    @ananthu6 Před 5 lety

    hmm... when you pointed that finger multiple times , it reminded me of supreme leader allladin talking about enriched uranium... sorry i dunno why.. lol
    k will subscribe now... even though i dont understand much of it... a vid put together well... n u know yiur thing.. looking forward learn more from you.. :)

  • @92trdman
    @92trdman Před 5 lety

    The processing power never get enough (Because a multi-million market bussiness pushing it) But for how long ? We may never know....And BTW, the picture remind me of "Terminator 2" the movie !

  • @bikerbones4615
    @bikerbones4615 Před 5 lety

    What about Ray Tracing? From what I've heard it can only be done with tensor cores.

  • @FarrellMcGovern
    @FarrellMcGovern Před 5 lety +4

    Nice video! You might find it interesting that about a fifth of the Top 500 supercomputers in the world use Nvidia based Tesla GPUs, as I guess you would call them "math co-processors" for the supercomputers. Two of the systems use the new Volta cards. What *I* find really interesting is that they have Tesla Voltas running on IBM POWER CPU based systems, not Intel or AMD. Except for 26 systems, all of today's top 500 supercomputers run on Intel Xeon CPUs. Of the non-Intel CPUS, 22 are based on IBM POWER CPUs, 6 on the Sparc64 CPUs, two AMD Opterons, and two weird ones, named ShenWei, made by China, and said to originally be based upon the DEC Alpha CPU.
    I love digging around the Top 500 Supercomputer list's spreadsheet each time they are released...and the most recent one is notable for the total dominance of Linux as the operating system of Supercomputing. All of the Top 500 run Linux....No AIX, No Solaris, no HP-UX, and certainly no Windows!

    • @allenqueen
      @allenqueen Před 5 lety +1

      Farrell McGovern interesting digging. Anything else which you found?

    • @FarrellMcGovern
      @FarrellMcGovern Před 5 lety +1

      Well, there is another co-processor that is starting to be used, it's the Intel Xeon Phi. It is basically a version of the Xeon CPU that has added around it math units that were designed for GPU use, giving it superior specialized math capability, around an x86 core which makes it easier (in theory) to program. Only 7 of the Top 500 list use Xeon Phi co-processors, while 97 use Nvidia Tesla, 2 use Nvidia Volta, and 2 use a combined Tesla/Xeon Phi setup. There is one other co-processor, a proprietary accelerator made by a Japanese company PEZY Computing / Exascaler Inc., called the PEZY-SC2 and PEZY-SCnp, which I can find very little about in a language I can read.

    • @rudysal1429
      @rudysal1429 Před 5 lety

      Farrell McGovern gotta love those corrupt companies

  • @4.0.4
    @4.0.4 Před 5 lety

    I believe 1180 will have Tensor cores, for the same reason CUDA cores were introduced before all games used them. Machine learning is starting to become quite mainstream, maybe not in games but who knows, that could also happen!

    • @Rafael96xD
      @Rafael96xD Před 5 měsíci

      Aaaah yeah, the good old 2000 series, a lot has changed and its amazing.

  • @dhruvmini01
    @dhruvmini01 Před 5 lety

    Did the mathematics behind your petroleum engineering degree help you in understanding concepts like these within computer hardware technologies?

  • @space302
    @space302 Před rokem

    3:21 : " Tensor cores can handle 64 floating-point mixed precision operations per second ".
    How do you not realize how that's clearly false because it would be incredibly slow ?

  • @Natewu188
    @Natewu188 Před 5 lety +23

    The Engineering Explained of tech.

  • @springbok4015
    @springbok4015 Před 5 lety

    You need to add a correction for your mistake at 2:33 +Science Studio

  • @ivailomanolov6908
    @ivailomanolov6908 Před 3 lety

    FP 16 and FP 32 is not a flop . It is a floating point that tells you how precise your calculation will be . What are you talking about ?

  • @funbucket09
    @funbucket09 Před 5 lety

    Nice shirt Greg. It suits you. Lovely Colour :)

  • @tiagooliveira95
    @tiagooliveira95 Před 5 lety

    i want to game, but i also want to train my models, my CNNs and RNNs will appreciate every tensor core, soo, yep, bring more tensor cores to regular Gpus, not only to 2k+ gpus

  • @texasdeeslinglead2401
    @texasdeeslinglead2401 Před 5 lety

    Would love to see more graphical demonstrations on micro architecture.

  • @marvelv212
    @marvelv212 Před 5 lety

    So you can replace tensor cores with cuda cores when needed although not as efficient.

    • @bitscorpion4687
      @bitscorpion4687 Před 5 lety

      yep, there is a major difference in speed
      www.nvidia.com/en-us/data-center/tensorcore/

  • @PaulAllenPiano
    @PaulAllenPiano Před 5 lety

    Thanks for simplifying all this for us "Dummies"!

  • @STRAGGLER36
    @STRAGGLER36 Před 3 lety

    So, you mean efficiency of software can impact the efficiency of hardware? What are unique and interesting idea. Also I guess what you're saying is read nvidia's blog on what a tensor core actually is. Thanks

  • @brentgreeff1115
    @brentgreeff1115 Před 4 lety

    For someone starting to learn ML would you recommend a 2060 or 1070. - this seems to indicate that for ML, the tensor cores really count- towardsdatascience.com/rtx-2060-vs-gtx-1080ti-in-deep-learning-gpu-benchmarks-cheapest-rtx-vs-most-expensive-gtx-card-cd47cd9931d2

  • @B4dD0GGy
    @B4dD0GGy Před 5 lety

    32GB Titan V CEO Edition please, so I can open DOS prompt and type dir

  • @mcericx
    @mcericx Před rokem +1

    5 minutes video translated into one sentence: they are fixed function units

  • @junakkibria6380
    @junakkibria6380 Před 3 lety

    Only if Greg knew Nvidia's going to drop a bomb (DLSS) with this tensor cores a year later then he would be more emphasizing on this for sure XD

  • @cappuccino-1721
    @cappuccino-1721 Před 5 lety

    what about Ray Tracing? doesn't that need tensor core?

  • @woolfel
    @woolfel Před 5 lety

    The newer tensor cores are different now and supports 8bit and 4bit numbers.

  • @rickgrimezz3942
    @rickgrimezz3942 Před 4 lety

    got curious... and looked this up... came across this video... and then I heard him say @4:37 .... Seemed silly since all we get now are Nvidia cards with Tensor cores. Queue the SUPER SERIES.... the follow up to the 20 series cards all mainstream cards with tensor Cores... hehe! Found this comment silly!

  • @Mr.VaSimple
    @Mr.VaSimple Před 5 lety

    We need games with a great AI that will utilize Tensor Cores, so this cores may appear in consumer GTX videocards later.

  • @samuelkrusiewicz2828
    @samuelkrusiewicz2828 Před rokem

    This video needs its own breakdown. I understand like 10% of it

  • @captainconcernedsr.5360

    Doesn't stop me from wanting tensor cores to enhance AI's for shooters and Sim's while the reduced workload on the CPU can provide more room to do other things.

    • @GregSalazar
      @GregSalazar  Před 5 lety +1

      Don't count on mainstream devs jumping on board any time soon.

    • @wing0zero
      @wing0zero Před 5 lety

      The new Metro game is the first to use ray traced lighting, maybe they are using tensor cores as they are perfect for it.

  • @adonistopofmen2571
    @adonistopofmen2571 Před rokem

    Great technology ...

  • @Cuplex1
    @Cuplex1 Před měsícem +1

    This didn't age well. Welcome to the world of Nvidia RTX. 😳

  • @texasdeeslinglead2401
    @texasdeeslinglead2401 Před 5 lety

    But can it djent

  • @IgordasTunfisch
    @IgordasTunfisch Před 5 lety +1

    Battery is filling up

  • @darthtahguy
    @darthtahguy Před 5 lety

    How about the new snapdragon 845 it has tensor cores

  • @IVI00101
    @IVI00101 Před 5 lety

    I thought, they were planning to use tensor cores to axelerate ray tracing.

  • @xXxJackalxXxPGxXx
    @xXxJackalxXxPGxXx Před 5 lety +3

    I think the Tensor cores for gaming only make sense if they are used to do the RTX lighting effects

    • @brandonn.1275
      @brandonn.1275 Před 5 lety +1

      Ruben Robres ehh tensorcores will probably really bad in graphics application most calculations would be in the realm of f32 calculations instead of the realm of f16 calculations that machine learning loves. Since machine learning isn't about high precision calculations as it is high volume low precision calculations.

    • @xXxJackalxXxPGxXx
      @xXxJackalxXxPGxXx Před 5 lety +1

      RTX lighting effects are done with an AI (FP16) if im right

    • @brandonn.1275
      @brandonn.1275 Před 5 lety +1

      from the looks of it, 16 bit float isn't gonna be enough to run ray tracing since the math for it appears to use higher precision calculation fitting somewhere near the ballpark of the kind of calculations that AVX does with most calculations sitting around 256 bit - 512 bit. Which is kind of the main reason why ray tracing was mostly done on processors until now since gpus had poor support for doing high volumes of high precision math quickly until recently. My guess for how RTX support started exist in nvidia was that they started either embedding higher precision calculators into their CUDA architecture or found away to link their f32 calculators together in order to start emulating support for AVX math.

    • @brandonn.1275
      @brandonn.1275 Před 5 lety

      after taking another look it does look like AVX (256-bit) is what's used for accelerating ray-tracing in graphics applications. With this paper showing major performance increases in ray-tracing of around 15%-25% on the AVX 2 (512-bit) standard.
      graphics.cs.ucdavis.edu/~hamann/FuetterlingLojewskiPfreundtHamannEbertHPG2017PaperFinal06222017.pdf

    • @xXxJackalxXxPGxXx
      @xXxJackalxXxPGxXx Před 5 lety

      im not talking about normal Ray Tracing, im talking about the new NVIDIA tech called RTX, Ray Tracing with AI algorithms apparently