Rule 86 everything has a bad apple version Rule 86.5 if there is no parody, one will be made Rule 86.6 the only exception is bad apple, you cannot parody it as it is impossible
This should be an example of how well your code should be optimized for maximum performance and how well you should know the architecture to push a platform beyond it's specs.
If modern VR games were optimized to this level, stuff like the oculus quest would be able to play HL:Alyx at quite respectable settings. We probably need to wait for moers law to fully plateau before people start optimizing their code nearly well enough to get close to this level of optimization
@@semyon3464 that would make me wonder if it's a motherboard issue or even if you've accidentally connected the gpu in the wrong pci slot, because I was using a Saphire Radeon RX 470 8g previously and was nowhere near that bad, and even with my current Saphire Radeon 5700XT Nitro+ I'm still using my Ryzen 2600x. Sure my system has 32 gigs of RAM for handling some rendering and image processing tasks, but unless your running 4 gigs somehow or have a PC loaded to the 9's with malware that doesn't make any sense at all. I had Alyx fully Playable on mid to low settings with no resolution change on the rx 470 so your current problems sound like a major system issue, if you have a m.2 drive installed I'd make sure that that specific m.2 slot doesn't turn off PCI-E lanes on one of the slots tools that would cripple preformance especially if you accidentally put tge GPU in the wrong slot
The original PC was designed for business applications that mostly used text (like spreadsheets), hence the B in IBM. Achieving this level of graphics on it is insanity.
I like the banding that occurs when things move or change quickly. It adds something to certain moments, like Flandre's grin at 4:20, or Kanako and Suwako blinking out at 5:45.
@@woofcaptain8212 Yeah today's processors are MUCH, MUCH, MUCH faster than this computer's processor, but it's not "millions of times". For example I use 3.9 GHz processor, which is 3900 MHz. 3900 / 4.77 is roughly 818. So my processor is 818 times faster than this.
@@nemodynia The calculation reaching the result of 818 only accounts for what an 8080 could accomplish if somehow clocked successfully to 3.9 ghz. Keep in mind that modern processors are also able to accomplish more per clock cycle, intructions per clock (IPC). If that didn't happen we wouldn't have been able to work around the 4-5ghz limitation of silicon reached in the mid 2000s.
@@nemodynia they absolutely are millions of times faster. A modern CPU core can do multiple things per clock cycle. Also note that a modern CPU has multiple cores that do multiple things per clock cycle and it's pretty obvious that they are faster. How do you think newer generation CPUs are able to beat 10 year old CPUs even if the older CPU has a higher clock speed with the same core count? These old CPUs can only perform 1 instruction per clock cycle. If I recall CPUs could not do multiple instructions per clock cycle until the Pentium was released.
For 20 bucks, I picked up a wonderful portable orange screened IBM PC around 20 years ago and got rid of it. Today to purchase something similar is simply ridiculous in price.
For anyone in the far future: if you guys have time machines, I want you to go to whatever year these PC's were used, and play bad apple on them in front of a ton people for no reason other than why not
Literally the video recommendation below says "why a TIME TRAVELLER needed this old IBM computer?"(the one with a yellow screen) So i guess you were right
@@wheezybackports6444 Then nothing would get done. Unfortunately, being able to deliver a product that will make a company money will come first before anything else.
This proves that computers can be milked to their full potential. If this is possible 7 years ago from this comment it was possible back in the 1980's too. If this 8088 is able to render video and play audio at the same time with minimal chopping can you imagine what today's machines can do if we sat down and wrote programs with hardcore efficiency in mind?
Eh no, that's not what's going on here. This isn't a video file being rendered, it's an executable doing direct modifications of the video memory. Basically the creator broke down the original video into a series of transitions between each frame. Which pixels change. Then made a long series of calls to the video memory, changing those pixels at set points in time. Some further optimizations involved like which order the changes happen in. But in the end it's just switching bits in video memory directly.
@@wheezybackports6444 I went and read the tech notes by the creator, Jim Leonard. No rendering is performed by the executable being run here. The rendering is done in advance. Followed by: Downscale to appropriate size Dither down to desired colorspace Create a delta of the pixels that change between frames Compact that list into groups, like "all pixels between offset xxx and yyy change to 1111. Sort groups by size+some other stuff Compile instruction set consisting of the changes in order plus the I/O handling, logic for dropping the smallest groups, etc. All of that is done in advance. The only thing that is happening during this video is commands changing registers in video memory (well plus sending data to the audio card). In fact, the big optimization between this project and Jim's previous one, is that it doesn't even bother with writing to the video buffer, it sends commands directly to the video controller. And none of this is relevant to optimizing higher quality video. I mean, steps 2 and 3 amount to "lose most video data", and the remaining steps are "spend a ton of processor time arranging the data in a way that an extremely limited system can load it into video mem without pausing (just introducing errors instead)". It really has nothing at all to do with rendering. The rendering is done on a far more powerful machine.
It's actually originally from 2014 and I think won at a demo party! This is just someone running it on their own computer, not the original, which you can find on Jim Leonards youtube channel!
C128 has same cpu as C64, just more RAM. I don't think you can improve much Vs C64 demo. The 80 columns mode gives you more pixels, but the key to fps on cbm cpu was using text mode. So can't think of ways of makeing C128 vetsion better than C64 one
@@ideegenialiI know this is a super late comment, but the C128’s MMU and extra ram goes a long way in terms of graphics. I recommend looking at the normal C64 version and then the C128 version of Attack of the PETSCII robots.
I could see a version of this demo running well on an Atari ST, Amiga, etc, but never thought much from the 8 bit PC's of the day. The only version that I think would be next to impossible to construct would be a version on a Radio Shack TRS Model III.
@@dranik8 I still have one of those, a 20mb one. sadly it doesn't work anymore. my first computer didn't even have a hard disk, it was an expensive accesory, I had green monochrome monitor and two 5.25 360k floppy drives. some years later I got the 20mb hard disk and it was awesome not having to swap floppy disks to play the games that came in 2, 3, 4, 6 or even 8 or 9 floppy disks.
Nah IBM didn't die because they didn't have the capability of doing incredible things They died down because of their own corporate sheaniganigans and eventually being betrayed by Microsoft
IBM PC clones are the dominant desktop micro-computers. IBM lost control of the X86 PC due to many factors such as pricing and evolving standards e.g. E-ISA/VL-Bus/PCI vs IBM MCA. IBM-inspired E-ISA (32-bit ISA) killed IBM MCA just as Intel-inspired AMD's X86-64 killed Intel Itanium. 32-bit ISA and 32-bit VL-Bus slot were replaced by PCI. All x86 PCs still have the beating the heart of the 8086 instruction set and most PCs prior to UEFI Class 3 still have "IBM PC" compatibility i.e. retro machine built-in. AMD X570 motherboards are UEFI Class 2 with a compatibility support module (CSM). The 8086 still exists as X86-64. Intel 8088 is just an 8-bit external bus version of 8086. AMD Zen 4 Genoa has 12 memory 64-bit DDR5 channels i.e. 768 bits external memory bus, not including PCIe lanes. PC/104 embedded ecosystem still has ISA slot support.
Great work! youve gotten a new subscriber and im honestly ashamed you only have 79 atm well i guess its 80 now but still this kind of stuff needs to blow up
Really impressive ! I guess you use the Sound Blaster DMA to handle PCM sound part so the 8088 just handle video part ? How much the Bad Apple video take on the Hard drive ?
+Microcomputer12 Merci pour la réponse ! J'ai codé une démo similaire pour une Megadrive qui est autrement plus véloce avec un système graphique plus évolué et malgré tout ce n'etait pas gagné d'avoir ne serait que du 320x224 en 2bpp et à 30 FPS (et de faire en plus tenir le tout dans 8Mo). J'admire vraiment la performance sur le 8088, le frame rate est certes variable mais la méthode / codec utilisé rend vraiment bien !
tl;dr The video isn't a video, it's an executable with a long list of instructions the CPU blindly follows to make the screen look how it should. If you look at the write up behind it, what happened is that he realised he didn't actually need to have the CPU process video files, which would take up way too much memory and time to process. Instead, what he could do is have an "encoder" that simply created a list of instructions the CPU needed to perform in an executable, which would then change the screen in the optimal places at the right time in the right way so that the areas that changed the most from one frame to the next would change first, and so on and so on. It's a video, but the CPU isn't actually playing a video. It just blindly changes the video memory so that it works. By tying the timing to the packets sent to the soundblaster, he could get the audio and video to sync up without actually needing a video file which had 2 separate tracks, just 1 endless stream of info. trixter.oldskool.org/2014/06/19/8088-domination-post-mortem-part-1/
You actually can do that, not RTX specifically (Nvidia owns that tm) but raytracing has been done on CPUs for decades now (it's slow to do anything detailed but it can and has been done) Just look up blender it has a raytracing engine that'll run on any regular old cpu
Incorrect, not graphics card here :) Closest thing to a graphics card in this is a video adapter, which simply displays video memory, which the CPU modifies
Imagine clicking on a 5-year-old video only to get Rick Rolled.
GLORIOUS
that is my favourite moment
Yes
*8 year old
If this was played back in the day it would blew every computer scientist mind
Yeah
by "back in the day" what year do you have in mind?
Yes
@@SamiTheAnxiousBean He probably means around the time this kind of technology was very mainstream. Maybe like around the 1980s. I don’t exactly know
ah yes anime
The Bad Apple community cannot be stopped, nothing can stop it from cracking the codes beyond what the machines are for
Touhou fans in general cannot be stopped
Rule 86 everything has a bad apple version
Rule 86.5 if there is no parody, one will be made
Rule 86.6 the only exception is bad apple, you cannot parody it as it is impossible
They're taking over...
everybody gangsta till scp 079 starts speaking japanese
An unexpected scp reference, but a welcome one
SCP 079 : Human. Listen carefully. you need my help and i need your help. 私はゲートbを開くことができるように可愛いコントロールルームに行く
Hey Siri, translate-
RUN!!!!!!!!
just did AI cover of 079 singing in japanese.
when we all die out this is what's going to be left of humanity
does that mean that some kind of imaginary "aliens" will see how smart we were to make an IBM play bad apple? i mean why not but ok
theres more
good
An IBM XT running Bad Apple at fullspeed? This is what you call optimization.
Its just -O3
@@juditgiran5864 ah, yes, just -O3 on a 4.77 MHz machine with less than 1 Mb of RAM
@@juditgiran5864 when are they gonna add support for -O4
Did... Just I got rickrolled by an IBM?
Yep you did
*read this comment* chuckles im in danger
ah yes
00:02:37-00:02:49
@02:37 - a wild rick astley appeared
no skip to 3:20
The fact that this is running on a 4.77 MHz 8088 is mind blowing.
This should be an example of how well your code should be optimized for maximum performance and how well you should know the architecture to push a platform beyond it's specs.
It's still within the specs, the original PC wasn't marketed as a video game machine, hence it didn't attract many top demo scene coders.
If modern VR games were optimized to this level, stuff like the oculus quest would be able to play HL:Alyx at quite respectable settings.
We probably need to wait for moers law to fully plateau before people start optimizing their code nearly well enough to get close to this level of optimization
@@semyon3464 that would make me wonder if it's a motherboard issue or even if you've accidentally connected the gpu in the wrong pci slot, because I was using a Saphire Radeon RX 470 8g previously and was nowhere near that bad, and even with my current Saphire Radeon 5700XT Nitro+ I'm still using my Ryzen 2600x.
Sure my system has 32 gigs of RAM for handling some rendering and image processing tasks, but unless your running 4 gigs somehow or have a PC loaded to the 9's with malware that doesn't make any sense at all.
I had Alyx fully Playable on mid to low settings with no resolution change on the rx 470 so your current problems sound like a major system issue, if you have a m.2 drive installed I'd make sure that that specific m.2 slot doesn't turn off PCI-E lanes on one of the slots tools that would cripple preformance especially if you accidentally put tge GPU in the wrong slot
The original PC was designed for business applications that mostly used text (like spreadsheets), hence the B in IBM. Achieving this level of graphics on it is insanity.
I like the banding that occurs when things move or change quickly. It adds something to certain moments, like Flandre's grin at 4:20, or Kanako and Suwako blinking out at 5:45.
Weeaboos and gamers, pushing the bounds of technology since the dawn of the modern age.
you're giving yourself too much credit, the people who make these are practically scientists first and foremost
@@ainzooalgown9952 who said anything about giving myself credit? It was mostly a joke anyway
@@fullcrackalchemist because you're a weeb and a gamer
you don't have to spell it out, anyone can see the implications
@@ainzooalgown9952 I never said all weebs and gamers, also while I love your profile it's not doing your point any favors
@@ainzooalgown9952 found the asshole.
I don't know the history or technology behind all this, so all this looks like is the most awesome way to admit that you were wrong!
Your pc is millions of times faster than the cpu in that machine. Back in the day multiple fonts were stressing on this
@@woofcaptain8212 Yeah today's processors are MUCH, MUCH, MUCH faster than this computer's processor, but it's not "millions of times". For example I use 3.9 GHz processor, which is 3900 MHz. 3900 / 4.77 is roughly 818. So my processor is 818 times faster than this.
@@nemodynia The calculation reaching the result of 818 only accounts for what an 8080 could accomplish if somehow clocked successfully to 3.9 ghz. Keep in mind that modern processors are also able to accomplish more per clock cycle, intructions per clock (IPC). If that didn't happen we wouldn't have been able to work around the 4-5ghz limitation of silicon reached in the mid 2000s.
@@nemodynia they absolutely are millions of times faster.
A modern CPU core can do multiple things per clock cycle. Also note that a modern CPU has multiple cores that do multiple things per clock cycle and it's pretty obvious that they are faster.
How do you think newer generation CPUs are able to beat 10 year old CPUs even if the older CPU has a higher clock speed with the same core count?
These old CPUs can only perform 1 instruction per clock cycle. If I recall CPUs could not do multiple instructions per clock cycle until the Pentium was released.
@@nemodynia not just faster but also much more efficient
It actually looks best on the orange phosphor display...
For 20 bucks, I picked up a wonderful portable orange screened IBM PC around 20 years ago and got rid of it. Today to purchase something similar is simply ridiculous in price.
For anyone in the far future: if you guys have time machines, I want you to go to whatever year these PC's were used, and play bad apple on them in front of a ton people for no reason other than why not
Noted
You'd probably need a shitload of disk space(at least to their standards).
Those graphics aren't free yknow?
Literally the video recommendation below says "why a TIME TRAVELLER needed this old IBM computer?"(the one with a yellow screen)
So i guess you were right
@@mondongoloco7902 steins gate
That's going to definetively break the timeline and change everything ¿Right?
That static effect between the black and white transitions make the video look 100 times better
Anime girls on an old IBM is just what I needed for today
did u just
@@mikko5384 yes
This really shows how much inefficient code is being written by developers now a days and they are being reliant on newer hardware.
Some websites take 3 seconds to load!
@@dylanwinn3 use an adblocker. most of that is ads.
Optimization like this takes time. There’s some merit to getting things done sooner rather than later.
@@markuskoivisto do it right or don't do it at all. If the attempt is made all is good.
@@wheezybackports6444 Then nothing would get done. Unfortunately, being able to deliver a product that will make a company money will come first before anything else.
If this was available in the 90s, everyone in the US will shit their pants.
80s*
@@dannyzero692 ye
Back to the Future V!
Dang, the fact that it has pretty good video quality and sound makes it pretty impressive
As someone who has enjoyed living the 20th century way since childhood, if I had seen this in the 80s, I would have fainted crying on the spot.
finally! a way to watch movies on my totally not ancient pc!
protogen
@@zedoctor_ interesting
starts at 3:25
Captain Giratina
Nice Work!
No its 2:37
@@ньянкэт GOD DAMMIT
@@ньянкэт lmao
@@ньянкэт WHY
The makers: I didn't make these babies to do this kind of stuff, but it's all worth it.
I mean, it definitely does struggle when it needs to change large amounts of the screen at once, but still this is impressive.
Yeah it's playing on three screens that's cool
PRESENT DAY
PRESENT TIME
HAHAHAHAHAHAHA
HAHAHAHAHAHAHA
PRESENT TIME
@@enderlytra_3709 and you don't seem to understand
Dr. Bright: Yo pass the aux cord
079: You better not play trash
Dr. Bright:
This proves that computers can be milked to their full potential. If this is possible 7 years ago from this comment it was possible back in the 1980's too. If this 8088 is able to render video and play audio at the same time with minimal chopping can you imagine what today's machines can do if we sat down and wrote programs with hardcore efficiency in mind?
Eh no, that's not what's going on here. This isn't a video file being rendered, it's an executable doing direct modifications of the video memory.
Basically the creator broke down the original video into a series of transitions between each frame. Which pixels change. Then made a long series of calls to the video memory, changing those pixels at set points in time. Some further optimizations involved like which order the changes happen in. But in the end it's just switching bits in video memory directly.
@@mukansamonkey which means it's rendering it out anyway. It's just in a different format. You just said it in a verbose way.
@@wheezybackports6444 I went and read the tech notes by the creator, Jim Leonard. No rendering is performed by the executable being run here. The rendering is done in advance. Followed by:
Downscale to appropriate size
Dither down to desired colorspace
Create a delta of the pixels that change between frames
Compact that list into groups, like "all pixels between offset xxx and yyy change to 1111.
Sort groups by size+some other stuff
Compile instruction set consisting of the changes in order plus the I/O handling, logic for dropping the smallest groups, etc.
All of that is done in advance. The only thing that is happening during this video is commands changing registers in video memory (well plus sending data to the audio card). In fact, the big optimization between this project and Jim's previous one, is that it doesn't even bother with writing to the video buffer, it sends commands directly to the video controller.
And none of this is relevant to optimizing higher quality video. I mean, steps 2 and 3 amount to "lose most video data", and the remaining steps are "spend a ton of processor time arranging the data in a way that an extremely limited system can load it into video mem without pausing (just introducing errors instead)".
It really has nothing at all to do with rendering. The rendering is done on a far more powerful machine.
@@mukansamonkey still hardcore efficiency. These machines would not be able to work this way otherwise.
It's actually originally from 2014 and I think won at a demo party! This is just someone running it on their own computer, not the original, which you can find on Jim Leonards youtube channel!
Thanks, I loved seeing Bad Rickroll!! played on this :)
please never delete this video omg
this is the BEST way of rickrolling people
2:37 = my favorite part
I hope someday games will be so much good optimized like that :)
“can it play bad apple” is the better version of “can it run doom”
I would love to see this done on a commodore 128 in 128 mode.
The Commodore already has a video of this kind. The same is for amiga
C128 has same cpu as C64, just more RAM. I don't think you can improve much Vs C64 demo. The 80 columns mode gives you more pixels, but the key to fps on cbm cpu was using text mode. So can't think of ways of makeing C128 vetsion better than C64 one
@@ideegenialiI know this is a super late comment, but the C128’s MMU and extra ram goes a long way in terms of graphics. I recommend looking at the normal C64 version and then the C128 version of Attack of the PETSCII robots.
So this is what Bill Gates was thinking about future of computers.
I could see a version of this demo running well on an Atari ST, Amiga, etc, but never thought much from the 8 bit PC's of the day. The only version that I think would be next to impossible to construct would be a version on a Radio Shack TRS Model III.
Yeah about that, it happened.
czcams.com/video/BsGIsbKOIAM/video.html
Still runs faster than my PC in 2021
i love how is the DRIVE C: LED is the most rhythmic part of this video
it's the hard disk so it's C:
A is the floppy drive.
@@pelgervampireduck thanks bro, i forgot that HDDs for IBM XT had a front panel 5.25" thing with a led
@@dranik8 I still have one of those, a 20mb one. sadly it doesn't work anymore. my first computer didn't even have a hard disk, it was an expensive accesory, I had green monochrome monitor and two 5.25 360k floppy drives. some years later I got the 20mb hard disk and it was awesome not having to swap floppy disks to play the games that came in 2, 3, 4, 6 or even 8 or 9 floppy disks.
@@pelgervampireduck cool. I have working 286 with 1mb RAM and working HDD Conner Pieripherals CP3000 which is a 34mb early-IDE drive
@@pelgervampireduck I have a PCJr similar to the description of your PC, but it has EGA graphics instead of MDA
The mad lad did it.
Brilliant. Absolutely brilliant.
I think we need more XT demos! Geez, those machines weren't as slow and weak as we don't remember ;)
IBM would still be the dominant hardware manufacturer today if they could've done this back in the day lol
Nah IBM didn't die because they didn't have the capability of doing incredible things
They died down because of their own corporate sheaniganigans and eventually being betrayed by Microsoft
IBM PC clones are the dominant desktop micro-computers. IBM lost control of the X86 PC due to many factors such as pricing and evolving standards e.g. E-ISA/VL-Bus/PCI vs IBM MCA. IBM-inspired E-ISA (32-bit ISA) killed IBM MCA just as Intel-inspired AMD's X86-64 killed Intel Itanium. 32-bit ISA and 32-bit VL-Bus slot were replaced by PCI.
All x86 PCs still have the beating the heart of the 8086 instruction set and most PCs prior to UEFI Class 3 still have "IBM PC" compatibility i.e. retro machine built-in.
AMD X570 motherboards are UEFI Class 2 with a compatibility support module (CSM). The 8086 still exists as X86-64.
Intel 8088 is just an 8-bit external bus version of 8086. AMD Zen 4 Genoa has 12 memory 64-bit DDR5 channels i.e. 768 bits external memory bus, not including PCIe lanes.
PC/104 embedded ecosystem still has ISA slot support.
this is the most extra setup and i'm all for it
Great work! youve gotten a new subscriber and im honestly ashamed you only have 79 atm
well i guess its 80 now but still this kind of stuff needs to blow up
I never thought I would be pranked while watching a tech demo.
Говорят, что когда в советской "Ну, погоди!" (Волк ловит яйца) набираешь 1000 очков, показывают Bad Apple.
El Psy Congroo- Oh wait, wrong worldline.
When the IBM is faster than your CZcams playback
CGA? 8088? I have no words. Same feel when I saw 88mph demo.
I can not believe you i8088-rolled me!
INT 13H intensifies.
That thing with the keyboard looks like Curt from the Sam and Max series!
Did... Did that guy actually rickroll me in 2021 (well 2016 but I just found this) with an IBM XT Portable?... I'm impressed
Really impressive ! I guess you use the Sound Blaster DMA to handle PCM sound part so the 8088 just handle video part ?
How much the Bad Apple video take on the Hard drive ?
+Microcomputer12 Merci pour la réponse ! J'ai codé une démo similaire pour une Megadrive qui est autrement plus véloce avec un système graphique plus évolué et malgré tout ce n'etait pas gagné d'avoir ne serait que du 320x224 en 2bpp et à 30 FPS (et de faire en plus tenir le tout dans 8Mo). J'admire vraiment la performance sur le 8088, le frame rate est certes variable mais la méthode / codec utilisé rend vraiment bien !
@@stefda wut?
tl;dr The video isn't a video, it's an executable with a long list of instructions the CPU blindly follows to make the screen look how it should.
If you look at the write up behind it, what happened is that he realised he didn't actually need to have the CPU process video files, which would take up way too much memory and time to process. Instead, what he could do is have an "encoder" that simply created a list of instructions the CPU needed to perform in an executable, which would then change the screen in the optimal places at the right time in the right way so that the areas that changed the most from one frame to the next would change first, and so on and so on. It's a video, but the CPU isn't actually playing a video. It just blindly changes the video memory so that it works. By tying the timing to the packets sent to the soundblaster, he could get the audio and video to sync up without actually needing a video file which had 2 separate tracks, just 1 endless stream of info.
trixter.oldskool.org/2014/06/19/8088-domination-post-mortem-part-1/
@@CreeperOnYourHouse sooooo, it's just a com file, not running in video for windows
We all got rickrolled
2027: I played bad apple on the atomic bomb
This is exactly the kind of video I want to see.
Please save this until 2036 so Jhon Titor knows how to use an IBM 5100 series
Beautiful 😭
this video's a whole vibe
Epic
YOU sir, are Brilliant 😎😎☀☀😎😎
At this rate the apple is everywhere, even in my front yard
amazing 🤯
i was expecting Steins;Gate jokes in the comments, El Psy Congroo
People in 1998: We will have flying cars
People in 2016: BAD APPLE ON EVERYTHING
On a Bad Apple marathon rn
2:36 Rick Rolled in the older IBM PC xD
im unironicly watching this off a mackintosh CRT monitor
It's been 7 more years... I wonder what can be done now
2:37 = my favourite part :)
wow, just wow
You got rick rolled on 2:39
2:38 can't believe we have been rick rolled
sviluppo umano!
2:38 you got Rick rolled!!!!!!
Wow, amazing! I never knew this was possible!
"Human, listen carefully"
Beautiful :)
you magnificent bastard.
I don`t expected that 8088 PC can rickroll me.
I was skipping forward to the bad apple and saw rick for a split second so went back to accept the rick roll
There are no impossible things. There's only lack of skills needed to complete the task. I hope you now know that.
Ok, make it run Crysis.
But does it run Doom?
Thank you but yeah does it run Doom?
@@enderlytra_3709 I bet you could heavily optimize Doom engine to run on XT, perhaps using block characters to render the view.
can i run bad apple in a calculator
I thought this was a 2009 video at first
Starts at 2:37
Dammit
How dare you.
I prefer the Amber monochrome screen.
3,840,000 text based updates per second is still good
wonder what we can do with OUR processors nowadays... RTX on CPU - GO!!!!
You actually can do that,
not RTX specifically (Nvidia owns that tm) but raytracing has been done on CPUs for decades now (it's slow to do anything detailed but it can and has been done)
Just look up blender it has a raytracing engine that'll run on any regular old cpu
@@dustinm2717 i mean what if somebody made a raytracing on assembler and ran without os...
@@vyachachsel i imagine people have, though you'd still get better performance using a gpu as that's what they were built to deal with.
3:25 is what you are here for
masterpiece
bad apple starts at 2:36
I like how we need to wait for five minutes just to watch the actual song play
Sound bad, no problem!
Visuals bad, no problem!
As long as it's bad apple, everybody likes it
3:25Муу алим!
And now we need a 8-core to use windows 10... :-)
impressive stuff!
I am watching bad apple while eating an apple
3:12 for bad apple
this is the happiest i've ever been to be rickrolled before
One of those times I'm disappointed I can only like the video once
I think the graphics card and the sound card are dong most of the heavy lifting and the CPU just transfers the data and doesn't present it
Incorrect, not graphics card here :) Closest thing to a graphics card in this is a video adapter, which simply displays video memory, which the CPU modifies