How RAID Works

Sdílet
Vložit
  • čas přidán 7. 10. 2022
  • What is RAID and how does it work? Why would you use RAID and what are the benefits in terms of performance and reliability? Dave Explains mirroring, striping, JBOD, RAID levels, and ZFS.
    For my book "Secrets of the Autistic Millionaire": amzn.to/3diQILq
    Upcoming LIVESTREAM: Tentative for Sunday the 9th at 10AM PST, 1 PM EST.
    ERRATA: RaidZ2 can accommodate 2 failures, RaidZ3 can accommodate 3. Whoops!
    Discord Chat w/ Myself and Subscribers: / discord
    Primary Equipment (Amazon Affiliate Links):
    * Black and Decker Stud Finder - amzn.to/3fvEMuu
    * Camera: Sony FX-3 - amzn.to/3w31C0Z
    * Camera Lens: 50mm F1.4 Art DG HSM - amzn.to/3kEnYk4
    * Microphone: Electro-Voice RE 320 - amzn.to/37gL65g
    * Teleprompter: Glide Gear TMP 100 - amzn.to/3MN2nlA
    * SD Cards: Sony TOUGH - amzn.to/38QZGR9
    Experiment with Houston Command Center or Learn More: www.45drives.com/blog/uncatego...
  • Věda a technologie

Komentáře • 323

  • @amonynous9041
    @amonynous9041 Před rokem +143

    calling his storage unit lil nas is such a power move

    • @DavesGarage
      @DavesGarage  Před rokem +38

      Not as much as "Let's ask the big organ" :-)

    • @eadweard.
      @eadweard. Před rokem +6

      Mine's called "illmatic".

    • @user-yr1uq1qe6y
      @user-yr1uq1qe6y Před rokem +24

      My teen thinks I’m cool because I told her I was watching a little nas video.

    • @mikkelbreiler8916
      @mikkelbreiler8916 Před rokem +11

      @@user-yr1uq1qe6y Dad humour. Right there.

    • @shez666
      @shez666 Před rokem +6

      I hope he mounts it as the X: drive in Windows

  • @kevinshumaker3753
    @kevinshumaker3753 Před rokem +66

    The one thing I've learned from experience:
    DO NOT USE ONLY DRIVES FROM A SINGLE BATCH / DATE CODE & MANUFACTURER.
    Mix your drive manufacturers. Mix your date codes. We had a 500TB Server go south within days due to the drives being all from the same manufacturer and date code, and a lot of drives failed and cascade failed. They were under warranty, and we did have a good tape backup, but it took a LONG time to do the swap, mixing, and restoring, the whole time we were down (about 10 days, tape is SLOW).

    • @eadweard.
      @eadweard. Před rokem +3

      Was it a bad batch?

    • @talibong9518
      @talibong9518 Před rokem +9

      And when using random drives, you'll naturally end up with the best drives for the job anyway as the shit ones get replaced over time and you know what to avoid.

    • @Doesntcompute2k
      @Doesntcompute2k Před rokem +8

      ^^^ THIS. But I've got to add, too: FLASH YOUR FIRMWARE ON THE DRIVES TO THE LATEST, before adding to a RAID volumeset. You'll thank yourself. Put a sticker on the drive AND RECORD the version, date of FW, and date of flashing. It's saved me a lot whether it's four drives in a WS or 400 drives in an EMC. 😁👍

    • @DrakiniteOfficial
      @DrakiniteOfficial Před rokem +1

      Good to know, thanks!

    • @kevinshumaker3753
      @kevinshumaker3753 Před rokem +8

      @@eadweard. Yep, but the manufacturer wouldn't admit it and do a recall until I found a bunch of users reporting the same problem. When confronted, they issued the recall notice. I had used the same drives in regular PCs and other servers, and they all got replaced.

  • @kevintrumbull5620
    @kevintrumbull5620 Před rokem +9

    The ZFS equivalents to raid are RAIDZ1 for RAID5 (Single parity), RAIDZ2 for RAID6 (Double parity), and RAIDZ3 for triple parity RAID. The number after "RAIDZ" is the number of drives (parity) that you can lose before you have data loss.
    It's worth noting that ZFS contains a large number of advancements beyond what traditional RAID offers, such as data integrity guarantees, transactional operation (which keeps the write hole from corrupting your array), and more.

  • @muchosa1
    @muchosa1 Před rokem +19

    When I worked for VzB, I learned unless you test a backup, consider it bad. We had clients that did weekly restores of their data to confirm if it was good.

    • @Doesntcompute2k
      @Doesntcompute2k Před rokem

      Always one of the BEST ideas ever invented. If you cannot restore when you want to, you cannot restore when you need to.

    • @stapedium
      @stapedium Před rokem

      Problems come up when you spend 4 days of the week restoring and only 1 day working. Problems like the business making enough money to pay someone to do all those restores.

    • @SomeMorganSomewhere
      @SomeMorganSomewhere Před rokem +2

      Yup. it's not a viable backup until you prove you can restore it...
      In a previous job the morons who set up the infrastructure I was building systems on had "set up backups", but after one of them somehow managed to trash an ENTIRE SAN we discovered that the backup directory which was SUPPOSED to store the backup of the MSSQL server database only had file called "placeholder.txt" in it...

    • @rickgreer7203
      @rickgreer7203 Před rokem +2

      Reminds me of a place I found, where (in ~1999) for years an exec would put a new tape in the server, and take ejected tape home each night. Daily backups stored offsite in their home safe. Except they never noticed the tape was ejected shortly after inserting it. The tapes had been used for so long a few were worn completely clear, and all were bad. (And prior tech hadn't bothered to check it, since things seemed to be working...and of course, back then, central monitoring and such was less ideal than now. When I did get it fixed, the tape/drive size was far too small for what it was trying to backup as well...)
      Within a month after I left a few years later, the new folks did an Exchange upgrade, lost the CEO's mail...and they hadn't verified their new backup scheme either. I got paid to come in and rebuild the old server from old backup...getting back everything less the most recent few month's was better than nothing, and they'd given me the old hardware.

    • @rickgreer7203
      @rickgreer7203 Před rokem +2

      @@stapedium Automate the restore and verification each day of a prior daily, and weekly of a prior weekly/monthly/yearly. Set a quarterly or so human audit process for sanity....same for walking the cages and looking for bad blinky lights that didn't get reported. Push failures to a well defined alert/ticket pipeline that isn't among the noise. Overall save time and have better reliability. And if the tools don't exist, write them, and publish them open source with the company name attached, for some free content marketing too.

  • @meggrobi
    @meggrobi Před rokem +38

    I do like Dave's "conversational" presentation, it's interesting yet not over the top. I also like the trip down memory lane as I relate to many of his anecdotes. The loss of two drives, in a large array, is a thing when you take into account mean time failure.

  • @katbryce
    @katbryce Před rokem +21

    I don't know how this storinator thing works, but in FreeBSD (and Solaris), which have had zfs for a lot longer than Linux, RAID-Z1 is sort-of equivalent to RAID 5, RAID-Z2 is sort of equivalent to RAID-6, and RAID-Z3 can cope with three drive failures.
    One important thing about zfs is that it should always have access to the raw drives. You should not create a zfs file system out of "drives" that are RAIDed using some other method.
    One big advantage that zfs has over standard RAID is that it can deal with non-catastrophic disc failures - where the disk is apparently still working, but giving back wrong data. In standard RAID, you get a parity error and you know something is wrong, but because you don't know which drive has failed, you don't know what the right answer is. In zfs, you will have a much better chance of knowing which drive is reporting the incorrect data.

    • @alessandrozigliani2615
      @alessandrozigliani2615 Před rokem +5

      Confirmed. I don't think Storinator works differently than freebsd, linux or solaris for that matters if it is ZFS. RAID-Z1 is similar to RAID-5. RAID-Z2 is similar to RAID-6 and RAID-Z3 is a level further. The main difference with traditional (hardware) RAID is that if you have only partial failures of some sectors over multiple drives, there is a chance even in raid-z1 that you can still recover the data by mounting the pool readonly. IN traditional RAID, the disk is taken offline and is lost even with minimal problems.

    • @LiraeNoir
      @LiraeNoir Před rokem +3

      Storinator is just the hardware: the case, backplanes, etc. Otherwise, it's a regular "basic" server PC.
      And yes, in the age of ZFS, raid is a bit... past century :) RAID doesn't protect against bit rot, doesn't really know which bits are the bad ones in case of conflict (and worse, doesn't tell you that explicitly), and the list gores on for quite a while.
      Which is why you'll see a LOT of people put something like TrueNAS on their storinator. Plenty of storage for cheap, and a robust software to handle the zfs of disks and turn the whole thing into a storage appliance if one so wishes.

    • @butstough
      @butstough Před rokem +4

      came down here to find this comment. was worried my 15 disk wide raidz2 had only single disk fault tolerance for a second lmao.

  • @marcfruchtman9473
    @marcfruchtman9473 Před rokem +3

    The "Count" was absolutely glorious!

  • @ferna2294
    @ferna2294 Před rokem +8

    Watched the whole thing. I love the thorough way you explain everything. Even though I work as a technician, I have 0 experience with RAID machines, so this is always useful. Thank you ♥

  • @SB-KNIGHT
    @SB-KNIGHT Před rokem +5

    This was perfect timing for this episode on RAID. I am currently working on updating my old antiquated RAID-0 Striped array which has been serving me well for the past decade, but, its time to update to something a bit more reliable. Despite that, I have been using the same disks in it for over a decade and have not once had any data loss or disk failures (dispite living on the edge LOL). Always kept the data mirrored to another disk using SyncToy. BTW, Who made that wonderful app and why did it never gain traction? Its a great utility and has always worked great for copying large amounts of data from one place to another with different modes and having a way to verify the operations. Keep up the great work Dave! This is now one of my new favorite channels. Cheers!

  • @LanceMcCarthy
    @LanceMcCarthy Před rokem +3

    I just bought a 12 disk, 4U JBOD. It uses a USB 3.2, but is architecture allows for parallel access to each internal SATA interface. It arrives tomorrow, can't wait to load it up.

  • @TractorWrangler01
    @TractorWrangler01 Před rokem +6

    Your channel is great. Lots of information from behind the user perspective that is never seen by the end user. Love your content.

  • @SauvikRoy
    @SauvikRoy Před rokem +2

    Great topic of discussion. Thanks Dave!

  • @ianjharris
    @ianjharris Před rokem +2

    Dave you bring these subjects to practical reality from your experience. An obvious pro.

  • @LogicalNiko
    @LogicalNiko Před 9 měsíci +1

    Back in the day I had to write a recovery mechanism for 720MB SCSI-2 drives from a large raid set whose hardware controller failed and was basically unobtanium. IIRC we built a linux box with a big JBOD, and then basically selecting offsets from dd images of each drive. Then reconstructing the partitions in the right order. I remember we ended up using box fans and steel desk filing folder organizers to hold each of the drives and give them a decent heat sink. IIRC it took something like 7-8 days in those days to process such "massive" drives 🤣

  • @skooby180
    @skooby180 Před rokem +1

    Thanks Dave. Informative and interesting. Love your work.

  • @ryanwalker4660
    @ryanwalker4660 Před rokem +8

    Thanks for the tips Dave, I've only ever done RAID 0 for the read and write speeds and I'm sure like many stories you've heard that lead to a failed drive or corrupt data down the line. I've started off on my path for my AAS with a emphasis is cyber security and this is one of the first topics we hit on. I'm okay with tech, I might not get every aspect when in classes, I'm more of a hands on person and reading and learning from books is not my strong suit. Your videos are great and you have a good way of explaining the material. Thanks again.

    • @rawsaucerobert
      @rawsaucerobert Před rokem

      I got two used (although hardly) hard drives from FB marketplace, made sure there were no errors, they were legit, etc etc and now I just use 2 externals to back up all my stuff. I back up my pc to one, clone it or just run the backup again on the other drive, store one off site and switch them out every so often. I don't think the wear and tear or energy consumption of a NAS suited me when I backup maybe once per month or less but it's still essential for me to store all my photos and such once in a while. Cheers and good luck.

    • @c128stuff
      @c128stuff Před rokem

      @@rawsaucerobert you could still use a raid based solution for that (raid1, mirroring). It will just take care of the duplication part you do manually now, saving you some time, and reducing risk on both 'human errors' and technical failures.
      My 'big array' which I use for backups is turned off most of the time, and only gets turned on for making or restoring backups. Using 'wake on lan' you could even automate the turning on part of that.
      There are companies (icybox being one) which sell external enclosures for this specific use case, with raid1 support built into the enclosure.

  • @michaelpezzulo4413
    @michaelpezzulo4413 Před rokem +3

    Far better explanation of RAID then the Linkedin learning video course. Everybody's gotta love a man who knows what he is talking about.

    • @stevendavies-morris6651
      @stevendavies-morris6651 Před rokem +2

      I love how Dave doesn't use 10 words when 3 will do. And (being very techie myself who taught loads of computer classes) I appreciate how he uses a KISS approach to his presentation so not-very-techie people can learn from and apply what he explains.

  • @kc5402
    @kc5402 Před 8 měsíci

    I remember a *long* time ago when I was running an AMD Athlon 1.2GHz PC, ABIT came out with a mobo called the "KT-7A RAID". At the time, CD-ROM drives were still fairly new, and the technology hadn't yet matured. Hooking up an IDE CD drive to the same cable as your hard drive(s) could slow down your system. The KT-7A board had four IDE connectors instead of the normal two, so you could give all your IDE devices their own cable and DMA channel, and it speeded things up a lot. After setting up a new PC and running it with conventional hard drives for a while, I decided to go even further and try out a RAID setup (RAID 0 for speed, rather than RAID 1 for data security). It worked very well.
    Thanks for another great video Dave. I think you've now got me hooked on nostalgia, and you're my dealer! 😄👍

  • @moorera364
    @moorera364 Před 9 měsíci +1

    Another outstanding, informative and entertaining video. Keep them coming Dave, really do enjoy your content! Love to see something on Networking if you do it!

  • @tullyal
    @tullyal Před rokem +7

    Another super video - I haven't heard RAID explained as well in a long time.

  • @itsyaboikirbo
    @itsyaboikirbo Před rokem +1

    Thank you so much for explaining this! I look forward to more quality content from you! :)

  • @homemedia4325
    @homemedia4325 Před rokem +5

    @12:18 ... I was waiting to hear you say this... haha ... "RAID is not a backup" - I learned the hard way many years ago :D
    Edit: I use Ceph via Proxmox... it is like raid spanned across 2 or more servers (3 is the recommended minimum) - however, I have backups to a RAID Mirror and important things go to the cloud! - great vid :)

    • @mikkelbreiler8916
      @mikkelbreiler8916 Před rokem

      RAID is peeing in your ski suit at the top of the piste. By the time you get to the ski hut you better have a plan B or you'll be shitting yourself next.

  • @silentjohn80
    @silentjohn80 Před rokem +16

    I love ZFS, it has a lot of great features. And the zpool and zfs tools are very well documented. My config is 1 zpool with 2 raidz2-volumes of 6 disks each.

    • @Doesntcompute2k
      @Doesntcompute2k Před rokem +2

      Good config. With tuning, ZFS is really a monster with large datasets. If Oracle just would release the FULL ZFS build vs. the "open sourced" build, it would really be great.

    • @SyBernot
      @SyBernot Před rokem +4

      It's always a struggle to come to a good config because with every choice you gain something and loose something else. More vdevs means faster drives but less space. Bigger vdevs means less redundancy but more overall space. You really have to know what you need your storage to do before you ever start using it. I think Dave will have regrets on going with raidZ3 and 3 spares, modern drives don't fail as much as they used to but if his goal was to have bulletproof storage I think he's achieved it.

    • @georgH
      @georgH Před rokem +4

      @@Doesntcompute2k Actually, shortly after Oracle closed their ZFS, most of the ZFS developers (including historical and leads), quit and went on collaborating on OpenZFS (either on their own or with other companies), which, to this date, it gets the newest features and improvements (and Oracle can't get those back :)

    • @Doesntcompute2k
      @Doesntcompute2k Před rokem +2

      @@georgH Thank you very much! I completely forgot about Oracle closing down their division. The say they dumped on Solaris, ZFS, etc., is as bad as what HP did with their purchases. I didn't see the name of the devs in the OpenZFS maintainers, but I miss so many things these days. Glad they are contributing. I do know several of the maintainers worked closely with the TrueNAS teams.

    • @belperite
      @belperite Před rokem +1

      @@georgH Indeed, it's just a shame it can't get included in the Linux kernel because of the CDDL/GPL licensing issues* (and certain Linux kernel devs are ambivalent about ZFS at best). It's not even clear if Oracle could change the license agreement to GPL even if it wanted to (possible implications for the Sun / Netapp settlement and other issues). For most linux distros it has to be compiled by the user as a kernel module (although Ubuntu distribute it precompiled and haven't got sued yet).
      Personally though I've got a manually-built NAS based on Debian that uses ZFS and I'd never use anything else!
      *Discussed to death elsewhere so I'm not gonna respond to it here ;)

  • @BoldRam
    @BoldRam Před rokem

    My favourite episode, cheers Dave!! 👌👌

  • @grottyboots
    @grottyboots Před rokem

    I love your ability to explain failry complicated topics life parity bits or Hamming codes. Brilliant!

  • @daishi5571
    @daishi5571 Před rokem +2

    Back in 1999-2000 I worked for a bank in London (city of) and was put in charge of the servers in the disaster recovery building across the river (south London) I don't remember how many servers and it it fluctuated all the time, but it was always full and took up 3.5 acres. I would constantly be coordinating with the main building for data recovery testing because although I had 3.5 acres of servers it was only a mirror of what was in the main building and it was also mirrored to Germany. One day the bank of England wanted to know what the data delay was between the main building and my site and due to this being unacceptable (if memory serves it was up to 45 mins by the end of the day) I was tasked to decide how this could be remedied. My options were take up almost all the copper cables across the river (they did not want it bounced off other sites/exchanges) or fiber optics for which I was told there was plenty that was still dark so I chose fiber. It turns out the a lot of the fiber was damaged so they still ended up digging up the roads to put in more fiber (apparently the Bank of England is a really good stick to hit telecoms companies and local governments with because it was started almost immediately) So anyone who was stuck in traffic near either London or Southwark Bridges in the later half of 2000 I'm sorry.
    TLDR RAID across building/river is cool :-)

  • @marksterling8286
    @marksterling8286 Před rokem +10

    Loved this video, reminded me of my first production netware server, an eisa 486 with 2 scsi cards and 1gb scsi disks mirrored. The disks were full height 5 1/4 inch monsters. 2 years later I installed compaq proliant 1500s with the hot plug 2gb drives and the smart array raid5 with 4 active disks and a hot standby. At that office we ended up with 6 servers in that configuration. And with the 30 drives we would get at least 1 amber alarm on a disk every month or so.

    • @meggrobi
      @meggrobi Před rokem +2

      Ahh, Netware and a 486 server, fond memories. If they had implemented a true TCPIP solution, they may have put up a fight against Windows NT. As a file, print server and directory services, it was definitely a superior system to NT.

    • @marksterling8286
      @marksterling8286 Před rokem +2

      @@meggrobi I remember with netware 3.12 you could apply an tcp/ip stack but could only really use it to tunnel ipx/spx to a client. I had a huge amount of fun in those days with kit that seemed massive back then but tiny by todays standards. An example those proliant servers came with 80mb of ram that seemed massive back in the day.

    • @meggrobi
      @meggrobi Před rokem

      @@marksterling8286 yes, ipx/spx was not routable so you needed a bridged network or router to carry ipx/spx. I still have a set of NetWare 4 discs. It was major flaw not to use TCPIP as the internet was just taking off outside universities etc in the 80s.

    • @marksterling8286
      @marksterling8286 Před rokem

      @@meggrobi you could route ipx/spx but you needed to setup the router with ipx subnets you could also use the netware server as a ipx router if you wanted to save some money, back in the day I used to have both Ethernet and token ring at home and a netware 3.12 box did the ipx routing. At work we used the 3com net builder 2 routers. At the time I worked for a large steel company and we had a very large ipx network with about 200 servers visible and about 300 ipx subnets

    • @marksterling8286
      @marksterling8286 Před rokem

      Although we also had to bridge some network subnets because of netbios and netbeui for lotus notes

  • @infocentrousmajac
    @infocentrousmajac Před rokem

    Very informative video. Thank you

  • @stevendavies-morris6651

    Outstanding low key roll up yer sleeves and get to work presentation about RAID, Dave. Thanks for showing us the Houston Command Center. That looks great! I want that for my Ubuntu based media center server (8*3TB drives in RAID6 plus 2*128gb mirrored boot pair) that serves the music for my internet radio show. Subscribed and liked.

  • @TheInternetHelpdeskPlays

    I remember my first raid setup, I got an IDE expansion card for a cd writer and it supported RAID so I gave it a try out. 2x10gb spare drives I had in a stripe and used as my windows drive. It was amazing how much faster it was seeking in windows XP than a single drive.

  • @NeverlandSystemZor
    @NeverlandSystemZor Před 9 měsíci

    RAID is amazing... it's THIS old yet THIS solid and so reliable... just amazing logic behind it.

  • @MrJugsstein
    @MrJugsstein Před rokem

    really informative thanks Dave.

  • @thumbwarriordx
    @thumbwarriordx Před rokem +2

    Throwing a raid into a desktop PC I always felt the RAID 10 striped and mirrored approach with 4-6 drives made way more sense than any of the RAID parity options.
    Keep that CPU usage down at the cost of an extra drive maybe.
    The big gains for RAID 5/6 are on the upscaling. Whew boy.

  • @colinmaharaj
    @colinmaharaj Před rokem

    I've been doing mirrors in my dev drives for over a decade. no regrets, plus backups on external drive, and a backup computer for my dev work, and I'm doing a 3rd back up PC to try new stuff.

  • @georgH
    @georgH Před rokem +2

    I just love ZFS, use it with redundancy on both my main and backup pools. The zpool/zfs command line is very clear and simple to use.
    Thanks to the checsumming and periodic scrubbing, I discovered a fault on one drive on the backup pool, it was writing rubbish once in a while!
    Using znapzend to automate snapshots, snapshot thinning, and send it to the backup server. All the configuration it needs is stored using zfs properties of each filesystem, really neat having the configuration along with the pool itself.
    I created this pool using OpenSolaris back in 2008, then migrated to FreeBSD and, finally, Linux, which I feel more comfortable using :)

  • @LZeugirdor
    @LZeugirdor Před rokem

    Lately I've been studying and freaking out over filesystems and stuff like that using Linux with btrfs for raid as little jokes on my flash drives and whatnot. It's super cool and I love this sort of content

  • @muddyexport5639
    @muddyexport5639 Před rokem

    Thanks! Keep up the keeping on. Look forward to each new (evening saving some ;~).

  • @sigmb6
    @sigmb6 Před rokem +1

    I now finally understand RAID. Thank you.

  • @kencreten7308
    @kencreten7308 Před rokem

    I appreciate your logic, Dave!

  • @everythingquads
    @everythingquads Před rokem

    As always, enjoyed your video, thanks.

  • @HDSpaceFox
    @HDSpaceFox Před rokem

    first time somebody explains this issue in 100 percent informative manner. i am using a sync tool to sync my personal files (mostly vacation photos and also other pics from my photo camera hobby) from a single drive on the pc to a 4 drive raid5 NAS. but what if that also fails both the same time? this made me think about also adding a cloud subscription. its not that expensive as it was years before...

  • @wskinnyodden
    @wskinnyodden Před rokem +3

    Yeah, rebuild times are a pain... Disk capacities grown much faster than their speed. A 40Mb disk was faster to do a full disk scan/write than a 4Tb does and on a much slower interface. This is a stupid gap HDDs have had for a while. SSDs mostly fix it but I don't like the SSD longevity odds when powered down for long periods of time.

  • @glitchout0137
    @glitchout0137 Před rokem +1

    informative and entertaining
    thanks for another great one Dave.
    question for any who know:
    can you have an array of "Drives" made of a raid them selves. e.g. 2 banks(mirrored) of 6 drives in raid 5
    may be stupid question but came to me when I was watching this

  • @powerpower-rg7bk
    @powerpower-rg7bk Před rokem +3

    @15:54 RAIDZ2 is akin to RAID6 as that permits two drive failures while maintaining data in the array. RAIDZ (or RAIDZ1) is similar to traditional RAID5 in that a single drive can fail and the array is still functioning. RAIDZ3 can have three drives fail while data is still accessible.
    One thing also worth noting about RAIDZ2 is that the two parity calculations are different and when scrubbing data single bit errors can be detected and corrected.
    And for those curious, RAID3 did get some usage historically. It was where the parity information was put on a dedicated disk. The only modern implementation of RAID3 I've seen 'recently' were in some Texas Memory System DRAM based arrays about a decade ago. The reasoning for this is that since the drives were based on DRAM, rotating parity between memory channels impacted latency too much. The DRAM channels themselves had ECC and would only invoke the parity channel for recovery. The combination of ECC and parity did permit modules in these DRAM arrays to be swapped out while the system was running. Texas Memory Systems was purchased by IBM years ago and their technology has been implemented in various mainframes. Uniquely, IBM mainframes effect support RAID5 across memory multiple channels which their marketing calls RAIM. Pretty good timing for Texas Memory systems to exit the market as in a few years time, PCIe based SSDs would have displaced their entire product line offering fast, low latency storage at higher capacities and lower prices.

    • @blazewardog
      @blazewardog Před 9 měsíci

      Unraid's storage array works like Raid3. The 1-2 Parity drives contain all the metadata from the array with each logical sector matching up directly across all the data drives in the array. Also both Parity drives contain the exact same Parity data, so really Unraid is a raid 3 with a raid 1 drive set for Parity.

  • @c128stuff
    @c128stuff Před rokem +1

    'As long as the second drive doesn't fail before the first is recovered'
    If you get to deal with arrays often enough, you will run into this. For that matter... I have encountered this situation at least twice on the raid storage I use at home. 2 solutions:
    - use smaller drives. Rebuild times for 1-2TB drives are quite reasonable still, reducing the risk significantly.
    - use raid6

  • @geohanson
    @geohanson Před rokem +1

    Great video as always, but I think your ZFS descriptions are out slightly at 15:59
    ZFS1/RAIDZ1 = RAID 5 (1 parity drive)
    ZFS2/RAIDZ2 = RAID 6 (2 parity drives)
    ZFS3/RAIDZ3 = RAID ‘7’ (3 parity drives)
    RAID 7 doesn’t really exist as far as I know though other than in ZFS

  • @HelloKittyFanMan....
    @HelloKittyFanMan.... Před rokem

    "...So I will 'SPARE' you today." Haha, I saw what you did there; nice work!

  • @tubbydrmmr
    @tubbydrmmr Před rokem

    Dave, I'm just starting to dive into your content and I am really liking it. Thanks for another great video.
    Question: You said that with your 45 Drives server setup, you could in theory lose 9 drives before you lost data. In that scenario, if you lost two drives in each of your vdevs, you would need something to add to rebuild that data. If you then lost the 3 hot spares as well, you wouldn't have any drive available to rebuild any of the pools, and so you would lose the data unless you manually replaced the 2 bad drives (6 total) in each pool. Does this make sense and am I thinking of it correctly? Thanks, Tim

  • @FreonUWP
    @FreonUWP Před rokem

    LilNas, and the Storinator are my new favorite tech terms!

  • @DiegoPereyra
    @DiegoPereyra Před rokem +3

    Master Class, Master...!

  • @neorandy
    @neorandy Před rokem +2

    Gladly done with RAID and supporting Windows users, but enjoy your videos. Have you done one on 3.11/Windows for Workgroups?

    • @neorandy
      @neorandy Před rokem +1

      I installed my first SCSI drive in a Tandy 4000. It was a Hard Card SCSI from Tandy.

  • @butstough
    @butstough Před rokem +2

    also you should disable access time, or set relatime, and change the record size to 1MB, especially since youre storing video. having atime set sends your drive activity to the moon, as zfs is constantly writing updated access times to the disk as you read files. theres no downside to setting record size to 1MB, as zfs will use the smallest record size possible for small writes, like single text files or metadata changes.

  • @rcstl8815
    @rcstl8815 Před 10 měsíci

    I am in awe that you would have any mental inventory of that many drives. I can't even find an email from last week!

  • @yellowcrescent
    @yellowcrescent Před rokem +1

    Nice overview! Have seen RAID6 fail many times at work... although usually from hardware RAID controller and disk issues (Samsung EVO SSDs + Dell PERC). When you start a fsck and it segfaults, or everything is getting moved to lost+found, you've got a long night ahead.. switching to Intel DC SSDs solved the issue, so makes sense to use proper datacenter drives with hardware RAID controllers :P Also, the RAID5 scenario where you lose an additional disk while rebuilding is happening is not uncommon, since the rebuild process really stresses out the remaining disks. At home, I use a single-node Ceph setup (osd level), since it makes expanding much easier than ZFS, and I can use CephFS + SMB for sharing.

  • @VraccasVII
    @VraccasVII Před rokem

    Holy hell, I did not expect that torching!

  • @Neeboopsh
    @Neeboopsh Před 11 měsíci

    the question in the thumbnail reminds me of a question asked of one of the science consultants on star trek, and his answer
    "how do heisenberg uncertainty compensators work?"
    "quite well, thanks"
    ;)

  • @davidberndt6275
    @davidberndt6275 Před rokem

    i was reminiscing about sub-block allocation and how we don't even concern ourselves with slack space these days let alone head count, clusters, etc

  • @KirkDickinson
    @KirkDickinson Před 9 měsíci

    That Houston Command Center looks like it does about what TrueNAS does. I have two TrueNAS servers set up in my office. One with 6-8TB drives, and a backup of 6-8TB drives. Both running ZFS-Z2. The production server takes snapshots every hour and every day copies to the backup server every evening. The servers are set up identically so if the production server goes down, I can remap my SMB's and be up and running on the backup in short order.
    All workstations back up documents, photos, emails, and a boot drive image to the production server on a regular basis.
    I am only using about 35% of my storage capacity on the production server, so there is a lot of room for snapshots.
    Critical DB data is also backed up in the cloud. Every document, photo, email archive, etc is synced back to extra drives in individual workstations in my office.
    My plan within a couple months is to put together another TrueNAS box and install it in a building offsite that will connect with a ubiquity wireless so I can have a totally offsite backup.
    I wish that there were affordable tape backup solutions now days for another layer of backup.

  • @stapedium
    @stapedium Před rokem

    More details on you backup and restore regimen please

  • @smeuse
    @smeuse Před rokem +4

    ST-251? I would have sworn it was an ST-225 :) I loved the sound of that era of disks. You can almost hear each byte being written....

    • @Doesntcompute2k
      @Doesntcompute2k Před rokem +1

      So many of those old Seagates. And Miniscribes--before they shipped bricks, I mean. The 225 was indeed a classic.

    • @wesley00042
      @wesley00042 Před rokem +1

      I remember that little ST-251 seek test on power up.

    • @benz-share9058
      @benz-share9058 Před rokem +1

      I had (still have somewhere?) an ST-251. They were popular classics around about 1988. When a MB was a bunch of data! Great to see and hear one once again.

  • @mobilephone4045
    @mobilephone4045 Před rokem

    Would love to hear your thoughts about snapraid

  • @shephusted2714
    @shephusted2714 Před rokem +1

    setup writing to a couple of nas each with its own raid arrays - flash is changing the game - value tier smb needs to go to 100g to complement faster speeds and dirt cheap storage - you need 10g/s to saturate 100g but expect 6-8gb/s - you can upgrade to 200g with bonding/bridge. flash/ssd/nvme nas should help with some reliability - no moving parts

  • @tropicalspeedbmw
    @tropicalspeedbmw Před rokem +1

    Haha the Count segment was great

  • @timlong7289
    @timlong7289 Před rokem

    @Dave have you looked at Unraid at all? It has some nice advantages over Raid but probably some disadvantages, too. I like that you can easily extend the array forever and that you can just unplug a disk and still read the filesystem on it.

  • @ddhanlin
    @ddhanlin Před rokem

    @Dave's Garage. I heard you mention in the video that you have OneDrive cloud storage in your mix. I also have a small Raid 1 NAS setup with my files existing on both my NAS and OneDrive.
    In addition to the "regular" Onedrive... I also use OneDrive Personal Vault. The only way I could figure out to sync the Personal Vault portion of OneDrive to my NAS was to manually run a Powershell script. Do you have any thoughts/ideas that could automate the syncing? Any of Dave's Garage listeners familiar with Personal Vault have thoughts/ideas? The Personal Vault sits behind 2FA access and some pretty heavy encryption and is therefore not "seen" by any syncing software that I could find.

  • @thinkpadBentnoseTheBlind
    @thinkpadBentnoseTheBlind Před 11 měsíci

    I seem to recall having to install scuzzy drivers to make the raid work on a desktop build in the early late 90's

  • @NovaLand
    @NovaLand Před rokem +1

    I've read that in theory, the MTBF is smaller than the time it takes to restore a modern sized drive (even with hotswap), making raid actually not living up to.. anything. Sure, I'm using a raid 6 configuration myself, but it's good to be aware that when comparing numbers of MTBF and time to restore a drive, it could be a problem. Another issue that could be handy to know is that modern production is so streamlined, that if a drive fails and you have a bunch of drives from the exact same batch of manufacturing, it is very possible that more drives could fail VERY soon. So a good way to solving that problem is to buy drives from different batches to make sure they don't fail at the same time.

  • @cyclic2696
    @cyclic2696 Před rokem

    Please can you put a link in about the LED devices glowing and moving in the background - they're such a tease and always out of focus! Thanks. 👍😃

  • @jackeriksen6753
    @jackeriksen6753 Před rokem

    Hey Dave! When you say non recoverable, does that meen secure aswell?
    Lets pretend I had a RAID 0 stripe, but to back it up I sent the "data" on one disk to a datacenter and the other disk to another, would anything be interceptable or readable without access to both locations?

  • @larzblast
    @larzblast Před rokem

    Interesting that you chose Houston over TrueNAS Scale. Did you examine TrueNAS Scale and if so, what made you choose Houston as a result?

  • @dennisfahey2379
    @dennisfahey2379 Před rokem +1

    So each NAS is created to back up the last NAS because the size is so big it will not fit on anything else. So much for the 1,2,3 backup paradigm. I miss the early days of PC backup: audio cassette, VHS, 9-track - even paper tape. I recall a paper tape storage system company running ads indicating that paper was the only medium that would survive in an EMP attack. Backup has always been a fear tactic business.

    • @DavesGarage
      @DavesGarage  Před rokem

      You miscalculate. Multiple older small NAS units always get backed up to the new, largest, so there are always 2-3 copies. If I were filling the big NAS with nowhere to back it up to, that'd be silly.

  • @zzstoner
    @zzstoner Před rokem

    The Roosters of Lake Bill. Now there's some nostalgia.

  • @chbruno79
    @chbruno79 Před 5 měsíci

    Hi Dave. I have a data recovery business and I'm wondering if I can use this video on my website (of course, giving you the credits), because this is the best explanation I ever found on CZcams about how RAID works. Please let me know if you are ok with that.

  • @KROUTCHY
    @KROUTCHY Před rokem

    nice, thnx

  • @RaymondDay
    @RaymondDay Před rokem

    Aliso wow how much power is all them drives taking? So do you run off green power?

  • @annieworroll4373
    @annieworroll4373 Před rokem

    I've got a 4TB RAID 1 set up. Quite a bit cheaper than 4TB of SSD storage. Quite a bit slower, of course, but given the fault tolerance of mirroring it's probably a bit safer for my data. I've got a couple NVMe drives for things that need to be fast.

  • @stapedium
    @stapedium Před rokem +3

    A question I’ve always wondered about mirrored systems. How does the computer know which of the two is broken?

    • @1971merlin
      @1971merlin Před rokem +2

      Relies in the drive itself returning an error code saying the data could not be read (or written).

    • @giornikitop5373
      @giornikitop5373 Před rokem

      yes, check hamming code and you will find that's both easy to understand and kind of genius for it's time. although it was primarily made for communications, the method is used in all sort of different areas. basically, with a bunch of XOR operations, which even the most pathetic chip can do fast, you can find if and where an error is present.

    • @1971merlin
      @1971merlin Před rokem +1

      @@giornikitop5373 there is no hamming data in a mirror.

    • @giornikitop5373
      @giornikitop5373 Před rokem

      @@1971merlin how are mirrors checked for errors?

    • @wesley00042
      @wesley00042 Před rokem +2

      @@giornikitop5373 The drive has its own ECC and reports sectors as bad if it can't get a good read.

  • @JesseHughson
    @JesseHughson Před rokem

    Do you have any issues with WiFi conflicting with your LED data signal using FastLED on ESP32?

  • @Doesntcompute2k
    @Doesntcompute2k Před rokem

    Another GREAT video Dave! I always enjoy how you go through the subject matter, step by step.

    • @DavesGarage
      @DavesGarage  Před rokem +1

      I appreciate that! Thanks!

    • @Doesntcompute2k
      @Doesntcompute2k Před rokem

      @@DavesGarage Dave you attracted a copy-cat spammer! 😁 Congrats, it's the big-time for you! 😁

  • @GeorgeZ213
    @GeorgeZ213 Před 11 měsíci

    How do you connect 15-60 drives to the mother board? I had raid 5 on my old desktop, plus 3 video cards about 12 years or so ago.

  • @dregenius
    @dregenius Před rokem

    Recently built a storage server using some hacked up pieces of HP server drive cage and backplane and an LSI 9271-8i, first time building a PC or working with PC hardware in a while, and I was tempted to just go for RAID5 as I'd done in the past, however I noticed a lot of people online saying that RAID5 is not useful, basically at all, with large drives... something about the chances of recovering from a drive failure on big drives like 12 or 14TB being in the 1-3% range... I avoided RAID5 but still wish there was a good explanation of exactly why the chances of successful recovery plummet so severely!

  • @wskinnyodden
    @wskinnyodden Před rokem +2

    Yeah... I've lost data before stupidly... Still can't get 12Tb backup online at a decent price, and even if I did, synching that at about 20Mbps upload speeds would be like watching paint dry.

  • @chad_levy
    @chad_levy Před rokem

    I have an 8-bay Synology and went with RAID 6 for 2-drive fault tolerance. Once you get to a sufficiently-large pool size, the act of rebuilding the index can be quite stressful on the drives. Having no fault tolerance during one of the more stressful things you'll put your drives under is a risk I prefer not to take. Also, another aspect of disaster recovery in general is file versioning. Without versioning you're susceptible to crypto locker attacks or just accidental overwrites which can result in data loss. It's easy to do on Windows with File History. I think Mac's Time Machine does this too. I use File History for work stuff - mainly source code - and rely on Backblaze's versioning for larger, lesser-important files like my media collection. Also, in terms of cloud backup, I prefer Backblaze as it's got among the best bang for the buck and Synology's backup software allows me to encrypt my files before upload so they're reasonably safe.

  • @RudysRetroIntel
    @RudysRetroIntel Před rokem +2

    Excellent explanation of RAID!! Do you use read and write cache? Thanks for sharing.

  • @wskinnyodden
    @wskinnyodden Před rokem +1

    RAID in my home country is a fly killing spray (bug bomb).
    That said, I have a ZFS dual redundant pool (custom built) using recovered SAS disks (not the best idea, but got lucky I suppose, but better shut up as I need to replace one already, I do have them spares but its a pain to track which one)

  • @paulrobertson3326
    @paulrobertson3326 Před 8 měsíci

    Did you do a video on RAID using windows storage mgr?

  • @fattomandeibu
    @fattomandeibu Před rokem

    I only have a single NAS. Nothing fancy, just a Raspberry Pi and 8tb USB3 HD. It has image files for every system drive for every system in the house, and is also used as a media server for the smart TV and various game consoles.
    The most important files(family photos etc. that I consider truly irreplaceable) get backed up to everything I can get my hands on, using the "if I back it up enough times, surely they won't all fail at once" method. Off the top of my head, I use USB flash drives, USB hard disks and optical discs. I'd probably still use floppies if they had them in useful capacities. For the USB drives, I have a sorta "3 rolling auto-save" where I always have 3 USB sticks and hard drives with my 3 newest backups, but I have redundant optical disc backups going back over 20 years.
    For hard disk and flash storage that rarely get used, good practice of refreshing disks to help guard against bit rot might be a good topic for another video. It always seems like something a lot of people aren't aware of, at least for flash devices.

  • @BlackHoleForge
    @BlackHoleForge Před rokem

    I'm so glad that today's video sponsor wasn't raid Shadow Legends. Good job Dave

  • @bjre.wa.8681
    @bjre.wa.8681 Před rokem

    can all the drive be put in a new NAS box and maintain the data? Im not a big fan of any raid. I like to know what drive a particular file is on. Maybe I'd like to take that drive out for security reasons.

  • @archivushka
    @archivushka Před rokem

    I think the raid 0 is perfect for loading data that in case of a data loss can be renewed from a server. Such as PC games with cloud saves

  • @Muzer0
    @Muzer0 Před rokem

    I was always taught that "RAID nought is naughty" because it doesn't do redundancy, as a convenient way of remembering it

  • @RaymondDay
    @RaymondDay Před rokem

    I got some LVM drives on my Ubuntu server. World houston make it easy to use them? If so how do I install it?
    Thought it be as easy as this:
    apt install houston
    Reading package lists... Done
    Building dependency tree... Done
    Reading state information... Done
    E: Unable to locate package houston
    Thank you for your grate videos.
    -Raymond Day

  • @guilherme5094
    @guilherme5094 Před rokem

    Nice👍

  • @sleve_mcdichael_
    @sleve_mcdichael_ Před rokem

    will you share any info on your softwareonline business?

  • @HelloKittyFanMan....
    @HelloKittyFanMan.... Před rokem

    Since striping and bunching are so risky, then it's interesting that those were even made as options.

  • @bluehatguy4279
    @bluehatguy4279 Před rokem +1

    I'm a bit curious how ZFS compares with BTRFS.
    Between the two, I've really only tried BTRFS. When it works, BTRFS is like the best thing ever, and when it breaks, BTRFS is kinda the worst thing ever.

    • @travis1240
      @travis1240 Před rokem

      And it does break.... I went back to EXT4 because I wanted to keep my data.

  • @LinasR
    @LinasR Před rokem

    Do HDDs go sleep if unused to save power?

  • @UncleMinecraft
    @UncleMinecraft Před 9 měsíci

    I'm really careful about backups. Not got a 24 disk array though...😢

  • @Rennu_the_linux_guy
    @Rennu_the_linux_guy Před 9 měsíci

    damn ive never seen someone explain the hamming code thing in such an intuitive way

  • @talibong9518
    @talibong9518 Před rokem +1

    What's better for a typical PC for a storage partition? 3 HDD's in RAID 5, or 2 striped disks with a 3rd to back up important data to? I'm leaning towards RAID 5 but don't wanna waste CPU cycles on parity checks.

    • @Doesntcompute2k
      @Doesntcompute2k Před rokem

      Striped volumes will always fail; Two, three, 20--doesn't matter how many and you'll loose (likely) 100% of your data. Striped volumes are good for really fast temporary storage. Building a huge video image, .iso, etc., on a WS (workstation) before copying to NAS or a server? Stripe. RAID5 in three volumes, the same sizes, is a good solution. You're speed is better with four or five. You can--if you really need some speed and can spare the drives--stripe (RAID0) two volumesets with are mirrored (RAID1) of two drives each. So a two volume stripeset in this case would have four drives in two mirrors "under it." I do this all the time with SSD/NVMe for speed on a WS. Just a thought. :)

    • @LiraeNoir
      @LiraeNoir Před rokem

      We're coming to the age of saturating PCIE 4 NVME drives. In that age, do you *really* need more bandwidth? Some servers do, but those are quite rare. On the desktop, I'm struggling to find a use case. Maybe extremely tuned raw 8K video edits? Maybe...