back to article Why are enterprises being irresistibly drawn towards SSDs?

SSDs have been the subject of hype, hype and more hype. Any of us who have used them in our personal computers know the benefits SSDs bring, but personal experiences and hype alone don't explain the robustness of enterprise flash adoption. Flash has a lot of naysayers. Any article on The Register about flash inevitably invites …

Page:

  1. Borg.King

    Change in Flash technology to eliminate finite write lifetime?

    I would imagine that eliminating the lifetime contraints of a flash cell has to be one of the goals current and future manufacturers are striving for. The recent Aluminium / Graphite rechargeable battery work gives me some hope that a similar approach might be feasible for long term storage solutions.

    Of course, once your storage device has a long life time, the number of replacement units you sell is going to decline and reduce your income.

    1. Trevor_Pott Gold badge

      Re: Change in Flash technology to eliminate finite write lifetime?

      A) The Al/C battery work doesn't port to silicon chips. It is unlikely we will ever see flash chips without write limits.

      B) You'll sell just as many new units even if your units last forever because our demand for data is insatiable. What use is a SATA 32GB SSD today, excepting in some very niche applications? Hell, what use is a 120GB? Would you buy a 240GB for your notebook?

      Flash write lives aren't being artificially suppressed. It's just physics.

      1. frank ly
        Happy

        @Trevor_Pott Re: Change in Flash technology to eliminate finite write lifetime?

        "What use is a SATA 32GB SSD today ....."

        As the root, /home, swap and /{data} partitions of my desktop computer. If I want 'big data' I just push a 1TB spinning rust or 128GB SSD into one of the front panel slots (or use a USB-3 adapter cable to connect a big SSD to my laptop). I don't anticipate this situation changing for me for many years and I'm sure most people at home (a big market) would be able to do the same.

        1. Archaon

          Re: @Trevor_Pott Change in Flash technology to eliminate finite write lifetime?

          "What use is a SATA 32GB SSD today ....."

          Even taking frank ly's *Nix example as a given, I've got a machine with a pair of 30GB (not even 32GB) SSDs in RAID 1 which runs Server 2012 R2 Standard quite happily. Believe it typically sits at around 11GB free.

          1. Trevor_Pott Gold badge

            Re: @Trevor_Pott Change in Flash technology to eliminate finite write lifetime?

            "Even taking frank ly's *Nix example as a given, I've got a machine with a pair of 30GB (not even 32GB) SSDs in RAID 1 which runs Server 2012 R2 Standard quite happily. Believe it typically sits at around 11GB free."

            And I've got OS-only installs of Server 2012 R2 Standard that eat the better part of 80GB.

            You *might* be able to convince me if you tried to make a case for 32GB SSDs as an ESXi disk, except that's probably useless since there are USB keys that are better fits for that job, and just plug directly onto the motherboard (or into the SATA plug).

            Dragging along 32GB SSDs is an exercise more in being spectacularly cheap than anything else. I get it - I am an SMB sysadmin, we have to do this all the time . But the hassle of migrating components from system to system as everything else dies (or the system isn't worth the electricity it consumes) gets old fast.

            A dirt cheap thumb drive solves the problem of a place to put a hypervisor, and the ancient SSD from the beforetimes isn't going to help me run my datacenter. It might be useful to the poorest of the poor consumers, or people in some extreme niches, but as a general rule storage devices aren't much use to general market past about 5, maybe 6 years. After that they're just too small.

            A great example is the 1TB magnetic disk. I have an unlimited number of these things. I can't and won't use them. It costs me more to power up storage devices to run those drives for the next three years than it would to just go get 4TB drives. To say nothing of space, cooling, OPEX, etc.

            Even if all our storage devices lasted forever, they would eventually stop being used. Just like my Zip drive. Just like my Blu-ray. Newer devices hold more, and they are less of a pain in the ASCII to use.

            1. Archaon

              Re: @Trevor_Pott Change in Flash technology to eliminate finite write lifetime?

              "And I've got OS-only installs of Server 2012 R2 Standard that eat the better part of 80GB. etc"

              There's shades of grey (not 50), not just black and white, you know? The question, as always, is what is correct for the customer and their environment. 30GB SSDs are not wrong. 80GB SSDs are not wrong. USB drives are not wrong. It depends.

              I understand that just because I have Server 2012 R2 running at 19GB on a 30GB RAID 1 set does not mean everyone can or wants to do that. If you need 80GB, then you need 80GB, it's that simple. You should also understand the reverse is true, that if you only need 30GB there's no point paying for 80GB or 120GB, especially if the disks are going to go in the bin (or the dark corner of a cupboard) once the server is decommissioned. Call it 'spectacularly cheap' if you like, because yeah, in this instance that was the point.

              What I suspect you really mean is that relative to the cost of buying a slightly larger disk it's not worth the hassle to watch over machines with small system disks and maintain them. After all the time, effort and wage/consultancy costs of IT staff are a finite resource just like anything else, and I'm sure there's better things to spend on than worrying about whether a cached Windows Update has taken one of your systems over the edge. And that's a completely understandable and completely correct/justified point of view in many organisations.

              Many (not all) customers prefer redundant (i.e. RAID 1) boot drives. That kicks out USB drives but you'd get away with a RAIDed dual SD card module. Of course a USB key is fine if the customer doesn't mind.

              As I said, it all depends on the customer and the environment. The point is to take a step back, look at it objectively and not be blinkered into a default position.

              1. Trevor_Pott Gold badge

                Re: @Trevor_Pott Change in Flash technology to eliminate finite write lifetime?

                No, Archaon, in your non-objective, blinkered position on things you've missed the whole thrust of my argument: namely that there is no value - except in some very niche situations, including outright poverty - in recovering 10 year old drives from systems and reusing them, even if their lifespan was infinite.

                Just because you can take a 32GB SSD out of some ancient system and reuse it in a newer one (with a whole metric ****pile of TLC and babying) doesn't mean it's sane, rational, profitable or otherwise a good idea. It's also something that the majority of individuals or businesses will do.

                You, personally, may do it. That doesn't make it a good plan> It doesn't make it what the majority will do, would do, or even should do. And that, right there, is the whole damned point. Which you seem to be unable to grok.

                1. Archaon

                  Re: @Trevor_Pott Change in Flash technology to eliminate finite write lifetime?

                  "No, Archaon, in your non-objective, blinkered position on things you've missed the whole thrust of my argument: namely that there is no value - except in some very niche situations, including outright poverty - in recovering 10 year old drives from systems and reusing them, even if their lifespan was infinite."

                  What you don't seem to "grok" from your own blinkered position is that I pretty much agreed with you. Just making the point that there's alternatives. As I said, shades of grey. Not just Mr Pott's opinion and wrong.

                  1. Trevor_Pott Gold badge

                    @Archaon

                    I fairly explicitly stated in my original comment that there were other possible alternatives. I also made it pretty clear that they were niche and not very relevant. You then decided that the alternatives had to be spelled out and attempted to make them seem relevant.

                    That was not only pointless, you did not succeed in making them seem relevant at all. Which has now become the point of this thread.

                    If I said "for the purposes of creating the circle used in $company logo, the value of pi used was 3.14159{irrelevant additional numbers}" you'd be the guy not only explaining "pi is more than 3.14159, and in sometimes it matters that you use {long string of numbers}". And you'd be explaining that to the guy who owns http://www.tastypi.com.

                    Rock on!

                    1. Anonymous Coward
                      Anonymous Coward

                      Re: @Archaon

                      On Trevor's articles, the only pedantry allowed is Trevor's. Anyone else's pedantry will be mocked until either you give up or become so enraged that you lose objectivity.

                      1. Archaon

                        Re: @Archaon

                        "On Trevor's articles, the only pedantry allowed is Trevor's. Anyone else's pedantry will be mocked until either you give up or become so enraged that you lose objectivity."

                        That would appear to be the case. The former, that is.

                      2. Trevor_Pott Gold badge

                        Re: @Archaon

                        The tears of enraged commenters power my happiness.

          2. Steven Raith

            Re: @Trevor_Pott Change in Flash technology to eliminate finite write lifetime?

            @Archeon

            ""What use is a SATA 32GB SSD today ....."

            Even taking frank ly's *Nix example as a given, I've got a machine with a pair of 30GB (not even 32GB) SSDs in RAID 1 which runs Server 2012 R2 Standard quite happily. Believe it typically sits at around 11GB free."

            Come back to me in a couple of years when WinSXS has eaten up the rest of that ;-)

            Steven R

            1. Archaon
              Trollface

              Re: @Trevor_Pott Change in Flash technology to eliminate finite write lifetime?

              "Come back to me in a couple of years when WinSXS has eaten up the rest of that ;-)"

              Oh no. The terror. You mean I have to set up a StartComponentCleanup scheduled task to run once in a blue moon? Oh no. Save me from this tragedy please. The world hath ended and the fiery gates of hell have opened to swallow me up.

              Come off it.

              1. Trevor_Pott Gold badge

                Re: @Trevor_Pott Change in Flash technology to eliminate finite write lifetime?

                StartComponentCleanup does not prevent WinSXS from growing unchecked. It just slows the progression somewhat.

        2. Trevor_Pott Gold badge

          Re: @Trevor_Pott Change in Flash technology to eliminate finite write lifetime?

          "As the root, /home, swap and /{data} partitions of my desktop computer. "

          Desktop linux is pretty goddamned niche. From my original comment:

          "What use is a SATA 32GB SSD today, excepting in some very niche applications?"

          Funny how when you quote it you leave off the last bit.

          Also: " I'm sure most people at home (a big market)" won't be using Linux on the desktop. Doubleplus when we talk about putting different directories on different drives. Sorry, mate. You're not so much in a class by yourself as homeschooling from a tree in the middle of the Yukon.

    2. Archaon

      Re: Change in Flash technology to eliminate finite write lifetime?

      "I would imagine that eliminating the lifetime contraints of a flash cell has to be one of the goals current and future manufacturers are striving for."

      Proper enterprise SSDs are rated for 20+ full drive writes per day (DWPD) over a 5 year period. Endurance is not a problem in most, it's just a case of choosing the right SSD for the job. A good quality SSD being used for the correct task for the SSD type shouldn't hit it's endurance limit within the usable life of a system; and in terms of manufacturing defects etc I believe I'm right in saying that the failure rate of SSDs is considerably lower than of spinning disk.

      If you put a consumer grade SSD (normally rated well under 1 DWPD) into a server that's constantly caching or writing small data (as per one of Trevor's examples) then yeah sure, it will murder it in short order. For the exact opposite reason you're not going to put one of those 20+ DWPD enterprise grade drives in your desktop PC because they're expensive and there's no benefit to you in doing it.

      A part of me dies every time I see someone decide to try and be clever and save money by putting Kingston SSDNow drives* in their production servers or - god forbid - a storage array.

      * No hate for Kingston, just an example of a cheap consumer drive that sprang to mind.

      1. Anonymous Coward
        Anonymous Coward

        Re: Change in Flash technology to eliminate finite write lifetime?

        A Kingston SSD Now has about a 10th of the write endurance of a good enterprise SSD but it is also about a 10th of the price. It is perfectly suitable for some uses.

    3. prof_peter

      Re: Change in Flash technology to eliminate finite write lifetime?

      Until recently the *only* goal of flash designers was to deliver more bits for less money - until maybe 3 years ago maybe 95% of the flash that was produced went into iPods, cell phones, thumb drives, and SD cards, with SSDs accounting for a total of 3% of flash production. In the 10 years before that point, flash performance went down by nearly a factor of 10, and lifespan of low-end flash went down by far more.

      A lot can change in 3 years, though, and there is now enough demand for flash in performance-critical roles to at least halt this decline. (whether enough customers will actually pony up the money for higher-performance chips to make them profitable is debatable, though.)

      The reason we don't see flash devices failing left and right due to wear-out is because they've been getting bigger more rapidly than they've been getting faster. (or more accurately, than computer I/O workloads have been getting faster) The internal wear-leveling algorithms distribute writes evenly over the chips, so if you build an SSD out of consumer-grade chips with a 3000-write lifetime, you have to over-write the *entire* device 3000 times before the first chips start failing. (Sort of. A small number will die earlier, but there's spare capacity that can deal with that.) For a 500GB laptop drive, that's 1.5PB of writes, or 35GB/hr, 24/7, for 5 years. For a 1TB enterprise drive full of 30,000-cycle eMLC chips, you'd have to run it at 700GB/hr (200MB/s) for those 5 years to wear it out.

      And wear-out isn't just a problem for flash anymore - the newest high-capacity drives are only rated for a certain total volume of reads and writes, due to head wear effects, which works out to a shorter lifespan than flash. (and reads count, too, while they're free on flash) See http://www.wdc.com/wdproducts/library/other/2579-772003.pdf for the gory details...

  2. Anonymous Coward
    Anonymous Coward

    I like spinning disks. They may not be as fast, but in case of failure a disk has the opportunity to get some info off it 90% of the time.

    SSD's can go from good to bad in a microsecond taking all data with it. I use SSD's on my gaming rig, but keep my long term data on a pair of traditional RAID-1 4TB SATA drives.

    1. This post has been deleted by its author

      1. Rebecca M

        OK, you knew that, but did you realise that the risk of silent data corruption is actually higher with a RAID-1 array than it is with a single disk? No? Well, it is.

        You now have two disks, each with it's own on-board electronics, cache RAM, and firmware bugs, storing the same data. The same data is read from one drive one time you access it, and another driver another time. So the chance of any block of data being silently affected by a drive fault it doubled. Your RAID controller won't even notice, and will pass the bad data up to the application level twice as often.

        Bollocks. The overwhelming majority of hard drive errors (>99%) are between the platter and the head. Controller gremlins are so rare that you can essentially ignore them from a statistical perspective. Look at where the real errors happen - either the head didn't record the right thing in the first place (e.g. undervoltage coil), spontaneous corruption (i.e. a bit is toggled on the disk surface) or the head is unable to read valid data from the disk as the result of some issue after writing.

        In all of these cases the read will fail the CRC check built into any hard drive for decades, and the drive firmware will typically retry the operation a few times. If it manages to get the data subsequently the sector is mapped out and re-written in a spare block. This is standard single drive stuff, no need even for RAID at this point. The error only gets reported when the drive abandons the read. If that happens in a system with no redundancy you have a problem. With RAID it is not an issue - the sector is reconstructed from parity information. The possibility that data can be misread from the disk, passes the integrity tests even though it is invalid and passed on to the application without comment is remote in the extreme - the error detection mechanisms built in as standard work. When people talk about silent errors on the hard drive they are generally talking about spontaneous transitions that prevent the data being read as these checks fail, not bad data coming out of the drive without comment.

        As said before that eliminates well over 99% of errors but we'll call it 99% for ease of analysis. If we eliminate 99% of errors through RAID1 but double the remaining 1% does that mean the system as a whole is more or less reliable? You can't simply pretend that 60 years of research in data storage hasn't happened. Basic ignorance of hard drive integrity checking here, Blurays as the backup gold standard last week, a robust backup strategy that had no redundancy the week before. Perhaps it is time to stop pontificating and start learning.

        1. This post has been deleted by its author

          1. Rebecca M

            Secondly, as for the idea that these problems are so rare that you can essentially ignore them, just a few weeks ago, Corsair released a firmware update for one of their 60 Gb drives, (and I'm not picking on Corsair particularly, but it was an issue that affected me, so I have the details to hand):

            Great, so you've managed to prove that controller errors occur. Great, but it doesn't get you anywhere. You've also accepted that notional 1% error rate from the electronics which is what you need to bear in mind when defending the point you originally made:

            You now have two disks, each with it's own on-board electronics, cache RAM, and firmware bugs, storing the same data. The same data is read from one drive one time you access it, and another driver another time. So the chance of any block of data being silently affected by a drive fault it doubled.

            With that 1% figure in mind you have to show not that controller errors exist or even that they increase with the number of drives. You have to show that doubling the number of drives increases the error rate of the associated electronics not by a factor of two but by a hundredfold, simply to get back to where you were, or increase 200x to get to that claimed doubling of risk.

            That's a big claim to make. Arguing about whether the errors attributable to that circuitry is 1% or 2% is most of the statistical insignificance to which I was referring. The rest is of course a read error that causes a CRC to pass even after corruption. Yes that's always possible too, but generally a one in 2^16 chance even for a completely garbaged sector (we're not talking one or two bit errors here). Even combined the two effects are small enough to make nonsense of the entire argument.

            1. This post has been deleted by its author

          2. the spectacularly refined chap

            Even if it was 99.9999999999%, then statistically you would experience 110 corrupted bytes in every terabyte of data you move. Doesn't that worry you at all? When you rebuilt the imaginary disk array you speak about above, you'll have no redundancy, (unless you're running RAID 6, and even then you could only hope to IDENTIFY the error, not correct it).

            You are using the wrong set of figures in your calculations. If 99% of errors are eliminated you can do nothing to extrapolate the number of errors remaining without reference to the starting error rate - here you assume that all reads even from a good disk are errors. Use the same set of figures consistently and the position becomes a lot more difficult to justify.

            1. This post has been deleted by its author

              1. the spectacularly refined chap

                I know, and I didn't make the original claim, Rebecca did. I simply took her argument and pointed out that even changing 99% to 99.9999999999% still did not produce the result she wanted to, thereby reducing her argument to absurdity.

                Your original position is clear enough - RAID1 increases the scope for corruption. You have twice as much data to go wrong, which copy is read is essentially random so if data is constantly read and written back without checks the scope for corruption across both disks is increased. This is all fine and I would agree with it.

                Rebecca's point is that checks are made in the form of the CRC attached to each and every sector. For any read that results in an error (detected or not) there is a good chance that it will indeed be detected at the CRC checking stage - that is the 99% figure being bandied about. The remaining 1% are the cases that either slip through CRC checking or occur subsequently to it.

                In the 99% cases the error is reported to the RAID controller (hardware or software) so the data is retrieved from the mirror. No data loss results thanks to the presence of that mirror. It is only the other 1% of cases that go undetected that can spread corruption from one disk to the other. This is what happens when an error occurs: it says nothing of the probability of an error occuring in the first instance.

                1. This post has been deleted by its author

                  1. Solmyr ibn Wali Barad

                    "OK, we agree up to a point - great."

                    Good call. It's an interesting discussion, and there's no need to get too emotional about it, because everybody had valid points to make.

                    Please consider two topics that were not mentioned yet:

                    1) PRML encoding method, which was introduced around 1991, and has been prevalent ever since. With major updates of course.

                    PRML means that reading bits off the magnetic media is not a straightforward process, it's a highly sophisticated guesswork. So an uncertain amount of uncertainty is to be expected.

                    2) T10-PI is a system of additional checksums that is designed to combat silent corruption problems discussed here. PI-capable drives can be formatted with 520-byte sectors, 8-byte checksum is kept inside the same data frame as 512 bytes of payload, and consistency checks are performed along the whole data path.

                  2. Rebecca M

                    I know, we all know that. One of the incorrect assumption made by some people, is that these checks always either correct the error or return an error flag to the RAID controller or OS, or that the cases where they do not, (bad data is passed upwards as good), are so rare as to be ignorable.

                    Go back to my original post and you see that I acknowledge their existence but point out that is the uncommon case. You have accepted that uncommon nature. Now let's consider your different error types:

                    1. Complete and sudden failiure. The classic, 'doesn't spin up'.

                    An irrelevant distraction from your original point which specifically excluded drive failure. Let's move on immediately.

                    2. Media errors which enter the drive's onboard controller, are corrected by the controller logic, and good data is passed upwards to the RAID controller or OS. You don't even know this is happening, unless you notice reduced performance or monitor SMART statistics.

                    Well, yes you do since the block remap table is available to any software that asks for it - this isn't even a SMART feature. Again, it doesn't alter the analysis one jot since it happens in both single drive and RAID configurations.

                    3. Media errors which enter the drive's onboard controller, which is detects but cannot correct, an error flag is passed upwards to the RAID controllor or OS. The classic 'Error reading drive C:, (r)etry, (i)gnore, (a)bort, or (f)ail?'.

                    Now this is the key point - this is the instance a single drive configuration can't recover from but a RAID1 setup can, by reading the other disk.

                    4. Controller firmware bugs, (or other rare causes), that pass BAD data upwards to the RAID controller or OS, as if it was GOOD data. Rebecca originally claimed that this never happens. Now she is claiming that it is a very low percentage of errors.

                    I never claimed anything of the sort, just that the effect is so small we can ignore it when considering your claim. Remember your claim: that a RAID1 configuration suffers from more silent corruption than a single drive setup. We have already established that the case 3 errors are the vast majority of errors of this kind and that RAID1 virtually eliminates them. Fiddling around with this tiny percentage of errors does nothing to make that original claim correct unless the rate of them goes up by several orders of magnitude. That is what you have to show and what you have failed to do.

                    It is your job to show where all those extra errors come from. In the absence of that I consider this closed, I'm not letting you continue to redefine and clarify everything you say and misrepresent every argument to the contrary.

                    1. This post has been deleted by its author

                      1. the spectacularly refined chap

                        Show me the code to retrieve this information from the controller of a standard SATA or SAS drive.

                        What, you mean like bad144 included as standard on this very (NetBSD) system? Call yourself a data storage expert? It is one of the things in our routine monitoring.

                        READ THE CERN PAPER AND UNDERSTAND IT, then try to find a solution.

                        I've read both papers you cite. Neither support you. You can quote the specific sentence or paragraph if you like but I'll tell you now: it isn't there. You made a very specific claim about a head to head comparison between RAID and a single disk. Neither makes such comparisions because that is not the goal of either paper. Both are concerned with the system as a whole rather than the specific characteristics of the array - indeed the CERN paper ackowledges that the errors they recorded where primarily attributable to memory, i.e. before it even got to disk.

                        It is not enough to show that data corruption exists. It is not enough to show that the most widely used error detection and correction systems are not bulletproof. Rebecca has kept a laser focus on the original claim: each time you have introduced irrelevances and distractions that bring in factors that affect the legitimacy of your claim neither one way or the other. Neither deals with RAID1 at all.

                        Although neither paper addresses your point the NEC paper supports the argument to the contrary:

                        "Disk drives commonly experience errors that are recoverable via the ECC..."

                        "At some point, data becomes unreadable leading to “read failures,”..."

                        "RAID can also catch and address errors flagged by hard drives..."

                        The silent corruption case not reported by the drive is restricted a small fraction of cases. It doesn't tackle that head on but does hint at the reduced magnitude of the problem:

                        "However, there is a set of disk errors not caught by the hard drive..."

                        Please do comment back, I'm enjoying watching how you can dance on the head of a pin for so long withotu admitting the mistake.

                        1. This post has been deleted by its author

                          1. Solmyr ibn Wali Barad

                            @1980s_coder

                            "Go ahead, upvote Rebecca and let her upvote you"

                            For what it is worth, I upvoted every long comment on this subthread, because they contained one or more reasonable points.

                            Which does not mean I condone personal attacks, or unconditionally agree with every claim made.

        2. Paul Crawford Silver badge

          @Rebecca M

          The majority of HDD errors are indeed detected by the controller and/or reported by the disk itself when a read request cannot be honoured. That is what classical RAID protects against.

          With a periodic "scrub", where the system attempts to real all HDD sectors so errors are seen and re-written to hopefully fix the problem via sector reallocation, you get a good chance of not ever suffering from known RAID failure under normal conditions (data read, or more commonly when a HDD is replaced and a rebuild is needed).

          But today where you might have massive data sets you can't ignore the problems of "silent errors" where the HDD's correction/detection system, or any one of a number of other sub-systems, has mess with your data. You might want to read this paper on the subject:

          http://research.cs.wisc.edu/wind/Publications/zfs-corruption-fast10.pdf

          (There is another from CERN but I don't have the link to hand)

          1. This post has been deleted by its author

          2. Rebecca M

            Re: @Rebecca M

            With a periodic "scrub", where the system attempts to real all HDD sectors so errors are seen and re-written to hopefully fix the problem via sector reallocation, you get a good chance of not ever suffering from known RAID failure under normal conditions (data read, or more commonly when a HDD is replaced and a rebuild is needed).

            Yes, I'm familiar with ZFS but that actually makes the point for me. ZFS's integrity checks really come into there own for data that is not accessed for years (decades?) at a time and ensuring that it remains readable by correcting any errors as they occur and hopefully while they are still correctable, rather than remaining undetected for years by which time things have decayed to the point that you don't have enough left from which to reconstruct the original data.

            However, there is an implicit assumption made, namely the errors that result from that additional processing are more than offset by the reduction in errors caused by the underlying storage. There is always an outside chance that the maintenance process itself introduces errors as a result of bugs or disturbances somewhere along the path - electronics, firmware, interconnects, system software etc. The scrub process is still regarded as a good thing to do because those errors are rare enough to be not worth considering when set against against the much greater risks of corruption on the underlying media.

        3. Tom 13

          Re: Bollocks. The overwhelming majority of hard drive errors

          Admittedly a very small sample size, but my personal experience suggests otherwise. Granted it was a solution dreamed up by a cheap CIO* on non-server systems, but 50% of our we've lost this system because of a drive problem were the results of bad data on both drives. IIRC it was 6 systems in two years with a user group of about 30 people. Part of the reason you're missing is point is you're being hyper-technical in saying "hard drive error". Yes, you're technically correct about that. But data corruption that gets replicated to both disks comes from other sources as well. The first is some damn fool who thinks holding the power button for 4 seconds is the same thing as Shut Down. Data corrupts on one drive as a result and gets copied to the other after reboot. The most common of course is a malware infection that corrupts the boot sector. RAID tech being what it is, even at the cheap level we were using (PCI IDE solution from Adaptec) this was immediately copied to the mirrored drive. So it wasn't just a matter of breaking the array then using the good drive to recover the system. I do think in all cases we were able to recover the data. And yeah, the cheap CIO was using it as a backup, not a drive failed protection. The previous backup solution was IDE from some now thankfully defunct outfit who couldn't properly engineer their tape back up to comply with the IDE spec.

          *In fairness to the cheap CIO, he was saddled with a horrendous system by the CEO of the company. The CEO had sold a government agency a solution for using a statistical programming solution running on desktops instead of on a proper server. Yes it was true when the department was 3 people, but not once they got to 30.

      2. Daniel B.

        RAID

        If you scribble random data all over just one of the drives, your RAID controller won't notice, and will return that data 50% of the time, when it reads the relevant sectors from the corrupted drive.

        Um... That only applies to RAID1. RAID5/6 does actual parity check on stuff and thus won't return corrupted data. Even better if you're using ZFS, which actually has data integrity checks. ZFS+raidz1 is the best option out there, if you really care that much about corruptible data.

        1. This post has been deleted by its author

    2. Natalie Gritpants
      Boffin

      Hope you have a backup system too. That you've tested recovery on.

    3. Paul Crawford Silver badge

      You have to start with the assumption that if a storage device fails, you won't ever/economically get any/trusworthy data back off it.

      From that starting point, you ought to have enough paranoia to assume the worst, so you begin with the question of what happens when (not if) your device fails/corrupts?

      RAID save you down-time, both use (machine keeps working) and admin (no need to restore your backup) but RAID!=Backup as we are always told.

      Also most RAID & file systems don't have integrity checks so you can have data corruption and not know until something starts playing up. Once you realise this and the vast amount of data you may need to store (comparable to the 10^14 bits of HDD error rate) you might want that, so you then invest in ECC memory and a file system like ZFS or GPFS that has checks. They also support snapshots, a vastly under-rated feature that can save a lot of hassle in restoring a just deleted/modified file, or simplifying a consistent backup point-in-time.

      And there there is your backup, which ought to be in another building and not on-line as a mounted file system or you might get randsomeware screwed (something that snapshots can also help with, if you notice soon enough).

      Really the arguments for SSD vs HDD that matter are cost/GB and IOPS, and smarter systems will use both to give to lots of storage at good price and responsiveness.

  3. Hellcat

    Why is enterprise loving SSDs.

    Simple. Money.

    An SSD easily knocks 2 minutes (it's probably closer to 5 minutes) off our fat client bootup and login to usable desktop.

    2 minutes (1/30th hour) x 5 days x 48 weeks x average hourly rate > Cost of buying an SSD.

    And that's only over 1 year. Multiply by 3 or more years, and over 10,000 client devices and the productivity savings are massive*. Even the bean counters have to agree with that one.

    *Yes I know it probably doesn't save the company real money as the users go for a smoke/coffee/chat at Dave's desk about that holiday he's just got back from but... when they get back it's ready for them, not sitting at 'applying your settings'.

    1. Daniel B.

      Well...

      An SSD easily knocks 2 minutes (it's probably closer to 5 minutes) off our fat client bootup and login to usable desktop.

      Most companies where I've worked keep all PCs turned on. Desktop boot times don't matter if you aren't booting up that much.

      1. Archaon

        Re: Well...

        Most companies where I've worked keep all PCs turned on. Desktop boot times don't matter if you aren't booting up that much.

        If they're like most companies where I've worked then the actual company policy is to turn all PC off to save power, however staff don't do it because they're working on old junker PCs connecting to a convoluted environment (in our case in a different country) and take 20 minutes to boot up and get logged in each morning.

        When someone else is footing the electricity bill most people - myself included - slip into the habit of just leaving their machine on for convenience.

  4. Howard Hanek

    Not Fit For Production?

    I imagine that there are some data centers where spinning wheels sit in the corner just waiting for that mending job......

  5. Tom 38
    Headmaster

    RAID controllers that are not from before-time-began also know how to talk to SSDs so as not to wear a hole in them. As you move higher up the chain into enterprise SSDs, you find that the individual drives have supercapacitors and thus can do a lot of this directly at the drive level, saving further on wear

    Supercaps are a feature of enterprise SSDs, but have FA to do with wear levelling.

    All SSDs, enterprise or not, have wear levelling in their firmware - on an SSD an LBA does not refer to a fixed storage block, it refers to an internal pointer to a block lookup table, wear levelling rejigs the table according to use.

    However, this doesn't mean that enterprise SSDs are a con - an SSD is a small computer of its own, and the quality of the firmware on the SSD operates greatly impacts the performance of the device.

    Consumer SSDs can do *daft* things that are merely daft when they happen in your home PC, but cost money when they happen in your server - one example is the firmware changing its allocation approach based upon free capacity in the device, so going above 70% usage causes it to stop responding until it has restructured its internal tables, which can take several minutes. This might make sense in a home PC - users expect devices to have good performance right up until completely full, so a little lockup once is acceptable.

    1. Trevor_Pott Gold badge

      "Supercaps are a feature of enterprise SSDs, but have FA to do with wear levelling."

      I apologize for not being more explicit in my article. Supercaps - and the functionality they provide - allow write buffering and write coalescing to be handled by the drive itself, rather than relying entirely on the controller or OS. Because of the supercaps, writes can be stored in buffer on the drive until there is enough to write a full block.

      SSDs without supercaps do not all do this. Some do, some don't, and there is some debate about whether or not those that do should.

      So you are partly correct: supercaps do not directly have anything to do with wear leveling. What they enable is write coalescing which enables a more efficient form wear leveling than would be otherwise possible.

      1. Duncan Macdonald

        Write buffering and coalescing can be done without supercaps

        For example the Samsung 840 EVO and 850 EVO SSDs use part of the array in SLC mode which allows writes to be combined reducing the write amplification for the main TLC array.

        Also a number of SSD controllers combine writes by initially buffering them in the controller RAM even without supercaps. (A major investigation into SSD power fault handling proves that this happens - see https://www.usenix.org/system/files/conference/fast13/fast13-final80.pdf for more details)

Page:

POST COMMENT House rules

Not a member of The Register? Create a new account here.

  • Enter your comment

  • Add an icon

Anonymous cowards cannot choose their icon

Other stories you might like