Single SSD PCIe vs HDD RAID Reliability
-
So the real question is this....
What makes 400K IOPS without RAID worth $600 - $800 when 300K IOPS with RAID is just $300 for this specific use case?
-
And, it should be pointed out, that a $300 RAID 5 array here is likely safer (both in terms of continuous uptime as well as in terms of dataloss) than the PCIe SSD + the SAN replication. If it were me, and I had to choose between the RAID array and the async replication to an external SAN I'd take the SSD RAID 5 array because it is fully consistent, not just crash consistent.
-
@scottalanmiller said:
@Francesco-Provino said:
As I said, I can restart the VMs almost immediately on the SAN (they are replicated, so ready to be restarted), or restore them, both from replication pool or from backup, to one of the other two servers local storage.
What is the manner of replication?
VMware Replication to the SAN, Veeam to the NAS.
-
@scottalanmiller said:
@Francesco-Provino said:
We will do both async replication from SSD to SAN (direct attach Fibre Channel, already in place, our main storage pool as of today) and backup to a NAS unit (QNAP, big SATA drives).
So the failover to the SAN is risky in that data could be lost because it is only crash consistent and the filesystem and/or databases might be corrupted when attempting to use it.
What is the time of dataloss if you need to go to the QNAP to do a restore?
That's always true with async replication. QNAP is in the same building, connected with gigabit network. In my tests, I can retrieve the backup of our biggest VM in almost an hour and a half. Totally ok for us.
-
@scottalanmiller said:
So the real question is this....
What makes 400K IOPS without RAID worth $600 - $800 when 300K IOPS with RAID is just $300 for this specific use case?
@scottalanmiller said:
And, it should be pointed out, that a $300 RAID 5 array here is likely safer (both in terms of continuous uptime as well as in terms of dataloss) than the PCIe SSD + the SAN replication. If it were me, and I had to choose between the RAID array and the async replication to an external SAN I'd take the SSD RAID 5 array because it is fully consistent, not just crash consistent.
Unfortunately, this is not my case: OEM SSD aren't supported with our RAID cards in the servers, and VMware can't do software RAID (apart from, well, sort of, uhm, VSAN).
IBM's SAS SSD are still incredibly expensive. -
If you can compare Samsung drives like this one: http://www.amazon.com/Samsung-2-5-Inch-Internal-MZ-75E500B-AM/dp/B00OBRE5UE/ref=sr_1_1?ie=UTF8&qid=1447070361&sr=8-1&keywords=samsung+ssd+500GB
And the details on the Intel PCIe card: http://www.thessdreview.com/our-reviews/intel-ssd-dc-p3700-nvme-ssd-enthusiasts-report/ (that's the p3700, not the p3500)
It seems like the PCIe card is difficult to choose in this case. You can get more IOPS for less money and more protection from the SATA SSDs still.
-
@Francesco-Provino said:
Unfortunately, this is not my case: OEM SSD aren't supported with our RAID cards in the servers, and VMware can't do software RAID (apart from, well, sort of, uhm, VSAN).
IBM's SAS SSD are still incredibly expensive.Ah... the devil is in the details. You are using VMware and lack enterprise software RAID options so can't do super high performance SSD without having a RAID card to support it. Yet another VMware caveat. They screw you at every turn. So many limitations that you would never guess would be there.
Are you sure that "unsupported" is the case, though? Of course it is not supported by IBM, neither is the Intel PCIe board. So both cases are equally without support. The question is "do they work?"
-
@scottalanmiller said:
@Francesco-Provino said:
Unfortunately, this is not my case: OEM SSD aren't supported with our RAID cards in the servers, and VMware can't do software RAID (apart from, well, sort of, uhm, VSAN).
IBM's SAS SSD are still incredibly expensive.Ah... the devil is in the details. You are using VMware and lack enterprise software RAID options so can't do super high performance SSD without having a RAID card to support it. Yet another VMware caveat. They screw you at every turn. So many limitations that you would never guess would be there.
Are you sure that "unsupported" is the case, though? Of course it is not supported by IBM, neither is the Intel PCIe board. So both cases are equally without support. The question is "do they work?"
This consideration is interesting, but I don't even know if it's possible to put OEM disks in those hot-swap slots…
-
@Francesco-Provino said:
This consideration is interesting, but I don't even know if it's possible to put OEM disks in those hot-swap slots…
Should be, people do it all of the time. It's very standard. There are problems sometimes and the RAID monitoring tools mostly don't work. But blocking non-OEM drives is illegal in many countries (like the US and I presume the EU) so they normally work.
-
Putting a non-OEM disk into a hot swap bay is no different then putting on into a PCIe bay.
-
@scottalanmiller said:
Putting a non-OEM disk into a hot swap bay is no different then putting on into a PCIe bay.
Errr, no, It will lack the caddy, and I don't think they sell it as spare parts.
-
Oh, I see. I don't use IBM servers (especially now that they don't make them anymore, but even before that as even IBM doesn't use their own servers) and forgot that they might be pulling the caddie trick on you. HP does this as well, Dell and SuperMicro do not.
You are right, you might be stuck. In the future, I would use this as a solid reason to avoid both IBM and VMware (IBM is gone now, so it matters little) as both are causing you to:
- Have to spend extra to get less.
- Avoid standard best practices.
- Work around basic system limitations.
- Go to unsupported designs.
I can see why you are interested in the PCIe SSD approach. It isn't because it is cheap or fast or reliable - it is a workaround to the IBM and VMware decisions I think, when we look at it from that perspective, it starts to make a lot more sense. From purely a technology standpoint, I don't think that it makes sense.
-
I think with all of that info that the PCIe SSD approach makes sense. It will be seriously fast and pretty easy to use. And with the sync and backup options you are pretty decently protected. If you can handle the associated downtime to flip over to the SAN while waiting for the SSD to be replaced you will be fine.
-
Assuming the IBM RAID controller will allow the use of non OEM drives, I'd buy a bunch of tiny drives on ebay, rip out the old drive, mount the SSDs and you should be good.
If you have to to no RAID card (@scottalanmiller - wouldn't this mean he'd have to install a SAS/SATA controller? I'm guessing the system doesn't have onboard support) yet another reason to move to to Hyper-V now.
-
He does have on board hardware RAID. The problem is not his card but his hot swap bays.
-
@scottalanmiller said:
He does have on board hardware RAID. The problem is not his card but his hot swap bays.
You mean because of the lack of empty caddies? Sure I understand that, but I addressed that in my post.
-
@Dashrender said:
@scottalanmiller said:
He does have on board hardware RAID. The problem is not his card but his hot swap bays.
You mean because of the lack of empty caddies? Sure I understand that, but I addressed that in my post.
oh, meaning rip the caddies off. I didn't understand what you were implying. That makes more sense.
-
@scottalanmiller said:
@Dashrender said:
@scottalanmiller said:
He does have on board hardware RAID. The problem is not his card but his hot swap bays.
You mean because of the lack of empty caddies? Sure I understand that, but I addressed that in my post.
oh, meaning rip the caddies off. I didn't understand what you were implying. That makes more sense.
Exactly. I've considered doing the same for my HP server.