IBM FlashSystem

IBM FlashSystem

Find answers and share expertise on IBM FlashSystem

 View Only

Subtleties between Marketing versus Engineering

By Tony Pearson posted Wed May 13, 2009 01:18 PM

  

Originally posted by: TonyPearson


Continuing my ongoing discussion on Solid State Disk (SSD), fellow blogger BarryB (EMC) points out in his [latest post]:
Oh – and for the record TonyP, I don't think I ever said EMC was using a newer or different EFDs than IBM. I just asserted that EMC knows more than IBM about these EFDs and how they actually work a storage array under real-world workloads.

(Here "EFD" is refers to "Enterprise Flash Drive", EMC's marketing term for Single Layer Cell (SLC) NAND Flash non-volatile solid-state storage devices. Both IBM and EMC have been selling solid-state storage for quite some time now, but EMC felt that a new term was required to distinguish the SLC NAND Flash devices sold in their disk systems from solid-state devices sold in laptops or blade servers. The rest of the industry, including IBM, continues to use the term SSD to refer to these same SLC NAND Flash devices that EMC is referring to.)

The disagreement resulted from his earlier statement from his post[IBM's amazing...part deux]:

Although STEC asserts that IBM is using the latest ZeusIOPS drives, IBM is only offering the 73GB and 146GB STEC drives (EMC is shipping the latest ZeusIOPS drives in 200GB and 400GB capacities for DMX4 and V-Max, affording customers a lower $/GB, higher density and lower power/footprint per usable GB.)

Here is where I enjoy the subtleties between marketing and engineering. Does the above seem like he is saying EMC is using newer or different drives? What are typical readers expected to infer from the statement above?

  • That there are four different drives from STEC, in four different capacities. In the HDD world, drives of different capacities are often different, and larger capacities are often newer than those of smaller capacities.

  • That the 200GB and 400GB are the latest drives, and that 73GB and 146GB drives are not the latest.

  • That STEC press release is making false or misleading claims.

Uncontested, some readers might infer the above and come to the wrong conclusions. I made an effort to set the record straight. I'll summarize with a simple table:

Raw capacity128 GB256 GB512 GB
Usable (conservative format)73 GB146 GB300 GB
Usable (aggressive format)100 GB200 GB400 GB

So, we all agree now that the 256GB drives that are formatted as 146GB or 200GB are in fact the same drives, that IBM and EMC both sell the latest drives offered by STEC, and that the STEC press release was in fact correct in its claims.

I also wanted to emphasize that IBM chose the more conservative format on purpose. BarryB [did the math himself] and proved my key points:

  • Under some write-intensive workloads, an aggressive format may not last the full five years. (But don't worry, BarryB assures us that EMC monitors these drives and replaces them when they fail within the five years under their warranty program.)

  • Conservative formats with double the spare capacity happen to have roughly double the life expectancy.

I agree with BarryB that an aggressive format can offer a lower $/GB than the conservative format. Cost-conscious consumers often look for less-expensive alternatives, and are often willing to accept less-reliable or shorter life expectancy as a trade-off. However, "cost-conscious" is not the typical EMC targeted customer, who often pay a premiumfor the EMC label. To compensate, EMC offers RAID-6 and RAID-10 configurations to provide added protection. With a conservative format, RAID-5 provides sufficient protection.

(Just so BarryB won't accuse me of not doing my own math, a 7+P RAID-5 using conservative format 146GB drives would provide 1022GB of capacity, versus 4+4 RAID-10 configuration using aggressive format 200GB drives only 800GB total.)

In an ideal world, you the consumer would know exactly how many IOPS your application will generate over the next five years, exactly how much capacity you will require, be offered all three drives in either format to choose from, and make a smart business decision. Nothing, however, is ever this simple in IT.

technorati tags: , , , , , , , , , , , , , ,

4 comments
12 views

Permalink

Comments

Tue May 19, 2009 06:03 PM

BarryB,You make a good argument why RAID-5 should be the best choice for both protection and performance balance for SSD. I am glad to see both IBM and EMC recommend RAID-5 for SSD. This brings up two questions in my mind:
1) Given that RAID-5 is best choice for SSD, why did EMC invest additional development investment to also offer RIAD-10 and RAID-6 which are not recommended for use with SSD?
2) Why would you poke fun that IBM only offers RAID-5 on SSD when even EMC suggests this is the best recommendation?
--- Tony

Tue May 19, 2009 05:59 PM

Khue,Thanks for your two cents. Both IBM and EMC offer the latest technology from STEC, both will stand behind drive replacements if/when they occur. The key difference I believe is philosophy.
EMC prefers that entire LUNs be migrated from HDD to SSD, and as such offering the largest capacities and providing more usable TB per drive with more aggressive formatting serves their purpose well. While a bigger drive reduces the $/GB, it increases the $/IOPS, as each drive can only do so many IOPS regardless of capacity or formatting.
IBM prefers smart placement of data sets, files and individual database tables that make the most use of SSD's unique capabilities. IBM has shown that with only 10 percent of the SSD, you might get as much as 90 percent of the performance benefits. IBM offers a variety of tools to help identify the right data to move, and tools to perform the migration.
Where EMC wants its disk systems to work harder for you, IBM wants them to work smarter.
-- Tony

Fri May 15, 2009 09:01 AM

Just my 2 cents guys:
I'd just like to say I appreciate these debates. The information being presented by both sides merits research on my own part to actually get facts straight between the two vendors. While we are IBM locked (which I have no problem with as IBM has been fantastic partners to my organization) Barry's comments, while strangely over passionate, are relevant and interesting. Tip to Barry, try to present your information a little more professionally because I have a tough time taking you seriously when you harp on a single phrase and finish various statements with triple exclamation points like my sister does on Facebook. It does make me curious though, why IBM is so cautious on the formatting of such drives? I would tend to say it's mostly for CYA purposes if they really are behind on understand SSDs which is not a bad thing and being a second mover on the format issue does filter them from any catastrophic issues that may occur with EMC's offering (although the chance of that is most likely slim).
Anyway, I look forward to reading more from the Barrys and Tony. Thanks guys.

Thu May 14, 2009 07:34 AM

Again, Tony, if you would stick to the facts, you'd have had a pretty good post, even though you are probably now in full violation of your claimed restrictions against engaging in "blogger wars" a couple of posts back.
But still it seems you just can't resist twisting facts to fit your distorted perspective on the world of Flash.
And you make my point again that you really don't understand flash drive technology.
First mistake: Given the significantly higher MTBPR of EFDs vs. HDDs and the faster rebuild times, EMC recommends using RAID 5 for the flash drives we sell, not RAID-10 (nor RAID 6).
For you to suggest otherwise is obviously an intentional attempt to mislead your readers.
Of course, most readers will recognize that RAID-10 would in fact obviate the benefit of the larger capacity (as you demonstrate). EMC's recommended best practice for EFDs is to use RAID 5, and thus those same 8 200GB drives will provide 1,400GB of usable capacity, vs. the paltry 1,022 on the DS8K with the 146GB drives.
Thus, customers get about 37% more usable capacity from EMC for the same STEC drive, providing a lower $/GB usable.
Second mistake: as you point out so clearly, IBM is not selling the STEC ZeusIOPS drive with 512GB of raw capacity - neither in its 300GB format nor its 400GB format (EMC sells the 400GB).
Technically, the 512GB ZeusIOPS drive is both "newer" and "different," as it is the latest capacity point added to the ZeusIOPS product line. So no, my statement is truthful: IBM isn't selling the identical driveS, we only have the 256GB (raw) device in common.
And EMC customers alone can take advantage of the ADDITIONAL cost benefits of the larger capacity, while IBM customers are stuck requiring more than 2 times the number of flash drives to get to the same usable capacity as say, 8 400GB EFDs in RAID 5 7+1.
More capacity with fewer drives = lower acquisition AND operating costs.
Third: you denigrate EMC's use of "EFD" (Enterprise Flash Drives", but it is far from a marketing term. Just as the industry refers to enterprise disk drives to summarize the technical features of HDDs optimized for the 24xforever workloads our storage arrays support, EFD identifies SSDs that meet the requirements of our storage platforms as well. In both cases, the term "enterprise" serves to differentiate these devices from those designed for less demanding environments.
Noting that IBM eventually selected the same STEC drives that EMC calls "enterprise" would seem to underscore your agreement that these drives alone meet the requirements of a storage array - especially since there are cheaper alternatives in the market. If you didn't care about the "enterprise" features, I'm sure you would have chosen to resell/OEM a lesser drive.
BTW - EFD is a descriptor that IBM should support rather than attack...it is in our entire industry's best interests to expand the set of drives that meet the more stringent requirements of the enterprise. It is not a term to be used lightly...
And fourth: you still haven't shown us the math that justifies IBM's conservative approach.
I've shown that EFDs formatted to 200/400 will easily meet the life expectancy requirements of the warranty in any real-world application environment - when used in an EMC array, that is.
You and BarryW, on the other hand, offer nothing other than hand-waving, innuendo and unsubstantiated FUD in response.
So I'll ask once again: SHOW US THE MATH that explains why customers have to pay for so much unusable capacity to use flash drives on the DS8K. GIVE US THE MATH that shows the drive would suddenly fail under ANY projected real-world workload if IBM were to be less conservative.
I honestly don't expect you to do so, because you would then have to admit one or more of the following is true:
a) That the DS8K can't really generate sufficient write workload to exceed the P/E life of the drive under any conditions;
b) That the drive performance characteristics under mixed read/write workloads means that it will never actually experience the pathological write load you assert puts data at risk;
c) That the DS8Ks paltry write cache is insufficient to delay writes significantly to minimize P/E wearing on flash drives;
d) That the the DS8Ks I/O block sizes and cache deficiencies actually exacerbate the "write amplification" factor and therefore significantly accellerate P/E wear, even under relative light workloads;
e) That the DS8K does not do ANY SSD monitoring to detect or protect against premature wear-out, and updating the code of the soon-to-be-retired dinosaur to do so is out of the question. Thus the only option is to be conservative (and make your customers pay more to cover for the inadequacies of the DS8K platform);
f) That you never actually DID the math,
g) That your DS8K developers were in fact totally unaware that the capacity was tunable on the ZeusIOPS until the day EMC announced the new larger capacities;
h) That you realize now that if your engineers had actually done their homework, you wouldn't be caught in the competitive disadvantage selling smaller, less space-efficient SSDs - costing customers more $/GB, $/IOPS and $/power+cooling;
i) that IBM not only doesn't get flash, it is incapable of admitting or correcting its misleading white papers on the subject. (Last I checked, Clod Berrara's Feb 2009 white paper STILL claimed that SLC NAND would tolerate 1,000,000 P/E cycles - now some 3 months later and still the facts are wrong in that paper).
j) all of the above.
I truly suspect that the answer is j), and I'm sure most of our readers do as well. You've given them no reason to think otherwise.
And your misrepresentations of the facts make it clear you are trying to hide something.
Or you just don't know.
Either way, Tony - cut out the baloney, and just DO THE MATH!!!