IBM FlashSystem

IBM FlashSystem

Find answers and share expertise on IBM FlashSystem

 View Only

Does Size Really Matter for Performance?

By Tony Pearson posted Fri January 18, 2008 06:35 PM

  

Originally posted by: TonyPearson


Fellow Blogger BarryB mentions "chunk size" in his post [Blinded by the light],as it relates to Symmetrix Virtual Provisioning capability. Here is an excerpt:
I mean, seriously, who else but someone who's already implemented thin provisioning would really understand the implications of "chunk" size enough to care?

For those of you who don't know what the heck "chunk size" means (now listen up you folks over at IBM who have yet to implement thin provisioning on your own storage products), a "chunk" is the term used (and I think even trademarked by 3PAR) to refer to the unit of actual storage capacity that is assigned to a thin device when it receives a write to a previously unallocated region of the device.
For reference, Hitachi USP-V uses I think a 42MB chunk, XIV NEXTRA is definitely 1MB, and 3PAR uses 16K or 256K (depending upon how you look at it).

Thin Provisioning currently offered in IBM System Storage N serieswas technically "implemented" by NetApp, and that the Thin Provisioning that will be offered in our IBM XIV Nextrasystems will have been acquired from XIV. Lest I remind you that many of EMC's products were developed by other companies first, then later acquired by EMC, so no need for you to throw rocks from your glass houses in Hopkington.

"Thin provisioning" was first introduced by StorageTek in the 1990's and sold by IBM under the name of RAMAC Virtual Array (RVA). An alternative approach is "Dynamic Volume Expansion" (DVE). Rather than giving the host application a huge 2TB LUN but actually only use 50GB for data, DVE was based on the idea that you only give out 50GB they need now, but could expand in place as more space was required. This was specifically designed to avoid the biggest problem with "Thin Provisioning" which back then was called "Net Capacity Load" on the IBM RVA, but today is now referred to as "over-subscription". It gave Storage Administrators greater control over their environment with no surprises.

In the same manner as Thin Provisioning, DVE requires a "chunk size" to work with. Let's take a look:

DS4000 series

On the DS4000 series, we use the term "segment size", and indicate that the choice of a segment size can have some influence on performance in both IOPS and throughput. Smaller segment sizes increase the request rate (IOPS) by allowing multiple disk drives to respond to multiple requests. Large segment sizes increase the data transfer rate(Mbps) by allowing multiple disk drives to participate in one I/O request. The segment size does not actually change what is stored in cache, just what is stored on the disk itself.It turns out in practice there is no advantage in using smaller sizes with RAID 1; only in a few instances does this help with RAID-5 if you can writea full stripe at once to calculate parity on outgoing data. For most business workloads, 64KB or 128KB are recommended. DVE expands by the same number of segments across all disks in the RAID rank, so for example in a 12+P rank using 128KB segment sizes, the chunk size would be thirteen segments, about 1.6MB in size.

SAN Volume Controller

On the SAN Volume Controller, we call this "extent size" and allow it to be various values 64MB to 512MB. Initially,IBM only managed four million extents, so this table was used to explain the maximum amount that could be managedby an SVC system (up to 8 nodes) depending on extent size selected.

Extent SizeMaximum Addressable
16MB64TB
32MB128TB
64MB256TB
128MB512TB
256MB1PB
512MB2PB

IBM thought that since we externalized "segment size" on the DS4000, we should do the same for the SANVolume Controller. As it turned out, SVC is so fast up in the cache, that we could not measure any noticeable performance difference based on extent size. We did have a few problems. First, clients who chose 16MB andthen grew beyond the 64TB maximum addressable discovered that perhaps they should have chosen something larger.Second, clients called in our help desk to ask what size to choose and how to determine the size that was rightfor them. Third, we allowed people to choose different extent sizes per managed disk group, but that preventsmovement or copies between groups. You can only copy between groups that use the same extent size. The generalrecommendation now is to specify 256MB size, and use that for all managed disk groups across the data center.

The latest SVC expanded maximum addressability to 8PB, still more than most people have today in their shops.

DS8000 series

Getting smarter each time we introduce new function, we chose 1GB chunks for the DS8000. Based on a mainframebackground, most CKD volumes are 3GB, 9GB, or 27GB in size, and so 1GB chunks simplified this approach. Spreadingthese 1GB chunks across multiple RAID ranks greatly reduced hot-spots that afflict other RAID-based systems.(Rather than fix the problem by re-designing the architecture, EMC will offer to sell you software to help you manually move data around inside the Symmetrix after the hot-spot is identified)

Unlike EMC's virtual positioning, IBM DS8000 dynamic volume expansion does work on CKD volumes for our System z mainframe customers.

The trade-off in each case was between granularity and table space. Smaller chunks allow finer control on the exact amount allocated for a LUN or volume, but larger chunks reduced the number of chunks managed. With our advanced caching algorithms, changes in chunk size did not noticeably impact performance. It is best just to come up with a convenient size, and either configure it as fixed in the architecture, or externalize it as a parameter with a good default value.

Meanwhile, back at EMC, BarryB indicates that they haven't determined the "optimal" chunk size for their newfunction. They plan to run tests and experiments to determine which size offers the best performance, and thenmake that a fixed value configured into the DMX-4. I find this funny coming from the same EMC that won't participate in [standardized SPC benchmarks] because they feel that performance is a personal and private matter between a customer and their trusted storage vendor, that all workloads are different, and you get the idea. Here's another excerpt:

Back at the office, they've taking to calling these "chunks" Thin Device Extents (note the linkage back to EMC's mainframe roots), and the big secret about the actual Extent size is...(wait for it...w.a.i.t...for....it...)...the engineers haven't decided yet!

That's right...being the smart bunch they are, they have implemented Symmetrix Virtual Provisioning in a manner that allows the Extent size to be configured so that they can test the impact on performance and utilization of different sizes with different applications, file systems and databases. Of course, they will choose the optimal setting before the product ships, but until then, there will be a lot of modeling, simulation, and real-world testing to ensure the setting is "optimal."

Finally, BarryB wraps up this section poking fun at the chunk sizes chosen by other disk manufacturers. I don't knowwhy HDS chose 42MB for their chunk size, but it has a great[Hitchiker's Guide to the Galaxy]sound to it, answering the ultimate question to life, the universe and everything. Hitachi probably went to theirDeep Thought computer and asked how big should their "chunk size" be for their USP-V, and the computer said: 42.Makes sense to me.

I have to agree that anything smaller than 1MB is probably too small. Here's the last excerpt:

Now, many customers and analysts I've spoken to have in fact noted that Hitachi's "chunk" size is almost ridiculously large; others have suggested that 3PAR's chunks are so small as to create performance problems (I've seen data that supports that theory, by the way).

Well, here's the thing: the "right" chunk size is extremely dependent upon the internal architecture of the implementation, and the intersection of that ideal with the actual write distribution pattern of the host/application/file system/database.

So my suggestion to EMC is, please, please, please take as much time as you need to come up with the perfect"chunk size" for this, one that handles all workloads across a variety of operating systems and applications, from solid-state Flash drives to 1TB SATA disk. Take months or years, as long as it takes. The rest of the world is in no hurry, as thin provisioning or dynamic volume expansion is readily available on most other disk systems today.

Maybe if you ask HDS nicely, they might let you ask their computer.

technorati tags: , , , , , , , , , , , , , , , , , , , , , , , , , , , ,

6 comments
10 views

Permalink

Comments

Sat January 26, 2008 05:35 PM

Thanks for the good article. The author points to a number of factors that will help move company for the next phase of the company's development.If you are interested in balanced scorecard, KPI and metrics in business, check this web-site to learn more about Metrics and development metrics.
http://www.business-development-metrics.com

Wed January 23, 2008 09:56 AM

My DS4000 colleagues have pointed out that with 16 drives per drawer, 7+P is not recommended as it would not leave room for spares, and offers an awkward 896K data stripe. More typical configurations are 4+P or 12+P. DS4000 can support even larger RAID5 configurations across multiple drawers. See the above IBM Redbook for additional considerations.

Tue January 22, 2008 11:31 AM

For more on DS4000 Best Practices for segment size selection, or other tuning guidelines, see the IBM Redbook here:http://www.redbooks.ibm.com/redbooks/pdfs/sg246363.pdf

Sat January 19, 2008 10:21 PM

BarryW - not sure how you are able to measure DMX performance with thin provisioning, since EMC's implementation for the DMX-3 & 4 isn't shipping yet. If you're using SVC thin provisioning, I suspect the problem lies in your product, or possibly something is simply mis-configured.
TonyP - Let's review those results - maybe I'm missing something (all quotes direct from the IBM press release):
First, the "highlights" at the top of the release:
Diluted earnings of $2.80 per share, up 24 percent as reported; Total revenues of $28.9 billion, up 10 percent; Global Technology Services revenues up 16 percent; pre-tax income up 26 percent; Global Business Services revenues up 17 percent; pre-tax income up 9 percent; Services signings of $15.4 billion; short-term signings up 8 percent; Software revenues up 12 percent; pre-tax income up 21 percent; 65 percent of revenues from outside the U.S.; E/ME/A revenues up 16 percent; Asia Pacific up 15 percent.
Odd...there's no mention of STORAGE, or ANY HARDWARE for that point, in the "highlights" section of the press release.
Pretty clear that most of the revenue gains came from outside the US:
From a geographic perspective, the Americas' fourth-quarter revenues were $11.7 billion, an increase of 5 percent as reported (2 percent, adjusting for currency) from the 2006 period. Revenues from Europe/Middle East/Africa were $10.8 billion, up 16 percent (6 percent, adjusting for currency). Asia-Pacific revenues increased 15 percent (9 percent, adjusting for currency) to $5.5 billion. OEM revenues were $894 million, down 13 percent compared with the 2006 fourth quarter.
Seems perhaps folks in the US (and some of your OEMs) aren't shopping at the Big Blue Super Mart as much...makes you wonder if they opted for the specialty stores instead?
Revenues from the Systems and Technology segment totaled $6.8 billion for the quarter, down 4 percent (8 percent, adjusting for currency).
Revenues from System z server products decreased 15 percent versus the year-ago period.
Ahh..THERE'S the culprit. The MAINFRAME business was off (or was that just a tough compare?)
But WAIT:
Revenues from System Storage increased 11 percent
Ahh - storage to the rescue.
That would be impressive, I guess, if you actually told us how big the number for just Storage was, but since it's just part of the 6.8Bn Systems & Tech revenus, it's pretty clear that the 11% gains were on a small number. And since the press release already admitted that the majority of the gains where due to currency exchange benefits of a weak dollar, methinks I can guess where you sold the majority of your storage products.
And as we know from past reports (and TonyP blogs) a rather significant component of "System Storage" is TAPE.
All nets out that you really aren't doing all that much business selling storage...
=======To complete this discussion, we need another couple of weeks worth of earnings reports, to see how IT (and DISK storage) spend in the US fared. Let's check in again at the end of the month.
BTW -it sure must be scary to see demand for "zee" Cash Cow tail off so quickly, though, huh?

Sat January 19, 2008 07:53 PM

One of these days, that sofa will get round that bend and make it into the apartment (dirk gently book1)
Thin provisioning has is place, its uses, and its issues. If you are happy to accept 30% of your performance (what we have measured on DMX) when you thin provision - fair enough - if you want more - watch this space...

Sat January 19, 2008 09:02 AM

BarryB,IBM is a supermarket, some of the products on our shelves are made by others, some by IBM itself. EMC is a disk/archive/security specialty shop, some of your technology was acquired, while others are post-acquisition. If you want to limit your discussions to post-acquisition developments, that is up to you.
However, just because IBM leverages the smart people in other companies does not imply we do not have smart engineers ourselves. Have you seen our 4Q07 numbers yet? They are quite impressive!