Start a Conversation

Unsolved

This post is more than 5 years old

3414

December 16th, 2015 08:00

RAID 10 Read Performance

Hi,

We are in a weird situation.

As my understanding the read performance on RAID 10 is high, and would be higher that write speed.

We have a RAID 10 LUN (12 disks) which is connected to Windows Server 2012 (Through 4 Fiber Connection using EMC PowerPath)

The read performance is worse. (13 MBPS)!!!!!

I've attached the IO-Meter snapshot to this post to review.

Thanks for any help and idea.

Mohammad

IO-Meter Random Read.PNG.png

1.2K Posts

December 16th, 2015 10:00

A few things to note -

  1. You didn't state the disk type (SATA, FC, EFD).
  2. You didn't mention the type of VNX (5200?  5300?  7600?)
  3. Your test is set for 100% random reads.

While this performance seems low, there could be extenuating factors that are causing issues.  For starters, write performance will be consistently high on RAID10.  Why?  The VNX SP cache can quickly capture incoming writes and commit them to the backend very quickly, since there's no RAID penalty with RAID10.  However, random reads are coming from all over the disk, and you may be seeing a (typically) low random read hit rate on SP cache.

What's going on with the backend during your testing?  Have you looked at Unisphere during the test and examined performance from the VNX side? Do you have other workloads on these 12 drives?

The more details you can provide, the more we can comment and assist.

Thanks!

Karl

3 Posts

December 16th, 2015 11:00

1. Disk type: 600GB / SATA 15k

2. VNX type: 5300

Regarding to the bellow table, I should get excellent performance on R/W on RAID 10.

The random write speed is 131 MBs/sec which is almost 10 time faster that read.

RAID Level.PNG.png

They way I've created the LUN is: I created a RAID10 pool using all 12 disks and the I created a 3 TB LUN.

I know that the recommendation of creating RAID 10 pool is using 8 disks but the way I created is putting all disks into the pool. Could is be the root of issue?

I've enabled the performance log to send to EMC. I don't know how to examine the performance on the VNX.

Thanks,

Mohammad

65 Posts

December 17th, 2015 02:00

Hello Mohammad,

There's a lot more data I'd like to look at in order to determine the cause of the speed difference between reads and writes, such as:

- In the random writes test, did you use the same IO size?

- Is FAST Cache enabled?

- Was the pool only recently created?

This information, and more useful points, would be available from the SP Collects and NAR files you'll gather (or already have) to contact support, which is good.

However for your knowledge; yes, its better to have RAID 1/0 RGs on a 4+4 configuration, and with high (>30%) random writes expected.

Useful document:

https://www.emc.com/collateral/white-papers/h12682-vnx-best-practices-wp.pdf

At the end of page 16:

"AVOID a drive count that results in a small remainder when divided by the preferred drive count"

Adham

3 Posts

December 17th, 2015 07:00

Hi Adham,

1. I did the all the tests (read & write) with the same IO size

2. Looks we don't have FAST-Cache installed

3. The pool created couple of days ago.

What's the best way to create a RAID10 pool using 12 disks? I saw an article saying to create the pool with 6 disks and then add 6 more disks.

Mohammad

4.5K Posts

December 29th, 2015 10:00

When you create a R10 pool, the default is 4+4. If you select 12 disks, it will create within the Pool two private raid groups:

1. 4+4 (the default)

2. 2+2 (the remaining four disks)

Ideally you would want to add disks in multiples of 8 disks (4+4). With 12 disks you get one private raid group using 4+4 and the remaining are in a 2+2. While each disk has the same performance characteristics, the individual raid groups will have a different performance level. As a LUNs is composed of slices taken from each raid group, you could see slower performance when the slices are on the 2+2 compared to the slices on the 4+4.

Each 15K SAS disk can handle between 12MB/s to 30MB/s depending on the randomness of the IO, the IO size, and the mix of reads and writes. For a 100% Read, 100% Random, 64KB IO Size, the bandwidth that each disk can handle will be closer to 15MB/s. For very random data, you'll get no benefit from pre-fetching on the array. Try changing your test to 0% Random (100% Sequential) and see what you get. Also, try 32KB IO Size.

If you have 12 disks at 15MB/s each, that means the total bandwidth should be in the range of 12 * 15 = 180MB/s. This is a raw number. It can be affected by other activity on the array, the configuration of the paths to the array, the HBA settings on the host, the failover software on the host, the load balancing on the host, and the two private raid groups within the pool.

If you only getting 13MB/s from the LUN, then there is something else occurring that is causing this. I see that you've already opened a case with EMC, they should be able to determine what is going on. It's possible that the LUN was still initializing when you ran the tests as it can take a day or more to finish.

glen

No Events found!

Top