Start a Conversation

Unsolved

This post is more than 5 years old

9204

October 15th, 2012 16:00

Larger LUN size considerations

As VMware environments are getting bigger and bigger and VMware clusters need more storage capacity we are creating bigger Luns. Recently we decided to increase our Lun sizes for VMware to 4TB from 1TB for our vSphere environment. We are using Symmetrix with the latest 5875 code and Fast pools.

With VAAI we are not seeing any lock contention or performance impacts. I wonder what other functions such as replication and backup might have with bigger LUN sizes. What are other people seeing and have you see any impact to your environments as you move to larger Lun sizes?

40 Posts

October 15th, 2012 16:00

Any datastore you want to replicate via RecoverPoint has to be (2TB - 512 bytes) or smaller.  Like you we're using larger LUNs / datastores but for replicated datastores just under 2TB.

Note: That info is as far as RP 3.4 - 3.5 is the latest and may be different.

J

57 Posts

October 16th, 2012 02:00

Interesting point.

Best practices seem to point to increasing LUN sizes. I get that.  Bigger LUNs = Less LUNs = Less administration = better for admins. However, there used to be quite some discussion about queue depths when picking a LUN size, as there is a queue per LUN. What i want to know is if we somehow need to think different of queues these days ? Less and bigger LUNs means more VMs will share the same queues, given VMs dont get much bigger most of the time. Is this less of an issue these days with the current versions of VMware?

26 Posts

October 16th, 2012 05:00

Hi,

if you wouldn't mind let me add my 2 cents here.

Keep in mind what virtualisation do and don't do, any one of the things it doesn't do is virtualizing IO's.

What I mean is the following, keep in mind that your infrastructure has to handle the total IO load, regardless if you're using VM's or not.

In good old days each server had multiple disk, (and sometimes for good reasons ), now many VMware customers keep this design but place all vmdk's on one datastore which is contra productive.

VAAI itself is cool, but not often used when you compare it to normal IO's.

SDRS (and storage array technologies like FAST) tries to get rid of the problems with IO performance issues, but they do require analysis time and are therefor reactive tools.

In general, a good design is priceless.

So you should keep an eye on the amount of IO's your environment produces and then present enough LUN's to handle that load, all other features will than take care that the load will be moved when needed.

Hth,

Ralf

40 Posts

October 16th, 2012 06:00

Clustor,

Just remember you're able to do multiple LUNs presented as 1 datastore - so if you find that your LUN queue is a bottleneck, there's little downside to creating 4 1 TB LUNs and presenting them as 1 4TB datastore.

Ralf's point is in general good advice, but really has nothing to do with our conversation. Assuming you're talking about whether it makes sense to put the 4x1TB or 1x4TB LUNs all on the same RAID group (going with traditional LUNs to make the conversation easy), they both provide the same amount of IOPs. As mentioned you do get a queue by LUN, so there is that.

VMware performance can frequently be a bit of a whack-a-mole situation. A good design is very important, no argument, but understanding how to identify and resolve bottlenecks when they occur is vitally important.. So monitor the queues for your Datastores / LUNs and as necessary re-design your layout to reduce the hot spots.

I'd suggest checking out the joint EMC / VMware session from VMworld this year on Storage best practices.

http://virtualstorageguy.com/2012/09/10/vmworld-2012-sto2980-vsphere-storage-best-practices/

40 Posts

October 17th, 2012 11:00

Although no one mentioned heap size anywhere here, you do bring up a good point. Heap size, as I understand it, only matters when you're talking about the aggregate amount of storage that makes up the VMs that are attached to the host (aka, if you have a 100G VM on a 4TB datastore attached to an ESX host, the heap on the ESX host only cares about the 100G)

Having said that, although the defaults and maximums  for vSphere 4 and vSphere 5 changed, there is a workaround

http://www.boche.net/blog/index.php/2012/09/12/monster-vms-esxi-heap-size-trouble-in-storage-paradise/

Also check out KB article

KB article 1004424: An ESXi/ESX host reports VMFS heap warnings when hosting virtual machines that collectively use 4 TB or 20 TB of virtual disk storage.

The key is to change VMFS3.MaxHeapSizeMB

Note also this only is an issue with VMFS datastores, not NFS.

17 Posts

October 17th, 2012 11:00

Heap size isn't affected by 1 or 4 TB LUNs. But heap size is another interesting number to keep in mind for ESX servers with large amounts of storage.

So a single ESX host is limited as to how much capacity it can address because of heap size. We ran into issues where we had a large amount of storage assigned to a single ESX host and it started randomly dropping LUNs. So in our case, this issue presented itself on a large VM running Exchange. The Exchange VM would randomly lose connectivity to the LUN that had it's E: drive or F: drive or fill-in-the-blank drive.

v4 and v5 have different maximum heap sizes. But be careful. Eventhough v5 has a larger heap size, it also has a larger block size (4 mb). So v5 actually made the problem worse. I haven't heard of any movement/progress/patches for this issue yet.

Maybe the best case scenario is where you start out with v4, create your volumes, and then upgrade to v5. This method uses the v4 block size (1 mb) so you get that benefit. And then v5 has the larger heap size. So you get that benefit. 

Here's some info on it: http://blogs.vmware.com/vsphere/2012/08/vmfs-heap-considerations.html

57 Posts

October 17th, 2012 12:00

BTW, i don't think your numbers are correct on the blocksizes for v4 and v5.

VMFS5 uses 1MB default. Earlier versions VMFS used 1,2,4 or 8MB

17 Posts

October 17th, 2012 12:00

Clustor,

I think you're right about the block size. The main point I wanted to make is that maximum addressable open storage went down between ESX v4 to v5.  So be careful.

◦In ESXi/ESX 4.x and earlier releases, max heap size is 128 MB. This allows a maximum of 32 TB of open storage.

◦In ESXi 5.x, max heap size is 256 MB. This allows a maximum of 25 TB of open storage.

17 Posts

October 17th, 2012 12:00

Jay,

Thanks for the additional info. If people are using 4 TB LUNS, they may have already followed the workaround to bump up their ESX hosts to the maximum heap size.  I know we have. We'd love to go way beyond what ESX allows. It's kind of a thorn right now and the kb article and workaround link attest to that. You have to be really careful when you put one of your ESX servers in maintenance mode. You could easily go beyond the max heap size on the other ESX servers in the cluster when everything gets vmotioned off the ESX server that you put in maint mode.

286 Posts

February 14th, 2013 08:00

Jay just to correct a statement you made earlier. The RP limit is 32TB for a lun not 2TB assuming you are using one of the array based spitters.

No Events found!

Top