Professional Documents
Culture Documents
Understanding how pool LUN ownership and trespassed LUNs can impact performance as well as uptime
The following article applies to the VNX 05.31.x.x.x and 05.32.x.x.x. families of code only. The introduction of Pool LUNs provided users with many benefits not possible with traditional LUNs. Pools support Thin LUNs, and Thin LUNs offer customers more flexibility when configuring and allocating storage since Thin LUNs allocate their space from the pool on demand, and can be over -provisioned. Overprovisioning allows a system administrator to present more potential storage space to users than is actually available in the pool. Since the space is allocated and consumed on demand, the users enjoy more flexibility in their planning and allocating of storage space. Pools can be easily expanded on an as needed basis. Along with the increased flexibility and simplicity offered by Pools comes some underlying changes in the architecture that have very real implications on design and performance that not everyone is familiar with. This article will discuss the impact of trespassed Pool LUNs, one of the most common configuration problems associated with Pool LUNs, and will provide tips on how to avoid unexpected performance problems. Pools allow for the creation and deployment of Thin LUNs. Thin LUNs differ from traditional LUNs in several key ways. When a traditional LUN is created, all of its configured space is carved out and allocated up front. The traditional LUN is assigned a default owner, and a current owner. Initially, these are both set to the same storage processor (SP). During the life of a traditional LUN, circumstances may cause the LUN to trespass over to the peer SP. When this happens, the LUNs current owner will change. When a traditional LUN resides on its non-default SP, there is no significant performance impact other than the increased load on the peer SP associated with the extra LUN or LUNs that it now owns. When a Pool is created, a large number of private FLARE LUNs are bound on all the Pool drives, and these drives are divided up between SP-A and SP-B. When a Pool LUN is created, it uses a new category of LUN ownership called the allocation owner to determine which SPs private LUNs should be used to store the Pool LUN slices. For example, a Pool LUN created with SP-A as its allocation owner will be allocating its space from private FLARE LUNs owned by SP-A. When a Pool LUN is created, its allocation owner is the same as its default and current owner. A Pool LUNs default owner should never be changed from its allocation owner. If a Pool LUNs current owner differs from its allocation owner, I/O to that LUN will have to pass over the CMI bus between SPs in order to reach the underlying Pool private FLARE LUNs. This is inefficient and may introduce performance problems. When Pool LUN ownerships (default, allocation, and current) do not match, it creates a potentially sizeable performance bottleneck. If enough LUNs have inconsistent ownerships, performance can bottleneck to the point where host I/Os can timeout and even result in data unavailability in extreme cases. Every effort should be made to maintain consistent ownership of Pool LUNs. EMC recommends avoiding prolonged periods of having trespassed Pool LUNs. For more information about Pool LUN ownership settings, see knowledge-base article 88169: Setting the Pool LUN default and allocation owner correctly. For information about several other performance related concerns specific to Pool LUNs, see knowledge-base article 15782: How to fix Pool performance issues.
VOLUME 6
September 2013 Pool LUN ownership concerns. 248 Day reboot issue reminder. Vault drive replacement procedure tips. Array power down tips. Target code revisions and key fixes. Storage pool reserved space requirements by software revision. VNX Block OE 32.207 key fixes. 1 2 2
2 3 4
Follow us on the web! Search the VNX Series page for Uptime Bulletin here: https://support.emc.com/prod ucts/12781
WED LIKE TO HEAR YOUR Customer DocumentaCOMMENTS ABOUT THE UPtion TIME BULLETIN. SEND US YOUR IDEAS FOR FUTURE https://mydocs.emc.com/VNX ISSUES AT : UptimeforVNX@emc.com
http://emc.com/vnxesupport
Common mistakes and serious ramifications of incorrect or unplanned power downs of VNX storage systems.
Past VNX Uptime Bulletins have shared details of many bug fixes or best practices which can help avoid outages and improve uptime statistics on your VNX storage system. It may surprise you to learn that the most common, ongoing cause of outages on the VNX platform is the improper shutdown of the VNX storage system. Never power down your system simply by pulling cables. Pulling cables to achieve a power down of your system can bypass the systems ability to protect its cache. Detailed VNX storage array power down procedures can be found here: https://mydocus.emc.com/VNX/requestMyDoc.jsp.
Within EMC Online Support, users can subscribe to advisories for individual products from the Support by Product pages. Under Advisories on the left side of a given product page, click Get Advisory Alerts to subscribe. You can also view and manage subscriptions for multiple
products within Preferences > Subscriptions & Alerts > Product Advisories.
VNXe OS VERSION 2.4.1.21171 2.4.1.21171 UNIFIED VNX CODE VERSIONS (7.0 & R31) 7.0.54.5 (for File OE component) 7.0.54.6 (for File OE component) 05.31.000.5.726 (for Block OE component) 05.31.000.5.727 (for Block OE component) UNIFIED VNX CODE VERSIONS (7.1 & R32) 7.1.72.1 (for File OE component) 7.1.72.1 (for File OE component) 05.32.000.5.207 (for Block OE component) 05.32.000.5.207 (for Block OE component)
Note: VNX OE 05.32.000.5.201 or later requires PowerPath 5.7 or later in order to make full use of the ODX and Thin Provisioning features. It will otherwise function acceptably with older versions of PowerPath. on low level correctable errors.
RELEASE DATE 05/29/13 05/29/13 RELEASE DATE 10/12/12 07/15/13 08/24/12 07/15/13 RELEASE DATE 05/16/13 08/15/13 05/16/13 08/15/13
STATUS Target Latest Release STATUS Target Latest Release Target Latest Release STATUS Target Latest Release Target Latest Release
NDU adaptive timer feature. Driver fix to prevent iPv6 issues noted. New LCC firmware, CDES 1.33/7.05. Fixes various issues noted on LCCs. Fix for AR 497033, coherency errors may be reported after drive probation.
Why is it important to leave some free space in a storage pool, and how does the recommended amount of free space vary by software revision?
While a storage pool can operate at 100% consumed space, it is best practice to leave some space available as free overhead. The amount of recommended space left free varies by VNX OE software revision. In VNX OE releases in the R31 family, we recommend leaving 10% free space, so that a storage pool is no more than 90% full. In VNX OE releases in the R32 family, some changes were made to pre-allocate some space in the background, so best practice dictates that you only need to leave 5% free space. Two specific areas of functionality can be negatively impacted if pools do not leave enough free space. The FAST VP feature requires free space in the pool in order to perform its slice relocations. When there are not enough free slices remaining in the pool, it could limit the amount of relocations that can run at once. This can impact pool LUN performance, or can cause slice relocations and auto-tiering migrations to fail. Although it is not common, LUNs in a storage pool with 0% free space can become offline to a host. If Pool LUNs are taken offline for any reason such that they require EMC to run a recovery on them using EMCs recovery too ls, those recovery tools also use the free space remaining in a pool. If there is not enough free space for recovery tools to run efficiently, a pool may need to be expanded in order to give it more free space. However, expanding a pool is a time consuming option and can prolong the duration of an outage. Pools can be configured to warn the user as they approach various thresholds of available space consumption. For more information about the free space requirements for FAST VP, see knowledge base solution 00078223: Storage Pool does not have enough free space for FAST
VP to relocate slices.
Upgrading to R31.727 as an interim step prior to upgrading from R31 to R32 can avoid a small risk of an outage:
Should this panic occur during the NDU, it may or may not cause a data unavailable The latest VNX OE release for the R31 family, event, depending on the timing of the panic. 05.31.000.5.727, has a fix for an uncommon If the panic occurs on one SP while its peer is bug which could impact a very small percent- rebooting to install software, then it can Note that EMC has not made this interim NDU age of NDU upgrades from the R31 family cause a brief data unavailable situation to step mandatory at this time because of the into the R32 family. The bug can cause a any attached hosts. very low rate of occurrence of the problem. storage processor (SP) to panic at some
EMC believes the information in this publication is accurate as of its publication date. The information is subject to change without notice. THE INFORMATION IN THIS PUBLICATION IS PROVIDED AS IS. EMC CORPORATION MAKES NO REPRESENTATIONS OR WARRANTIES OF ANY KIND W ITH RESPECT TO THE INFORMATION IN THIS PUBLICATION, AND SPECIFICALLY DISCLAIMS IMPLIED WARRANTIES OF MERCHANTABILITY OR FITNESS FOR A PARTICULAR PURPOSE. Use, copying, and distribution of any EMC software described in this publication requires an applicable software license.EMC 2, EMC, E-Lab, Powerlink, VNX, VNXe, Unisphere, RecoverPoint, and the EMC logo are registered trademarks or trademarks of EMC Corporation in the United States and other countries. All other trademarks used herein are the property of their respective owners. Copyright 2013 EMC Corporation. All rights reserved. Published in the USA, September, 2013.