Date post: | 20-Feb-2016 |
Category: |
Documents |
Upload: | mihail-gheorghe |
View: | 16 times |
Download: | 1 times |
© IntelliMagic 2012 Page 1
The IntelliMagic
White Paper:
Storage Performance
Analysis for an IBM Storwize V7000®
Summary:
This document describes how to analyze
performance on an IBM Storwize V7000.
© IntelliMagic 2012 Page 2
This white paper was prepared by:
IntelliMagic B.V.
Leiden, The Netherlands
Phone: +31 71 579 6000
IntelliMagic, Inc.
Texas, USA
Phone: +1 214 432 7920
Email: [email protected]
Web: www.intellimagic.net
Disclaimer
This document discusses storage performance analysis for IBM Storwize V7000® storage
systems.
IntelliMagic products can be used to support all phases of the storage performance management
processes. Appropriate usage and interpretation of the results of IntelliMagic products are the
responsibility of the user.
Support
Please direct support requests to [email protected].
Please direct requests for general information to [email protected].
Trademarks
All trademarks and registered trademarks are the property of their respective owners.
© 2012 IntelliMagic B.V.
© IntelliMagic 2012 Page 3
Table of Contents
Section 1. Introduction .......................................................................................................................................... 4
1.1 I/O Path ...................................................................................................................................................................... 4
Section 2. IBM Storwize V7000® Architectural Overview and Measurements ................................. 5
2.1 IBM Storwize V7000® Architecture Overview ............................................................................................. 5
Measurement Overview ...................................................................................................................................... 6
2.2 I/O Group .................................................................................................................................................................. 7
Measurements ......................................................................................................................................................... 7
2.3 Nodes .......................................................................................................................................................................... 7
Measurements ........................................................................................................................................................ 8
Cache ........................................................................................................................................................................... 9
Cache Measurements .......................................................................................................................................... 10
2.4 Ports ........................................................................................................................................................................... 11
2.5 Internal RAID Groups or External Managed Disks .................................................................................... 11
Managed Disk Measurements........................................................................................................................... 11
2.6 Internal Disk .......................................................................................................................................................... 14
Disk Measurements .............................................................................................................................................. 14
2.7 Storage Pool ........................................................................................................................................................... 15
Storage Pool Measurements ............................................................................................................................. 16
2.8 Volume .................................................................................................................................................................... 16
Measurements ....................................................................................................................................................... 17
2.9 Thin Provisioning ................................................................................................................................................ 18
2.10 Easy Tier ................................................................................................................................................................ 18
Section 3. Conclusion ........................................................................................................................................... 20
Additional Resources .................................................................................................................................................... 21
© IntelliMagic 2012 Page 4
Section 1. Introduction The purpose of this paper is to provide a practical guide for conducting performance analysis on
an IBM Storwize V7000®. This paper will discuss the end-to-end I/O path, the Storwize V7000
architecture, and key Storwize V7000 measurements. In addition, it will provide guidance in
diagnosing and resolving performance issues using IntelliMagic products such as IntelliMagic
Vision and IntelliMagic Direction.
IntelliMagic Vision and IntelliMagic Direction are part of the IntelliMagic Storage Performance
Management Suite. For additional information on these software products, please refer to
http://www.intellimagic.net/intellimagic/products.
1.1 I/O Path
Figure 1: End-to-End View illustrates how I/Os traverse the I/O path. At its simplest level, any
host initiated I/O request is either a read or a write. The host device driver instructs the host bus
adapter (HBA) to initiate communication with the IBM Storwize V7000® fibre channel ports. The
connectivity equipment, such as the SAN switches and directors, confirm access and send the
packet to the destination fibre ports on the IBM Storwize V7000. If the data requested by the
host resides within the V7000 cache, then the data is sent back across the fabric and to the host.
If the data requested by the host does not reside within the IBM Storwize V7000’s cache, the IBM
Storwize V7000 requests it from its local disks, or the back-end storage array associated with the
volume. A detailed discussion of the read and write paths for different types of I/O will be
discussed in the cache section.
Figure 1: End-to-End View
© IntelliMagic 2012 Page 5
Section 2. IBM Storwize V7000® Architectural Overview and
Measurements
This section contains a brief overview of the IBM Storwize V7000 components and their relevant
measurements. For an in-depth discussion on the IBM Storwize V7000 architecture and
performance considerations see the references in the Additional Resources section of this paper.
2.1 IBM Storwize V7000® Architecture Overview
The IBM Storwize V7000 is a storage system residing in the I/O path between the host and the
back-end storage as illustrated in Figure 1: End-to-End View. It leverages software from the IBM
SAN Volume Controller® (SVC), IBM XIV, and IBM DS8000® series to provide block level storage
virtualization, and automated storage tiering for either internally or externally managed drives.
Typically the IBM Storwize V7000 includes internal disk drives, but in some environments it may
also be used to virtualize external storage. The IBM Storwize V7000 can support up to 32 PB of
externally managed storage.
The advantage to virtualizing the back-end storage arrays is that it facilitates centralized
migrations, provisioning, and replication. The virtualization and pooling of traditional storage
may also lead to improved capacity utilization, as well as performance improvements, as I/O can
be easily balanced across the back-end resources.
An IBM Storwize V7000 consists of two or four hardware components called nodes or node
canisters. Each pair of nodes is known as an I/O group, and it may contain up to twenty disk
storage enclosures. The I/O group consists of two redundant, clustered nodes. At the time that
this document was published, each node consists of an Intel chipset, 8 GB of memory, four 8
Gbps fibre channel ports, two Ethernet ports, two USB 2.0 ports, two 6 Gbps SAS ports, and either
12, 3.5 inch drives, or 24, 2.5 inch drives. The Storwize V7000 supports up to eighteen disk
expansion enclosures each containing either 12, 3.5 inch drives or 24, 2.5 inch drives. The physical
components of IBM Storwize V7000 are commodities. The uniqueness of IBM Storwize V7000 is
in the software and the logical components that will be discussed in the remainder of this
section.
Figure 2: IBM Storwize V7000® Components illustrates the IBM Storwize V7000 components.
Working our way from the bottom of the diagram to the top of the diagram, the LUNs can be
from internal storage system RAID groups or external storage system LUNs. The IBM Storwize
V7000 manages these objects as managed disks (mdisks). Each mdisk consists of a number of
extents of a specified size (default 256 MB). The mdisks are grouped together to form storage
pools. Once part of the storage pools, the extents are grouped to form volumes. Hosts are zoned
to the nodes which have access to all the volumes within the storage pools. The volumes are
assigned to the hosts. The individual components, how they relate to each other, and their
associated measurements will be discussed in more detail in the remainder of this section.
© IntelliMagic 2012 Page 6
Figure 2: IBM Storwize V7000® Components
Measurement Overview
Both front-end and back-end measurements are available for read response times, write
response times, read I/O rate, write I/O rate, read throughput, and write throughput. Front-end
only metrics include write cache delays, and read hit percentage for the I/O groups and nodes.
Back-end only metrics include read queue time and write queue time. Users who are familiar
with the measurements available for the IBM SAN Volume Controller® (SVC) will find nearly
identical measurements for the IBM Storwize V7000®. The V7000 also includes statistics on the
physical disks internal to the controller.
Figure 3: V7000/SVC Multi-Chart Performance Overview illustrates how one might track key
performance indicators for their IBM Storwize V7000.
© IntelliMagic 2012 Page 7
Figure 3: V7000/SVC Multi-Chart Performance Overview
2.2 I/O Group
The I/O group is a logical entity that refers to a pair of redundant clustered controller nodes or
controller node canisters. If one node fails within the I/O group, its workload will be transferred
to the other node. Hosts should be zoned to both nodes within the I/O group so that a node
failure does not cause a host’s volumes to become inaccessible.
Within an I/O group, a volume is associated with a preferred node. In normal operations, the
preferred node services all the I/Os for a given volume. The preferred node can be selected at
volume creation. By default, the IBM Storwize V7000® attempts to distribute the volumes evenly
across the nodes. If a preferred node is not manually selected, the IBM Storwize V7000 will
assign the volume to the node with the fewest volumes. As with host workloads, volume
workloads can vary dramatically. If one node is more heavily utilized than the other node, the
preferred node can be manually changed for a specific volume.
Measurements
The IBM Storwize V7000 aggregates volume measurements such as response times, throughput
and I/O rates to the I/O group. The read response time, and write response times for the volumes
associated with the I/O group represent the average amount of time required for IBM Storwize
V7000 to service the workload. Acceptable I/O response times will vary depending on application
and user requirements, the IBM Storwize V7000 hardware, the IBM Storwize V7000 firmware,
and back-end storage hardware and configuration. In addition to monitoring the I/O response
times, the read and write I/O rates and throughput should be monitored to understand if the
workload is balanced across the I/O groups and the nodes and if the workload is approaching the
limits of the I/O group.
2.3 Nodes
The nodes run the IBM Storwize V7000 software and provide I/O processing, memory buffering,
and connectivity. The IBM Storwize V7000 nodes utilize off the shelf Intel processors, memory,
network, and fibre channel ports. The processors provide the compute power for processing all
© IntelliMagic 2012 Page 8
the I/O operations. The memory serves as both a read and write I/O cache. The Ethernet ports
enable iSCSI host connections while the fibre ports provide fibre connectivity between the
attached hosts, IBM Storwize V7000®, and externally managed storage systems. The selected
connectivity medium can also enable communication between peer IBM Storwize V7000 clusters
for replication activities.
Measurements
The IBM Storwize V7000® provides a robust set of measurements for both the front-end, and
back-end measurements, as well as individual node CPU utilization. The node CPU utilization is
illustrated in Figure 4: Node Utilization.
Perhaps the only shortcoming of the IBM Storwize V7000 node metrics is the lack of visibility into
the internal bandwidth utilization; however, IntelliMagic
Direction may be used to estimate these metrics as
shown in Figure 5: IntelliMagic Direction Internal
Components.
Data throughput and I/O rate per port are required to
understand the port utilizations. The IBM Storwize
V7000 provides response times for each of the volumes
associated with the node from the viewpoint of the host
to the IBM Storwize V7000. These can be rolled up at the
node level to provide a view of the latency for any
particular node.
Tip #1: When planning connectivity requirements ensure that there are adequate bandwidth
and processor resources on the IBM Storwize V7000 ports to handle all the read hits, writes, and
the read misses. The read miss payload and the write payload must be staged and destaged to
Storwize V7000 cache from the internal disks or external storage controllers, and sent out to the
host over the same ports, effectively doubling the port bandwidth requirement for read miss
workloads and write workloads.
Figure 4: Node Utilization
Figure 5: IntelliMagic Direction Internal Components
© IntelliMagic 2012 Page 9
Cache
The primary purpose of cache is to reduce the host I/O response time. This is achieved by
providing cache hits for read requests, and buffering of fast cache writes.
The entire cache on an IBM Storwize V7000® node can be used for read or write activity. If all the
cache is consumed by write activity, and draining writes to the back-end storage is slow, the
system will go into write-through mode. This allows for unwritten write data to be drained to
internal storage in the case of a power outage.
Cache on an IBM Storwize V7000 is segmented into 4 KB segments or pages. A track is used to
describe the unit of locking and de-stage granularity. There are up to eight 4 KB segments in a
track. The IBM Storwize V7000 attempts to coalesce multiple segments into a track destage
when the segments to be written are within the same track.
Read I/O requests are either read cache hits or read cache misses. If the requested data is
resident in cache, the data is immediately transferred to the requestor from memory. This is
called a read cache hit and avoids back-end disk access. If the requested data is not resident in
cache, the data is requested from the internal disk drives or the external storage systems. This is
called a read cache miss. After the data is loaded into cache from internal drives or the external
storage system, the front-end adapter sends it to the host that initiated the request.
For write I/O requests, the data is written to cache on the host’s preferred node. It is then
mirrored to the partner node to provide resilience in the event of a node failure. Subsequently,
the preferred node sends an acknowledgement to the host that the I/O has been completed. This
is called a fast write (FW). At some point after the acknowledged completion of the I/O, the write
is destaged to the internal drives or external storage system. The tracks are marked unmodified
at this point and will stay in cache until they are moved out of cache by the controller using an
LRU algorithm. Assuming there is sufficient free write cache on the external storage system, the
write response times to external storage systems should be fast writes. This latency is not
accumulated on the front-end response time; rather it is associated with the mdisk latency as
will be discussed further in the mdisk section.
In the event of a node failure, all the modified cache from the remaining node will be drained to
the internal drives or external storage system. The behavior of write I/Os in this scenario will
change to write through mode. The acknowledgement from an IBM Storwize V7000® to the host
that a write has been completed will only be sent upon confirmation that the write has been
completed to the internal drives or acknowledged as completed by the back-end storage system.
Tip #2: Consider running your nodes and their associated components at no more than 50%
utilization during online periods. That way if you have a node failure, your cluster will not
severely impact the performance of your online applications.
© IntelliMagic 2012 Page 10
Cache is also managed at the storage pool level. This is referred to as partitioning. There is an
upper cache limit set for each of the storage pools. This prevents a single pool from getting more
than its fair share of cache. The upper cache limit depends on the number of storage pools. If the
cache limit is reached for I/Os to a particular storage pool, write I/Os will behave in a write
through mode. This behavior will only continue while the amount of cache consumed for write
I/Os to a specific storage pool exceeds the upper limit. As discussed previously, this has the effect
of requiring write I/Os to be completed to internal drives or acknowledged by the external
storage system prior to informing the initiator that the write has been completed. It is rare to
encounter this behavior unless there is a problem draining writes to an external storage system
due to overutilization, or a problem within the storage system.
Cache Measurements
From a performance analysis perspective, it is important to understand the effectiveness of the
cache management. Key cache measurements include the read cache hit ratios, and the write
cache delays. Consistent low read cache hit ratios may indicate that the storage system has
insufficient cache for the workload. These metrics are available at the IBM Storwize V7000® I/O
group, node, partition, and volume level. They can be rolled up at the storage pool level as
illustrated in Figure 6: Read Hit % vs I/O Rate.
Figure 6: Read Hit % vs I/O Rate
Tip #3: For open systems workloads, an average read cache hit ratio greater than 50% is desired.
IntelliMagic Direction can help identify whether additional cache will help improve the response
time or throughput for a particular workload. For certain workloads that have very random read
access patterns it may be completely acceptable to have a read cache hit ratio much lower than
50%.
The number of write cache delays per second indicates whether or not the IBM Storwize V7000 is
able to destage I/Os quickly enough to the back-end storage system. A write delay event
indicates that cache is full, and that new write I/O operations may only be completed after a
physical destage to disk frees up space. This typically only happens when the internal drives or
external storage system is saturated.
© IntelliMagic 2012 Page 11
Tip #4: A small number of write delays can significantly increase write response times to hosts.
2.4 Ports
There are fibre channel and Ethernet ports on the nodes. All traffic from hosts, between nodes,
and to external storage systems passes through the node’s ports. Measurements include read
and write throughput, and read and write operations to and from hosts, other SVC nodes, and
external controllers. It is important to keep the node port throughput balanced and less than
50% utilized in case there is a node failure. Figure 7: Fibre Host Read MB/sec Balance Chart
illustrates how balanced the ports are on a V7000 cluster. Each bar represents a single fibre
channel port. The average read mb/sec, standard deviation and minimum/maximum values
provide the analyst a quick view of how well balanced port activity is across the entire cluster.
Figure 7: Fibre Host Read MB/sec Balance Chart
2.5 Internal RAID Groups or External Managed Disks
A managed disk has a one-to-one relationship with a back-end storage system LUN or internally
managed RAID group. There should also be a 1:1 relationship between externally managed LUNs
and their supporting RAID groups. This section uses mdisk to refer to both the internally
managed RAID groups and the external managed disks.
Managed Disk Measurements
You can monitor the performance of your managed disks using the managed disk statistics that
include both the internal RAID group and external mdisk read and write response times. These
statistics measure the amount of time required to perform a read or write operation from the
Storwize V7000 to internal RAID groups or the external mdisk. For externally managed mdisks,
read and write queue times measure how long read or write operations are waiting to be sent to
the back-end storage system or internally managed drives.
Tip #5: Average external mdisk queue times should be less than a couple of milliseconds as the
Storwize queue times only measure the amount of time the I/O request spends waiting to be
© IntelliMagic 2012 Page 12
sent to the internal disk or external storage systems. If this is over 1.0 ms, it is a sign of
contention on the fabric or on the back-end internal disks or external storage system.
Figure 8: External Write Queue Time illustrates a situation where the external write queue time
is exceptionally high.
Figure 8: External Write Queue Time
Tip #6: For increases in front-end response time that do not correlate to increases in the RAID
group or external mdisk response times, the performance issue is with the path from the host to
the Storwize V7000 or the Storwize V7000’s ports or processors.
Tip #7: Average mdisk response times greater than 12.0 ms for fibre channel drives and 15.0 ms
for SATA drives indicate some sort of constraint within the Storwize V7000 to the internally
managed drives, or for externally managed mdisks it indicates a problem along the path or
within the external storage system. The expected response times will vary depending on the
exact configuration.
© IntelliMagic 2012 Page 13
Figure 9: Managed Disks External Read Response Time SLA Chart illustrates a situation in which
the read response times to a significant number of the mdisks is exceeding desirable service
levels.
Figure 9: Managed Disks External Read Response Time SLA Chart
In order to understand if high back-end response times on the Storwize V7000 are a result of
saturated internal disk RAID groups or an external storage system component, you will need to
have visibility into the storage system’s components. For supported platforms, IntelliMagic
Vision provides end-to-end visibility into both the internal RAID groups and the external storage
system components.
When vendors describe I/O response times they typically mean weighted averages similar to
what is described in Example 1: Average Response Time.
Example 1: Average Response Time
Workload Type: OLTP
Read / Write Ratio: 80/20
Read Hit Ratio: 50%
Write Hit Ratio: 100%
Write Hit Response Time: 1.0 ms
Read Hit Response Time: 1.0 ms
Read Miss Response Time: 6.0 ms
Read I/Os per Second: 800
Write I/Os per Second: 200
Average Response Time =
((400*1.0) + (200*1.0) + (400*6.0))/1000 = 3.0 ms per I/O
© IntelliMagic 2012 Page 14
2.6 Internal Disk
The IBM Storwize V7000® can have up to 240 3.5 inch drives or 480 2.5 inch drives. Table 1: Drives
Supported shows the drives currently supported:
Table 1: Drives Supported
Drive Type Speed (RPM) Size
2.5-inch form factor SSD N/A 200, 300, and 400 GB
2.5-inch form factor SAS 10,000 300,450, and 600 GB
2.5-inch form factor SAS 15,000 146 and 300 GB
2.5-inch form factor Nearline SAS 7,200 1 TB
3.5-inch form factor Nearline SAS 7,200 2 TB
Note: For the latest supported disk drives, please consult the IBM web site.
Disk Measurements
For each drive, the IBM Storwize V7000 provides read I/O, write I/O, read throughput, write
throughput, read response time, and write response time. From these metrics, disk utilization
can be calculated. Depending on the type of drive, you can expect different types of response
times. For an estimation of expected service times, you can refer to Table 2: Disk Service Time
Estimates:
Table 2: Disk Service Time Estimates
Drive
IOPS
(No
Queue)
Average Service
Time (ms)
SATA 7200 RPM 120 8.5
SAS/FC 10K RPM 167 6
SAS/FC 15K RPM 250 4
© IntelliMagic 2012 Page 15
Figure 10: Internal Disk Drive Utilization SLA Chart illustrates the utilization of all the internal
drives of a six-node SVC cluster. This is a good way to quickly identify and disk hot spots within
the internal drives.
Figure 10: Internal Disk Drive Utilization SLA Chart
2.7 Storage Pool
A storage pool is a grouping of more than one managed disk. When planning a storage pool it is
important to remember that when a single managed disk experiences a failure, it brings the
entire storage pool offline. As a result, one of the primary design goals is to limit the hardware
failure boundaries. General performance guidelines should also be addressed as part of the
design, and will also be discussed in this section. With these thoughts in mind, there are several
best practices to consider when creating a storage pool as enumerated in Table 3: Storage Pool
Best Practices.
Table 3: Storage Pool Best Practices
Best Practice Availability Performance A storage pool should utilize managed disks from one storage system. X Each external storage system should provide managed disks to a single
Storwize V7000 cluster. X
A V7000 can present storage to SVC, but a V7000 cannot present storage to
another V7000.
Each internal or external RAID group must be included in only one storage
pool. X
Rather than adding capacity to an existing storage pool, create a new
storage pool. X X
Implement striped volumes for all workloads except 100% Sequential. X For externally managed disks, utilize storage pools with at least eight
managed disks to take advantage of the round-robin distribution of I/O
workload to the external storage controller. X X
Select an extent size that balances cluster capacity and volume granularity.
Testing has shown good results with 128 MB and 256 MB extents. X
© IntelliMagic 2012 Page 16
Storage Pool Measurements
The storage pool measurements provide a good means for monitoring the performance of both
the front end (host to port) and the back end (cache to disk). The storage pool measurements
provide a combination of measurements that are aggregated from both the front-end volume
operations, as well as the back-end managed disks operations. On the front end, the response
times, I/O rates, and I/O throughput provide an excellent means for understanding if the storage
pools are responsive and balanced. The response times include the read cache hits as well as the
read cache misses. On a system with a normal read cache hit percentage, the average front-end
read response time will be understandably lower than the back-end read response times for the
same storage pool. The front-end write response times should generally be 1.0 ms or less as they
are only measuring the amount of time required to write to the preferred node’s cache and
mirror to the secondary node.
Figure 11: V7000 Read Response Time by Storage Pool
2.8 Volume
Storage pools are created from the storage provided by externally managed disks (mdisks) or
internally managed RAID Groups. The storage from the managed volumes are grouped together
and divided into extents. A volume is a discrete grouping of storage extents can be made
addressable to a host. The size of the extent can be selected during the storage pool creation but
the default size is 256 MBytes. When volumes are created the capacity and the layout of the
volume are selected. For IBM Storwize V7000® managed disks, the data layout can be striped or
sequential. In all cases but 100% sequential workloads, the volumes should be created as striped
volumes. Striped volumes consist of extents that are spread across the managed disks in a
round-robin fashion. Non-managed Storwize V7000 volumes or image-mode disks are not
covered in this paper as they not typically part of steady state configurations.
Table 4: vdisk Best Practices
Best Practice Management Availability
New vdisks should go to least utilized storage pool within cluster. X
© IntelliMagic 2012 Page 17
Vdisk size should be appropriate for hosts. For most hosts, fewer
larger vdisks work better from a performance and management
perspective. This should be balanced with the size and activity level
of the vdisk in relationship to the storage pool in which it resides. A
rule of thumb is the vdisk should not consume more than 10% of
the capacity of the storage pool nor should it consume more than
20% of the performance bandwidth of the storage pool.
X X
Use striped volumes even if application or host has its own striping
as long as stripe sizes are dissimilar. Fine-grained host LVM
striping can be beneficial for performance.
X
Set Thin Disk to auto-expand, so if allocated space is used the disk
doesn’t go offline if additional writes occur. X X
Measurements
In order to detect whether performance problems exist, it is important to ignore inactive
volumes, since those volumes have an insignificant impact on the overall performance of the
storage system. It is common for some of the volumes with the highest response time to have
little I/O activity. Sorting the volumes based on their response time alone, therefore, is not a good
way to find the volumes that cause bad overall performance. A good way to find the volumes
that have the biggest impact on the overall performance is to sort the volumes by their I/O rate
time's response time, which is called “I/O Intensity” as illustrated in Figure 12: Top Volume by I/O Intensity.
Figure 12: Top Volume by I/O Intensity
Response times for volumes are provided from the view of the front end of the IBM Storwize
V7000®, so they include both cache hits and cache misses. On a well balanced system that is not
overloaded, 100% of writes should be satisfied from cache. The response time for writes satisfied
from cache should be no more than the amount of time required to write to the target node’s
cache, mirror to the secondary node, and then send the host an acknowledgement.
© IntelliMagic 2012 Page 18
Read hits should also take very little time to process and transfer the data. For internally
managed disks, the read misses will require a physical disk access. For externally managed disks,
the read miss may be served from the externally managed controller’s cache or its back-end disk
drives if the requested tracks are not in cache. In either case, a read miss is significantly slower
than a read hit. Figure 13: v7000 vdisk Response Time illustrates the average vdisk response
time for the top 30 volumes.
Figure 13: v7000 vdisk Response Time
2.9 Thin Provisioning
Thin provisioning provides a mechanism to assign more logical capacity than physical capacity.
In a thin provisioning environment, only the portion of the LUN that is written to is actually used.
This means that if 1 GB of a 20 GB LUN is written to, then only 1 GB is in use. To accomplish this,
the extents are divided into smaller elements called grains. The grain size is configurable at the
storage pool or volume level. The default is 32 Kbytes. In a thin provisioned environment, only
the grains that are written to will be counted towards the amount of storage actually used by the
volume. The IBM Storwise V7000® tracks both the allocated and the used.
Tip #8: When you create a thin-provisioned volume, set the cache mode to readwrite to cache
metadata. This will reduce latency associated with I/Os to thin provisioned volumes as the
metadata will be cached.
2.10 Easy Tier
Easy Tier provides a mechanism to transparently migrate hot extents to the most appropriate
storage tier within an IBM Storwise V7000 environment. It evaluates the read miss activity of the
volumes within an Easy Tier enabled storage pool and moves the active extents to a higher disk
tier. The current implementation is designed to take advantage of a storage pool containing
mixed drive technology such as SSD and FC. Hot extents will be migrated to the SSD drives and
extents with low activity will be migrated to the slower and less expensive disk technology. The
migration does not take place between storage pools.
© IntelliMagic 2012 Page 19
The SSDs should be placed internally within the IBM Storwise V7000® even if the FC or SATA
drives are located externally. This will reduce the fabric overhead.
Tip #9: Set grain size for non flash copy thin volumes to 256 KB. For flash copy volumes, set the
grain size to 64 KB. Default is 32 KB! If you leave the grain size at 32 KB, EasyTier will assume that
it is not sequential I/O even when it is. This may result in sub-optimal data placement and
performance.
© IntelliMagic 2012 Page 20
Section 3. Conclusion In this paper, we highlighted some of the architectural components of the IBM Storwize V7000®.
We also discussed some of their associated measurements. Finally, we utilized IntelliMagic
Vision to examine several of the key performance metrics for the IBM Storwize V7000 storage
system performance. Using IntelliMagic Vision we were able to easily identify imbalances and
service level exceptions.
We realize that many subjects were greatly simplified, and understand that becoming an expert
in storage performance management requires significant real world experience that cannot be
obtained by reading a white paper. Here at IntelliMagic, we strive to make storage performance
management easier by providing world class solutions, support, training, and services. Next
time you need some guidance on storage performance issues, feel free to contact us for a free
performance analysis at [email protected].
© IntelliMagic 2012 Page 21
Additional Resources
Implementing the IBM Storwize v7000®, IBM Redbook SG247938
IBM Storwize V7000 Information Center,
http://publib.boulder.ibm.com/infocenter/storwize/ic/index.jsp