Storage Technology Primer
This section recaps the various storage technologies previously touched on and places them in a matrix before moving on to the actual implementation discussion.At the disk level, there are four connectivity options:Direct-attached storage (DAS) configured in a one-to-one relationship with the hostNetwork-attached storage (NAS) residing on an IP-based networkStorage Area Network (SAN)-attached storage shared between hosts via a private Fibre Channel networkSAN-attached storage shared between hosts over a routable IP network
As noted earlier, in addition to being expensive, inefficient, and difficult to scale, DAS presents multiple single points of failure. NAS allows storage to be accessed remotely and utilizes capacity on the IP network infrastructure. Fibre Channel (FC) SANs allow sharing of storage resources over a private Fibre Channel network, increasing allocation efficiency. IP SANs allow for low-cost networked storage as well as the extension of SANs over long distances. Consequently, both FC and IP SANs offer significant advantages over DAS and NAS.A Fibre Channel SAN infrastructure can be comprised of small independent SAN islands built on fixed Fibre Channel switches (16- and 32-port switches). A SAN infrastructure can also be built by joining multiple larger switches together Chapter 5, "Maximizing Storage Investments," simplifies the management of storage devices by abstracting the devices themselves to increase operational efficiencies. Virtualization also reduces capital expenditures by increasing utilization rates. Software products in this category include MonoSphere Storage Manager, VERITAS Storage Foundation for Networks, and IBM SAN Volume Controller.Table 4-1 highlights the solution types and their application in the enterprise.
Data storage | Hardware | DAS | Expensive disk |
Shared data storage | Hardware | FC SANs | Small Fabric, fixed Fibre Channel switch |
Shared data storage | Hardware | FC SANs | Large Fabric, modular Fibre Channel switch |
Shared, low-cost data storage | Hardware | NAS | Inexpensive, redundant, shared disk |
Shared, low-cost data storage | Hardware, Protocol | iSCSI storage network | Small fixed device, blade, multiprotocol modular switch |
SAN Extension | Hardware, Protocol | CWDM, DWDM, SONET/SDH, and dark fiber | Optical, long-distance carry |
SAN Extension | Hardware, Protocol | iSCSI storage network | Blade, multiprotocol modular switch between hosts, SCSI storage |
SAN Extension | Hardware, Protocol | FCIP storage network | Blade, multiprotocol modular switch between Fibre Channel devices |
SAN Extension | Hardware, Protocol | IFCP storage network | Gateway device between Fibre Channel devices |
SRM | Software | Resource and asset management | Agent devices on hosts |
SAN Management | Software | Device management | Centralized device administration and management |
Virtualization | Software | Virtualization of resources | Abstraction of resources to increase utilization |
TCO, Tiered Storage, and Capacity Planning
Calculating TCO is covered in Chapter 2; however, it is important to take that discussion to the next level. Discussions of TCO eventually lead to the discussion of a tiered storage architecture as part of an Information Lifecycle Management (ILM) framework. In an ILM framework, processes and procedures dictate the movement of information through storage tiers as its criticality and frequency of access decrease over time.The U.S. Securities Exchange Commission (SEC) regulations and new requirements for compliance with the Sarbanes-Oxley Act and the Health Insurance Portability and Accountability Act (HIPAA) have lengthened retention times for many types of data up to seven years or more. It is not cost effective to store all types of data on the same format for such long periods of time because the criticality of information typically decreases over time. This is not to say that the information becomes worthless as it ages (if a certain record is required after a long period of time, the crucial nature of that information increases tenfold), but the frequency of its access decreases and therefore the nature of the storage solution should change as well. This is a simple cost-benefit analysis. Even much of the information that is considered mission-critical changes little after it is written, and might be accessed infrequently if at all. Abstraction or virtualization of the storage through software can allow an application to transparently access the information even after it is moved to a different tier.
Information Lifecycle Management
ILM is not a new concept; processes similar to those now being introduced as part of an ILM framework have been used by business for years to manage the storage of information. As requirements for retention change and the frequency of access declines, the information moves through a tiered infrastructure. In this manner, the TCO is reduced as hardware solutions at each tier have different cost structures. Ultimately, the information is archived to tape for long-term, offline storage, converted to some type of flat file format for access by any application, or deleted entirely as appropriate. Figure 4-2 outlines a generic tiered, ILM infrastructure.
Figure 4-2. Tiered Storage and Information Lifecycle Management

Performance Planning
Critical to supporting a tiered infrastructure and managing information through its lifecycle is the ability to accurately size an application based on its performance traits (such as heavy writes in online transaction processing, a lot of reads in data warehousing, or bursts of reads and writes in batch processing). In addition, it is necessary to match the storage subsystem to the functional requirements of the application (high availability, near line, length of retention).Estimating an application's performance requirements can be done simply by running systems tools such as, system activity reporter (SAR), to find IO rates over a set period of time, then sizing appropriately to handle periods of peak throughput based on the block size of the application. Similarly, the same type of data can be collected at the database level with regard to cache hits and misses.Alternately, capacity planning and application sizing can be done with proprietary capacity planning tools or by using capacity planning frameworks, such as those offered by Teamquest or BMC Software. Depending on the nature of the environment, if one of these larger frameworks is not already installed, it is highly unlikely that it is worth buying and installing one just to size a solution prior to investing in a SAN. Ultimately, upon maturity, SRM software and SAN management software greatly simplify the process of capacity planning for SAN environments.Using system tools and doing the math required to accurately size and plan for disk and IO throughput prior to installing a SAN are relatively straightforward processes. The final requirement for performance planning is a holistic understanding of the application environment, which is where corporate knowledge plays a big role. When are the backup windows? When are the periods of peak processing? When was the last time the system was upgraded? This type of data is crucial to solid analysis and cannot be completely obtained through the use of software. Although this information might not be inherently quantifiable, it can help increase the granularity of your quantitative analysis and help make a better business decision.
Oversubscription
Another aspect of performance planning is oversubscription. Oversubscription in the SAN arena, just as in LAN or WAN networking, is a capacity planning technique that guarantees a minimum rate of throughput to an application or set of applications, while at the same time provisioning for an estimated maximum rate of throughput. In other words, the capacity of the network is oversold under the assumption that not all applications burst and reach maximum throughput at the same time. In this way, provisions are made for maximum bandwidth demand with significantly less infrastructure cost.The following is a generic example designed to illustrate the concept of oversubscription exclusive of the concept of Quality of Service (QOS). Also note that oversubscription, in the context of host to switch connectivity, is often referred to as the fan-in or fan-out ratio.A small SAN-island backbone (64 2-Gbps ports on a director-class switch) is oversubscribed to five applications on each of six hosts. Each application's demand bursts at 2 Gbps on each channel for a total application demand of 60 Gbps (6 hosts x 5 applications x 2 Gbps), as shown in Figure 4-3.
Figure 4-3. Oversubscription

Figure 4-4. Oversubscription with ISLs

Core Topology
The previous examples demonstrated a core topology using a single FC switch at the core of the architecture. Core topology SANs may also use a large SAN fabric built on multiple core switches linked together by ISLs. Both of these scenarios are demonstrated in Figure 4-5.
Figure 4-5. Core Topology

Core-Edge Topology
A core-edge topology, as shown in Figure 4-6, works well in an environment with a large number of hosts accessing small amounts of data. To provide adequate service to all hosts, as well as face the challenge of locality and many fiber runs across the datacenter back to the core, smaller departmental, fixed, or edge switches are inserted at fixed points in the topology to provide fan-in to many hosts.
Figure 4-6. Core-Edge Topology

Redundancy and Resiliency
In the discussion of topologies, it is important to understand the nature of redundancy and resiliency in building a SAN architecture as well as the cost differences between these two features.Previous examples glossed over the topic of redundancy to simplify the concepts of core and core-edge topologies. In these examples, redundant hardware is needed to provide for maximum availability and uptime, while a resilient architecture protects against a single component failure.To implement full redundancy, there must be at least one additional switch with separate power, separate networking, and multiple paths from the hosts as shown in Figure 4-7. This design also utilizes a separate fabric at each core that prevents failures associated with single-fabric service outages. Multipathing software such as VERITAS DMP or EMC PowerPath provides for multiple paths to the disk at the operating system level, which provides path failover capability at the host level. Obviously, with redundant hardware at each level from the host bus adapter (HBA) to the disk, this solution has the highest price tag. But this solution also provides the best protection against failure and interruption in service, and is a suitable architecture for a Tier 1 infrastructure requiring maximum availability.
Figure 4-7. Redundant Architecture

Figure 4-8. Resilient Architecture
