Chapter 25. Utilizing Storage Area Networks

Few items in the data center invoke more awe and respect than the shiny black monolith that is the SAN. SANs, or Storage Area Networks, are extremely high-performance collections of disks that can be sliced and diced dynamically and attached to remote systems as though they were directly attached. SANs differ from traditional DAS, or Direct Attached Storage, in that the disks are no longer attached to the local system through SCSI or IDE connections. The SAN is viewed as a cloud and is literally a separate high-speed network with the sole purpose of connectivity between hosts and high-speed disks. From the server’s point of view, the remote disk acts exactly the same as the locally attached disk. By consolidating the disks into a central location, you are able to take advantage of situations that just weren’t possible in the past.

This chapter compares and contrasts Storage Area Networks, Direct Attached Storage, and Network Attached Storage (NAS), and shows you when to take advantage of one technology over another. This chapter will explain the requirements of each technology to help you avoid common mistakes when choosing a storage technology. It will also touch on industry best practices for using NAS and SAN technologies with specific applications.

Defining the Technologies

To understand how and when to use technologies like NAS or SAN it is important to understand what they are and what they offer. The technologies differ in how they are used and what advantages they provide. Many administrators assume that they need SAN when often a NAS will suffice. Because IT budgets are far from limitless, it is to your advantage to know that you aren’t overbuying for your solution.

What is a SAN?

A Storage Area Network is a high-speed special-purpose network or subnetwork that connects various data storage devices with associated data servers on behalf of a larger network of users. Typically, a SAN is but part of an overall network of computing resources for an enterprise. A SAN is usually located in relative proximity to other computing resources, such as databases and file servers, but might also extend to remote locations for backup and archival storage. These remote locations are traditionally connected via wide area network carrier technologies such as asynchronous transfer mode (ATM) or Synchronous Optical Networks (SONET).

It is very important to understand that the SAN is more than just the chassis that contains the disks. It includes the RAID controllers for the disks, the Fiber Channel switching fabric, and the Host Bus Adapters that reside in the data servers. SANs are traditionally connected to hosts via Fiber Channel. Although it can be fairly easy to support dual arbitrated fiber loops in a corporate environment, keep in mind that one of the primary benefits of SAN is the ability to do block-level mirroring to another SAN. If this SAN is located remotely, up to 1,000km away with current Fiber technology, a company needs to have fiber between the two locations. A fiber connection across those kinds of distances can be quite expensive.

SAN technologies excel in the area of disk performance. Fiber channel networks regularly push 2GB/sec of throughput. Although SCSI technologies can move data at up to 320MB/sec, they are limited to less than 25 feet of distance. SAN, not unlike SCSI, is seen by the host system as RAW disk. This is also referred to as a block-level technology. In the past, database applications required block-level access to the disk as well as the “near 0 latency” offered by SAN.

Zero Latency

Although most SAN manufacturers refer to the performance of their products as having zero latency it is important not to misinterpret this. Zero latency refers to the fact that Fiber Channel has extremely low overhead and doesn’t add additional latency. The laws of physics, on the other hand, are still in effect. A 1,000km fiber run will still take 7 milliseconds roundtrip.

What is NAS?

Network attached storage (NAS) is a hard disk storage technology that uses its own network address rather than being attached directly to the host computer that is serving applications or data to a network’s users. By removing storage access and its management from the host server, both application programming and files can be served faster because they are not competing for the same processor time. The network-attached storage device is attached to a local area network via Ethernet and given an IP address. File requests are mapped by the host server to the NAS device.

Network-attached storage consists of hard disk storage, including multidisk RAID systems, and software for configuring and mapping file locations to the network-attached device. NAS software can usually handle a number of network protocols, including Microsoft’s Internetwork Packet Exchange, Common Internet File System and NetBEUI, Novell Netware Internetwork Packet Exchange, and Sun Microsystems Network File System. Configuration, including the setting of user access priorities, is usually possible using a Web browser though many NAS offerings require command line configuration. Most NAS manufacturers include specialized software for allowing specific applications such as SQL or Exchange to take advantage of special functions provided by the NAS. These functions include things like mirroring, failover, automated recovery, and snapshotting.

A Single MAC Address Conversation Cannot Span Multiple Network Interfaces

When considering network architecture to support a NAS device, remember that when bonding multiple Ethernet interfaces, a single MAC address conversation cannot span multiple network interfaces. If the NAS is on Gigabit Ethernet and a server has a quad fast Ethernet card, it will only talk to the NAS device at 100MB/sec even though it has 400MB/sec of potential throughput.

NAS has the advantage of using existing Ethernet technologies that are much less expensive than Fiber technologies. With the availability of 10GB Ethernet, NAS is able to compete with Fiber Channel–based technologies even with the added overhead of Ethernet over Fiber Channel.

What is DAS?

Direct Attached Storage (DAS) is the traditional “local disk” that most administrators are accustomed to. Technologies like IDE and SCSI are used to connect drives or RAIDs to the host computer. Direct Attached Storage has many advantages in the area of performance and cost. With SCSI technologies like Ultra 320, which can transfer 320MB/sec of data across the bus and controllers with large amounts of cache memory, DAS is well situated to provide fast and stable disk performance. The disadvantage of DAS is that the management load of each of these DAS subsystems increased linearly with the number of servers. This can eventually become unwieldy as there is a lack of centralized management.

With the advent of Serial ATA and Serial ATA RAID controllers, IDE devices are becoming more and more viable for systems that need fast access to large amounts of data that won’t see huge numbers of transactions. Systemlike backup servers that will cache to disk before spooling to tape can benefit from the speed and low cost of Serial ATA devices. IDE drives almost always lead SCSI drives in the area of size and are always less expensive. IDE RAID controllers offer excellent performance at much lower costs than their SCSI equivalents. IDE-based RAID controllers don’t handle large numbers of requests as well as SCSI-based RAID controllers. IDE-based RAIDs are still limited to four devices. This isn’t always an issue as SATA drives of 250GB or more are commonly available. A 1 Terabyte SATA RAID can be built for well under $1,500. With throughput of over 180MB/sec, Serial ATA RAID can be a very tempting DAS technology.

IDE Technologies Can Compare Favorably to SCSI Technologies

Although IDE technologies can compare favorably to SCSI technologies in the area of seek time and bus bandwidth, they can’t compete in the area of IO transactions per second. Most applications that require large amounts of disk are IO bound, not bandwidth bound. Be sure to understand the IO requirements of an application before choosing disk technologies.

When is the Right Time to Implement NAS and SAN Devices?

There are many reasons to implement a NAS or SAN solution in favor of DAS. If the requirements for storage consolidation, reduction in server count, centralized management of disk resources, SLA recoverability times, or near real time mirroring of data justify the cost of a SAN or NAS solution it is time to explore those options. To make an informed decision about when to make the switch it is important for you to pass through several phases:

  1. Analyze—. Gather usage metrics and performance metrics. Determine how storage is being used and how it affects the business processes.

  2. Plan—. Determine the current limitations of your storage solutions. Prioritize the problems and determine if there is a better way. Don’t fall into the trap of doing things just because they were always done a particular way.

  3. Develop—. Build the proposed solution for testing. Perform benchmarking to show improvements over the old methods.

  4. Pilot—. Test the solution and improve it based on user feedback. Educate the user population on how to take full advantage of the new functions and determine the improvements in efficiencies.

  5. Deploy—. Deliver the solution to the masses.

Following this methodology will not only streamline the process of implementing new and more efficient storage technologies but also the process will provide valuable data to help upper management buy into the upgrades and support the storage program.

Analyzing Your Storage Needs

The first phase of any good project is an in-depth analysis of the environment and its needs. In the case of Storage Systems it is critical to identify any systems with special requirements. This would include systems that require multiple layers of redundancy, systems that are under extremely tight Service Level Agreements, and systems that cannot tolerate a loss of data.

Another key area to understand is the capacity requirements of the enterprise. If an investment is going to be made in storage it is a good idea to plan for several years of growth. Look at the number of servers in the environment. If additional servers have been added simply because that is the way things were always done, it is time to look at shifting the philosophy to doing things because they are the right way to do them.

Disk Drives Get Larger, Faster, and Cheaper Each Year

Disk drives get larger, faster, and cheaper each year. When planning for the future, keep expandability in mind. By buying a partially filled chassis now and adding additional disks later, you can take advantage of falling disk prices and save money over the long run and still get the full capacity they need and the benefits of fewer chassis.

Planning the Storage Solution

Storage technologies can be very confusing. In most situations, valid arguments can be made for using any of the available technologies. This is a situation in which it makes a lot of sense to get your vendors involved. Contact your potential vendors and let them know what your storage requirements are. Often times they have worked with other companies with similar needs and can provide valuable insight into what worked and what didn’t. Given the costs of a large storage system, you can’t afford to do it wrong.

After you have an idea of what you’d like to implement, find out if there are reference accounts that you can contact to determine if they were happy with the solution they implemented. Some companies will try to get you to commit to the latest greatest versions of their software and firmware. Large storage environments are a big investment and business processes will depend heavily on it. Ensure that you are implementing a stable and well-tested solution.

Does This Decision Support the Goals of the Project?

There are a tremendous number of options when it comes to storage solutions. When in doubt about a decision, always refer back to the original goals of the project and ask yourself “Does this decision support the goals of the project?”

Developing the Storage Solution

After you have determined the needs, explored the options, and come up with a plan, the real fun can begin. Any solution that will become part of the critical path of business must be developed and tested in a controlled lab environment. This is the part of the project where policies and procedures start to take form. Practice runs of mirroring, failing over of resources, and recovery of systems will ensure that the solution will be able to support the needs of the company.

This development phase will identify several requirements that are not usually thought of during the planning phase. Most specifically these requirements are in the area of facilities. Most SAN devices are fairly large. An EMC Symetrix and Connectix, for example, will take up a full rack each. With heat generation more than 3,000BTUs, HVAC resources will need to be considered. Also keep in mind that most SAN and NAS solutions require 220V to run them. Ensure that planned data center locations have appropriate space, cooling, and power. Power should include not only the standard AC feed, but battery backup as well. Be aware of any special requirements of the SAN or NAS. Some SAN devices on the market will void their warranty if they are placed within five feet of any solid objects.

Piloting the Storage Solution

After the solution has been well tested in the lab, it’s time to break it in limited production. Even extensive testing in a lab cannot always fully test all possible scenarios that a storage system will see in production. The best way to try to break the solution is to let users have access to it. Rolling the storage solution into a limited Pilot program will enable you to gather feedback about the performance of the system as well as feedback on the policies and procedures that will support the system. It’s also a great time to get support staff trained on a system that they will likely end up supporting in full production.

Deploying the Storage Solution

After the system has gone through a final feedback loop from the pilot phase it is time to put the system into production. After following these steps you should have a well thought out solution that meets the current and future needs on the company. The solution will have been well tested and will encompass tweaks and improvements that came from the various testing phases. Performance metrics will be known and monitoring points determined. Now the company can begin to enjoy the benefits of the new storage solution.

Designing the Right Data Storage Structure

A code vault that stores source code and enables programmers to check code in and out so that they can work on it locally on their systems probably doesn’t justify a SAN with 28 platters. An Exchange 2003 Cluster that supports 3,000 users probably could. When designing a storage structure it is critical to understand your own needs.

Choosing the Right Connectivity

All the high-speed disks in the world won’t amount to much if you can’t get the data to and from the servers quickly. In a NAS environment, the network itself is the biggest concern for performance. Most NAS devices on the market use very fast heads that are literally dedicated computers with high performance processors and loads of memory. With SCSI RAID controllers on board, they can easily saturate multiple 100MB Ethernet connections. Attaching such a device to a low-end switch would result in the NAS running in an extremely restricted manner. Strongly consider using a switch that will enable you to use a gigabit connection.

Consider creating a separate network for the NAS environment. Say, for example, that the NAS is going to support a number of Exchange servers. By multi-homing the Exchange servers, they would have one Ethernet connection that faced the users and provided connectivity to the mail clients, whereas the other interface would be dedicated to NAS traffic. This would allow each interface to run unfettered by the traffic associated with the other network. This also enables you to upgrade only a subset of the network to improve performance and save money. The traffic of the database transaction back to the NAS device by Exchange would be much greater than the traffic associated with users viewing their mail because the traffic that would normally go to the local disk would now be traveling across the Ethernet via the Virtual Disk driver that connects the NAS to the Exchange server.

When selecting network gear for a NAS out-of-band network, focus on packets per second. Whenever possible, build this NAS network with multiple switches that are cross-connected. Connect each server to both switches with the NICs in a teamed mode. This will not only add bandwidth but also will create redundancy for the network layer. Odds are if the application warranted the use of a NAS device, it deserves redundancy at the network level as well.

Slicing and Dicing the Available Disk

Simple physics tells you that you’ll get improvements in performance as you add more disks to an array. Because each drive’s read/write head can operate simultaneously you get a fairly linear improvement as drives are added. NAS and SAN offer the advantage of dynamically increasing the size of a volume without taking the volume offline. This allows for the addition of even more spindles.

Although it’s possible to later resize a volume from a NAS or SAN, you must be careful not to over subscribe the device. Devices that support snapshots of the data reserve twice the volume size that they claim for capacity. So in order to make 100GB available to a server, the NAS would reserve 200GB on itself. This ensures that it is able to complete all transactions. This function can be disabled on most devices but it is not recommended. This removes the protection from oversubscription of the disks.

Never Underestimate the Power of Multiple Spindles

Never underestimate the power of multiple spindles. For applications like Exchange that are very IO-intensive, a quad processor Xeon 1.6 GHz with 4GB of memory and seven local disks can support up to 1,000 heavy users. When the number of attached disks was increased to 112, the system was able to support more than 11,000 heavy users.

Adding in Fault Tolerance for External Storage Systems

When implementing centralized storage solutions, you are often placing a large number of very important eggs into a single basket. SAN and NAS manufacturers understand this and have spent a lot of research and development dollars on building in Fault Tolerance into their offerings. There are many options available to the end user; some of the fault tolerance options are as follows:

  • RAID Configurations—. RAID levels 0+1 and 5 are most common. RAID level 6 offers the ability to lose two drives at a time and not lose data.

  • Triple Mirroring—. This enables you to snap off a mirror so that data becomes static for purposes of backup. Meanwhile the system still has mirrored drives for fault tolerance. This is most commonly used with databases.

  • Log Shipping—. Most SAN and NAS devices can copy log files in near real time to another SAN or NAS so that databases can be copied regularly and log files can be kept in synch remotely.

  • Geographic Mirroring—. SAN and NAS devices offer in-band and out-of-band options for mirroring data across wide distances. Whereas SCSI has a 25-foot limitation, Fiber Channel can locate a device up to 1,000km away.

  • Snapshotting—. By flagging disk blocks as being part of a particular version of a file and writing changes to that file on new blocks, a NAS or SAN device can take a snapshot of what the data looked like at a point in time. This enables a user to roll back a file to a previous version. It also enables you to roll back an entire system to a point in time almost instantly.

  • Clustering—. NAS devices that use heads to serve data offer dual heads so that if one fails, the other will continue to serve data from the disks.

  • Redundant power systems—. Any good SAN or NAS will offer multiple power supplies to protect against failure. Take advantage of the separate power supplies by attaching them to separate electrical circuits.

  • Redundant back-planes—. Many NAS and SAN devices will offer redundant back-planes to protect against hardware failure.

  • Hot Standby Drives—. By having unused drives available in the chassis, the device can replace a failed disk instantly with one that is already present and ready for use. Be sure to monitor the SAN or NAS device to see if a disk has failed and been replaced. It can be easy to miss because there would be no interruption to service.

RAID 5 Is Not Recommended

RAID 5 is not recommended for any application that performs write transactions more than about 15% of the time. This is due to the fact that each write transaction requires reading multiple disks and recalculating and writing of parity bits.

Combining Hardware Fault Tolerance with Windows Server 2003 Technologies

Windows Server 2003 possesses several technologies that improve system availability and recoverability with Direct Attached Storage. These same technologies work very well with NAS and SAN devices as well.

Distributed File System with NAS or SAN

Distributed File System has come a long way with Windows Server 2003. The ability to pre-stage data, the ability to mirror the DFS root, and improvements in control of replication topologies have all helped companies to adopt and take advantage of Distributed File System. Although things have gotten better, many administrators still find that there are too many limitations to the functionality of File Replication Services. FRS can’t deal properly with roaming profiles and it often replicates data unnecessarily. For example, most antivirus products flag files that they have recently scanned. FRS will misinterpret this as a modification to the file and will queue it up for replication to all of its partners.

Many administrators are finding that the native replication functions of NAS and SAN devices are a much better way to maintain replication for DFS replicas. By having Windows Server 2003 servers mount those replicas and present them as nodes in the DFS tree, they are able to achieve enhanced usability in their DFS structures. This adds an additional layer of redundancy because NAS or SAN mirrored replicas are automatically presented to users if a local copy were to fail.

Leveraging Logical Disk Manager

Most administrators are very familiar with the Logical Disk Manager that comes built into Windows. This is the interface that is regularly used to mount new disks, create partitions, and format drives. Administrators who are a bit more advanced have no doubt learned the benefits of being able to mount disks to folders rather than drive letters. Very large powerful servers that serve multiple tasks no longer have to worry about running out of drive letters.

Most SAN and NAS devices on the market offer management plug-ins that enable them to be managed through the existing Logical Disk Manager. Through hardware abstraction the remote disk is made to look like a local disk. This enables you to perform much of the management of the remote resources through an interface that they already know. This reduces the requirements to retrain administrative staff on a new technology.

Remote Storage Management

Remote Storage, the Windows version of Hierarchical Storage Management (HSM), enables you to increase disk space on a server without physically adding more hard disks. Remote Storage automatically moves data back and forth between high-cost, faster disk drives and low-cost, high-capacity storage media. This traditionally meant moving data from Direct Attached Storage to either a lower performance server or a very low cost but high capacity storage media such as a tape library or rewritable DVD or DC media. Remote Storage monitors the amount of space available on local NTFS volumes, and when the amount of free space dips below the low water line, it transfers eligible files from the primary storage to the secondary storage. Users can still see and access archived files but they are actually seeing links to the files stored on the lower cost media. This frees up storage on the file server without requiring the purchase and installation of additional hard disks. The Remote Storage service also works with the Removable Storage service to access any removable media used for remote storage.

In an environment that uses NAS or SAN for storage, this concept can be moved up one tier. In most cases, the NAS or SAN device will replace a number of servers. These are servers that would normally be retired. Rather than scrap all the servers, you can take a subset of the servers and salvage the disks from all the systems and distribute them across the remaining servers. Those servers would be relegated to storing low demand data only. This would enable you to still lower the overall number of servers in the enterprise that require management.

Optionally these “lower tier” servers could be low-end servers with IDE RAID devices attached to them. Given the current prices and sizes of IDE drives, a 1-TB server could be built for well less than $5,000. Because it wouldn’t be holding “primary” data, it would be acceptable to take the hit in performance and reliability that comes with IDE drives. This could allow for an environment where high demand and highly important data could live on redundant and high performance SAN/NAS devices. Data that hadn’t been accessed in some time could be moved to the IDE RAID systems where the data would still be retrieved quickly. Data that hadn’t been accessed in long periods of time would be moved to tape or rewritable media for long-term storage.

Integrating Backups with NAS and SAN

Many administrators have faced the challenge of dealing with narrow windows of opportunity for performing backups. Given a choice, every department would request that their systems only be backed up between 2 a.m. and 3 a.m. so that they wouldn’t suffer a performance hit while users might be accessing them. This is especially difficult in environments where systems are used globally. The challenge with narrow backup windows is that data cannot be spooled to tape fast enough to meet the time commitments. One of the easiest ways to improve this performance is to perform the remote backups to disk.

By using a NAS or SAN device to buffer the backups, you create an environment where the NAS or SAN becomes the backup device. Backups are written directly to the NAS/SAN at impressive speeds. Outside the backup windows, the data is backed up from the NAS/SAN to tape for long term storage. The added benefit of this model is that 90% of restore requests are for data from the previous evening. By having that data on disk on the SAN/NAS, the restore process is also much faster than it would have been from tape. SAN devices often support what is known as serverless backups. This means that the SAN itself controls the tape device and writes the data to it without the need for a traditional backup server.

SAN-attached Tape Devices

SAN-attached tape devices range from single drive single tape devices all the way up to office-sized backup systems with dozens of drives and literally thousand of available tapes. Robotic arms drive on tracks around the device scanning and moving tapes back and forth between drives, storage slots, and mail slots.

Leveraging Disk Quotas on NAS or SAN Devices

Disk quotas were introduced as a long overdue feature in Windows. For a long time, you had to worry about users storing inordinate amounts of data in their user directory and filling up the entire volume. Any other users whose home directories were on the same volume would suffer through no fault of their own. In an environment that makes user directory space available on a NAS or SAN will appreciate the safety that comes with being able to enforce quotas through Windows Server 2003. Imaging having to explain to your management that your brand new 4-TB SAN is already full because your users have filled it with “personal” data.

Using Encrypted File System to Protect Files on the SAN or NAS

One of the greatest concepts ever put into use is the OSI Model. In a nutshell, the concept is that if you have a series of layers that work together than anything that can be abstracted to look like one of those layers will automatically work with the rest of the layers. In the case of SAN or NAS, a driver is used to make the remote storage look just like regular attached storage. After that is done, the rest of the system will treat the abstracted drive like another locally attached drive. This means that technologies like EFS will automatically work with the remote disk. Users can continue to encrypt data on the SAN or NAS and administrators can continue to manage EFS keys and agents the way that they always have.

Best Practices for SAN and NAS

SAN and NAS manufacturers have provided a number of technologies that make it easier to integrate their products with specific software products. Because these products having been available for a number of years, best practices around these implementations have come about and can help you avoid common pitfalls with SAN and NAS usage.

Exchange with NAS/SAN

When implementing a NAS or SAN solution in a Microsoft Exchange environment, there are many different interpretations on the best way to implement the solution. Some of the recommended best practices are as follows:

  • Run multiple Host Bus Adapters in each Exchange server with each HBA connected to a different Fiber Channel switch. This will allow for failover if one of the Fiber Channel switches should fail.

  • Backups should be performed at the storage group level rather than the mailbox level. Mailbox-level backups are very processor-intensive for the Exchange server.

  • Separate logs files from databases onto different drive sets. This will improve overall throughput as well as improve recoverability in the case of a NAS/SAN failure.

  • Replicate databases hourly to another device for disaster recovery. Logs should be replicated every few minutes. This will limit potential mail loss to one log replication interval.

  • Always use integrated tools if they are available, such as Network Appliance’s SnapManager for Exchange 2000. They will greatly simplify management and recoverability of the product for which they were designed.

  • Always plan for space reservation on a volume. If the database will grow to 80GB and will have snapshots taken for recoverability, reserve 160GB of space on the device.

  • Avoid placing multiple Virtual Logical Disks or LUNs on the same volume. This could result in databases and log files being on the same volume. This would complicate system recovers if the volume were to fail.

SQL with NAS/SAN

A Microsoft SQL environment can leverage NAS and SAN technologies to improve storage management and storage solution implementation. Some of the recommended best practices for implementing a NAS or SAN device in a SQL Server environment are

  • Separate database from each other by placing them on separate VLDs or LUNs on different volumes to maximize read/write performance.

  • Place databases and log files in separate VLDs or LUNs on different volumes to not only improve read/write performance but to enhance recoverability.

  • Run multiple Host Bus Adapters in each SQL Server with each HBA connected to a different Fiber Channel switch. This will allow for failover if one of the Fiber Channel switches should fail.

  • Depending on the access pattern, a single hard disk can support roughly 150 I/O operations per second (IOPS). It is important to understand the I/O rate during peak load to estimate the number of disk drives needed to support the I/O load. For example, a RAID group size of eight disks (seven data and one parity) can at most support 7 * 150 IOPS = 1,050 IOPS

  • Always take into account database growth to avoid having to constantly expand the volume. When possible, operate with extra free volume space. Maintaining extra free volume space will decrease a possible volume fragmentation rate and will also prevent sudden “out of space” events. When a volume’s free space is very limited I/O performance could suffer.

When implementing a NAS solution, some of the technological solutions that can improve operational performance need to be taken into account. As an example, the following fictitious organization with the following requirements for a SQL Server can be modeled for a NAS configuration:

  • Initial database size of 80GB

  • Database growth rate is estimated to be 15% per month

  • Change rate of the current database is estimated to be 15% per month

  • Snapshot requirement is four Snapshots per day with a total of 12 Snapshots (three days)

  • Default RAID group size is 72GB * 8 devices

  • The administrator wants to expand the volume at most every six months

  • The growth and change rates are estimates, so extra volume space has been requested, and the customer realizes that a volume always has to operate with some free space to decrease the possible fragmentation rate, or I/O performance could suffer; therefore, an extra 20% free space per disk drive will be allocated as a free-space buffer

  • The average I/O rate is about 1.5MB per second and the peak rate is about 3MB per second

Based on the statistical information for this organization, the resulting analysis and projections can be as follows:

  • The database size after six months will be about 144GB.

  • About 1.6GB of the database will change every month after six months, which is equal to 0.12GB per four hours.

  • The minimum space requirement after six months will be (144GB * 2) + (0.12GB * 12) = 290GB.

  • A 72GB disk drive has 68GB usable file space, and because 20% is allocated as extra free space, only 55GB is usable per disk drive. Hence, six disk drives are needed for data and one disk drive for parity, for a total of seven disk drives. However, it is important for performance reasons to always configure complete RAID groups; therefore, the volume will be created with eight disk drives. If the rate of growth is consistent over time the volume will have to be expanded with another RAID group after six or seven months.

  • The estimated peak load was 3MB per second, which is equal to about 770 IOPS. Because seven data disk drives can support 1,050 IOPS, a RAID group size of eight will be sufficient to support both space requirements and I/O load requirements.

File Servers with NAS/SAN

File servers are based typically on small individual files as opposed to large blocks of structured information as in a SQL or Exchange environment. Best practices found in creating a NAS or SAN environment for file servers are as follows:

  • Allocate only the space needed for a particular server. If users have access to infinite space, they’ll find a way to fill it.

  • Use third-party file comparison tools to determine when duplicate information is being stored on a separate file server.

  • Snapshot the data regularly to enable users to recover their own files across multiple versions of the file.

  • Consolidate file servers to a location where they can have very high-speed access to the NAS device. If possible, plug in the servers to the same blade on the network device to avoid sending traffic across the backplane.

  • Enforce quotas on user data areas to prevent users from consuming all available space.

  • Critical data should be mirrored to another location to ensure rapid recoverability.

  • Data should be backed up to long-term storage media such as tape to maintain sufficient data history

  • Run multiple Host Bus Adapters in each file server with each HBA connected to a different Fiber Channel switch. This will allow for failover if one of the Fiber Channel switches should fail.

  • Users should be trained on any tools that will be available to them for tasks such as recovering deleted data or recovering an older version of an existing folder.

Backup Systems

When configuring NAS or SAN devices as part of a backup, fault tolerance, disaster recovery, or business continuity structure, the configurations and optimization are much different than for file servers or database servers because the information is primarily stored for redundancy purposes. Some of the best practices for building a backup system NAS or SAN environment are as follows:

  • Use configurations that maximize data throughput. Because data will only stay on the NAS or SAN for short periods of time and will be immediately spooled to tape, there is not a large need for redundancy at the SAN or NAS. RAID 0 is a good choice for this.

  • The backup system should be connected to the high demand hosts via a dedicated network. This allows the backup to occur without interference from user-facing traffic.

  • If using a SAN for the backups, consider using a SAN-attached tape device to spool the data for long term storage.

  • For databases that cannot be acquiesced for purposes of backup, use a triple mirror configuration. This allows the third mirror to be broken and brought to a static state. This mirror can then be backed up to tape. After the backup occurs, the third mirror is reattached and synchronized with the system. During the backup process, the system is still protected by the second mirror.

  • Run multiple Host Bus Adapters in each file server with each HBA connected to a different Fiber Channel switch. This will allow for failover if one of the Fiber Channel switches should fail.

Active Directory Integration

Building a NAS or SAN environment in a Windows environment can leverage Active Directory and be integrated into the Directory for authentication, security, and management purposes. Some of the best practices for integrating Active Directory for NAS and SAN devices are as follows:

  • NAS devices authenticate against the domain by which they are accessed. To ensure that this still works when a domain is upgraded to Active Directory, be sure to point the NAS at a DNS that holds the appropriate service records (SRV).

  • Windows Server 2003 domains default to using only NTLM v2 or Kerberos and require the digital signing of communications. Ensure that any NAS devices present in the enterprise can support these requirements.

  • NAS devices ACL resources based on SIDs. If a domain upgrade will involve migrating objects from another domain, be sure to either maintain the SID history or re-ACL the NAS device.

Terminal Servers

Although a service that runs on Windows 2003, Windows Terminal Services operates similar to multiple workstations as opposed to a limited number of database servers. Some of the best practices in optimizing Windows Terminal Services in a NAS or SAN environment are as follows:

  • Store user profiles on a NAS or SAN device to ensure fast access to the profile from any Terminal Server in the farm.

  • Store applications on the NAS or SAN device. This allows the new Terminal Servers to be pointed to the existing applications for rapid deployment.

  • Store user data on a NAS or SAN device. Due to the high IO requirements of multiple terminal servers potentially needing to reach a given user’s data, only a NAS or SAN device spanning a large number of drives will be capable of providing adequate performance.

  • Build VLDs or LUNs based not only upon space requirements but on IO requirements as well. Depending on the access pattern, a single hard disk can support roughly 150 I/O operations per second (IOPS). It is important to understand the I/O rate during peak load to estimate the number of disk drives needed to support the I/O load. For example, a RAID group size of eight disks (seven data and one parity) can at most support 7 * 150 IOPS = 1,050 IOPS.

  • Store user data and applications on separate VLDs or LUNs. This enables you to easily lock down application files across the entire disk without affecting user-by-user access to their unique data or profiles.

Booting from NAS/SAN

Some network solutions require certain network functions to load upon server boot-up whereas other applications are independent of server boot functions. When configuring a NAS or SAN device for system boot-up, some of the best practices for configuration are as follows:

  • If the Host Bus Adapter supports it, booting a server directly from the SAN without any local hard drives is an easy way to replace failed servers.

  • Terminal access type computers can be built by booting via BOOTP and receiving their operating system image from a NAS device. This greatly simplifies management of specialized stations. Using a NAS device for this will greatly improve scalability due to its ability to support higher IO rates.

  • Replicate live systems to a test lab SAN by mirroring the data from servers that boot via the SAN. This allows isolated testing with data that is identical to the live system.

Recovering from a System Failure

Having data stored on a SAN or NAS device can greatly reduce recovery time for failed systems. By having replicas of data on other devices, data can be reattached to a system in case of a system failure. In the situation where a server was booting its operating system off a SAN device, the Host Bus Adapter can simply be placed in similar hardware and the system will come right back up.

In the case of system failure due to data corruption, a SAN or NAS will enable you to roll the system back to a point in time before the corruption occurred based on the number of snapshots available.

If a system fails due to actual disk failure on the VLD or LUN that it was attached to, a system can be brought back up quite quickly if the data on that VLD or LUN was mirrored to another NAS or SAN device. By simply remapping the VLD or LUN the system can be made functional again in a very rapid manner.

Mirroring and snapshotting are the keys to using NAS or SAN to make a system more easily recoverable than it would have been with DAS. Simple planning like keeping servers standard in their brand and configurations, always using the same brand and model of HBA or NIC, and keeping Fiber Channel switches standard across the enterprise enables you to have an environment where failed components can be replaced from a standard set of spare parts and the system brought up immediately.

Even in cases where entire sites fail due to storm, earthquake, or fire, systems can be brought back up quite rapidly in a failover location if the data was mirrored. Using SAN or NAS to perform this mirroring ensures a much greater level of confidence that the data was replicated correctly, even for databases that remain hot on a 24 by 7 schedule. Recoverability from both a site and a system level make SAN or NAS integrated environments much more resilient and reliable.

Leveraging NAS and SAN Solutions for Server Consolidation

One of the most popular uses for NAS and SAN devices is to reduce the number of servers in the environment by consolidating servers and server functions. Rather than have dozens and dozens of file servers with locally attached storage, a smaller number of servers with NAS- or SAN-attached storage can serve the same purpose.

Consolidating the Number of Exchange Servers

Exchange servers were traditionally sized based not only on performance potential but also on the time needed to recover a system. Administrators knew that if they had a four-hour Service Level Agreement for system recovery they could count on using half that time to recover data from tape and half that time to perform the recovery tasks. This meant that they could only have as much local storage as they could recover in two hours. So if a backup/restore system could restore 16GB of data in two hours and each user was allowed 100MB of storage, the maximum number of users on the system would be 160. For a company of 1,600 users, this would mean 10 Exchange servers would be required to support the four-hour SLA.

By placing the mailbox stores onto a NAS or SAN device that can be mirrored and snapshotted, the recoverability time for a 16GB database would drop to mere minutes. Now the bottleneck would become the performance of the server itself and possibly the IO rate of the NAS or SAN. Odds are that the systems that had been purchased for the ability to support 160 users would be dual processor systems with a Gig or two of memory. By reducing the server count to two and fully populating those two systems with memory taken from the retired systems, the two systems with NAS- or SAN-based mailboxes could easily support the 800 users each and still meet the four-hour recovery time required by the SLA.

This would result in the reduction of eight Exchange servers which would free up OS licenses and hardware as well as reduce the effort required to manage the data center.

Consolidating the Number of File Servers

Most companies grow their file server environment in an organic manner. This is to say that as different groups in the company come up with needs for data storage, additional file servers are brought up. This can be an expensive process in that each of these file servers requires not only hardware but also an operating system, antivirus software, management software, space in the data center, facilities like cooling and electricity, and many other expenses. Users take storage for granted because drive prices are relatively inexpensive and they believe that disks can just be added and added forever so that they can fill it.

This results in a difficult to manage environment for the IT professional. When system patches become available it can be quite an event to ensure that potentially hundreds of file servers are up to date. Companies that enforce a life cycle for their servers find themselves replacing more and more servers each year.

NAS and SAN devices offer a solution to this problem. By consolidating the storage of data and presenting this data to the user community through a smaller number of high-performance file server clusters, you can ensure that users have reliable access to the files they need. At the same time, you can greatly reduce the management overhead of the vast number of servers and reduce costs across the board.

In environments where file servers are brought up simply to provide better access for local users, you can leverage technologies like DFS and mirroring to make data available to each location. Through DFS and site locality, users can ensure that they are reaching the closest copy of a file without any sort of user interaction. By leveraging a SAN or NAS at each major location administrators can still reduce the number of file servers at each location and maintain the same if not better level of support.

Summary

This chapter has introduced us to the concepts of Network Attached Storage and Storage Area Networks as options to improve performance and manageability over traditional Direct Attached Storage. We’ve seen how SAN and NAS can be used to manage data more effectively through the reduction of servers.

Applications like Exchange can be made to support much larger numbers of users through the leveraging of large numbers of disks. Performance scales nearly 1:1 as additional disks are added. Plus adding additional disks allows an application to support more IO operations per second, which is critical to database applications.

You’ve seen how advanced technologies like snapshotting enable you to back up data regularly on the device itself so that users can recover their own data without having to involve administrators.

NAS implementations work with existing Ethernet infrastructures and impart additional loads on them that must be planned for. You’ve seen that a strong network is the key to good NAS performance.

This chapter discussed some common Microsoft applications that work well with both NAS and SAN storage. You’ve learned that SAN provides block-level access to the disk, whereas NAS provides file-level access. Some applications require SAN but most can work with NAS.

SAN and NAS offer you greater performance and enables you to perform geographic mirroring that was previously impossible for the application itself. As Ethernet and Fiber Channel technologies continue to improve and as prices continue to fall you will find NAS and SAN becoming more and more common in the IT world.

..................Content has been hidden....................

You can't read the all page of ebook, please click here login for view all page.
Reset