Minding the Store
        As SAN and NAS storage drops in price, more companies are implementing them. What's your role in managing these devices?
        
        
			- By Bill Heldman
- October 01, 2002
        Users are a disk-hungry bunch, aren't they? Consider a few scenarios:
      
        - Users need a place to back up their critical data—preferably 
          on some server disk that you currently back up with your formal server 
          backup program.
- The e-mail system is so heavily utilized that your e-mail database 
          is growing by leaps and bounds. You just can't get your boss (and others) 
          to give up their multi-megabyte inboxes!
- The Home and Shared directories on your file servers are gargantuan.
- You've introduced a geographic information system and the spatial 
          images take up massive quantities of disk space.
- Your Oracle and SQL Server databases are colossal.
Where are you going to put all this data? You could keep buying servers 
        with a few hard drives in them, but at some point your servers begins 
        to procreate like rabbits and your farm overrunneth. The biggest question 
        you need to ask yourself in this situation: "Am I maximizing the usage 
        of all this disk space?" Some studies that show that only about 30 to 
        50 percent of disk space is utilized on all these servers. And that unused 
        disk is expensive!
      Also, even though RAID array controller cards have brought a tremendous 
        measure of fault tolerance to servers, there are two problems associated 
        with RAID systems:
      
        - If more than one drive in a RAID 5 array fails, you're toast—you've 
          got to rebuild the array and restore from tape.
- Admins often don't get it when it comes to RAID. They don't have hot 
          or cold spares available and at the ready or they incorrectly rebuild 
          the mirror or stripe set with parity.
- Additionally, even though most Admins routinely back up their systems, 
          they don't often validate the backups by performing routine restoration 
          tests to make sure the data on the tape can actually be restored. (I'm 
          raising my hand as I write this paragraph—been there, done that, 
          bought the soundtrack.)
On top of all that, a $1,700 hardware RAID array controller card, while 
        better than nothing at all, is still a single point of failure (SPOF). 
        While your users' data is certainly safer when on a drive hooked to a 
        RAID card, it still isn't completely safe.
      So, is there a solution that can help offset some of these issues? Is 
        there a way that we can provide ample, inexpensive disk space for our 
        users, a place where all of our clients—Linux, Unix, Windows, even 
        mainframe—can go to get their data?
      Storage Area Network (SAN)
        A SAN is actually a collection of equipment that makes possible 
        the consolidation of disk storage. Following are some of the critical 
        elements in a SAN:
      
        - Storage array—This is a device that has a lot of disks 
          in it. Typically these devices are rated in the number of supported 
          Terabytes (TB). Most high quality (expensive) storage arrays provide 
          some sort of fault-tolerance software in them that continuously monitors 
          the disks, proactively moves data off of bad sectors and even "phones 
          home" if a disk is going to go bad so that the vendor's service department 
          can bring a new one out.
- Network Attached Storage (NAS)—A NAS device (discussed 
          below) is essentially a server with a lot of disks in it using a highly 
          scaled down OS. 
- Fiber-channel—The fibre-channel standard is used to connect 
          devices to the storage array. Typically servers that run applications 
          will connect to the storage array to access their data. Most connections 
          are done through some form fibre-channel.
- Host Bus Adapters (HBAs)—Think of these as fibre-channel 
          network cards, HBAs go into your servers so that you can connect them 
          to the storage array.
- Fibre-channel switches—There are two categories of fibre-channel 
          (FC) switch: normal and "director-class". The director-class switch, 
          invented and patented by McData (www.mcdata.com) 
          allows for more intelligent monitoring and traffic management of the 
          FC switch fabric.
- Software—The value-add and the most important feature 
          of a high-quality storage array installation is the software that's 
          used to run, manage and report on the array. Big companies like EMC 
          really shine in this area, but their software is pricey. Other big players 
          such as Legato and Veritas have gotten into the storage array software 
          game. There are a variety of needs: reporting, representing NetBIOS-type 
          share information to Windows clients, monitoring disk health, managing 
          disk space, real-time copying to a remote device backup software, etc.
The combination of all of the above constitutes what most people call 
        a SAN. The idea is that you have users who connect to the SAN via software 
        running on the storage device that allows them to do so, even though they're 
        not fibre-channel-attached and then to access data that's managed by servers 
        that are fibre-channel-attached. This is a SAN in its idealized form—there 
        are other variations such as servers that connect to a FC switch, which, 
        in turn, talks to the storage array; NAS devices that connect to the storage 
        array and so forth. The reason we call a SAN what we do is that with a 
        little bit of architectural expertise you can set up what are called "islands 
        of storage" that provide highly available and fault-tolerant data space 
        and that can act as backups to one another.
      Network Attached Storage (NAS)
        There are simple solutions out there that allow you to snap on 
        a piece of disk space for your users without having to get into the whole 
        SAN-box (ha-ha). The solution is called Network Attached Storage (NAS), 
        and a wide variety of vendors today offer NAS boxes. NAS solutions typically 
        utilize a mini-OS with enough intelligence to get the data to users (it 
        ain't Windows folks) although big players such as EMC provide very robust 
        NAS solutions. Admins use a set of tools (which vary from vendor to vendor) 
        to manage the NAS. 
      NAS devices support a wide variety of file systems and generally support 
        all "open systems" clients. When we say "open systems" we mean Linux, 
        all flavors of Unix, Windows, Netware and Macintosh along with the various 
        file systems that these OSes require. We do not mean mainframe clients. 
        If you want a storage device that can host both mainframe and open systems 
        clients you need a larger, much more expensive class of device such as 
        one that IBM or EMC could provide.
      Some NAS devices use the Network Data Management Protocol (NDMP) to allow 
        administrators to back up the device. NAS devices do not support niceties 
        such as e-mail, file management or user authentication services.
      We're talking basic, inexpensive, reliable easy-to-manage storage. Basic 
        NAS boxes are extremely plug-and-play. (See www.networkappliance.com 
        as an example of a company heavily involved in this type of NAS device).
      Admins interested in getting started with SAN/NAS technology and who 
        have minor data consolidation issues to worry about should investigate 
        NAS architectures.
      Storage Arrays
        A larger storage array differentiates from its NAS cousin in the 
        software that it runs. If you purchase an EMC Symmetrix or Clariion or 
        a Hitachi Data Systems (HDS) 9900, for example, you are paying largely 
        in part for the software that makes that bad boy so highly fault-tolerant. 
        There are expensive up-front software costs, as well as equally expensive 
        ongoing software maintenance fees. However, if any component of the array 
        goes south, it is the vendor that makes sure that you continue operation. 
        EMC and IBM guarantee five-nines uptime (99.999%) while HDS (and its certified 
        reseller—Sun Microsystems) guarantee 100% uptime. How do they do 
        it? Redundancy, redundancy, redundancy. Plus good software. If there appears 
        to be a problem with a section of disk, the operating system will proactively 
        move the data to a good sector and mark the bad place as bad. It will 
        continuously monitor the array. If the bad sector gets worse—that 
        is, it spreads—the OS will actually phone home to tell the vendor 
        that it has a problem and to please bring out a new disk.
      One night, I was a server administrator on call in a shop that used EMC 
        Symmetrix. The night operator called me at about 2:00 a.m. to tell me 
        that there was an EMC technician at the door—our Symm had a bad disk 
        and the guy wanted to replace it. The operator asked me what I thought 
        she should do. "Let the guy in!" I said. We had no clue the disk was failing—all 
        of the data was accessible and just fine.
      Fibre-Channel Switches
        A storage array device typically comes with a certain amount of 
        fibre-channel ports (fibre-channel in the mainframe world is provided 
        through the FICON standard). If the number of hosts you plan to connect 
        to the array exceeds the amount of ports in your array, then you've got 
        to augment with a FC switch.
      A fibre-channel switch is used to connect a lot of devices to the SAN, 
        or to connect devices a long way away from the SAN to the SAN, or both. 
        FC switches aren't any different than IDF/MDF switches in the sense that 
        you buy them by the number and by the kind of ports they have in them. 
        There is one noticeable difference however: For what you pay for one FC 
        switch you could probably equip an entire IDF or MDF. FC switches are 
        priced in the tens of thousands of dollars.
      Host Bus Adapters
        HBAs are also quite expensive. Plan on spending in excess of $1,000 
        for a typical HBA. Note that you won't need HBAs for a little NAS device 
        that you purchase and hang on the network. You get into HBAs when you 
        have a sophisticated storage array that comes with some FC ports, but 
        not enough for you to hang all your hosts onto the SAN.
      Generally your mission-critical data will live on the storage array and 
        your servers that talk to the data will need to be attached to the SAN 
        via FC. We call such servers "channel-attached." As a rule of thumb, you 
        should plan on putting two HBAs in any such mission-critical server.
      Fiber-optic cable is generally utilized as the connection between the 
        SAN and the HBA, though copper is also supported in fibre-channel technology.
      Note that tape back-up devices such as tape libraries are completely 
        capable of connecting to a SAN via HBAs and are thus equipped to provide 
        tape backup of the data on the SAN.
      
      
Backin Up My Words
        SAN technology is complicated, high-tech and extremely expensive. 
        But it's also the best way to go if you're considering protecting your 
        user's mission critical data. Another reason to consider SAN/NAS solutions 
        revolves around the idea behind data consolidation. You want to get all 
        that file, home-directory and shared-directory off of your servers and 
        put it onto a central location. These are the main reasons you'd consider 
        a SAN/NAS solution.
      There are major players that support both mainframe and open systems 
        on the same storage array. If you don't have a requirement for mainframe 
        storage, you may be able to save significant dollars with a downscaled 
        storage array device (such as the EMC Clariion or offerings from Compaq, 
        HP and others).
      When you build a SAN, you're building a little storage network that is 
        comprised of several different hardware components that are managed with 
        software specially designed for the SAN.
      If you're considering designing a more complex SAN/NAS solution that 
        a simple 1U NAS box, you'd be wise to see experienced counsel before proceeding.