Importance of disk offset

SAN Guy at CLARiiON BLOGS has a good article that explains disk alignment. A very important subject regarding performance, this issue has recently come up at my data center: a EHR application is experiencing performance issues due to fragmentation (of course), and improper disk alignment on the LUN. Our DBA just ran a SQL 2005 best practices on the database that resides on this particular LUN, and it spit out the recomendation of a 64k offset. SAN Guy taught me a few things on this important subject.

Advertisements

An overview of MAID

ubuntu-serversSGI, a technical computing manufacturer, produces a line of storage systems that incorporate green technologies. The COPAN line of storage products use the MAID storage methodology that presents the ability of powering down unused disks in the array while not in use, conserving power as a result. This is managed via internal “Power Managed RAID software” (Sgi, 2010). MAID, the acronym for Massive Array of Idle Disks, is suited for the environment that consists of long-term storage that requires write-once, read-occasionally data (WORO). MAID can be used in disk (virtual) libraries, also known as EDL (Enterprise Disk Library). This is energy efficient storage. SGI’s COPAN solution could be some tough competition against the Centera line from EMC. As of the time of this writing, I have not yet seen MAID technology on the archive systems from EMC.

Massive Array of Idle Disks (MAID) consists of a large disk group, consisting of hundreds or thousands of disks configured into RAID groups. Through internal power management code, only drives that are needed are activated at a time. This reduces drive wear and power consumption.

Reference:

Sgi. MAID vs. Enterprise MAID. (2010). Retreived November 2, 2010 from http://www.sgi.com/products/storage/maid/what.html

Thesis research: SSD vs. FC drive benchark tests – part I

typewriterI am writing my graduate thesis on the subject of Solid State Drive (SSD). By the way, D stands for DRIVE, not DISK, as SSD does not use disk. Now, with that out of the way…

I have been benchmarking a new SSD array that I have added to my companies SAN: an EMC CLARiiON Cx4-480 system running on 4Gb/s fiber. It will be 8Gb/s soon, but we are waiting on the NAS code (an EMC NS-40G) to catch up so it will support 8Gb/s: the firmware on the NAS only supports up to 4G currently. The SAN is held together with two Brocade 4900 FC switches.

About the disks that I will be testing and comparing:

Disks used: (5) EMC 70GB SSD and (5) 300GB FC disks.

SSD:

66.639 raw capacity -FC SSD – Manufacturer: STEC – Model: ZIV2A074 CLAR72 – Serial: STM0000E9CFD – 4Gbps

FC:

268.403 raw capacity – FC – Manufacturer: SEAGATE – Model: STE30065 CLAR300 – Serial: 3SJ09XWW – 4Gbps

  • Created RAID5 (RAID group 100) on five SSD model ZIV2A074, 66.639GB each.
  • Creating RAID5 (RG 101) on five 300GB FC disks: Seagate 15K.7 ST3300657FC
  • LUN 104 is assigned drive letter W: (disk 3) (RG100) and named “SSD
  • LUN 108 is assigned drive letter X: (disk 4) (RG101) and named “FC

The test server was installed and set up with one dual-port 4Gb/s HBA. Windows Server OS Standard with 1GB RAM.

SAN Management: EMC Navisphere 6.28.21.

Network: 4Gb/s fiber. 2-Brocade 4900B FC switches. Host HBA: Emulex LightPulse FC2243.

Host connection via EMC PowerPath v5.2

Test I/O is generated by Microsoft SQLIOSim; I/O generation utility to similate I/O patterns found within versions of Microsoft SQL server. The versions simulated are Microsoft SQL Server 2005, SQL Server 2000, and Server 7.0. Brent Ozar, a SQL expert, has a good video on using SQLIO on his web site at brentozar.com. I have learned some things from him and am using the tips on SQLIO for my benchmarking.

The monitoring will be done with EMC Navisphere Analyzer and SUN StorageTek Workload Analysis Tool (SWAT).

Here is a preliminary test on SSD vs. FC data rates using SQLIOSim to generate I/O and SWAT to record the results:

SSD:

SSD performance survey data rate

FC:

FC performance survey data rate

So far there is not much of a difference. The Fiber Channel drives are keeping up with the SSD. Of course, this is a preliminary test and other tests at this time are giving similar results. I continue to plan my testing methodology.

Measuring disk performance

Here is a video on SearchDataCenter that I found that explains that noise can effect disk vibration:

EMC CLARiiON Cx4 integrated thin provisioning

virtualizationWe just upgraded to the new EMC CLARiiON Cx4 from the Cx3 over the weekend. The technician who arrived had done a good job: in about seven hours, the SAN was fully functional and we were back in business. We took this downtime to do some patching on host systems. The Cx4 has 8Gb fiber ports as compared to the Cx3 4Gb ports; however we could not use the 8Gb modules as the NAS, an EMC NS40g, contains code that is not 100% compatible with the Cx4 fiber module code. We will run it all on the 4Gb/s fiber modules for now until the NAS code catches up, and then the 8Gb modules will be installed. This is not a problem, as the 4Gb modules are still very fast and provide sufficient throughput for our electronic health records system (EHR).

One useful improvement about the Cx4 is the addition of thin pools – one can create RAID groups that use minimal storage and can expand dynamically when the need for more storage arises. See the white paper from EMC on virtual (thin) provisioning here.

LUN migration on EMC CLARiiON

dataminingSince we are upgrading our EMC CLARiiON CX3 to a CX4 this month, EMC has suggested that I get some LUN’s off of the vault (I didn’t put them there) and migrate them to another series of LUN’s. The vault is an array of disks used for storing the system cache – I never liked having other LUN’s on it. So, I am migrating LUN’s off of it, and this is taking some time due to the amount of data. This can be used to move the LUN to a bigger LUN or just to move it for performance reasons. To migrate a LUN is simple:

First, it is required to have some spare LUN’s on a RAID group(s) that are the same size or greater as the source LUN. Right-click on the LUN to migrate (source) and select migrate. A selection of prospective targets LUNs will be displayed. Select the target and the migrate. Very simple. The target will assume the identity (LUN number) of the source LUN when the migration is completed, and this is entirely transparent to the host that is using the LUN and there will be no interruption in LUN access on the host. The migration could take some time depending on the amount of data, of course.

If migrating to a larger LUN, it would be good if the disk were dynamic on the host, if possible. Then, one could just rescan the disks (on Windows Server) and the new larger space would be enabled. If the partition on the host is basic, one would have to use Microsoft Server’s DiskPart – an effective but risky utility.

A review of IP traffic on its effect within SAN

This brings iSCSI to mind. According to Network Magazine, iSCSI makes SAN data accessible to everyone, as this enables a node to access a storage target and set it up for access so that it appears to the OS as a local drive. The advantage of an IP transport in comparison to Fiber Channel is that it is possible to scale a SAN over a longer distance. “TCP/IP can run over any physical medium, so it offers a choice of topologies… and IP packets can travel across alternative LAN topologies just as easily ” (Hall, 2004). This can happen only if the network hardware is sufficient to support the additional Gigabit traffic that iSCSI causes, as iSCSI produces large amounts of traffic bursts. As TCP contains flow control that is very sensitive to loss and delay of packets which could result in negative performance impact, it is essential to implement the fastest switching and routing protocols such as Gigabit switches and router RIP. Currently, there are new 10Gbs Ethernet switches on the market that can more than handle the increase traffic bursts of iSCSI such as Intransa’s 10G Ethernet storage appliance that is designed for the demands of video on-demand, IPTV, digital imaging, video surveillance, and data mining” (Messmer, 2009) which demand those higher speeds. Let’s not forget the backup of data within the IP SAN. Data backup is the cause of a large amount of SAN traffic both on the wire and on the storage targets. Using a LAN-free approach, such as having a backup server within the SAN would eliminate much (IP) backup traffic so that the backup data is not flowing between the client node and the backup server, but only between the backup server and the storage target. As stated in Arkeia’s article on its new backup software, Arkeia Network Backup version 6.0, the “…Server for SAN option allows LAN-free backup for SAN environments, thus easing network congestion. It allows direct SAN connectivity of servers, storage arrays and tape libraries” (Connor, 2006).

References:

iSCSI: Storage Networking Sans the SAN? — Farewell to Fibre Channel? iSCSI lets storage traffic run over a regular TCP/IP network.(Internet SCSI)(Storage Area Networks)(Transmission Control Protocol/Internet Protocol).Eric A. Hall. Network Magazine (May 1, 2004): p62. (3170 words) 

Messmer, E. (2009). Intransa debuts 10G Ethernet storage appliance; Latest StorStac appliance offers up to 180TB of storage capacity. Network World. Academic OneFile. Gale. BCR Regis University. 
http://find.galegroup.com.dml.regis.edu/itx/start.do?prodId=AONE

Connor, D. (2009). Arkeia rolls out new backup software. Network World. Academic OneFile. Gale. BCR Regis University.
http://find.galegroup.com.dml.regis.edu/itx/start.do?prodId=AONE