Thursday, November 3, 2011

Cloud computing server architecture: Designing for cloud efficiency

A server is one of those industry terms whose definition is broadly understood yet at the same time ambiguous. Yes, "server" means a computing platform on which software is hosted and from which client access is provided. However, the generalizations end there. Not only are there many different vendors that manufacture servers, but there are also a variety of server architectures, each with its own requirements. A mail server, a content server, a Web server and a transaction server might all need a different mixture of compute, network and storage resources. The question for many providers is: What does a cloud computing server need?
The answer will depend on the target market for the cloud service and how that market is reflected in the applications users will run. Servers provide four things: compute power from microprocessor chips, memory for application execution, I/O access for information storage and retrieval, and network access for connecting to other resources. Any given application will likely consume each of these resources to varying degrees, meaning applications can be classified by their resource needs. That classification can be combined with cloud business plans to yield a model for an optimum cloud computing server architecture.
For a starting point in cloud computing server architectures, it's useful to consider the Facebook Open Compute project's framework. Facebook's social networking service is a fairly typical large-scale Web/cloud application, and so its specific capabilities are a guide for similar applications. We'll also discuss how these capabilities would change for other cloud applications.

Cloud computing servers needs may not align with Facebook Open Compute
The Open Compute baseline is a two-socket design that allows up to 12 cores per socket in the Version 2.x designs. Memory capacity depends on the dual inline memory modules (DIMMs) used, but up to 256 GB is practical. The design uses a taller tower for blades to allow for better cooling with large lower-powered fans. Standard serial advanced technology attachment (SATA) interfaces are provided for storage and Gigabit Ethernet is used for the network interface. Facebook and the Open Compute project claim a 24% cost of ownership advantage over traditional blade servers. Backup power is provided by 48-volt battery systems, familiar for those who have been building to the telco Network Equipment Building System (NEBS) standard.

The Open Compute reference has a high CPU density, which is why a higher tower and good fans are important. However, many cloud applications will not benefit from this high of a CPU density for several reasons:
  • Some cloud providers may not want to concentrate too many users, applications or virtual machines onto a single cloud computing server for reliability reasons.
  • The applications running on a cloud computing server may be constrained by the available memory or by disk access, and the full potential of the CPUs might not be realized.
  • The applications might be constrained by network performance and similarly be unable to fully utilize the CPUs/cores that could be installed.
If any of these constraints apply, then it may be unnecessary to consider the higher cooling potential of the Open Compute design, and shorter towers may be easier to install to support a higher overall density of cloud computing servers.

How storage I/O affects cloud computing server needs
The next consideration for cloud computing server architecture is storage. Web applications typically don’t require a lot of storage and don't typically make large numbers of storage I/O accesses per second. That's important because applications that are waiting on storage I/O are holding memory capacity while they wait.

Consider using larger memory configurations for cloud applications that are more likely to use storage I/O frequently to avoid having to page the application in and out of memory. Also, it may be difficult to justify the maximum number of CPUs/cores for applications that do frequent storage I/O, as CPU usage is normally minimal when an application is waiting for I/O to complete.
A specific storage issue cloud operators may have with the Open Compute is the storage interface. Web applications are not heavy users of disk I/O, and SATA is best suited for dedicated local server access rather than storage pool access.
Additionally, it is likely that a Fibre Channel interface would be preferable to SATA for applications that demand more data storage than typical Web servers-- including many of the Platform as a Service (PaaS) offerings that will be tightly coupled with enterprise IT in hybrid clouds. Software as a Service (SaaS) providers must examine the storage usage of their applications to determine whether more sophisticated storage interfaces are justified.
Cloud computing server guidelines to consider
Here are some summary observations for cloud providers looking for quick guidance on cloud computing server architecture:
  • You will need more sophisticated storage interfaces and more installed memory, but likely fewer CPUs/cores for applications that do considerable storage I/O. This means that business intelligence (BI), report generation and other applications that routinely examine many data records based on a single user request will deviate from the Open Compute model. Cloud providers may also need more memory in these applications to limit application paging overhead.
  • Cloud providers will need more CPUs/cores and memory for applications that use little storage-- particularly simple Web applications -- because only memory and CPU cores will limit the number of users that can be served in these applications.
  • Pricing models that prevail for Infrastructure as a Service (IaaS) offerings tend to discourage applications with high levels of storage, so most IaaS services can likely be hosted on Open Compute model servers with high efficiency.
  • PaaS services are the most difficult to map to optimum server configurations, due to potentially significant variations in how the servers will utilize memory, CPU and especially server resources.
  • For SaaS clouds, the specific nature of the application will determine which server resources are most used and which can be constrained without affecting performance.
The gold standard for server design is benchmarking. A typical mix of cloud applications running on a maximum-sized, high-performance configuration can be analyzed for resource utilization. The goal is to avoid having one resource type -- CPU capacity, for example-- become exhausted when other resources are still plentiful. This wastes resources and power, lowering your overall return on investment (ROI). By testing applications where possible and carefully monitoring resource utilization to make adjustments, cloud providers can sustain the best ROI on cloud computing servers and the lowest power consumption. That's key in meeting competitive price points while maximizing profits.

About the author: Tom Nolle is president of CIMI Corporation, a strategic consulting firm specializing in telecommunications and data communications since 1982. He is the publisher ofNetwatcher, a journal addressing advanced telecommunications strategy issues.

Friday, October 21, 2011

Four ways to reduce data center power Consumption

At a recent Association of Information Technology Professionals data center panel discussion, a seasoned group of IT admins discussed meeting customer power demands, with the consensus that demand is insatiable. Even as budgets seesaw from abundant to sparse, the demand curve never flattens, instead climbing skyward. The Jevons Paradox, the nineteenth century axiom of “the more we produce, the more we consume,” looms large in IT for the foreseeable future. Or as my colleagues say, “If you build it, they will fill it.”  The first panel warning was that virtualization is not a cure-all for reducing data center power consumption. Of course, there’s a clear advantage to high-density computing --cramming many virtual machines (VMs) into a single server -- but CPU demands for power and cooling still grow with each VM. In many cases, power and cooling costs shift from distributing power across lots of small servers to boosting power to cool red-hot VM-hosting systems.
 It’s not just about CPUs cranking out BTUs, which raises the next issue. The power needed for cooling, lighting, battery backup (UPS systems) and other environmental factors usually accounts for 35% or more of a data center’s total energy consumption, regardless of how efficient a building is built. Servers gobble watts, and keeping them happy is major overhead.
 Another important panel consideration was to shorten the return on investment. Returns have to show up fast -- within days or weeks. Nevermind three-to-five year returns -- in this economy, strained budgets can’t wait. All of the panelists insisted that IT managers have to show fast results before selling long-term solutions.
 So what are fast turnaround projects that deliver results quickly? The suggestions below are somewhat small, mostly single project efforts, or at least quick changes without high infrastructure costs. Combining them could create a synergy where the sum is greater than the parts, but doing so isn’t required to make efficiency gains. There were four main ideas presented for reducing data center power consumption, each of which can be implemented separately.

Switch to variable-speed fans
Recent research found that power consumption drops 30% for every 10% reduction in fan speed. As the name implies, these fans only consume power when needed, only running at the speed required, based on fairly sophisticated thermostatic measures. Since these fans slow down over long periods of time at low CPU utilization, they quickly decrease powerusage with each non-turning blade. And don't stop with servers; check cooling features of UPS devices and power supplies of various appliances on the same power grid, plus any other hot spots that may have a fan spinning for a while.

Raise the air temperature      
According to data center infrastructure suppliers, modern servers can perform well up to 77 degrees Fahrenheit. Yet many data centers have cooled servers down to mid-60 degrees F for years. By raising the ambient air temperature a few degrees, there can be an immediate drop in power usage by the cooling system with no server performance impact. There’s no overhead or investment needed, although close monitoring and a solid pilot program would be advisable to avoid unpleasant surprises. Granted, a slightly warmer server room can be a disconcerting change. For example, the dress code may have to be adjusted to allow for lighter clothes in warmer conditions.


             Use bigger, slower drives
 Of course, this should not be done for high-demand transactional processes, such as financial databases or critical 24-hour systems. But by delegating a percentage of mostly unused files to a lower tier of storage, big, low-energy demand drives can replace small, fast units. In turn, less drives burn less energy, creating less heat. This can be an expensive undertaking, but as most shops build out more storage every quarter, they should see it as a worthwhile investment.


Use hosted services                
Although moving IT workloads to a cloud or colocation provider externalizes the carbon consumption off to the host site, many will concede that big vendors are experts at squeezing the most out of a kilowatt. By using hosted services, you’ll be able to focus on delivering better value at a lower cost for your customers.

The risks of data center power consumption projects
IT organizations need to acknowledge the inherent risks in energy-efficiency projects. As one power company director put it, in a high-density, highly efficient environment, the data center can go thermal in seconds. Several recent high-profile outages started as a partial interruption, but cascaded to bringing down the entire facility. The catalyst – overheating that spread from rack to rack until all systems shut down for self protection.
The final warning: Spell out any risks before implementing changes to the data center and make sure to get executive support before pursuing any of these tactics of reducing data center power consumption.










By Mark Holt, Contributor, SearchDataCenter.com

Monday, October 17, 2011

Aberdeen Helps Out The U.S. Army With Its Stirling Storage Solutions

Storage is an important facet of any data center, but when the data center in question belongs to the U.S. Army, finding a secure, reliable storage solution is even more vital. With the U.S. Army in need of a large, fast storage solution for hosting user data and providing disk-based backups, Jeff Dupere, network administrator of the AATD (Aviation Applied Technology Directorate), turned to Aberdeen. “We also needed a complete turnkey solution to begin leveraging a virtualized environment,” Dupere says. “This included a storage-area network, hosts for running the virtual machines, and the Fibre Channel connectivity to tie the two together. Also, the entire solution needed to be VMware-certified.”
The Right Stuff
Aberdeen’s sales rep provided Dupere with the expertise of one of Aberdeen’s engineers, who helped to determine the best solution for the U.S. Army’s environment and needs. The solution included four Stirling 266 servers (2U SuperServers that support Intel’s Nehalem processors), an XDAS D-Series SAN (a 4U, 24-bay Fibre 8G/SAS 6G DAS), and a Stirling X888 (an 8U storage server). “The Stirling 266s provided us with the hosts for running our VMs, the DAS gave us the backend storage for hosting the data for those VMs, and the X888 met the storage requirements for our user data and disk backups,” Dupere says. The equipment offered all of the required features necessary for the job, such as full redundancy, high amounts of storage, and fast I/O.
Aberdeen’s solution was evaluated against several well-known competitors. “In the end, the solution Aberdeen offered was well-supported, exceeded our performance requirements, and was significantly less expensive than the other offerings,” Dupere says. “They are also one of the few companies I’ve seen to offer a full five-year warranty on virtually all of their equipment at no additional cost. Given our life cycle commonly exceeds three years on server hardware, this was a significant advantage.”
With its dual LGA 1366 sockets, the Stirling 266 is a 2U SuperServer 6026T-TF that’s capable of utilizing Intel’s high-end Nehalem processors. There are also 12 DDR3 sockets that can support up to 192GB of Registered ECC DDR3 or up to 48GB of unbuffered memory with ECC. Helpful features include support for KVM over LAN and virtual media over LAN through the integrated IPMI 2.0 and Realtek dedicated LAN. The U.S. Army will also enjoy peace of mind with built-in PC health monitoring. There are four onboard voltage monitors for the processors, as well as six fans with tachometer status monitoring. Environmental temperature monitors, including chassis and CPU  overheat alarms, provide further security. “[The Stirling 266] is what the engineer recommended, and based on previous experiences, I was happy to go with his suggestion,” Dupere says. 

Aberdeen also preloaded ESXi on the hosts of the server, which made the environment turnkey for the Army. “We more or less just had to install the equipment in the racks and turn it on,” Dupere says. “They also were able to provide all the extras we needed (additional drives, controllers, etc.) to make onsite repairs much quicker.”
The Stirling X888 storage server can provide up to 100TB of storage, and the U.S. Army took advantage of the X888’s dual SFF-8087 miniSAS connectors to connect with the Aberdeen XDAS D-Series SAN to deliver up to 196TB of storage. “We needed a ton of storage and very fast I/O,” Dupere says. “The Stirling X888 had this in spades—so much so that we’re in the process of buying another to supplement our environment.” For fast I/O when needed, the server features quad Gigabit Ethernet LAN and dual SAS expansion ports. And Aberdeen’s Teaming Technology offers transfer rates up to 430MBps with an added XDAS-iSCSI RAID enclosure. The Army also benefits from Intel’s QuickPath Interconnect Tech-nology, which can be found on Intel Xeon 5500 Series processors, providing them with 6.4GTps and 4.8GTps data transfer speeds.
The SAS RAID on the X888 includes dual IOP348 1,200MHz PCI-E controllers, and each controller features 512MB of DDR2-553 SD RAM with ECC protection. The controllers support RAID 0, 1, 5, 6, and 10. Overall, the Stirling X888 can provide up to 1,200MBps of internal transfer speeds. The RAID controllers can also support SATA disk drives and SAS hard drives at the same time. External SAS connectivity is also available via the SFF-8088 connector. The miniSAS backplane on the X888 provides the Army with 50 (48 front, 2 rear) hot-swap/hot spare SATA 3.0 drive bays.
The 8U chassis comes with a 1,760W 3+1 redundant hot-swap power supply. Fan maintenance is also a breeze with the eight hot-swappable 80mm cooling fans. Options for flexibility with the build include a DVD-R or CD-RW and software upgrades such as iSCSI, NAS, SAN, and backup software.
The XDAS D-Series is a 4U high-speed, high-availability SAN that features four 8GB Fibre Channel host ports on each of the controllers, which are ideal for the fast throughput and I/Ops needed by the U.S. Army. The XDAS also features full support for 6Gbps SAS drives to provide support for today’s fastest hard drives. The XDAS D-Series is also designed to be always available with fault-tolerant hardware modules, including redundant controllers, PSUs, and fans. As such, there’s no single point of failure for the U.S. Army to worry about.
Other helpful protection features include real-time problem detection and notifications through multiple monitoring capabilities on the XDAS. And intelligent firmware helps to protect against hardware failure to optimize performance and maintain data integrity. The XDAS D-Series uses a power supply that’s more than 80% efficient, and it can spin down the drive to save energy when the disks aren’t in use.
The U.S. Army benefits from the XDAS D-Series’ local replication abilities. XDAS storage provides the Army with both snapshot and volume copy/mirror capabilities. Full data copies let administrators quickly restore service if a RAID volume fails, and files can be restored or rolled back through the XDAS’ snapshot copies. Some upgrades the U.S. Army chose to add include increasing the bandwidth across the Fibre Channel fabric and adding larger drives.
Dupere says that the set of products the Army chose from Aberdeen has met or exceeded all of the group’s expectations. “Aberdeen is not a huge company like some of their competitors, as evidenced that few folks will recognize the name when you bring it up,” Dupere says. “That said, they offer solutions and equipment that are every bit on par with the commonly referenced brands in this genre. From sales, to the product itself, to the support you receive afterward, you will not be disappointed.”
A number of intelligent storage systems, including the Stirling 266 servers, Stirling X888 storage server, and the XDAS D-Series SAN, designed to offer high-capacity, high-performance solutions to organizations in need of fast and reliable storage.
“[Aberdeen offers] solutions and equipment that are every bit on par with the commonly referenced brands in this genre. From sales, to the product itself, to the support you receive afterward, you will not be disappointed,” says Jeff Dupere, network administrator of the U.S. Army’s AATD (Aviation Applied Technology Directorate).

Monday, September 12, 2011

The Petarack™ A Full Petabyte of Managed SAN Storage in a Single Rack


Powered by Nexenta’s incredible Nexentastor ZFS file system implementation, the Petarack can be configured under a single Namespace/Volume that doesn’t have to stop at its own petabyte of included storage. The Petarack’s host capacity can be scaled to over 5 petabytes of continuous, uninterrupted storage in a single Namespace/Volume.
The Petarack’s intuitive web based GUI management is employed for both configuring the array(s), and for the control of a wealth of enterprise level software features. These features include, but are not limited to, unlimited snapshots and clones, block level replication, file level replication and management of virtualized storage.
The Petarack is Cloud ready and Virtualization ready with features such as 128 bit addressing, huge > storage file sizes, nearly unlimited clones and support for multiple, concurrent server virtualization vendors. Virtualization enables thin provisioning, while improving performance via I/O pooling, which actually improves performance as more storage is added. I/O performance can be further enhanced through the strategic deployment of optional, fast SSD drives in hybrid storage pools.
“Make no mistake about it, our Petarack is a revolutionary, innovate design delivering massive amounts of self-contained SAN storage at a price point that quite frankly, we strongly believe will make the entire community sit up and take notice,” said Aberdeen president and CEO, Moshe Ovadya, who went on to say, “The Petarack takes the petabyte plus level of storage, and instantly makes it available to, and affordable by, small and medium size organizations that never believed it would be within their reach."
MODELPetarack
COMPONENTSDual 2U High Availability Head Units
8 x 4U Dual Expander JBODs
Industry Standard APC Rackmount Cabinet
LCD/KVM Control Panel
CAPACITY1080TB or Raw Storage
(over one Petabyte)
PROTOCOL SUPPORTNFS/CIFS/RSYNC/FTP/iSCSI
MAX RAID VOLUME
CAPACITY
ZFS supports unlimited single volume capacity
(limited only by amount of storage present)
SUPPORTED
RAID LEVELS
Stripe, Mirror, RAID-Z1 (Single parity), RAID-(Z2) (Double parity)
OPERATING SYSTEMEnterprise-Class NexentaStor 3 ZFS Platform
OpenSolaris kernel with the GNU/Debian user interface
HEAD UNIT PROCESSORDual Intel Xeon processor X5670
Westmere Microarchitecture
MEMORY48GB Per Head Unit
Up to 192GB ECC Registered DDR3 Supported
REDUNDANCYRedundant Head Units
Redundant Power Supplies
Redundant Cooling
Dual Ported, Dual Expanders, fully HA Ready
STORAGE IODual 10GbE iSCSI Ethernet Built-in
Fibre Channel Target Optional
WARRANTY5-Year Limited
Petarack Highlights:
  • A full petabyte of storage (1080TB) in a single rack, under a single namespace or volume
  • 360 x 3TB Enterprise SAS drives
  • Enterprise level iSCSI target built-in; fibre channel target ready
  • High availability with no single point of failure
  • Dual Head units, dual expander JBODs, redundant power supplies and fans on each unit
  • Each head unit comes with Dual Xeon Processors and 48 GB Memory by default.
  • Memory is upgradeable to 192GB on each unit
  • Dual 10GbE iSCSI Ethernet Ready
  • Cloud and virtualization ready
  • ZFS file system based, delivering fast, secure storage with unlimited scalability
  • Self healing and silent Data corruption healing capabilities
  • Unlimited snapshots and clones
  • Thin Provisioning
  • Integrated Search function
  • Both block based and file based replication
  • Intuitive web based GUI management with included LCD-KVM
  • Dedicated write log and Read SSD Cache options to speed up NFS
  • Dedicated IPMI ports for head units for direct hardware management and monitoring access
  • Powerful, redundant head units, each with 5 open expansion slots
  • All hard drives are in individual hot swap bays with in-rack access
  • 5U of rack space available for optional switches, a UPS, or another 135TB of storage
  • UPS connection is strongly recommended; or check your Datacenter for power redundancy options
  • Requires 7 Kilowatts of power when fully running
  • Nexenta certified, VMware certification pending
  • Onsite installation
  • Can be scaled to multiple Petabytes by additional racks
Target FC The Target FC Plugin provides a set of capabilities, including an intuitive user interface, that assist in the use of the AberSAN as a block-level target (Fibre Channel only). Target FC includes adaptive multi-pathing so that performance scales up as needed with additional threads -- and manages these threads over time including closing them if they are not needed. It also provides the ability to easily create logical groups of initiators and targets to enable such common block level management tasks as LUN mapping so that a particular application, for example, has unique access to a particular ZVOL, or virtual block device, and thereby inherits the storage management policies unique to that ZVOL such as replication and protection schemes.

High Availability Cluster Plugin (Included in Petarack)
The HA Cluster plugin enables two AberSAN instances to be configured as an active / active pair. When installed and configured, each instance of AberSAN is able to perform all regular functions while at the same time the Plugin provides high availability for shared storage accessible from both appliances. The storage jointly managed via HA Cluster is configured as shared storage. You can also establish RAID and replication policies to protect the data itself in a number of ways with the aid of the software.


VM DataCenter (VMDC)
The VM DataCenter application dramatically simplifies the experience of storage and system administrators in setting storage polices for their virtual environments and in performing common tasks such as stopping and starting VMs and performing template based clones as well. It is easily installed using standard system utilities. It provides a user interface that discovers all ESX or Xen servers and all Virtual Machines (VMs) on the servers. Users can establish replication policies for each VM and can perform common VM management tasks such as stopping and starting VMs. Also, the AberSAN with VMDC can be used for cloning via the 'template' functionality of VMDC to dramatically accelerate provisioning of identical VMs.


Delorean Plugin (Advanced Backup Utility for Windows)
Delorean is a high-powered Windows Backup, Restore and Search tool, with multiple advanced capabilities.
With Delorean you can extend the power of the AberSAN right to the your end users' computers, ensuring that your company's data residing on client machines, as well as servers, is protected. Delorean consists of a client -- that runs on end user Windows machines -- and a Plugin that adds capabilities to the AberSAN. End-users can select what local files and folders they want to protect via a built-in File Explorer interface. Delorean supports Windows VSS, combined backup + snapshot capability, and any backup schedule.


Cross-Platform Multi-Host SAN Target Sharing Support
metaSAN is a high-speed file sharing Storage Area Network (SAN) management software that sets new standards for cross-platform workgroup collaboration. metaSAN enables multiple users to share access to common data files in workgroups where heavy bandwidth requirements are the norm. With metaSAN, film and video editors, digital artists, healthcare specialists, and corporate users can simultaneously access a common pool of data files such as video clips, databases, satellite imagery, medical archives, or CAD files - as easily and transparently as if the content was stored on their local drive.



www.petrack.com

Tuesday, August 16, 2011

Aberdeen’s AberSAN Z30 Helps Out The River City Internet Group





The AberSAN Z-Series from Aberdeen is designed to provide SAN enterprise storage environments with a scalable storage system that’s as intuitive to use as network-attached storage. The file storage appliances can scale beyond 1PB and can be managed through a single Z-RAID array. River City Internet Group recently selected the AberSAN Z30 to help it keep up with the I/O demands of multiple virtual host machines that are constantly reading and writing production data.



High-Performance Storage
River City Internet Group was working with a hosting company that was looking for a high-performance storage system that’s designed to be used with a server farm. Another goal was that the virtual server farm needed to be extremely flexible and cost-efficient. “We have found the AberSAN Z30 to be exactly what we needed,” says Jeff Laird, systems architect for River City Internet Group. “It cost us 60% of what a similar system from another manufacturer would have cost, and it is providing us with performance we have not seen before from a storage system.” The Z30 combines Fibre Channel and iSCSI block-level connections with multiuser network sharing to overcome storage and partition-size limitations.

The Missouri-based company needed fast drives in order to handle the potential I/O load predicted in developing a multinode virtual machine farm. The AberSAN Z30 can be built with hardware and Stripe, Mirror, RAID-Z1 ( s i n g l e p a r i t y ) , R A I D - Z 2 ( d o u b l e parity), and combinations of these. The Z30 can also be built using single, double, or triple parity with hot spares. Multilevel data protection prevents silent data corruptions and supports block- and file-based replication. The AberSAN Z30 is also capable of scaling to work with the growth of diskbased storage. For example, the Z30 uses the ZFS file system, which allows for unlimited snapshots, making it easy for you to back up stored data more often and, if necessary, restore files q u i c k l y . I n c r e m e n t a l d a t a b a c k u p s use less network bandwidth than large daily updates.

For speed, the AberSAN Z30 can use dual Intel Xeon 5600 or 5500 series processors and up to 192GB of memory. The Z30 can reach up to 48TB of raw capacity without using any storage expansions. The AberSAN also meets the needs of River City Internet Group’s virtual machine farm through inherent I/O virtualization. There are hybrid storage pools with SSDs, as well as options for thin provisioning and cloud-ready storage capabilities. Also, unlike traditional solutions, IT staff will enjoy unlimited file system size.

The AberSAN Z30 also comes with the NexentaStor plug-ins. “The Nexenta software is very flexible and has allowed us to allocate and modify settings easily based on changing customer requirements,” Laird says. “It was also easy to learn and provides excellent informational feedback.” The NexentaStor software provides the high-performance features of the ZFS file system, including tools for in-line deduplication, data mirroring, WAN-optimized IP replication, and high-availability clustering.

The NexentaStor plug-ins are one of the main reasons that River City Internet Group selected Aberdeen. “[Aberdeen] offered a unique method for organizing and partitioning drives via the Nexenta software, which immediately made sense to us,” Laird says. NexentaStor’s VM Data
Center plug-in is available to simplify the management of storage for a variety of environments, including VMware vSphere, Citrix Xen, and Microsoft Hyper-V. Companies can even create replication policies for each VM and perform most common management tasks through the plug-in.

Delorean is another helpful NexentaStor plug-in that allows companies to create ZFS-powered Windows backups. IT staff can use the 100 client licenses in Delorean to select local files and folders they want to protect, which means they’ll be further replicated and protected through AberSAN.

Cost was also a major factor in the decision making process for the River City Internet Group. “The ROI for this project was unknown going in, and Aberdeen provided the best value for the equipment we needed,” Laird says.

Z30 Performance Options
The Z30 allows you to scale the hardware by increasing memory available or CPU performance to the number of targets available. With the 3U, 19-inch rackmount, “We chose to go with 10GbE fiber networking for its tremendous throughput capabilities and chose the fastest drives we could order so that the system could be used with an ever-expanding front-end environment,”
Laird says.

There is also a variety of NexentaStor plug-ins you can select to improve the capabilities of the storage system:

Delorean. This is a backup utility for Windows that functions as a backup, restore, and search tool. It includes a client that runs on all user machines and the company servers, while the plug-in delivers the same features to the AberSAN Z30. Delorean supports combined backup and snapshot capabilities, as well as Windows VSS and other backup schedules.

HA (High Availability) Cluster. The AberSAN Z30 can use the HA Cluster plug-in to create two AberSAN instances that will be configured as an active pair. Each instance of the AberSAN provides you with high availability, and the shared storage is available from both appliances. If you want to protect your data, the HA Cluster plug-in can work with RAID and replication policies.

metaSAN. This management software provides high-speed file sharing with cross-platform workgroup collaboration. Multiple users can share data files, which is ideal for film and video editors, healthcare specialists, and other environments where there’s a need to access a common pool of data files.

Target FC. This plug-in is designed to use the AberSAN Z30 as a block-level target when working only with Fibre Channel. Target FC can also create logical groups of initiators and targets. For example, a common block management could allow a ZVOL or virtual block device to inherit the storage management policies unique to the ZVOL or protection scheme.

VM Data Center. The VM Data Center plug-in makes it easier to set storage policies
for virtual environments and perform storage tasks, such as starting a VM. It can also help you discover all ESX or Xen servers on the VMs in the servers. The AberSAN Z30 and VM Data Center can close through the template function, which can quicken provisioning of identical VMs.

WORM. With this utility, IT staff can make any data folder into a Write Once, Read Many mode. The WORM utility works with all network content, including files, directories, and subfolders.http://www.rcig.net/













Tuesday, August 2, 2011

Aberdeen S-Series JBOD Expansion - 4U 45Bay SAS/SATA 6G

The NEW Aberdeen JBOD S-Series S41/S42 are 6Gb/s SAS-based expansion enclosures that can accommodate forty-five  (45) SAS/SATA drives in a 4U form factor and features a redundant 1400W Gold Level power supply. The expansion enclosure allows users to easily attach terabytes of capacity to any Aberdeen server with an external mini-sas raid port.  Hot-Swap drive bays are used for the front and back of the chassis.  Twenty-four (24) front and twenty-one (21) rear hdd bays are provided.  Single and dual expander models are available for standard or high availability scenarios. 

FEATURES
  • 45 HDD in 4U (extreme drive density)
  • 6Gb/s SAS connectivity
  • Single or Dual LSI SAS2X36 expander chips
  • 3.5-inch disk drives
  • Gold level high-efficiency 1400W power supply
  • Redundant, hot-swappable cooling fans
  • Quick Release, Tool-less rails
Host PortsFour or Two 6Gb/s SAS ports
Drive Support
  • 45x (24 front + 21 rear) 3.5" Hot-swap HDD bays

  • SAS or enterprise SATA HDD only recommended

  • Form Factor4U Rackmount
    System LED
  • Power LED

  • Hard drive activity LED

  • 2x Network activity LEDs

  • System Overheat LED

  • Power Fail LED

  • High-Availability Features
  • Redundant Power Supplies

  • Redundant Hot Pluggable Cooling System

  • Hot-Swap Drives

  • Power Supply1400W high-efficiency (1+1) redundant power supply with PMBus AC voltage:
    - 1100W: 100 - 140V, 50 - 60Hz, 9.5 - 13.5 Amp
    - 1400W: 180 - 240V, 50 - 60Hz, 7.0 - 9.5 Amp
    Certification 80 PLUS Gold Certified
    Environmental
  • Temperature: 10 to 35°C operating/ -40 to 70°C non-operating

  • Altitude: Sea level to 3660m (12,000ft) operating / sea level to 12,192m (40,000ft) non-operating

  • Relative humidity: 8 to 90% non-condensing, operating

  • DimensionsWithout chassis ears / protrusions:
    437mm (W) x 178mm (H) x 699mm (D)
    WeightWithout HDDs:36.3kg (80lbs)

    Tuesday, July 26, 2011

    Real-Time Data Compression's Impact on SSD Throughput Capability

    SSDs are one of the hottest technologies in storage. They have great throughput performance (read and write), great IOPS performance with the challenges of a limited number of rewrite cycles, and a much higher price than spinning media. In addition, there are some challenges to make the technology perform well, requiring new techniques to improve the overall behavior. Some of these controller techniques improve performance or longevity, or both. However, these techniques must be "tuned" so the best possible performance is extracted from the technology. One SSD controller company has a technique that improves both the performance and the longevity, but the improvements are solely based on your data.
    SandForce is a fairly new SSD controller company. Its products are used in a great number of SSDs, from affordable consumer SSDs to enterprise-class SSDs. The technique SandForce has developed is that its controllers use real-time data compression. They actually compress the data before it is written to the drive enabling the performance can be increased and the longevity of the drive can be improved.
    This article examines the concepts of using real-time data compression in SSDs by taking a consumer SandForce-based SSD for a spin. I test the throughput performance using IOzone, which allows me to vary the compressibility (dedupability) of the data so I can see the impact on the throughput performance of the drive. The results are pretty exciting and interesting as we'll see.

    Real-Time Data Compression in SSDs

    I talked about real-time data compression in SandForce SSD controllers in a different article, but the concepts definitely bear repeating because they are so fascinating.


    The basic approach SandForce has taken is to use some of the capabilities from its SSD controller for real-time data compression. Since the implementation is proprietary, I can only speculate on what is going on inside the controller. Most likely, the uncompressed data comes into the drive (the controller) and is likely stored in a buffer. The controller then compresses the data in the buffer in individual chunks or perhaps coalescing the data chunks prior to compression. This process takes some time and computational resources prior to writing it to the storage media.
    Once the data is compressed, it is placed on the blocks within the SSD. However, things are not quite this simple. To ensure the drive is reporting the correct amount of data stored, presumably the uncompressed data size is also stored in some sort of metadata format on the drive itself (maybe within the compressed data?). This means that if a data request comes into the controller with a request for the number of data blocks or size of a data block, the correct size is reported.
    But since the data has been compressed, the amount of data that is written is less than the uncompressed data. Less data is written to the storage media, which means less time is used, which means faster throughput. The amount of time used to write the data is proportional to the size of the compressed data (i.e., the compressibility of the data), which drives the throughput performance. But we also need to remember that the "latency" for a SandForce controller can be higher than a typical controller because of the time needed to compress the data. I'm sure SandForce has taken this into account so that not too much time is spent compressing the data. In fact, I bet it's a constant time compression algorithm.
    During a read operation, the compressed data is probably read into a cache and then uncompressed. After that, it is sent to the operating system as though the data was never compressed. Presumably, the performance also depends on the ability to uncompress the data quickly so the algorithm should have a fixed time.

    The really interesting and exciting part of this is that the compressibility of your data influences the performance of the storage media. If your data is very compressible, then your performance can be very, very good. If your data is as compressible as a rock then your performance may not be as good. But before you start thinking, "my data is very incompressible," note that I have seen lots of different data sets (even binary ones) capable of being compressed. Also keep in mind that the SandForce controller needs to compress the data only in a specific chunk, not the entire file. So it's difficult to say a priori with any certainty that your data will not perform well on a SandForce controller. I'm sure SandForce has spent a great deal of time running tests on typical data for the target markets and thus has a good idea of what it can and cannot do. Since the controller is quite popular, I'm sure the efforts have been successful.
    However, the fundamental fact remains that the performance of the SSD depends on the compressibility of the data. Thus, I decided to get a consumer SSD with a SandForce 1222 controller and run some IOZone tests with variable levels of dedupability (compressibility) to see how the SSD performs.


    Testing/Benchmarking Approach and Setup

    The old phrase of "if you're going to do it, do it right," definitely rings true for benchmarking. All too often storage benchmarks are nothing less than marketing materials providing very little useful information. In this article, I will follow concepts that aim to improve the quality of the benchmarks. In particular:
    • The motivation behind the benchmarks will be explained (if it hasn't already).
    • Relevant and useful storage benchmarks will be used.
    • The benchmarks will be detailed as much as possible.
    • The tests will run for more than 60 seconds.
    • Each test is run 10 times, and the average and standard deviation of the results will be reported.
    These basic steps and techniques can make benchmarking and testing much more useful.
    The benchmarks in this article are designed to explore the write, read, random write, random read, and fread, fwrite performance of the SSD. I don't really know what the results will be prior to running the tests so there are no expectations -- it's really an exploration of the performance.
    This article examines the performance of a 64GB Micro Center SSD that uses a SandForce 1222 controller. This is a very inexpensive drive that provides about 60GB of useable space for just under $100. ($1.67/GB). The specifications on the website state the drive has a SATA 3.0 Gbps interface (SATA II), and it has a performance of up to 270 MB/s for writes, 280 MB/s for reads, and up to 50,000 IOPS.


    The highlights of the system used in the testing are:
    • GigaByte MAA78GM-US2H motherboard
    • An AMD Phenom II X4 920 CPU
    • 8GB of memory (DDR2-800)
    • Linux 2.6.32 kernel
    • The OS and boot drive are on an IBM DTLA-307020 (20GB drive at Ultra ATA/100)
    • /home is on a Seagate ST1360827AS
    • The Micro Center SSD is mounted as /dev/sdd
    I used CentOS 5.4 on this system, but I used my own kernel -- 2.6.32. Ext4 will be used as the file system as well. The entire device, /dev/sdd was used for the file system so it is aligned with page boundaries.
    I used IOzone because it is one of the most popular throughput benchmarks. It is open source and written in very plain ANSI C (not an insult but a compliment), and perhaps more importantly, it tests different I/O patterns, which very few benchmarks actually do. It is capable of single-thread, multi-thread, and multi-client testing. The basic concept of IOzone is to break up a file of a given size into records. Records are written or read in some fashion until the file size is reached. Using this concept, IOzone has a number of tests that can be performed. In the interest of brevity, I limited the benchmarks to write, read, random write, random read, fwrite, and fread, all of which are described below.
    • Write: This is a fairly simple test that simulates writing to a new file. Because of the need to create new metadata for the file, many times the writing of a new file can be slower than rewriting to an existing file. The file is written using records of a specific length (either specified by the user or chosen automatically by IOzone) until the total file length has been reached.
    • Read: This test reads an existing file. It reads the entire file, one record at a time.
    • Random Read: This test reads a file with the accesses being made to random locations within the file. The reads are done in record units until the total reads are the size of the file. Many factors impact the performance of this test, including the OS cache(s), the number of disks and their configuration, disk seek latency and disk cache.
    • Random Write: The random write test measures the performance when writing a file with the accesses being made to random locations with the file. The file is opened to the total file size, and then the data is written in record sizes to random locations within the file.
    • fwrite: This test measures the performance of writing a file using a library function "fwrite()". It is a binary stream function (examine the man pages on your system to learn more). Equally important, the routine performs a buffered write operation. This buffer is in user space (i.e., not part of the system caches). This test is performed with a record length buffer being created in a user-space buffer and then written to the file. This is repeated until the entire file is created. This test is similar to the "write" test in that it creates a new file, possibly stressing the metadata performance.
    • fread: This is a test that uses the fread() library function to read a file. It opens a file and reads it in record lengths into a buffer that is in user space. This continues until the entire file is read.
    Other options can be tested, but for this exploration only the previously mentioned tests will be examined.
    For IOzone, the system specifications are fairly important since they affect the command-line options. In particular, the amount of system memory is important because this can have a large impact on the caching effects. If the problem sizes are small enough to fit into the system or file system cache (or at least partially), it can skew the results. Comparing the results of one system where the cache effects are fairly prominent to a system where cache effects are not conspicuous is comparing the proverbial apples to oranges. For example, if you run the same problem size on a system with 1GB of memory compared to a system with 8GB you will get much different results.
    For this article, cache effects will be limited as much as possible. Cache effects can't be eliminated entirely without running extremely large problems and forcing the OS to virtually eliminate all caches. However, one of the best ways to minimize the cache effects is to make the file size much bigger than the main memory. For this article, the file size is chosen to be 16GB, which is twice the size of main memory. This is chosen arbitrarily based on experience and some urban legends floating around the Internet.
    For this article, the total file size was fixed at 16GB and four record sizes were tested: (1) 1MB, (2) 4MB, (3) 8MB, and (4) 16MB. For a file size of 16GB that is (1) 16,000 records, (2) 4,000 records, (3) 2,000 records, (4) 1,000 records, respectively. Smaller record sizes took too long to run since the number of records would be very large so they are not used in this article.
    The command line for the first record size (1MB) is,
    ./IOzone -Rb spreadsheet_output_1M.wks -s 16G -+w 98 -+y 98 -+C 98 -r 1M > output_1M.txt
    
    The command line for the second record size (4MB) is,
    ./IOzone -Rb spreadsheet_output_4M.wks -s 16G -+w 98 -+y 98 -+C 98 -r 4M > output_4M.txt
    
    The command line for the third record size (8MB) is,
    ./IOzone -Rb spreadsheet_output_8M.wks -s 16G -+w 98 -+y 98 -+C 98 -r 8M > output_8M.txt
    
    The command line for the fourth record size (16MB) is,
    ./IOzone -Rb spreadsheet_output_16M.wks -s 16G -+w 98 -+y 98 -+C 98 -r 16M > output_16M.txt
    
    The options, "-+w 98", "-+y 98", and "-+C 98" are options that control the dedupability (compressibility) of the data. IOZone uses the phrase dedupe to describe if the data is capable of being deduplicated. This is basically the same as compressed, so I will use the phrases interchangeably. The number "98" in the options means the data is 98 percent dedupable, hence very compressible. These three options allow me to control the compressibility of the data, so I can examine the impact of data compressibility on performance.
    I tested three levels of data compressibility -- 98 percent (very compressible), 50 percent, and 2 percent (very incompressible). I wanted to get an idea of the range of performance with these levels of data compressibility but I didn't want to get unrealistic results with either 100 percent compressible data or 0 percent compressible. (Though I'm not really sure if either of those are really possible.)


    Results

    All of the results are presented in bar chart form, with the average values plotted and the standard deviation shown as error bars. For each of the three levels of compressibility (98 percent, 50 percent and 2 percent), I plot the results for the four record sizes (1MB, 4MB, 8MB, and 16MB).
    The first throughput result is the write throughput test. Figure 1 below presents the results for the four record sizes and the three levels of dedupability (compressibility). The averages are the bar charts and the error bars are the standard deviation.
    (Figure 1)Write Throughput Results from IOzone in KB/s
      The first obvious thing you notice in Figure 1 is that as the level of compressibility decreases (thus decreasing dedupability), the performance goes down. The best performance for the cases run was for a 1MB record size and 98 percent dedupable data, where the write throughput performance was about 260 MB/s, which is close to the stated specifications for the drive. As record size increases, the performance goes down, so that at a record size of 16MB, the write throughput performance was about 187 MB/s for 98 percent dedupability.
    For data that is 50 percent dedupable (compressible), the performance for a record size of 1 MB was only about 128MB/s, and for data that is almost incompressible (2 percent dedupable), the performance for a record size of 1 MB was only about 97 MB/s. So the performance drops off as the data becomes less and less compressible (as one might expect). Another interesting observation is that as the data becomes less compressible, there is little performance difference between record sizes.
    (Figure 2) Read Throughput Results from IOzone in KB/s
    These results differ from the write results in that the performance does not decrease that much as the level of data compressibility rises. For a 1MB record size, at 98 percent dedupable data, the performance was about 225 MB/s, for 50 percent dedupable data the performance was about 202 MB/s, and for 2 percent dedupable data, the performance was about 192 MB/s.

    In addition, as the level of compressibility decreases the performance difference between record sizes almost disappears. Compare the four bars for 98 percent dedupable data and for 2 percent dedupable data. At 2 percent the performance for the four record sizes is almost the same. In fact, for larger record sizes, the read performance actually improves as the data becomes less compressible.

    Figure 3 presents the random write throughput results for the four record sizes and the three levels of dedupability (compressibility).

    Random Write Throughput Results from IOzone in KB/s for the Four Record Sizes and the Three Data Dedupable Levels
    Figure 3
    Random Write Throughput Results from IOzone in KB/s
    Figure 3 reveals that just like write performance, random write performance drops off as the level of data compressibility decreases. For the 1MB record size, the random write throughput was about 241 MB/s for 98 percent dedupable data, 122 MB/s for 50 percent dedupable data, and about 93 MB/s for 2 percent dedupable data.
    In addition, just like the write performance, the performance difference between record sizes is very small as the level of compressibility decreases. So for small levels of dedupable data, the record size had very little impact on performance over the range of record sizes tested.
    Figure 4 presents the random read throughput results for the four record sizes and the three levels of dedupability (compressibility).

    Random Read Throughput Results from IOzone in KB/s for the Four Record Sizes and the Three Data Dedupable Levels
    Figure 4
    Random Read Throughput Results from IOzone in KB/s
    The general trends for random read performance mirror those of the read performance in Figure 2. Specifically:
    • The performance drops very little with decreasing compressibility (dedupability).
    • As the level of compressibility decreases, the performance for larger record sizes actually increases.
    • As the level of compressibility decreases, there is little performance variation between record sizes.
    Figure 5 below presents the fwrite throughput results for the four record sizes and the three levels of dedupability (compressibility).

    Fwrite Throughput Results from IOzone in KB/s for the Four Record Sizes and the Three Data Dedupable Levels
    Figure 5
    Fwrite Throughput Results from IOzone in KB/s
    These test results mirror those of the write throughput test (Figure 1). In particular,
    • As the level of compressibility decreases, the performance drops off fairly quickly.
    • As the level of compressibility decreases, there is little variation in performance as the record size changes (over the record sizes tested).
    Figure 6 presents the fread throughput results for the four record sizes and the three levels of dedupability (compressibility).

    Alt text
    Figure 6
    Fread Throughput Results from IOzone in KB/s
    Like the two previous read tests (read, random read), the general trends for fread performance has the same general trends:
    • The performance drops only slightly with decreasing compressibility (dedupability).
    • As the level of compressibility decreases, the performance for larger record sizes actually increases.
    • As the level of compressibility decreases, there is little performance variation between record sizes.

    Summary

    SandForce is quickly becoming a dominant player in the fast-growing SSD controller market. It makes SSD controllers for many consumer drives and increasingly, enterprise-class SSDs. One of the really cool features of its controllers is that they do real-time data compression to improve throughput and increase drive longevity. The key aspect to this working for you is that the ultimate performance depends on your data. More precisely -- the compressibility of your data.
    For this article I took a consumer SSD that has a SandForce 1222 controller and ran some throughput tests against it using IOzone. IOzone enabled me to control the level of data compressibility, which IOzone calls dedupability, so I could test the impact on performance. I tested write and read performance as well as random read, random write, fwrite, and fread performance. I ran each test 10 times and reported the average and standard deviation.
    The three write tests all exhibited the same general behavior. More specifically:
    • As the level of compressibility decreases, the performance drops off fairly quickly.
    • As the level of compressibility decreases, there is little variation in performance as the record size changes (over the record sizes tested).
    The absolute values of the performance varied for each test, but for the general write test, the performance went from about 260 MB/s (close to the rated performance) at 98% data compression to about 97 MB/s at 2% data compression for a record size of 1 MB.
    The three read test all also exhibited the same general behavior. Specifically,
    • The performance drops only slightly with decreasing compressibility (dedupability)
    • As the level of compressibility decreases, the performance for larger record sizes actually increases
    • As the level of compressibility decreases, there is little performance variation between record sizes
    Again, the absolute performance varies for each test, but the trends are the same. But basically, the real-time data compression does not affect the read performance as much as it does the write performance.
    The important observation from these tests is that the performance does vary with data compressibility. I believe that SandForce took a number of applications from their target markets and studied the data quite closely and realized that it was pretty compressible and designed their algorithms for those data patterns. While SandForce hasn't stated which markets they are targeting I think to understand the potential performance impact for your data requires that you study your data. Remember that you're not studying the compressibility of the data file as a whole but rather the chunks of data that a SandForce controller SSD would encounter. So think small chunks of data. I think you will be surprised at how compressible your data actually is. But it's always a good idea to test the hardware against your applications.
    April 12, 2011
    By Jeffrey Layton