Blade Server

A blade server is a stripped down server computer with a modular design optimized to minimize the use of physical space and energy. Whereas a standard rack-mount server can function with (at least) a power cord and network cable, blade servers have many components removed to save space, minimize power consumption and other considerations, while still having all the functional components to be considered a computer.[clarification needed][1] A blade enclosure, which can hold multiple blade servers, provides services such as power, cooling, networking, various interconnects and management. Together, blades and the blade enclosure form a blade system (although BladeSystem from Hewlett-Packard is a specific product name). Different blade providers have differing principles regarding what to include in the blade itself, and in the blade system altogether.

In a standard server-rack configuration, one rack unit or 1U—19 inches (480 mm) wide and 1.75 inches (44 mm) tall—defines the minimum possible size of any equipment. The principal benefit and justification of blade computing relates to lifting this restriction so as to reduce size requirements. The most common computer rack form-factor is 42U high, which limits the number of discrete computer devices directly mountable in a rack to 42 components. Blades do not have this limitation. As of 2014[update], densities of up to 180 servers per blade system (or 1440 servers per rack) are achievable with blade systems.[2]

Blade enclosure

Enclosure (or chassis) performs many of the non-core computing services found in most computers. Non-blade systems typically use bulky, hot and space-inefficient components, and may duplicate these across many computers that may or may not perform at capacity. By locating these services in one place and sharing them between the blade computers, the overall utilization becomes higher. The specifics of which services are provided may vary by vendor.

Power

Computers operate over a range of DC voltages, but utilities deliver power as AC, and at higher voltages than required within computers. Converting this current requires one or more power supply units (or PSUs). To ensure that the failure of one power source does not affect the operation of the computer, even entry-level servers may have redundant power supplies, again adding to the bulk and heat output of the design.

The blade enclosure’s power supply provides a single power source for all blades within the enclosure. This single power source may come as a power supply in the enclosure or as a dedicated separate PSU supplying DC to multiple enclosures.[3][4] This setup reduces the number of PSUs required to provide a resilient power supply.

The popularity of blade servers, and their own appetite for power, has led to an increase in the number of rack-mountable uninterruptible power supply (or UPS) units, including units targeted specifically towards blade servers (such as the BladeUPS).

Cooling

During operation, electrical and mechanical components produce heat, which a system must dissipate to ensure the proper functioning of its components. Most blade enclosures, like most computing systems, remove heat by using fans.

A frequently underestimated problem when designing high-performance computer systems involves the conflict between the amount of heat a system generates and the ability of its fans to remove the heat. The blade’s shared power and cooling means that it does not generate as much heat as traditional servers. Newer[update] blade-enclosures feature variable-speed fans and control logic, or even liquid cooling-systems[5][6] that adjust to meet the system’s cooling requirements.

At the same time, the increased density of blade-server configurations can still result in higher overall demands for cooling with racks populated at over 50% full. This is especially true with early-generation blades. In absolute terms, a fully populated rack of blade servers is likely to require more cooling capacity than a fully populated rack of standard 1U servers. This is because one can fit up to 128 blade servers in the same rack that will only hold 42 1U rack mount servers.[7]

Networking

Blade servers generally include integrated or optional network interface controllers for Ethernet or host adapters for Fibre Channel storage systems or converged network adapter to combine storage and data via one Fibre Channel over Ethernet interface. In many blades at least one interface is embedded on the motherboard and extra interfaces can be added using mezzanine cards.

A blade enclosure can provide individual external ports to which each network interface on a blade will connect. Alternatively, a blade enclosure can aggregate network interfaces into interconnect devices (such as switches) built into the blade enclosure or in networking blades.[8][9]

Storage

While computers typically use hard disks to store operating systems, applications and data, these are not necessarily required locally. Many storage connection methods (e.g. FireWire, SATA, E-SATA, SCSI, SAS DAS, FC and iSCSI) are readily moved outside the server, though not all are used in enterprise-level installations. Implementing these connection interfaces within the computer presents similar challenges to the networking interfaces (indeed iSCSI runs over the network interface), and similarly these can be removed from the blade and presented individually or aggregated either on the chassis or through other blades.

The ability to boot the blade from a storage area network (SAN) allows for an entirely disk-free blade, an example of which implementation is the Intel Modular Server System.

Other blades

Since blade enclosures provide a standard method for delivering basic services to computer devices, other types of devices can also utilize blade enclosures. Blades providing switching, routing, storage, SAN and fibre-channel access can slot into the enclosure to provide these services to all members of the enclosure.

Systems administrators can use storage blades where a requirement exists for additional local storage.[10][11][12]

Uses

Blade servers function well for specific purposes such as web hosting, virtualization, and cluster computing. Individual blades are typically hot-swappable. As users deal with larger and more diverse workloads, they add more processing power, memory and I/O bandwidth to blade servers.

Although blade server technology in theory allows for open, cross-vendor solutions, the stage of development of the technology as of 2009[update] is such that users encounter fewer problems when using blades, racks and blade management tools from the same vendor.

Eventual standardization of the technology might result in more choices for consumers;[13][14] as of 2009[update] increasing numbers of third-party software vendors have started to enter this growing field.[15]

Blade servers do not, however, provide the answer to every computing problem. One can view them as a form of productized server-farm that borrows from mainframe packaging, cooling, and power-supply technology. Very large computing tasks may still require server farms of blade servers, and because of blade servers’ high power density, can suffer even more acutely from the heating, ventilation, and air conditioning problems that affect large conventional server farms.

Blade models

Cisco UCS blade servers in a chassis

Though many independent professional computer manufacturers such as Supermicro offer blade solutions, the blade server market continues to be dominated by large public IT companies such as Cisco, which achieved No. 1 position on the blade server market with 40% share by revenue in Americas in the first quarter of 2014.[20] The remaining prominent brands in the blade server market are HP, Dell and IBM, though the latter sold its x86 business to Lenovo in 2014.[21]

Cisco offers a Unified Computing System in its blade server portfolio: a domain, consisted of 6U high chassis, up to 8 blade servers in each chassis, up to 20 chassis in a domain, all managed by a heavily modified Nexus 5K switch, rebranded as a fabric interconnect, which converge computing and networking, and manages the whole system as a single entity.[22] HP’s current line consists of two chassis models, the c3000 which holds up to 8 half-height ProLiant line blades (also available in tower form), and the c7000 (10U) which holds up to 16 half-height ProLiant blades. Dell‘s latest solution, the M1000e is a 10U modular enclosure and holds up to sixteen half-height PowerEdge blade servers or thirty two quarter-height blades.

http://en.wikipedia.org/wiki/Blade_server

Advertisements