12.07.2015 Views

Dell Power Solutions

Dell Power Solutions

Dell Power Solutions

SHOW MORE
SHOW LESS
  • No tags were found...

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

HIGH-PERFORMANCE COMPUTINGin a cluster are server type and network interconnect. The servers in acluster provide the computational power and are connected throughthe interconnect fabric.For coarse-grained parallel applications, communication is generallyless of a concern. Hence, organizations running a high-performing,coarse-grained parallel application can select high-performance servers,which offer good computation power. Gigabit networking canmeet the interconnect requirements of coarse-grained parallel applicationswhile maintaining a low price point. Using a traditional low-costinterconnect with a high-performance server can provide a low-costHPC solution for these applications.For medium-grained parallel applications, the communicationoverhead can be high. For such applications, organizations shouldconsider a high-performance interconnect like Myricom Myrinet 1or InfiniBand. Both interconnects offer low latency and high bandwidth.Latency is the amount of time taken to transmit a messagefrom the source server to the destination server. Bandwidth isthe rate at which information can be sent over the interconnect.A slow interconnect with a fast processing system will cause theinterconnect to become a bottleneck for the application.For fine-grained parallel applications running on Beowulf clusters,a fast interconnect like Myrinet or InfiniBand is recommended.Using a slower interconnect could cause the communication timeto overshadow the computation time, rendering the parallelizationof the application unsuccessful.Server types and interconnects are two high-level componentsof a cluster. Choosing the most appropriate server is in itself abroad topic with many components such as memory subsystem,processor speed, and cache sizes to be considered. <strong>Dell</strong> supports awide range of rack-mount servers—including the <strong>Power</strong>Edge 1850,<strong>Power</strong>Edge 1855, <strong>Power</strong>Edge 2850, and <strong>Power</strong>Edge 3250 servers—that are suitable for HPC clusters and offer varied architectures tosatisfy computational needs.For coarse-grained applications, choosing the appropriate industrystandardcomponents can enable organizations to create a low-costcluster that will still meet their application needs. A recently releasedserver from <strong>Dell</strong>, the <strong>Power</strong>Edge SC1425, is one such component thatcan help provide a viable low-cost alternative for HPC needs.Testing the <strong>Dell</strong> <strong>Power</strong>Edge SC1425 server as a cluster nodeIn October 2004, a team of <strong>Dell</strong> engineers tested cluster performanceusing a <strong>Power</strong>Edge SC1425 server as the building block for a low-costcluster. For this study, each <strong>Power</strong>Edge SC1425 server was configuredwith two Intel ®Xeonprocessors. However, the <strong>Power</strong>Edge SC1425 canbe configured with either single or dual CPUs and can run in either32-bit mode or Intel Extended Memory 64 Technology (EM64T) modefor 64-bit applications. For memory subsystem needs, the <strong>Power</strong>EdgeSC1425 supports dual-banked double data rate 2 (DDR2) memory onan 800 MHz frontside bus (FSB). It can be equipped with either serialATA (SATA) or SCSI hard drives. Because most of the data in this clusterperformance test was shared from the master node, a local SCSIhard drive on each compute node was not necessary. Two embeddedGigabit 2 Ethernet controllers were included in the base system, whicheliminated the need for any additional network hardware.The following sections demonstrate the performance of the<strong>Power</strong>Edge SC1425 cluster for different communication patterns.Using various well-known benchmarks—Pallas PingPong, NASParallel Benchmarks, IOzone, and STREAM—the team conductedtests on a single node, two nodes, and the entire cluster to profile the<strong>Power</strong>Edge SC1425 for cluster performance. The lab setup consistedof four <strong>Dell</strong> <strong>Power</strong>Edge SC1425 compute server nodes, each with2 GB of main memory and dual symmetric multiprocessing (SMP)processors at 3.6 GHz. Two types of interconnect were examined:Gigabit Ethernet and Myrinet. The <strong>Dell</strong> <strong>Power</strong>Connect 5324 switchwas used to form the Gigabit Ethernet fabric, while Myricom Myrinetswitches formed the Myrinet fabric.The impact of interconnects on cluster performanceUsing the Pallas 3 PingPong benchmark, the <strong>Dell</strong> test team examinedlatency at various message sizes for two <strong>Power</strong>Edge SC1425 clusternodes. The two nodes—one using Gigabit Ethernet and the otherusing Myrinet—sent data to each other via the Message PassingInterface (MPI). As shown in Figure 1, the latency for Myrinet withsmall messages of 4 bytes was about 6.5 microseconds (µs) comparedto the latency of 35 µs for 4-byte messages using Gigabit Ethernet.Figure 2 shows the MPI message throughput between thesetwo cluster nodes at various message sizes. In this study, GigabitLatency (microseconds)250200150100500MyrinetGigabit Ethernet0 1 2 4 8 16 32 64 128 256 512 1024 2048 4096 8192 16384Message size (bytes)Figure 1. Latency between two <strong>Power</strong>Edge SC1425 cluster nodes using different interconnects1For more information about Myrinet, see www.myri.com.2This term does not connote an actual operating speed of 1 Gbps. For high-speed transmission, connection to a Gigabit Ethernet server and network infrastructure is required.3For more information about Pallas MPI benchmarks, see www.pallas.com/e/products/pmb/index.htm.116POWER SOLUTIONS Reprinted from <strong>Dell</strong> <strong>Power</strong> <strong>Solutions</strong>, February 2005. Copyright © 2005 <strong>Dell</strong> Inc. All rights reserved. February 2005

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!