29.12.2014 Views

Magellan Final Report - Office of Science - U.S. Department of Energy

Magellan Final Report - Office of Science - U.S. Department of Energy

Magellan Final Report - Office of Science - U.S. Department of Energy

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

<strong>Magellan</strong> <strong>Final</strong> <strong>Report</strong><br />

Table 9.6: Comparison <strong>of</strong> STAR performance and cost across different Amazon instance types and <strong>Magellan</strong><br />

Instance type m1.large x2.4xlarge <strong>Magellan</strong><br />

cores 2 8 8<br />

EC2 units/core 2 3.25 -<br />

EC2 units/instance 4 26 -<br />

Cost ($)/instance hour $0.38 $2.28 -<br />

Cost ($)/hour/EC2 unit $0.10 $0.09 -<br />

Wall hours/job 2.119 1.242 1.269<br />

CPU hours/job 1.500 1.240 1.263<br />

Cost ($)/job 0.403 0.354 -<br />

cluster was 2 to 3 times slower than on bare metal. Due to the performance hit and the need to move data<br />

in and out <strong>of</strong> the virtual cluster, which did not have access to the NERSC Global File System, the user did<br />

not consider this a resource environment worth pursuing at the current time.<br />

9.5.5 <strong>Energy</strong>Plus<br />

Another type <strong>of</strong> application that was tested in the “Little <strong>Magellan</strong>” environment was a code called <strong>Energy</strong>Plus<br />

(E+) that does building simulations, such as estimating the energy needs <strong>of</strong> a building before it<br />

is built. The user had been tasked with improving the performance, and used OpenMP with great results<br />

when comparing serial with 1, 2, 4 and 8 OpenMP threads. Code ran in 10 different configurations (each<br />

<strong>of</strong> them 1, 2, 4 and 8 threads) on both virtual cluster and bare metal. Results for 1, 2 and 4 threads were<br />

identical. The results for 8 threads were 1015% slower.<br />

9.5.6 LIGO<br />

LIGO was another application that used the <strong>Magellan</strong> testbed through the “Little <strong>Magellan</strong>” setup. The<br />

application is described in greater detail in Section 11.2.3. LIGO users stated that they saw no difference<br />

between running through the serial queue on <strong>Magellan</strong> and on the virtual machines. The nature <strong>of</strong> the LIGO<br />

workload, i.e., minimal data and limited communication, makes it a suitable candidate to use virtualized<br />

cloud resources. There was negligible difference in CPU performance between bare metal and virtualized<br />

resources for this application.<br />

9.6 Workload Analysis<br />

Our benchmarking results show that applications with minimal communication and I/O do well in cloud<br />

environments. In order to further understand the range <strong>of</strong> HPC applications that might benefit from cloud<br />

environments, we analyzed the workflows <strong>of</strong> researchers who were using the <strong>Magellan</strong> batch queue system at<br />

NERSC.<br />

Workload pr<strong>of</strong>ile information was collected using the Integrated Performance Monitoring (IPM) analysis<br />

infrastructure. IPM is a low-overhead pr<strong>of</strong>iling infrastructure that provide a high-level report on the execution<br />

<strong>of</strong> a job. IPM reports hardware counters data, MPI function timings, and memory usage. It provides a lowoverhead<br />

means to generate scaling studies or performance data. IPM is available to users on NERSC<br />

systems using modules.<br />

On the <strong>Magellan</strong> systems, we force IPM pr<strong>of</strong>iling information to be collected for every job. The user gets<br />

normal output at the end <strong>of</strong> the job, and a log file is written to a shared space. To aid with the analysis,<br />

we built a generic infrastructure using MongoDB, also a cloud technology. MongoDB is a scalable, open<br />

source, document-oriented data store that supports dynamic schemas. An existing parsing script in IPM<br />

was modified to output a JSON-style document that was then used to populate MongoDB.<br />

78

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!