29.12.2014 Views

Magellan Final Report - Office of Science - U.S. Department of Energy

Magellan Final Report - Office of Science - U.S. Department of Energy

Magellan Final Report - Office of Science - U.S. Department of Energy

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

<strong>Magellan</strong> <strong>Final</strong> <strong>Report</strong><br />

the EBS corresponding volumes. This is especially true in the Amazon small instances that are bandwidth<br />

limited. At least during the duration <strong>of</strong> our tests, we noticed that the larger instance types showed lesser<br />

advantage with the local disks, possibly due to the increased network bandwidth available in the large instances.<br />

However, users need to do a performance-cost analysis <strong>of</strong> local disk vs EBS for their particular<br />

application [47].<br />

Instance Type. The anticipated I/O performance on the instance type is expected to get better with the<br />

larger better instances. However, in our limited testing, we encountered situations where we were able to get<br />

better I/O performance on the small instance local disk than the large instance. Our tests also show that<br />

the small instances do tend to show a fair amount <strong>of</strong> variability, and hence more extensive testing might be<br />

needed to capture these behaviors over time. The EBS performance appeared to improve with more capable<br />

instance types, possibly due to the better network available to the larger and/or the CC instances.<br />

Availability Regions. We observed that the west zone performance was better than the performance <strong>of</strong><br />

the east zone. The west zone VMs on Amazon have slightly higher price points, possibly resulting in better<br />

performance. However, our large-scale tests also show that the west zone has a higher standard deviation<br />

than the east zone.<br />

9.4 Flash Benchmarking<br />

Solid-state storage (SSS) is poised as a disruptive technology. This impact would likely affect both the<br />

cloud computing and scientific computing spaces. For these reasons, flash storage evaluation was included<br />

in the <strong>Magellan</strong> project. <strong>Magellan</strong> at NERSC first evaluated several products before deploying a larger<br />

storage system based on that evaluation. The technologies that were evaluated include three PCIe connected<br />

solutions and two SATA connected solutions. Table 9.5 summarizes the products that were evaluated.<br />

Table 9.5: Summary <strong>of</strong> flash-based storage products that were evaluated.<br />

Manufacturer Product Capacity<br />

PCIe Attached Devices - All use SLC Flash<br />

Texas Memory Systems RamSAN 20 450 GB<br />

FusionIO ioDrive Duo 320 GB<br />

Virident tachIOn 400 GB<br />

SATA Attached Devices - Both use MLC Flash<br />

Intel X25-M 160 GB<br />

OCZ Colossus 250 GB<br />

NAND flash devices have dramatically different performance characteristics compared with traditional<br />

disk systems. NAND typically delivers much higher random read rates. However, NAND chips must be<br />

erased prior to writing new data. This erase cycle can be extremely slow. For example, NAND may require<br />

several milliseconds to erase a block, yet can perform a read <strong>of</strong> block in several microseconds [79]. To help<br />

mask this impact, many high-end devices utilize background grooming cycles to maintain a pool <strong>of</strong> erased<br />

blocks available for new writes.<br />

Early benchmarking efforts focused on measuring the bandwidth the devices can deliver. Plots in Figure<br />

9.21 show the performance <strong>of</strong> the devices across a range <strong>of</strong> blocks sizes for both sequential read and write<br />

operations. As expected the PCIe attached devices outperformed the SATA attached devices. More interestingly,<br />

the solid-state devices are still sensitive to the block size <strong>of</strong> the IO operation. This is likely due to<br />

both overheads in the kernel IO stack, as well as additional transaction overheads in the devices. In general,<br />

the card devices provide more balanced performance for writes versus reads compared to the MLC based<br />

72

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!