Magellan Final Report - Office of Science - U.S. Department of Energy
Magellan Final Report - Office of Science - U.S. Department of Energy
Magellan Final Report - Office of Science - U.S. Department of Energy
You also want an ePaper? Increase the reach of your titles
YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.
<strong>Magellan</strong> <strong>Final</strong> <strong>Report</strong><br />
the EBS corresponding volumes. This is especially true in the Amazon small instances that are bandwidth<br />
limited. At least during the duration <strong>of</strong> our tests, we noticed that the larger instance types showed lesser<br />
advantage with the local disks, possibly due to the increased network bandwidth available in the large instances.<br />
However, users need to do a performance-cost analysis <strong>of</strong> local disk vs EBS for their particular<br />
application [47].<br />
Instance Type. The anticipated I/O performance on the instance type is expected to get better with the<br />
larger better instances. However, in our limited testing, we encountered situations where we were able to get<br />
better I/O performance on the small instance local disk than the large instance. Our tests also show that<br />
the small instances do tend to show a fair amount <strong>of</strong> variability, and hence more extensive testing might be<br />
needed to capture these behaviors over time. The EBS performance appeared to improve with more capable<br />
instance types, possibly due to the better network available to the larger and/or the CC instances.<br />
Availability Regions. We observed that the west zone performance was better than the performance <strong>of</strong><br />
the east zone. The west zone VMs on Amazon have slightly higher price points, possibly resulting in better<br />
performance. However, our large-scale tests also show that the west zone has a higher standard deviation<br />
than the east zone.<br />
9.4 Flash Benchmarking<br />
Solid-state storage (SSS) is poised as a disruptive technology. This impact would likely affect both the<br />
cloud computing and scientific computing spaces. For these reasons, flash storage evaluation was included<br />
in the <strong>Magellan</strong> project. <strong>Magellan</strong> at NERSC first evaluated several products before deploying a larger<br />
storage system based on that evaluation. The technologies that were evaluated include three PCIe connected<br />
solutions and two SATA connected solutions. Table 9.5 summarizes the products that were evaluated.<br />
Table 9.5: Summary <strong>of</strong> flash-based storage products that were evaluated.<br />
Manufacturer Product Capacity<br />
PCIe Attached Devices - All use SLC Flash<br />
Texas Memory Systems RamSAN 20 450 GB<br />
FusionIO ioDrive Duo 320 GB<br />
Virident tachIOn 400 GB<br />
SATA Attached Devices - Both use MLC Flash<br />
Intel X25-M 160 GB<br />
OCZ Colossus 250 GB<br />
NAND flash devices have dramatically different performance characteristics compared with traditional<br />
disk systems. NAND typically delivers much higher random read rates. However, NAND chips must be<br />
erased prior to writing new data. This erase cycle can be extremely slow. For example, NAND may require<br />
several milliseconds to erase a block, yet can perform a read <strong>of</strong> block in several microseconds [79]. To help<br />
mask this impact, many high-end devices utilize background grooming cycles to maintain a pool <strong>of</strong> erased<br />
blocks available for new writes.<br />
Early benchmarking efforts focused on measuring the bandwidth the devices can deliver. Plots in Figure<br />
9.21 show the performance <strong>of</strong> the devices across a range <strong>of</strong> blocks sizes for both sequential read and write<br />
operations. As expected the PCIe attached devices outperformed the SATA attached devices. More interestingly,<br />
the solid-state devices are still sensitive to the block size <strong>of</strong> the IO operation. This is likely due to<br />
both overheads in the kernel IO stack, as well as additional transaction overheads in the devices. In general,<br />
the card devices provide more balanced performance for writes versus reads compared to the MLC based<br />
72