11.12.2012 Views

NASA Scientific and Technical Aerospace Reports

NASA Scientific and Technical Aerospace Reports

NASA Scientific and Technical Aerospace Reports

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

60<br />

COMPUTER OPERATIONS AND HARDWARE<br />

Includes hardware for computer graphics, firmware <strong>and</strong> data processing. For components see 33 Electronics <strong>and</strong> Electrical Engineering.<br />

For computer vision see 63 Cybernetics, Artificial Intelligence <strong>and</strong> Robotics.<br />

20040111318 <strong>NASA</strong> Langley Research Center, Hampton, VA, USA<br />

Scalability of Parallel Spatial Direct Numerical Simulations on Intel Hypercube <strong>and</strong> IBM SP1 <strong>and</strong> SP2<br />

Joslin, Ronald D.; Hanebutte, Ulf R.; Zubair, Mohammad; [1995]; 54 pp.; In English<br />

Contract(s)/Grant(s): NAS1-19480; Copyright; Avail: CASI; A04, Hardcopy<br />

The implementation <strong>and</strong> performance of a parallel spatial direct numerical simulation (PSDNS) approach on the Intel<br />

iPSC/860 hypercube <strong>and</strong> IBM SP1 <strong>and</strong> SP2 parallel computers is documented. Spatially evolving disturbances associated with<br />

the laminar-to-turbulent transition in boundary-layer flows are computed with the PSDNS code. The feasibility of using the<br />

PSDNS to perform transition studies on these computers is examined. The results indicate that PSDNS approach can<br />

effectively be parallelized on a distributed-memory parallel machine by remapping the distributed data structure during the<br />

course of the calculation. Scalability information is provided to estimate computational costs to match the actual costs relative<br />

to changes in the number of grid points. By increasing the number of processors, slower than linear speedups are achieved<br />

with optimized (machine-dependent library) routines. This slower than linear speedup results because the computational cost<br />

is dominated by FFT routine, which yields less than ideal speedups. By using appropriate compile options <strong>and</strong> optimized<br />

library routines on the SP1, the serial code achieves 52-56 M ops on a single node of the SP1 (45 percent of theoretical peak<br />

performance). The actual performance of the PSDNS code on the SP1 is evaluated with a ‘real world’ simulation that consists<br />

of 1.7 million grid points. One time step of this simulation is calculated on eight nodes of the SP1 in the same time as required<br />

by a Cray Y/MP supercomputer. For the same simulation, 32-nodes of the SP1 <strong>and</strong> SP2 are required to reach the performance<br />

of a Cray C-90. A 32 node SP1 (SP2) configuration is 2.9 (4.6) times faster than a Cray Y/MP for this simulation, while the<br />

hypercube is roughly 2 times slower than the Y/MP for this application. KEY WORDS: Spatial direct numerical simulations;<br />

incompressible viscous flows; spectral methods; finite differences; parallel computing.<br />

Author<br />

Computers; Direct Numerical Simulation; Hypercube Multiprocessors; Parallel Computers; Parallel Processing<br />

(Computers); Supercomputers<br />

20040111377 <strong>NASA</strong> Langley Research Center, Hampton, VA, USA<br />

Experiences From <strong>NASA</strong>/Langley’s DMSS Project<br />

[1996]; 13 pp.; In English; 14th IEEE Symposium on Mass Storage Systems; No Copyright; Avail: CASI; A03, Hardcopy<br />

There is a trend in institutions with high performance computing <strong>and</strong> data management requirements to explore mass<br />

storage systems with peripherals directly attached to a high speed network. The Distributed Mass Storage System (DMSS)<br />

Project at the <strong>NASA</strong> Langley Research Center (LaRC) has placed such a system into production use. This paper will present<br />

the experiences, both good <strong>and</strong> bad, we have had with this system since putting it into production usage. The system is<br />

comprised of: 1) National Storage Laboratory (NSL)/UniTree 2.1, 2) IBM 9570 HIPPI attached disk arrays (both RAID 3 <strong>and</strong><br />

RAID 5), 3) IBM RS6000 server, 4) HIPPI/IPI3 third party transfers between the disk array systems <strong>and</strong> the supercomputer<br />

clients, a CRAY Y-MP <strong>and</strong> a CRAY 2, 5) a ‘warm spare’ file server, 6) transition software to convert from CRAY’s Data<br />

Migration Facility (DMF) based system to DMSS, 7) an NSC PS32 HIPPI switch, <strong>and</strong> 8) a STK 4490 robotic library accessed<br />

from the IBM RS6000 block mux interface. This paper will cover: the performance of the DMSS in the following areas: file<br />

transfer rates, migration <strong>and</strong> recall, <strong>and</strong> file manipulation (listing, deleting, etc.); the appropriateness of a workstation class of<br />

file server for NSL/UniTree with LaRC’s present storage requirements in mind the role of the third party transfers between<br />

the supercomputers <strong>and</strong> the DMSS disk array systems in DMSS; a detailed comparison (both in performance <strong>and</strong><br />

functionality) between the DMF <strong>and</strong> DMSS systems LaRC’s enhancements to the NSL/UniTree system administration<br />

environment the mechanism for DMSS to provide file server redundancy the statistics on the availability of DMSS the design<br />

<strong>and</strong> experiences with the locally developed transparent transition software which allowed us to make over 1.5 million DMF<br />

files available to NSL/UniTree with minimal system outage<br />

Author<br />

Supercomputers; Magnetic Disks; Computer Storage Devices; Peripheral Equipment (Computers); Distributed Processing<br />

233

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!