09.05.2014 Views

FY2010 - Oak Ridge National Laboratory

FY2010 - Oak Ridge National Laboratory

FY2010 - Oak Ridge National Laboratory

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

Director’s R&D Fund—<br />

Ultrascale Computing and Data Science<br />

vector fields at the boundary and (2) this approach admits different numerical discretizations, such as<br />

finite elements or finite differences, in each physical domain, enabling a wide range of applications to<br />

make use of this technique.<br />

This project proposed a new general optimization-based approach where we minimize the difference<br />

between scalar and/or vector fields at the domain interfaces subject to the constraining physics in each<br />

domain. This year, we demonstrated that this proposed approach is a generalization of the widely<br />

accepted mortar methods and will have applicability to a much wider class of problems. Additionally,<br />

with an appropriate choice of regularization (either explicit based on generalized cross-validation or<br />

implicit based on a premature solver termination dictated by an L-curve), this approach is as efficient and<br />

potentially more robust than the existing methods. To demonstrate this, we examined three different<br />

classes of problems. First, for a diffusion-reaction problem with an arbitrary number of domains and<br />

mismatched discretizations, we showed that the error convergence, as expected, was provably either<br />

second (in the H 1 norm) or third (in the L 2 norm) order. Second, we demonstrated the method accurately<br />

captured the fluid-structure coupling interactions when the fluid is governed by the nonlinear Navier-<br />

Stokes equations and the structure was linearly elastic. Finally, we showed that this method can<br />

efficiently represent disparate timescales for a conjugate heat transfer problem with a moderately large<br />

Rayleigh number (O(10 4 )). Lastly, this approach requires little additional implementation effort in<br />

existing software efforts, as was demonstrated by using the large-scale LifeV (www.lifev.org) package for<br />

these problems.<br />

05243<br />

MPI-3: Programming Model Support for Ultrascale Computer Systems<br />

Richard L. Graham, Thomas Naughton, and Chao Wang<br />

Project Description<br />

Upcoming generations of ultrascale computer systems promise an unprecedented level of computational<br />

capabilities and hand in hand provide a challenge to use these systems effectively. Hardware technology<br />

challenges are driving these systems to be many-core and multicore systems with immense component<br />

counts, and simulation codes, middleware, and system-level software need to be able to run in the face of<br />

errors. We will work with the application developer aiming to run on these systems to develop scalable<br />

strategies for dealing with such failures at the application and middleware level. We will investigate how<br />

to partition the solution between these two levels in the context of the ubiquitous communication<br />

standard, the Message Passing Interface (MPI) standard, and present proposed changes to this standard to<br />

the MPI Forum for inclusion in the MPI-3.0 standard. We will look at solutions that aim to avoid failures<br />

and at scalable mechanisms to recover once such failures have occurred.<br />

Mission Relevance<br />

DOE makes large investments in such areas as climate change, energy science and technology, and<br />

material science, with simulation playing an important role in the discovery process. To meet the<br />

simulation needs, DOE develops increasingly powerful and complex computer systems. This project aims<br />

to provide a solution to one of the pressing issues facing those trying to use these increasingly complex<br />

systems—harnessing the full potential of these systems in the face of component failures. With MPI<br />

being the ubiquitous parallel communications and process control library used by scientific simulation<br />

codes, providing fault tolerance in support of MPI is a key step in developing fault-tolerant applications.<br />

By adding fault-tolerant capabilities to the MPI 3.0 standard, adding uncoordinated checkpoint/restart<br />

71

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!