13.07.2015 Views

Direct Numerical Simulation of Autoignition in a Jet in a Cross-Flow ...

Direct Numerical Simulation of Autoignition in a Jet in a Cross-Flow ...

Direct Numerical Simulation of Autoignition in a Jet in a Cross-Flow ...

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

ANL/ALCF/ESP-13/3<strong>Direct</strong> <strong>Numerical</strong> <strong>Simulation</strong> <strong>of</strong> <strong>Autoignition</strong> <strong>in</strong> a <strong>Jet</strong><strong>in</strong> a <strong>Cross</strong>-<strong>Flow</strong> ConfigurationALCF-2 Early Science Program Technical ReportArgonne Leadership Comput<strong>in</strong>g Facility


About Argonne National LaboratoryArgonne is a U.S. Department <strong>of</strong> Energy laboratory managed by UChicago Argonne, LLCunder contract DE-AC02-06CH11357. The Laboratory’s ma<strong>in</strong> facility is outside Chicago,at 9700 South Cass Avenue, Argonne, Ill<strong>in</strong>ois 60439. For <strong>in</strong>formation about Argonneand its pioneer<strong>in</strong>g science and technology programs, see www.anl.gov.Availability <strong>of</strong> This ReportThis report is available, at no cost, at http://www.osti.gov/bridge. It is also availableon paper to the U.S. Department <strong>of</strong> Energy and its contractors, for a process<strong>in</strong>g fee, from:U.S. Department <strong>of</strong> EnergyOffice <strong>of</strong> Scientific and Technical InformationP.O. Box 62Oak Ridge, TN 37831-0062phone (865) 576-8401fax (865) 576-5728reports@adonis.osti.govDisclaimerThis report was prepared as an account <strong>of</strong> work sponsored by an agency <strong>of</strong> the United States Government. Neither the United StatesGovernment nor any agency there<strong>of</strong>, nor UChicago Argonne, LLC, nor any <strong>of</strong> their employees or <strong>of</strong>ficers, makes any warranty, expressor implied, or assumes any legal liability or responsibility for the accuracy, completeness, or usefulness <strong>of</strong> any <strong>in</strong>formation, apparatus,product, or process disclosed, or represents that its use would not <strong>in</strong>fr<strong>in</strong>ge privately owned rights. Reference here<strong>in</strong> to any specificcommercial product, process, or service by trade name, trademark, manufacturer, or otherwise, does not necessarily constitute or implyits endorsement, recommendation, or favor<strong>in</strong>g by the United States Government or any agency there<strong>of</strong>. The views and op<strong>in</strong>ions <strong>of</strong>document authors expressed here<strong>in</strong> do not necessarily state or reflect those <strong>of</strong> the United States Government or any agency there<strong>of</strong>,Argonne National Laboratory, or UChicago Argonne, LLC.


ANL/ALCF/ESP-13/3<strong>Direct</strong> <strong>Numerical</strong> <strong>Simulation</strong> <strong>of</strong> <strong>Autoignition</strong> <strong>in</strong> a <strong>Jet</strong> <strong>in</strong> a <strong>Cross</strong>-<strong>Flow</strong> ConfigurationALCF-2 Early Science Program Technical Reportprepared byAmmar Abdilghanie 1 , Christos E. Frouzakis 2 , Paul F. Fischer 31 Argonne Leadership Comput<strong>in</strong>g Facility, Argonne National Laboratory2 Federal Institute <strong>of</strong> Technology Zürich (ETH Zürich)3 Math and Computer Science, Argonne National LaboratoryMay 7, 2013


<strong>Direct</strong> <strong>Numerical</strong> <strong>Simulation</strong> <strong>of</strong> <strong>Autoignition</strong> <strong>in</strong> a<strong>Jet</strong> <strong>in</strong> a <strong>Cross</strong>-<strong>Flow</strong> ConfigurationAmmar Abdilghanie 11 , Christos E. Frouzakis 22 and Paul F. Fischer 331 Leadership Comput<strong>in</strong>g Facility, Argonne National Laboratory, Argonne,IL 604392 Aerothermochemistry and Combustion Systems Laboratory, SwissFederal Institute <strong>of</strong> Technology Zurich (ETHZ), Zurich, Switzerland3 Mathematics and Computer Science Division, Argonne NationalLaboratory, Argonne, IL 60439April 3, 20131 aabdilghanie@alcf.anl.gov2 frouzakis@lav.mavt.ethz.ch3 fischer@mcs.anl.gov


Abstract<strong>Autoignition</strong> <strong>in</strong> turbulent flows is a challeng<strong>in</strong>g fundamental problem due tothe <strong>in</strong>tricate coupl<strong>in</strong>g <strong>of</strong> different physical and chemical processes extend<strong>in</strong>g overmultiple flow and chemistry scales. At the same time, the improved understand<strong>in</strong>gand ability to predict autoignition <strong>in</strong> flows characterized by considerable fluctuations<strong>of</strong> velocity, composition, and temperature is essential for the development <strong>of</strong>novel low-emission concepts for power generation. The aim <strong>of</strong> this project is tostudy the fundamental aspects <strong>of</strong> autoignition <strong>in</strong> a fuel-air mix<strong>in</strong>g device directlyapplicable to mix<strong>in</strong>g ducts <strong>in</strong> gas turb<strong>in</strong>es. The NEK5000-based code for low Machnumber reactive flows is used to perform very large scale direct numerical simulations<strong>of</strong> autoignition <strong>of</strong> a diluted hydrogen jet ejected <strong>in</strong> a cross-flow<strong>in</strong>g stream<strong>of</strong> hot turbulent air <strong>in</strong> a laboratory-scale configuration. We report on our experiencerunn<strong>in</strong>g NEK5000 on the new BGQ system at ALCF mira dur<strong>in</strong>g the earlyscience period (ESP). First <strong>of</strong> all, the most efficient problem size per MPI-rank isobta<strong>in</strong>ed through core-level efficiency metric measured from the target simulation.Furthermore, the most efficient number <strong>of</strong> ranks is found through strong scal<strong>in</strong>g experiments.F<strong>in</strong>ally, low-level <strong>in</strong>sight <strong>in</strong>to the observed parallel efficiency is enabledthrough IBM’s HPC Toolkit libraries.


Contents1 Introduction 11.1 Overview <strong>of</strong> the <strong>Numerical</strong> Method . . . . . . . . . . . . . . . . 11.1.1 Summary <strong>of</strong> <strong>Numerical</strong> <strong>Simulation</strong>s . . . . . . . . . . . . 32 Parallel Scal<strong>in</strong>g and parallel efficiency metrics 42.1 Measurement <strong>of</strong> Parallel Efficiency . . . . . . . . . . . . . . . . . 42.1.1 Core-level MPI Thread<strong>in</strong>g Efficiency . . . . . . . . . . . 42.1.2 Strong Scal<strong>in</strong>g Experiment . . . . . . . . . . . . . . . . . 52.2 Pr<strong>of</strong>il<strong>in</strong>g and Hardware Performance Monitor<strong>in</strong>g . . . . . . . . . 52.2.1 IBM HPC Toolkit . . . . . . . . . . . . . . . . . . . . . . 62.2.2 Nek5000 Instrumentation . . . . . . . . . . . . . . . . . . 72.2.3 Sample MPI Pr<strong>of</strong>ile . . . . . . . . . . . . . . . . . . . . . 72.2.4 Hardware Performance . . . . . . . . . . . . . . . . . . . 83 Discussion <strong>of</strong> ESP Experience 10Bibliography 11i


Chapter 1IntroductionUnderstand<strong>in</strong>g the conditions under which autoignition <strong>of</strong> react<strong>in</strong>g mixtures occursis <strong>of</strong> primary importance for the design and operation <strong>of</strong> modern lean premixed(LP) and lean premixed prevaporized (LPP) combustion devices such as low-NO xstationary gas turb<strong>in</strong>es (Dobbel<strong>in</strong>g et al., 2007) and propulsion devices such as subsonicramjets and supersonic scramjets (Micka and Driscoll, 2012). In particular,the enhanced turbulent mix<strong>in</strong>g between the fuel and oxidizer streams <strong>of</strong> the jet <strong>in</strong>cross-flow (JICF) configuration makes it an essential component <strong>in</strong> the design <strong>of</strong>premix<strong>in</strong>g sections <strong>of</strong> modern high efficiency, low NO x combustors.The ma<strong>in</strong> objective <strong>of</strong> this study is to <strong>in</strong>vestigate the sensitivity <strong>of</strong> the JICF dynamicsto the cross-flow temperature. In particular we ask the follow<strong>in</strong>g questions:When and where do localized flame kernels form? Could the observed ignitiontime be correlated to ignition-delay time <strong>in</strong> the correspond<strong>in</strong>g homogeneous system?Do these kernels develop <strong>in</strong>to stable flame later on? How does the mixtureget prepared for ignition and what is the role <strong>of</strong> the vorticity and pressure <strong>in</strong> mix<strong>in</strong>gthe jet (fuel) with the cross-flow (oxidizer) fluid? What is the local combustionmode associated with the observed ignition scenarios?1.1 Overview <strong>of</strong> the <strong>Numerical</strong> MethodA weak formulation <strong>of</strong> the compressible reactive Navier-Stokes for ideal gas mixturesat the low Mach number limit were solved, us<strong>in</strong>g a parallel spectral-elementcode based on NEK5000 (Fischer et al., 2008). The spectral element methodis a high-order weighted residual technique that couples the rapid convergence<strong>of</strong> global spectral methods with the geometric flexibility <strong>of</strong> f<strong>in</strong>ite element methods(Deville et al., 2002).A high-order operator splitt<strong>in</strong>g technique is used to split the thermochemistry(species and energy equations) from the hydrodynamic subsystem (cont<strong>in</strong>uity andmomentum equations) (Tomboulides et al., 1997). The latter is <strong>in</strong>tegrated <strong>in</strong> timeus<strong>in</strong>g a 3rd-order semi-implicit method, with explicit treatment <strong>of</strong> the nonl<strong>in</strong>earterms and implicit treatment <strong>of</strong> the viscous and pressure terms, while the thermo-1


Figure 1.1: Snapshots <strong>of</strong> temperature and mass fractions <strong>of</strong> HO 2 , OH for the T cf =950 K case on a cross section through the symmetry plane.chemistry subsystem is solved fully implicitly by CVODE, a scalable BDF-basedstiff ODE solver (Cohen and H<strong>in</strong>dmarsh, 1996). Detailed chemistry, thermodynamicproperties and mixture-averaged transport properties are evaluated us<strong>in</strong>gChemk<strong>in</strong> (Kee et al., 1986).Tensor product factorization is the core computational kernel <strong>in</strong> NEK5000. Itis used to evaluate spectral element operators, <strong>in</strong>terpolation, <strong>in</strong>tegration and differentiation.It is cast <strong>in</strong> terms <strong>of</strong> a sequence <strong>of</strong> dense and non-square matrix products(mxm). Optimiz<strong>in</strong>g the mxm kernel was performed on the BG/P us<strong>in</strong>g loop til<strong>in</strong>gand architecture-specific assembly code (Kerkemeier, 2010). The code developmentsresulted <strong>in</strong> improved core throughput and cache utilization (through datareuse) compared to highly optimized BLAS libraries. F<strong>in</strong>ally, Kerkemeier (2010)developed a chemistry calculation kernel for the computation <strong>of</strong> the productionrates, transport and thermodynamic property evaluation. All expensive computationswere vectorized us<strong>in</strong>g the highly optimized math libraries MASS/MASSVwhich allowed best float<strong>in</strong>g po<strong>in</strong>t, load and store performance.2


1.1.1 Summary <strong>of</strong> <strong>Numerical</strong> <strong>Simulation</strong>sExploratory numerical simulations on experimental computational grids as well asproduction runs were performed dur<strong>in</strong>g the early science period on Mira. The computationalgrid is composed <strong>of</strong> N E = 1, 591, 752 spectral elements with<strong>in</strong> which thesolution is <strong>in</strong>terpolated us<strong>in</strong>g three-dimensional tensorial product <strong>of</strong> p=6-th orderLagrange polynomials.Post process<strong>in</strong>g <strong>of</strong> the solution was performed us<strong>in</strong>g NEK5000 on the twodatasets <strong>in</strong>clud<strong>in</strong>g computations <strong>of</strong> key species generation rates, mixture fraction,scalar dissipation rate, Takeno flame <strong>in</strong>dex and vorticity components. Analysis<strong>of</strong> the two large data sets was also ma<strong>in</strong>ly facilitated through visualizations withVisIt on the ALCF’s GPU system (Eureka). Two crossflow stream temperatures(T cf =930 and 950 K) were simulated at a Reynolds number, based on the frictionvelocity and the channel half-width, <strong>of</strong> Re τ = 180. Typical <strong>in</strong>stantaneous snapshots<strong>of</strong> the fully ignited case are shown <strong>in</strong> fig. 1.1.The simulations performed up till now employed close to 345 million gridpo<strong>in</strong>ts correspond<strong>in</strong>g to 4.8 billion degrees <strong>of</strong> freedom (14 unknowns per gridpo<strong>in</strong>t). We are currently develop<strong>in</strong>g the computational mesh for a planned highReynolds number simulation (Re τ = 590). It is estimated that the total number <strong>of</strong>elements will be close to 3 millions and the polynomial order will be at least 8.3


Chapter 2Parallel Scal<strong>in</strong>g and parallelefficiency metrics2.1 Measurement <strong>of</strong> Parallel EfficiencyParallel code scal<strong>in</strong>g focuses on one <strong>of</strong> two forms: strong scal<strong>in</strong>g or weak scal<strong>in</strong>g.The goal <strong>of</strong> strong scal<strong>in</strong>g is to reduce execution time for a fixed total problem sizeby add<strong>in</strong>g processors (and hence reduc<strong>in</strong>g problem size per worker/MPI rank). Onthe other hand, ideal weak scal<strong>in</strong>g behavior is to keep the execution time constantby add<strong>in</strong>g processors <strong>in</strong> proportion to an <strong>in</strong>creas<strong>in</strong>gly larger problem size (andhence keep<strong>in</strong>g problem size per worker fixed).Parallel efficiency,η, for a problem run on N 1 processes/MPI ranks is def<strong>in</strong>edrelative to a reference run on N 2 ranks asη p = N 1 t(N 1 )N 2 t(N 2 ) ,where t(N) is the execution time on N ranks.In order to ensure core utilization we measure the MPI-thread<strong>in</strong>g efficiencywhich we def<strong>in</strong>e asη t = N 1 t(N 1 )N 2 t(N 2 ) ,where N 1 is the number <strong>of</strong> cores for the one MPI rank/thread per core configurationand N 2 = 2N 1 <strong>in</strong> the two threads/core and N 2 = 4N 1 <strong>in</strong> the four threads/coreconfigurations. Note that a net speedup is obta<strong>in</strong>ed for thread<strong>in</strong>g efficiency <strong>of</strong>more than 50% <strong>in</strong> the two threads/core configuration and more than 25% <strong>in</strong> thefour threads/core.2.1.1 Core-level MPI Thread<strong>in</strong>g EfficiencyInstantiation <strong>of</strong> two and four MPI-ranks per core is enabled on Mira and hence fullcore compute power utilization is assured through the use <strong>of</strong> more than one rank per4


Figure 2.1: MPI thread<strong>in</strong>g efficiency for autoignition simulation us<strong>in</strong>g NEK5000.Note that the abscissa represents grid size per MPI rank.core. However, this may lead to performance degradation for very large problemsize because <strong>of</strong> the result<strong>in</strong>g on-chip resource contention. We have conducted acore-level parallel efficiency measurements on Mira for the target simulations us<strong>in</strong>gdifferent number <strong>of</strong> cores and either two or four ranks per code.Figure 2.1 compares the MPI thread<strong>in</strong>g efficiency for two and four ranks percore. It is clear that us<strong>in</strong>g two ranks per core is more efficient for same problemsize (i.e. grid po<strong>in</strong>ts) per rank and that the optimum grid po<strong>in</strong>ts per rank is 7,000-10,000. This size ensures maximum utilization <strong>of</strong> the core’s compute-power whilem<strong>in</strong>imiz<strong>in</strong>g on-chip resource contention among MPI threads.2.1.2 Strong Scal<strong>in</strong>g ExperimentA strong scal<strong>in</strong>g is performed under the two MPI-ranks per core configuration foran autoignition simulation with a total <strong>of</strong>≈ 345 million grid po<strong>in</strong>ts.. Figure 2.2shows that an ideal efficiency <strong>of</strong> a 100% is ma<strong>in</strong>ated up to approximately 130,000MPI-ranks (65,000 cores) and that up to 60% efficiency is susta<strong>in</strong>able us<strong>in</strong>g nearly500,0000 ranks.2.2 Pr<strong>of</strong>il<strong>in</strong>g and Hardware Performance Monitor<strong>in</strong>gPerformance monitor<strong>in</strong>g is enabled through a mechanism for obta<strong>in</strong><strong>in</strong>g <strong>in</strong>formationabout the use <strong>of</strong> MPI rout<strong>in</strong>es (pr<strong>of</strong>il<strong>in</strong>g) or wall clock time and hardware counters5


Figure 2.2: Parallel efficiency for autoignition simulation: strong scal<strong>in</strong>g with≈345 Million gridpo<strong>in</strong>ts and two MPI ranks per core.for a code as a whole or a segment that is usually bracketed with calls to <strong>in</strong>itiateand stop monitor<strong>in</strong>g libraries.2.2.1 IBM HPC ToolkitIBMs’ HPC Toolkit provides a mechanism for track<strong>in</strong>g the use <strong>of</strong> MPI rout<strong>in</strong>esdur<strong>in</strong>g a programs execution. This is done through the use <strong>of</strong> a library which <strong>in</strong>terceptscalls to MPI rout<strong>in</strong>es, records <strong>in</strong>formation about the call, and then cont<strong>in</strong>ueswith the MPI call. The primary types <strong>of</strong> <strong>in</strong>formation that are gathered are• MPI Pr<strong>of</strong>ile Data: A summary <strong>of</strong> MPI usage <strong>in</strong>formation typically consist<strong>in</strong>g<strong>of</strong> the number <strong>of</strong> times each MPI rout<strong>in</strong>e was called, how much time wasspent <strong>in</strong> each MPI rout<strong>in</strong>e, and the average size <strong>of</strong> a message for that rout<strong>in</strong>e.• MPI Trace Data: A detailed time-history <strong>of</strong> every <strong>in</strong>vocation <strong>of</strong> an MPI rout<strong>in</strong>ethat the program made.• Po<strong>in</strong>t-to-Po<strong>in</strong>t Communication Pattern: This <strong>in</strong>formation is derived from thecollected trace <strong>in</strong>formation and <strong>in</strong>dicates the number <strong>of</strong> bytes sent betweenranks by po<strong>in</strong>t-to-po<strong>in</strong>t MPI rout<strong>in</strong>es.IBM also provides a HPM Toolkit for hardware perfromance monitor<strong>in</strong>g thathas6


• A utility hpmcount, which starts an application and provides at the end <strong>of</strong>execution the wall clock time, hardware performance counters <strong>in</strong>foramation,derived hardware metrics, and resource utilization statistics• An <strong>in</strong>strumentation library libhpm, which provides <strong>in</strong>strumented programswith a summary output conta<strong>in</strong><strong>in</strong>g the above <strong>in</strong>formation for each <strong>in</strong>strumentedregion <strong>in</strong> a program (resource utilization statistics is provided onlyonce, for the whole section <strong>of</strong> the program that is <strong>in</strong>strumented). This librarysupports serial and parallel (MPI, threaded, and mixed mode) applications,written <strong>in</strong> Fortran, C, and C++.2.2.2 Nek5000 InstrumentationIn order to focus the analysis on the time stepper part <strong>of</strong> NEK5000, exclude restartfileread<strong>in</strong>g and recurr<strong>in</strong>g IO, the time stepp<strong>in</strong>g loop was bracketed by calls to therespective libraries as follows:call summary_start()call hpm_start("nek_advance")DO ISTEP=1,NSTEPSTIME STEPPING LOOPENDDOcall hpm_stop("nek_advance")call summary_stop()It should be noted that the code needs to be compiled with the debugg<strong>in</strong>g flag ”-g”enabled. F<strong>in</strong>ally, the follow<strong>in</strong>g l<strong>in</strong>k to IBM libraries needs to be appended to thelist <strong>of</strong> libraries (USER_FLAGS) used by NEK5000:USR_LFLAGS="${USR_LFLAGS} -L/s<strong>of</strong>t/perftools/hpctw -lmpihpm-L/bgsys/drivers/ppcfloor/bgpm/lib -lbgpm"2.2.3 Sample MPI Pr<strong>of</strong>ileA small section <strong>of</strong> a sample pr<strong>of</strong>ile dumped by rank 0 is shown below.Data for MPI rank 0 <strong>of</strong> 65536Times and statistics from summary_start() to summary_stop().-----------------------------------------------------------------MPI Rout<strong>in</strong>e #calls avg. bytes time(sec)-----------------------------------------------------------------MPI_Isend 508287 915.7 2.083MPI_Irecv 508287 916.1 0.416MPI_Waitall 162759 0.0 8.845MPI_Bcast 1 4.0 0.000MPI_Barrier 81 0.0 0.0027


MPI_Allreduce 19296 2716.7 17.573...etc.-----------------------------------------------------------------total communication time = 28.919 seconds.total elapsed time = 195.156 seconds.heap memory used = 103.090 MBytes.Although the MPI-time is dom<strong>in</strong>ated by synchronization (MPI_Allreducedur<strong>in</strong>g global <strong>in</strong>ner product evaluation for the l<strong>in</strong>ear system solve step for the velocity,pressure and thermo-chemistry), it is still represent<strong>in</strong>g only about 14% <strong>of</strong> thetotal execution time. The communication as well as total execution/elapsed timefor all MPI-ranks is plotted <strong>in</strong> figure 2.3. It can be seen that the communicationtime as well the total elapsed time (and hence the computation time) is approximatelyuniform across all MPI-ranks, an <strong>in</strong>dication <strong>of</strong> load balanc<strong>in</strong>g between theMPI ranks.Figure 2.3: Communication and elapsed time for MPI ranks collected with IBM’sHPC Toolkit2.2.4 Hardware PerformanceHardware performance metrics are written tohpm_summary.rank files. The filelists <strong>in</strong>formation about the overall performance <strong>of</strong> the whole code as well as <strong>in</strong>formationabout the bracketed segment <strong>of</strong> the code (the time stepper <strong>in</strong> this case).Table 2.1 shows that with the use <strong>of</strong> two MPI-ranks per core the float<strong>in</strong>g po<strong>in</strong>toperations per second per node has <strong>in</strong>creased from 4.398 to 6.057 GFLOPS ensur<strong>in</strong>ghigher utilization <strong>of</strong> the core compute-power. F<strong>in</strong>ally, it can be seen that there8


is a trend <strong>of</strong> <strong>in</strong>creas<strong>in</strong>g access to both the L1P, L2 Cache and ma<strong>in</strong> memory with<strong>in</strong>creas<strong>in</strong>g number <strong>of</strong> ranks per core. This is most likely the reason beyond performancedegradation for the four ranks per core configuration. Measurements froma four ranks per core were not possible due to the overhead <strong>of</strong> the measurement librariesthemselves, even with light-weight versions <strong>of</strong> the same libraries providedby IBM.Table 2.1: Summary <strong>of</strong> Hardware Performance Metrics Measured by hpm Library1 Rank/Core 2 Ranks/CoreFPU% , FXU % 20.6-79.4 17.7-82.3Inst/cycle/core 0.3586 0.5794GFLOPS/node 4.398 6.057% <strong>of</strong> max issue rate/core 35.9 47.7L1 d-cache hit 96.4 95.5L1P buffer hit 1.46 2.09L2 Cache hit 1.85 2.0DDR hit 0.32 0.4DDR total traffic(Bytes/cycle)2.264 3.5889


Chapter 3Discussion <strong>of</strong> ESP ExperienceThe highly-scalable low Mach number reactive flow solver based on NEK5000 isused to simulate turbulent autoignit<strong>in</strong>g flows <strong>in</strong> a laboratory scale jet-<strong>in</strong>-cross-flowconfiguration. The code relies heavily on two highly optimized mxm and thermochemistryPTTP kernels that were previously optimized on the BGP architecture(Kerkemeier, 2010).Scal<strong>in</strong>g results on Mira showed that ideal parallel efficiency <strong>of</strong> 100% can besusta<strong>in</strong>ed up to 130000 MPI ranks and up to 60% efficiency on nearly 500,0000ranks and hence no further tun<strong>in</strong>g is necessary for the current problem size.We are currently evaluation different design strategies for the computationalgrid that will be used for the high Reynolds number simulations. Further optimization<strong>of</strong> grid-generation rout<strong>in</strong>es for the Algebraic Multigrid solver is necessary fortackl<strong>in</strong>g the high Reynolds number grids and is the subject <strong>of</strong> ongo<strong>in</strong>g <strong>in</strong>vestigation.10


BibliographyCohen, S. and H<strong>in</strong>dmarsh, A. (1996) CVODE, a stiff/nonstiff ODE solver <strong>in</strong> C.Comp. Phys. 10(2), 138–143.Deville, M., Fischer, P. and Mund, E. (2002) High-order methods for <strong>in</strong>compressiblefluid flow. Cambridge University Press.Dobbel<strong>in</strong>g, K., Hellat, J. and Koch, H. (2007) 25 years <strong>of</strong> BBC/ABB/Alstom leanpremix combustion technologies. Trans.-ASME J. Eng. Gas Turb. Power 129(1),2.Fischer, P., J.W., L. and Kerkemeier, S. (2008) nek5000 Web page.http://nek5000.mcs.anl.gov.Kee, R., Dixon-Lewis, G., Warnatz, J., Coltr<strong>in</strong>, M. and Miller, J. (1986) Technicalreport sand86-8246. Sandia National Laboratories .Kerkemeier, S. (2010) <strong>Direct</strong> numerical simulation <strong>of</strong> combustion on petascaleplatforms. Ph.D. thesis, Swiss Federal Institute <strong>of</strong> Techonolgy Zurich (ETHZ),Nr. 19162.Micka, D. and Driscoll, J. (2012) Stratified jet flames <strong>in</strong> a heated air cross-flowwith autoignition. Combust. Flame 159(3), 1205–1214.Tomboulides, A., Lee, J. and Orszag, S. (1997) <strong>Numerical</strong> simulation <strong>of</strong> low machnumber reactive flows. J. Sci. Comp. 12(2), 139–167.11


Argonne Leadership Comput<strong>in</strong>g FacilityArgonne National Laboratory9700 South Cass Avenue, Bldg. 240Argonne, IL 60439www.anl.govArgonne National Laboratory is a U.S. Department <strong>of</strong> Energylaboratory managed by UChicago Argonne, LLC

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!