28.08.2016 Views

Iraqi Kurdistan All in the Timing

GEO_ExPro_v12i6

GEO_ExPro_v12i6

SHOW MORE
SHOW LESS

You also want an ePaper? Increase the reach of your titles

YUMPU automatically turns print PDFs into web optimized ePapers that Google loves.

Recent Advances <strong>in</strong> Technology<br />

Supercomputers<br />

for Beg<strong>in</strong>ners PART II<br />

LASSE AMUNDSEN, Statoil,<br />

MARTIN LANDRØ and BØRGE ARNTSEN, NTNU Trondheim<br />

Energy companies need ever larger computers. This article<br />

is <strong>the</strong> second of three articles giv<strong>in</strong>g an <strong>in</strong>troduction to<br />

supercomputers . Here, we look at <strong>the</strong> design of parallel software.<br />

Beta: Software undergoes<br />

beta test<strong>in</strong>g shortly before<br />

it’s released. Beta is Lat<strong>in</strong><br />

for ‘still doesn’t work’.<br />

Unknown<br />

Most supercomputers are multiple<br />

computers that perform parallel<br />

process<strong>in</strong>g, which is well suited for<br />

modell<strong>in</strong>g, simulat<strong>in</strong>g and understand<strong>in</strong>g<br />

many complex, real world phenomena.<br />

Historically, parallel comput<strong>in</strong>g has been<br />

considered ‘<strong>the</strong> high end of comput<strong>in</strong>g’,<br />

referr<strong>in</strong>g to high capability and high<br />

capacity comput<strong>in</strong>g, communication and<br />

storage resources. It has been used to<br />

model difficult problems <strong>in</strong> many areas of<br />

science and eng<strong>in</strong>eer<strong>in</strong>g. Today, however,<br />

a greater driv<strong>in</strong>g force <strong>in</strong> <strong>the</strong> development<br />

of faster computers is provided by<br />

commercial applications that require<br />

process<strong>in</strong>g of large amounts of data <strong>in</strong><br />

very advanced ways.<br />

Embarrass<strong>in</strong>gly Parallel Solutions<br />

In parallel comput<strong>in</strong>g <strong>the</strong> problem<br />

is divided up among a number of<br />

processors. The figure opposite shows<br />

<strong>the</strong> basic computer layout.<br />

The art of parallel programm<strong>in</strong>g<br />

is identify<strong>in</strong>g <strong>the</strong> part of <strong>the</strong> problem<br />

which can be efficiently parallelised. In<br />

many cases this can be straightforward<br />

and simple, such as <strong>in</strong> <strong>the</strong> process<strong>in</strong>g<br />

of seismic data. Seismic data consists<br />

of a (large) number of records, and very<br />

often a s<strong>in</strong>gle or a small number of<br />

records can be processed <strong>in</strong>dependently.<br />

A simplistic example of this k<strong>in</strong>d of<br />

process<strong>in</strong>g is <strong>the</strong> application of various<br />

filter operations. The problem is solved<br />

by distribut<strong>in</strong>g a small number of records<br />

to each processor which performs <strong>the</strong><br />

filter<strong>in</strong>g. This is usually refered to as an<br />

embarrass<strong>in</strong>gly parallel solution and, as<br />

<strong>the</strong> name suggests, does not require much<br />

sophistication, but a supris<strong>in</strong>gly large<br />

number of problems can be solved by<br />

this approach. The ga<strong>in</strong> <strong>in</strong> computer time<br />

needed to solve this k<strong>in</strong>d of problem can<br />

be measured by <strong>the</strong> speedup, which is <strong>the</strong><br />

ratio of computer time needed to solve<br />

<strong>the</strong> problem us<strong>in</strong>g a s<strong>in</strong>gle processor to<br />

<strong>the</strong> time needed to solve <strong>the</strong> problem on<br />

N processors. For embarrass<strong>in</strong>gly parallel<br />

problems <strong>the</strong> speedup is largely equal to<br />

<strong>the</strong> number of processors used. If each<br />

filter operation takes 0.001 second and we<br />

want to filter one million records, <strong>the</strong>n a<br />

s<strong>in</strong>gle processor would use 1,000 seconds,<br />

while a small supercomputer system<br />

with 1,000 processors would need only<br />

one second. This way of solv<strong>in</strong>g parallel<br />

problems is thus highly efficient and<br />

simple to implement.<br />

Message Pass<strong>in</strong>g Interface<br />

However, <strong>in</strong> many cases <strong>the</strong> problem<br />

at hand cannot be split <strong>in</strong>to completely<br />

Titan is a supercomputer built by Cray at Oak Ridge National Laboratory. In November 2012, Titan held<br />

number one position on <strong>the</strong> TOP500 list with 18,688 nodes and 552,960 processors. It relies on a comb<strong>in</strong>ation<br />

of GPUs and traditional CPUs to make it today <strong>the</strong> world’s second most powerful supercomputer, with peak<br />

performance of more than 27 petaflops. Titan also has more than 700 terabytes of memory.<br />

30 GEOExPro November 2015

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!