20.05.2014 Views

link to my thesis

link to my thesis

link to my thesis

SHOW MORE
SHOW LESS

Create successful ePaper yourself

Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.

230 CHAPTER 12. CONCLUSIONS & DIRECTIONS FOR FURTHER RESEARCH<br />

SPP <strong>to</strong>o. A central role in this approach is played by the notion of stable patterns,<br />

which are shifted along the 2D-grid. The same approach leads <strong>to</strong> partial results in<br />

the 3D-case (and higher dimensions).<br />

The results of Section 5.3.2 also underlie the design of the heuristic methods<br />

discussed in Section 5.3.3. These heuristic procedures are developed <strong>to</strong> arrive cheaply<br />

at suboptimal paths with acceptable performance. We have implemented five of these<br />

heuristics and compared their performance.<br />

Another way <strong>to</strong> improve the efficiency of an nD-system is <strong>to</strong> apply parallel computing<br />

techniques as described in Section 5.3.4. Here the ‘least-increments’ method<br />

is implemented in such a way that it computes each path in the nD-system on a<br />

separate processor. The parallel nD-system is applied in combination with the JDQZ<br />

and the JD methods. However, it turns out that parallelization is not useful in this<br />

application of the nD-system.<br />

The use of iterative eigenvalue solvers is described in Chapter 6. The iterative<br />

eigenvalue solver implementations we used in this <strong>thesis</strong> are JDQR, JDQZ and JD. An<br />

advantage of such an iterative eigenvalue solver, is that it is able <strong>to</strong> focus on a subset<br />

of eigenvalues of the involved matrix. Here the most interesting eigenvalue is the<br />

smallest real eigenvalue. Focusing on the smallest real eigenvalue of a matrix is not<br />

a usual parameter setting for an iterative eigenvalue solver. Therefore this option<br />

is developed and implemented as described in Section 6.3: the selection criterion is<br />

adapted such that it is no longer necessary <strong>to</strong> compute all the eigenvalues of the<br />

matrix but that it becomes possible <strong>to</strong> focus on the smallest real eigenvalues.<br />

An important observation is that the eigenvec<strong>to</strong>rs of the matrix at hand exhibit<br />

a large amount of internal structure, called the Stetter structure. In Section 6.4<br />

some procedures are studied <strong>to</strong> project an approximate eigenvec<strong>to</strong>r <strong>to</strong> a close-by<br />

vec<strong>to</strong>r with Stetter structure <strong>to</strong> speed up the convergence process of the iterative<br />

solver. The main bottleneck was the existence of negative and complex entries in the<br />

approximate eigenvec<strong>to</strong>rs in combination with the use of a logarithmic transformation.<br />

Subsequently, two methods are given <strong>to</strong> embed such a projection method in the<br />

Jacobi–Davidson implementation.<br />

The development of a Jacobi–Davidson eigenvalue solver for commuting matrices<br />

is described in Section 6.5. This Jacobi–Davidson method is called the JDCOMM method<br />

and is an extended version of the JD method. Its most important newly implemented<br />

feature is that it computes the eigenvalues of the matrix A p in the outer loop while<br />

iterating with a much sparser (but commuting) matrix A xi in the inner loop. Most<br />

of the computation time is spent in this inner loop, while the outer loop is only used<br />

occasionally, which results in a speed up in computation time and a decrease in the<br />

amount of required floating point operations.<br />

Chapter 7 presents the results of the numerical experiments in which the global<br />

minima of various Minkowski dominated polynomials are computed using the approaches<br />

and techniques mentioned in Part II of this <strong>thesis</strong>.<br />

Section 7.1 describes the computations on a polynomial of order eight in four variables.<br />

To compute the global minimum of this polynomial conventional methods are

Hooray! Your file is uploaded and ready to be published.

Saved successfully!

Ooh no, something went wrong!