link to my thesis
link to my thesis
link to my thesis
Create successful ePaper yourself
Turn your PDF publications into a flip-book with our unique Google optimized e-Paper software.
230 CHAPTER 12. CONCLUSIONS & DIRECTIONS FOR FURTHER RESEARCH<br />
SPP <strong>to</strong>o. A central role in this approach is played by the notion of stable patterns,<br />
which are shifted along the 2D-grid. The same approach leads <strong>to</strong> partial results in<br />
the 3D-case (and higher dimensions).<br />
The results of Section 5.3.2 also underlie the design of the heuristic methods<br />
discussed in Section 5.3.3. These heuristic procedures are developed <strong>to</strong> arrive cheaply<br />
at suboptimal paths with acceptable performance. We have implemented five of these<br />
heuristics and compared their performance.<br />
Another way <strong>to</strong> improve the efficiency of an nD-system is <strong>to</strong> apply parallel computing<br />
techniques as described in Section 5.3.4. Here the ‘least-increments’ method<br />
is implemented in such a way that it computes each path in the nD-system on a<br />
separate processor. The parallel nD-system is applied in combination with the JDQZ<br />
and the JD methods. However, it turns out that parallelization is not useful in this<br />
application of the nD-system.<br />
The use of iterative eigenvalue solvers is described in Chapter 6. The iterative<br />
eigenvalue solver implementations we used in this <strong>thesis</strong> are JDQR, JDQZ and JD. An<br />
advantage of such an iterative eigenvalue solver, is that it is able <strong>to</strong> focus on a subset<br />
of eigenvalues of the involved matrix. Here the most interesting eigenvalue is the<br />
smallest real eigenvalue. Focusing on the smallest real eigenvalue of a matrix is not<br />
a usual parameter setting for an iterative eigenvalue solver. Therefore this option<br />
is developed and implemented as described in Section 6.3: the selection criterion is<br />
adapted such that it is no longer necessary <strong>to</strong> compute all the eigenvalues of the<br />
matrix but that it becomes possible <strong>to</strong> focus on the smallest real eigenvalues.<br />
An important observation is that the eigenvec<strong>to</strong>rs of the matrix at hand exhibit<br />
a large amount of internal structure, called the Stetter structure. In Section 6.4<br />
some procedures are studied <strong>to</strong> project an approximate eigenvec<strong>to</strong>r <strong>to</strong> a close-by<br />
vec<strong>to</strong>r with Stetter structure <strong>to</strong> speed up the convergence process of the iterative<br />
solver. The main bottleneck was the existence of negative and complex entries in the<br />
approximate eigenvec<strong>to</strong>rs in combination with the use of a logarithmic transformation.<br />
Subsequently, two methods are given <strong>to</strong> embed such a projection method in the<br />
Jacobi–Davidson implementation.<br />
The development of a Jacobi–Davidson eigenvalue solver for commuting matrices<br />
is described in Section 6.5. This Jacobi–Davidson method is called the JDCOMM method<br />
and is an extended version of the JD method. Its most important newly implemented<br />
feature is that it computes the eigenvalues of the matrix A p in the outer loop while<br />
iterating with a much sparser (but commuting) matrix A xi in the inner loop. Most<br />
of the computation time is spent in this inner loop, while the outer loop is only used<br />
occasionally, which results in a speed up in computation time and a decrease in the<br />
amount of required floating point operations.<br />
Chapter 7 presents the results of the numerical experiments in which the global<br />
minima of various Minkowski dominated polynomials are computed using the approaches<br />
and techniques mentioned in Part II of this <strong>thesis</strong>.<br />
Section 7.1 describes the computations on a polynomial of order eight in four variables.<br />
To compute the global minimum of this polynomial conventional methods are