This work was supported by the Mathematical, Information, and Computational Sciences Division subprogram of the Office of Computational and Technology Research, U.S. Department of Energy, under Contract W-31-109-Eng-38.
The Toolkit for Advanced Optimization (TAO) focuses on the development of algorithms and software for the solution of large-scale optimization problems on high-performance architectures. Areas of interest include nonlinear least squares, unconstrained and bound-constrained optimization, and general nonlinear optimization.
The development of TAO was motivated by the scattered support for parallel computations and the lack of reuse of external toolkits in current optimization software. Our aim is to use object-oriented techniques to produce high-quality optimization software for a range of computing environments ranging from serial workstations and laptops to massively parallel high-performance architectures. Our design decisions are strongly motivated by the challenges inherent in the use of large-scale distributed memory architectures and the reality of working with large, often poorly structured legacy codes for specific applications.
This manual describes the use of TAO. Since TAO is still under development, changes in usage and calling sequences may occur. TAO is fully supported; see the the web site http://www.mcs.anl.gov/tao for information on contacting the TAO developers.
The initial development of TAO was funded by the ACTS Toolkit Project in the Office of Advanced Scientific Computing Research, U.S. Department of Energy. We gratefully acknowledge their support.
TAO owes much to the developers of PETSc. We have benefitted from their experience, tools, software, and advice. In many ways, TAO is a natural outcome of the PETSc development. TAO has also benefitted from the work of various researchers who have provided solvers, test problems, and interfaces. In particular, we acknowledge
Finally, we thank all TAO users for their comments, bug reports, and encouragement.
The Toolkit for Advanced Optimization (TAO) focuses on the design and implementation of component-based optimization software for the solution of large-scale optimization applications on high-performance architectures. Our approach is motivated by the scattered support for parallel computations and lack of reuse of linear algebra software in currently available optimization software. The TAO design allows the reuse of toolkits that provide lower-level support (parallel sparse matrix data structures, preconditioners, solvers), and thus we are able to build on top of these toolkits instead of having to redevelop code. The advantages in terms of efficiency and development time are significant.
The TAO design philosophy uses object-oriented techniques of data and state encapsulation, abstract classes, and limited inheritance to create a flexible optimization toolkit. This chapter provides a short introduction to our design philosophy by describing the objects in TAO and the importance of this design. Since a major concern in the TAO project is the performance and scalability of optimization algorithms on large problems, we also present some performace resuls.
The TAO design philosophy place strongs emphasis on the reuse of external tools where appropriate. Our design enables bidirectional connection to lower-level linear algebra support (e.g. parallel sparse matrix data structures) provided in toolkits such as PETSc [3] [4,2] and Trilinos [9] as well as higher-level application frameworks. Our design decisions are strongly motivated by the challenges inherent in the use of large-scale distributed memory architectures and the reality of working with large, often poorly structured legacy codes for specific applications. Figure 1.1 illustrates how the TAO software works with external libraries and application code.
The TAO solvers use four fundamental objects to define and solve optimization problems: vectors, index sets, matrices, and linear solvers. The concepts of vectors and matrices are standard, while an index set refers to a set of integers used to identify particular elements of vectors or matrices. An optimization algorithm is a sequence of well defined operations on these objects. These operations include vector sums, inner products, and matrix-vector multiplication. TAO makes no assumptions about the representation of these objects by passing pointers to data-structure-neutral objects for the execution of these numerical operations.
With sufficiently flexible abstract interfaces, TAO can support a variety of implementations of data structures and algorithms. These abstractions allow us to more easily experiment with a range of algorithmic and data structure options for realistic problems, such as within this case study. Such capabilities are critical for making high-performance optimization software adaptable to the continual evolution of parallel and distributed architectures and the research community's discovery of new algorithms that exploit their features.
Our current TAO implementation uses the parallel system infrastructure and linear algebra objects offered by PETSc, which uses MPI [15] for all interprocessor communication. The PETSc package supports objects for vectors, matrices, index sets, and linear solvers.
The TAO design philosophy eliminates some the barriers in using independently developed software components by accepting data that is independent of representation and calling sequence written for particular data formats. The user can initialize an application with external frameworks, provide function information to a TAO solver, and call TAO to solve the application problem.
The use of abstractions for matrices and vectors in TAO optimization software also enables us to leverage automatic differentiation technology to facilitate the parallel computation of gradients and Hessians needed within optimization algorithms. We have demonstrated the viability of this approach through preliminary interfacing between TAO solvers and the automatic differentiation tools ADIFOR and ADIC. We are currently working on developing TAO interfaces that use special problem features (for example, partial separability, stencil information) in automatic differentiation computations.
A major concern in the TAO project is the performance and scalability of optimization algorithms on large problems. In this section we focus on the GPCG (gradient projection, conjugate gradient) algorithm for the solution of bound-constrained convex quadratic programming problems. Originally developed by Moré and Toraldo [22], the GPCG algorithm was designed for large-scale problems but had only been implemented for a single processor. GPCG combines the advantages of the identification properties of the gradient projection method with the finite termination properties of the conjugate gradient method. Moreover, the performance of the TAO implementation on large optimization problems is noteworthy.
We illustrate the performance of the GPCG algorithm by
presenting results for a journal bearing problem
with over 2.5 million variables.
The journal bearing problem
is a finite element approximation to a variational problem
over a rectangular two-dimensional grid. A
grid with points in each direction, for example, is formulated
as a bound constrained quadratic problem with
variables.
The triangulation of the grid results in a matrix that has the
usual five diagonal nonzero structure that arises
from a difference approximation to the Laplacian operator.
The journal bearing problem contains an eccentricity parameter,
, that influences the number of active
variables at the solution and the difficulty in solving it.
Figure 1.2 shows the solution of the journal bearing problem
for
. The steep gradient in the solution
makes this problem a difficult benchmark.
The performance results in Table 1.1 are noteworthy is several
ways. First of all, the number of faces visited by GPCG is remarkably
small. Other strategies can lead to a large number of gradient
projection iterates, but the GPCG algorithm is remarkably efficient.
Another interesting aspect of these results is that due to the low
memory requirements of iterative solvers, we were able to solve these
problems with only processors. Strategies that rely on
direct solvers are likely to need significantly more storage, and thus
more processors. Finally, these results show that the GPCG
implementation has excellent efficiency. For example, the efficiency
of GPCG with respect to
processors ranges between
and
when
. This sustained efficiency
is remarkable since the GPCG algorithm is solving a sequence of linear
problems with a coefficient matrix set to the submatrix of the Hessian
matrix with respect to the free variables for the current iterate.
Thus, our implementation's repartitioning of submatrices effectively
deals with the load-balancing problem that is inherent in the GPCG
algorithm.
An important aspect of our results that is not apparent from Table 1.1 is that for these results we were able to experiment easily with all the preconditioners offered by PETSc. In particular, we were able to compare the diagonal Jacobi preconditioner with block Jacobi and overlapping additive Schwarz preconditioners that use a zero-fill ILU solver in each block. We also experimented with a parallel zero-fill incomplete Cholesky preconditioner provided by a PETSc interface to the BlockSolve95 [17] package of Jones and Plassmann. Interestingly enough, the diagonal Jacobi preconditioner achieved better performance on this problem.
The ability to experiment with various preconditioners is a direct result of our design philosophy that enables interfacing to the linear algebra infrastructure provided in toolkits such as PETSc, Trilinos, and ISIS++. In this spirit, we are working with the Common Component Architecture (CCA) [5] and Equation Solver Interface (ESI) [8] groups, and we have developed an ESI-compliant interface to enable the dynamic use within TAO of various ESI-compliant linear solvers.
TAO can be used on a personal computer with a single processor or within a parallel environment. Its basic usage involves only a few commands, but fully understanding its usage requires time. Application programmers can easily begin to use TAO by working with some examples provides in the package and then gradually learn more details according to their needs. The current version of TAO and the most recent help concerning the installation and usage of TAO can be found at http://www.mcs.anl.gov/tao/.
The current version of TAO requires an ANSI C++ compiler, an implementation of MPI, Version 2.3.0 of PETSc compiled with the C++ compiler, (PETSc must be configured with the -with-clanguage=C++ option) and at least 15 MB of free disk space. During the setup process, the user will have to set an environmental variable, TAO_DIR, indicating the full path of the TAO home directory. This variable will be used in this manual to refer to the location of files, and by computers that will compile TAO source code.
The examples throughout the library demonstrate the software usage and can serve as templates for developing custom applications. We suggest that new TAO users examine programs in
${TAO_DIR}/src/examples .Additional examples are available on our website and in
${TAO_DIR}/src/<unconstrained,bound,..>/examples/tutorials,where <component> denotes any of the TAO components, such as bound or unconstrained. The HTML version of the manual pages located at
${TAO_DIR}/docs/manualpages/index.htmlor
http://www.mcs.anl.gov/tao/docs/manualpages/index.htmlprovides indices (organized by both routine names and concepts) to the tutorial examples.
We suggest the following procedure for writing a new application program using TAO:
To help the user start using TAO immediately, we use a simple
uniprocessor example. The code in
Figure 2.1 is for minimizing the
extended Rosenbrock function
defined by
Note that while we use the C language to introduce the TAO software, the package is also fully usable from C++ and Fortran77/90. Section 5.14 discusses additional issues concerning Fortran usage.
The code in Figure 2.1 contains many of the components needed to write most TAO programs and thus, is illustrative of the features present in complex optimization problems. Note that we have omitted the code required for the routine FormFunctionGradient, that evaluates the function and gradient, and the code for FormHessian, that evaluates the Hessian matrix. for Rosenbrock's function. The following sections explain the components in Figure 2.1.
The C++ include file for TAO should be used via the statement
#include "tao.h"The required lower level include files are automatically included within this high-level file.
Most TAO programs begin with a call to
info = TaoInitialize(int *argc,char ***argv,char *file_name, char *help_message);This command initializes TAO (and also MPI and PETSc if these have not yet been initialized elsewhere). The arguments argc and argv are the command line arguments delivered in all C and C++ programs. The argument file_name optionally indicates an alternative name for an options file, which by default is called .petscrc and resides in the user's home directory. See the PETSc users manual for details regarding runtime option specification. The final argument, help_message, is an optional character string that will be printed if the program is run with the -help option.
As illustrated by the TaoInitialize() statements above, TAO routines return an integer indicating whether an error has occurred during the call. The error code is set to be nonzero if an error has been detected; otherwise, it is zero. For the C++ interface, the error variable is the routine's return value, while for the Fortran version, each TAO routine has as its final argument an integer error variable. Error tracebacks are discussed in Section 2.8.
All TAO programs should call TaoFinalize() as their final (or nearly final) statement
info = TaoFinalize();This routine handles options to be called at the conclusion of the program, and calls PetscFinalize() if TaoInitialize() began PETSc. If PETSc was initiated externally from TAO (by either the user or another software package), the user is responsible for calling PetscFinalize().
The primary commands for solving an unconstrained optimization problem using TAO are shown in Figure 2.2.
![]() |
The user first creates the TAO_SOLVER and TAO_APPLICATION contexts. He then sets call-back routines as well as vector (Vec) and matrix (Mat) data structures that the TAO solver will use for evaluating the minimization function, gradient, and optionally the Hessian matrix. The user then solves the minimization problem, and finally destroys the TAO_SOLVER and TAO_APPLICATION contexts. Details of these commands are presented in Chapter 3.
Note that TaoCreate() enables the user to select the solution method at runtime by using an options database. Through this database, the user not only can select a minimization method (e.g., limited-memory variable metric, conjugate gradient, Newton with line search or trust region), but also can prescribe the convergence tolerance, set various monitoring routines, indicate techniques for linear systems solves, etc.
Users of TAO are required to provide routines that perform function evaluations. Depending on the solver chosen, they may also have to write routines that evaluate the gradient vector and Hessian matrix.
Applications using the PETSc package for vectors, matrices, and linear solvers should include the appropriate header files. For example
#include "petscksp.h"includes the appropriate information for most TAO applications using PETSc.
The user can input control data at run time using an options database. The command
PetscOptionsGetInt(PETSC_NULL, "-n", &user.n, &flg);checks whether the user has provided a command line option to set the value of n, the number of variables. If so, the variable n is set accordingly; otherwise, n remains unchanged. A complete description of the options database may be found in the PETSc users manual.
In the example in Figure 2.1, the vector data structure (Vec) is used to store the solution and gradient for TAO unconstrained minimization solvers. A new parallel or sequential vector x of global dimension M is created with the command
info = VecCreate(MPI_Comm comm,int m,int M,Vec *x);where comm denotes the MPI communicator. The type of storage for the vector may be set with either calls to VecSetType() or VecSetFromOptions(). Additional vectors of the same type can be formed with
info = VecDuplicate(Vec old,Vec *new);The commands
info = VecSet(Scalar *value,Vec x); info = VecSetValues(Vec x,int n,int *indices, Scalar *values,INSERT_VALUES);respectively set all the components of a vector to a particular scalar value and assign a different value to each component. More detailed information about vectors, including their basic operations, scattering/gathering, index sets, and distributed arrays, may be found in the PETSc users manual.
Usage of matrices and vectors is similar. The user can create a new parallel or sequential matrix H with M global rows and N global columns, with the routine
info = MatCreate(MPI_Comm comm,int m,int n,int M,int N,Mat *H);where the matrix format can be specified at runtime. The user could alternatively specify each processes' number of local rows and columns using m and n. H can then be used to store the Hessian matrix, as indicated by the above routine TaoSetHessianMat(). Matrix entries can then be set with the command
info = MatSetValues(Mat H,int m,int *im,int n,int *in, Scalar *values,INSERT_VALUES);After all elements have been inserted into the matrix, it must be processed with the pair of commands
info = MatAssemblyBegin(Mat H,MAT_FINAL_ASSEMBLY); info = MatAssemblyEnd(Mat H,MAT_FINAL_ASSEMBLY);The PETSc users manual discusses various matrix formats as well as the details of some basic matrix manipulation routines.
Since TAO uses the message-passing model for parallel programming and employs MPI for all interprocessor communication, the user is free to employ MPI routines as needed throughout an application code. However, by default the user is shielded from many of the details of message passing within TAO, since these are hidden within parallel objects, such as vectors, matrices, and solvers. In addition, TAO users can interface to external tools, such as the generalized vector scatters/gathers and distributed arrays within PETSc, to assist in the management of parallel data.
The user must specify a communicator upon creation of any TAO objects (such as a vector, matrix, or solver) to indicate the processors over which the object is to be distributed. For example, some commands for matrix, vector, and solver creation are:
info = MatCreate(MPI_Comm comm,int m,int n,int M,int N,Mat *H); info = VecCreate(MPI_Comm comm,int m,int M,Vec *x); info = TaoCreate(MPI_Comm comm,TaoMethod method,TAO_SOLVER *tao);The creation routines are collective over all processors in the communicator; thus, all processors in the communicator must call the creation routine. In addition, if a sequence of collective routines is being used, the routines must be called in the same order on each processor.
Compilation of the TAO numerical libraries and TAO application codes requires three environmental variables to be set. These three variables, TAO_DIR, PETSC_ARCH, and PETSC_DIR, are discussed more fully in the TAO installation instructions.
TAO uses a portable makefile system provided by the PETSc [2,4] library, which is discussed further in Section 2.11. The TAO library can be compiled with the command
makefrom the TAO_DIR directory.
Running a TAO application on a single processor can be done in the usual way by entering the name of the executable and any command line options. Running programs in parallel, however, requires use of the MPI library. All TAO programs use the MPI (Message Passing Interface) standard for message-passing communication [23]. Thus, to execute TAO programs, users must know the procedure for beginning MPI jobs on their selected computer system(s). For instance, when using the MPICH implementation of MPI [14] and many others, the following command initiates a program that uses eight processors:
mpirun -np 8 tao_program_name tao_options
All TAO commands begin with the Tao prefix and return an integer indicating whether an error has occurred during the call. The error code equals zero after the successful completion of the routine and is set to a nonzero value if an error has been detected. The macro CHKERRQ(info) checks the value of info and calls an error handler upon error detection. CHKERRQ() should be used in all subroutines to enable a complete error traceback.
In Figure 2.3 we indicate a traceback generated by error detection within a sample program. The error occurred on line 1007 of the file ${TAO_DIR}/src/interface/tao.c in the routine TaoSetUp() and was caused by nonconforming local lengths of the parallel gradient and solution vectors, which must be identically partitioned. The TaoSetUp() routine was called from the TaoSolveApplication() routine, which was in turn called on line 229 of the main() routine in the program ex2.c. The PETSc users manual provides further details regarding error checking, including information about the Fortran interface.
When running the debugging version of the TAO software (PETSc configured with the -with-debugging option), checking is performed for memory corruption (writing outside of array bounds, etc). The macros CHKMEMQ and CHKMEMA can be called anywhere in the code to check the current status of the memory for corruption. By putting several (or many) of these macros into an application code, one can usually track down the code segment where corruption has occurred.
To manage code portability across a wide variety of UNIX systems, TAO uses a makefile system that is part of the PETSc software. This section briefly discusses makefile usage from the perspective of application programmers; see the ``makefiles'' chapter of the PETSc users manual for additional details.
To make a program named rosenbrock1, one may use the command
make PETSC_ARCH=arch rosenbrock1which compiles a debugging or optimized version of the example and automatically link the appropriate libraries. The architecture, arch, is one of solaris, rs6000, IRIX, hpux, etc. Note that when using command line options with make (as illustrated above), one must not place spaces on either side of the ``='' signs. The variable PETSC_ARCH can also be set as environmental variables.
Maintaining portable TAO makefiles is very simple. Figure 2.4 presents a minimal makefile for maintaining a single program that uses the TAO library. The most important line in this makefile is the line starting with include:
include ${TAO_DIR}/bmake/tao_commonThis line includes other makefiles that provide the needed definitions and rules for the particular base software installations (specified by ${TAO_DIR} and ${PETSC_DIR}) and architecture (specified by ${PETSC_ARCH}), which are typically set as environmental variables prior to compiling TAO source or programs. As listed in the sample makefile, the appropriate include file is automatically completely specified; the user should not alter this statement within the makefile.
Note that the variable ${TAO_LIB} (as listed on the link line in this makefile) specifies all of the various TAO and supplementary libraries in the appropriate order for correct linking.
Some additional variables that can be used in the makefile are defined as follows:
The home directory of TAO contains the following subdirectories:
Each TAO source code component directory has the following subdirectories:
TAO contains unconstrained minimization, bound constrained minimization, and nonlinear complementarity solvers. The structure of these problems can differ significantly, but TAO has a similar interface to all of its solvers. Routines that most solvers have in common will be discussed in this chapter. A complete list of options can be found by consulting the manual pages. Many of the options can also be set at the command line. These options can also be found in manual pages or by running a program with the -help option.
info = TaoInitialize(int *argc,char ***argv,char *file_name, char *help_message);This command initializes TAO, as well as MPI, PETSc, and other packages to which TAO applications may link (if these have not yet been initialized elsewhere). In particular, the arguments argc and argv are the command line arguments delivered in all C and C++ programs; these arguments initialize the options database. The argument file_name optionally indicates an alternative name for an options file, which by default is called .petscrc and resides in the user's home directory.
One of the last routines that all TAO programs should call is
info = TaoFinalize();This routine finalizes TAO and any other libraries that may have been initialized during the TaoInitialize() phase. For example, TaoFinalize() calls MPI_Finalize() if TaoInitialize() began MPI. If MPI was initiated externally from TAO (by either the user or another software package), then the user is responsible for calling MPI_Finalize().
A TAO solver can be created with the command
info = TaoCreate(MPI_Comm comm,TaoMethod method,TAO_SOLVER *newsolver);The first argument in this routine is an MPI communicator indicating which processes are involved in the solution process. In most cases, this should be set to MPI_COMM_WORLD. The second argument in this creation routine specifies the default method that should be be used to solve the optimization problem. The third argument in TaoCreate() is a pointer to a TAO solver object. This routine creates the object and returns it to the user. The TAO object is then to be used in all TAO routines.
The various types of TAO solvers and the flags that identify them will be discussed in the following chapters. The solution method should be carefully chosen depending upon problem that is being solved. Some solvers, for instance, are meant for problems with no constraints, while other solvers acknowledge constraints in the problem and solve them accordingly. The user must also be aware of the derivative information that is available. Some solvers require second-order information, while other solvers require only gradient or function information. The TaoMethod can also be set to TAO_NULL in the TaoCreate() routine if the user selects a method at runtime using the options database. The command line option -tao_method followed by an TAO method will override any method specified by the second argument. The command line option -tao_method tao_lmvm, for instance, will specify the limited memory variable metric method for unconstrained optimization. Note that the TaoMethod variable is a string that requires quotation marks in an application program, but quotation marks are not required at the command line. The method that TAO uses to solve an optimization problem can be changed at a later point in the program with the command TaoSetMethod(), whose arguments are a TAO solver and a string that uniquely identifies a method for solving the problem.
Each TAO solver that has been created should also be destroyed using the command
info = TaoDestroy(TAO_SOLVER solver);This routine frees the internal data structures used by the solver.
Although TAO and its solvers set default parameters that are useful for many problems, it may be necessary for the user to modify these parameters to change the behavior and convergence of various algorithms.
One convergence criterion for most algorithms concerns the
of digits of accuracy needed in the solution. In particular,
one convergence test employed by TAO attempts to stop when
the error in the constraints is less than
,
and either
Other stopping criteria include a minimum trust region radius or a maximum number of iterations. These parameters can be set with the routines TaoSetTrustRegionTolerance() and TaoSetMaximumIterates(). Similarly, a maximum number of function evaluations can be set with the command TaoSetMaximumFunctionEvaluations() .
The routine
int TaoSolveApplication(TAO_APPLICATION, TAO_SOLVER);will apply the solver to the application that has been created by the user.
To see parameters and performance statistics for the solver, the routine
int TaoView(TAO_SOLVER);can be used. This routine will display to standard output the number of function evaluations need by the solver and other information specific to the solver.
The progress of the optimization solver can be monitored with the runtime option -tao_monitor. Although monitoring routines can be customized, the default monitoring routine will print out several relevant statistics to the screen. The routine TaoView() (or the corresponding runtime option -tao_view) will display many statistics concerning the solver, including the number of function evaluations and the convergence tolerances used.
The user also has access to information about the current solution. The current iteration number, objective function value, gradient norm, infeasibility norm, and step length can be retrieved with the command
int TaoGetSolutionStatus(TAO_SOLVER tao, int* iterate, double* f, double* gnorm, double *cnorm, double *xdiff, TaoTerminateReason *reason)The last argument returns a code that indicates the reason that the solver terminated. Positive numbers indicate that a solution has been found, while negative numbers indicate a failure. A list of reasons can be found in the manual page for TaoGetTerminationReason().
The user set vectors containing the solution and gradient before solving the problem, but pointers to these vectors can also be retrieved with the commands TaoGetSolution() and TaoGetGradient(). Dual variables and other relevant information are also available. This information can be obtained during user-defined routines such as a function evaluation and customized monitoring routine, or after the solver has terminated.
The limited memory variable metric (LMVM) algorithm is a powerful method for solving unconstrained nonlinear optimization that requires only function value and gradient information. Although the number of iterations required by this algorithm is significantly higher than the number of iterations usually required by a Newton algorithm, this quasi-Newton method is especially useful for large-scale problems when the Hessian matrix is either unavailable or too expensive to compute. It keeps a limited history of previous points and previous gradients to approximate the second-order information. This algorithm is the default unconstrained minimization solver, and can be selected using the TaoMethod tao_lmvm.
The number of solutions and gradients at previous iterations that are kept to approximate the Hessian can be set with
info = TaoLMVMSetSize(TAO_SOLVER tao,int lm);The default number is
Nonlinear conjugate gradient methods can be viewed as extensions of the conjugate gradient method for solving positive definite linear systems. These algorithms require only function and gradient information as well as a line search. The three variations currently supported are the Fletcher-Reeves method, the Polak-Ribiére method, and Polak-Ribiére-Plus method[26]. These conjugate gradient methods can be specified using the TaoMethod tao_cg_fr, tao_cg_pr, and tao_cg_prp, respectively. Each of these conjugate gradient methods incorporates automatic restarts when successive gradients are not sufficiently orthogonal. TAO measures the orthogonality by dividing the inner product of the gradient at the current point and the gradient at the previous point by the square of the Euclidean norm of the gradient at the previous point. When the absolute value of this ratio is greater than eta, the algorithm restarts using the gradient direction. The parameter eta can be set using the command
int TaoCGSetRestartTol(TAO_SOLVER solver,double eta);or the runtime command -tao_cg_restart <eta>. For the best efficiency, the function and gradient evaluations in this method should be performed simultaneously.
The trust region method for unconstrained minimization, can be set using TaoMethod tao_ntr. The algorithm is based on the work of Steihaug [29]. This method also uses Hessian information, but applies a trust region and a preconditioned conjugate gradient method to minimize a local quadratic model function. In the PETSc package, the linear solver uses the KSP solver KSPQCG. This formulation requires the use of a symmetric preconditioner, where the currently available options are Jacobi, incomplete Cholesky, and the null preconditioners, which can be set with the options -pc_type jacobi, -pc_type icc, and -pc_type none, respectively. The preconditioner can be accessed by the application code with the commands TaoGetLinearSolver() and TaoLinearSolverGetKSP(). The initial trust region radius can be set using the command TaoSetTrustRegionRadius(). The initial trust region can significantly alter the rate of convergence for the algorithm and should be tuned and adjusted for optimal performance. Convergence tests for trust regions methods often impose minimum allowable trust region radius which can be set using the routine
info = TaoSetTrustRegionTolerance(TAO_SOLVER tao,double trtol);or the option -tao_trtol <trtol>. .
Bound constrained optimization algorithms
minimize
, subject to upper or
lower bounds on some of the variables.
These solvers also bounds on the variables as well as objective
function, gradient, and possibly Hessian information.
The TRON algorithm solves a reduced linear system defined by the rows and columns corresponding to the variables that lie between the upper and lower bounds. When running in parallel, these rows can either remain on their current processor or be redistributed evenly over all of the processors with the command TaoSelectSubset(). The TRON algorithm applies a trust region to the conjugate gradients to ensure convergence. The initial trust region can be set using the command TaoSetTrustRegionRadius(); and the current trust region size can be found using the command TaoGetTrustRegionRadius(); The initial trust region can significantly alter the rate of convergence for the algorithm and should be tuned and adjusted for optimal performance.
This method is the bound constrained variant of the LMVM method for unconstrained optimization. It uses projected gradients to approximate the Hessian - eliminating the need for Hessian evaluations. The method can be set using TaoMethod tao_blmvm. The command TaoLMVMSetSize(), which sets the number of vectors to be used in the Hessian approximation, also applies to this method.
This method calculates points satisfying the first-order necessary optimality conditions. The method uses the mixed complementarity problem solvers from Section 4.3 to calculate the solutions. The choice of complementarity solver is specified with the runtime option -tao_kt_method with the default being the tao_ssils method.
Mixed complementarity problems, or box-constrained variational inequalities,
are related to nonlinear systems of equations. They are defined by a
continuously differentiable function,
, and bounds,
and
, on the variables such that
. Given this information,
is a solution to
MCP(
,
,
) if for each
we have at
least one of the following:
Simple complementarity conditions arise from the first-order optimality
conditions from optimization [18,19].
In the simple bound constrained optimization case, these conditions
correspond to MCP(,
,
), where
is the objective function. In a one-dimensional setting these conditions
are intuitive. If the solution is at the lower bound, then the function must
be increasing and
. However, if the solution is at the
upper bound, then the function must be decreasing and
.
Finally, if the solution
is strictly between the bounds, we must be at a stationary point and
. Other complementarity problems arise in economics and
engineering [11], game
theory [25], and finance [16].
Evaluation routines for and its Jacobian must be supplied prior
to solving the application.
The bounds,
, on the variables must also be
provided.
If no starting point is supplied, a default starting point of all zeros
is used.
TAO has two implementations of semismooth algorithms [24,7,10] for solving mixed complementarity problems. Both are based upon a reformulation of the mixed complementarity problem as a nonsmooth system of equations using the Fischer-Burmeister function [12]. A nonsmooth Newton method is applied to the reformulated system to calculate a solution. The theoretical properties of such methods are detailed in the aforementioned references.
The Fischer-Burmeister function,
, is defined as,
The two semismooth TAO solvers both solve the system by applying
a nonsmooth newton method with a line-search. We calculate a direction,
,
by solving the system
where
is an element of the
-subdifferential [28] of
at
. If the
direction calculated does not satisfy a suitable descent condition, then
we use the negative gradient of the merit function,
, as
the search direction. A standard Armijo search [1] is
used to find the new iteration. Non-monotone searches
[13] are also available by setting
appropriate run-time options. See Section 6.3 for further
details.
The first semismooth algorithm available in TAO is not guaranteed to
remain feasible with respect to the bounds, , and is termed
an infeasible semismooth method. This method can be specified using the
TaoMethod tao_ssils. In this case, the descent test used is
that
An alternative is to remain feasible with respect to the bounds by using a
projected Armijo line-search. This method can be specified using the
TaoMethod tao_ssfls. The descent test used is the same as above
where the direction in this case corresponds to the first part of the
piece-wise linear arc searched by the projected line-search.
Both and
can be modified using the run-time
commands -tao_ssfls_delta <delta> and -tao_ssfls_rho <rho>
respectively. By default,
and
.
The recommended algorithm is the infeasible semismooth method,
tao_ssils, because of its strong global and local convergence
properties. However, if it is known that is not defined outside
of the box,
, perhaps due to the presence of
functions,
the feasible algorithm, tao_ssfls, is a reasonable alternative.
The solvers in TAO address applications that have a set of variables, an objective function, and constraints on the variables. Many solvers also require derivatives of the objective and constraint functions. To use the TAO solvers, the application developer must define a set of variables, implement routines that evaluate the objective function and constraint functions, and pass this information to a TAO application object.
TAO uses vector and matrix objects to pass this information from the
application to the solver. The set of variables, for instance, is
represented in a vector.
The gradient of an objective function
,
evaluated at a point, is also represented as a vector.
Matrices, on the other hand,
can be used to represent the Hessian of
or the Jacobian of a constraint
function
. The TAO solvers use
these objects to compute a solution to the application.
The PETSc package provides parallel and serial implementations of these objects and offers additional tools intended for high-performance scientific applications. The Vec and Mat types in PETSc represent the vectors and matrices in a TAO application. This chapter will describe how to create these an application object and give it the necessary properties. This chapter will also describe how to use the TAO solvers in conjunction with this application object.
TAO applications written in C/C++ should have the statement
#include "tao.h"in each file that uses a routine in the TAO libraries. All of the required lower level include files such as ``tao_solver.h'' and ``taoapp.h'' are automatically included within this high-level file.
TaoApplicationCreate(MPI_Comm, TAO_APPLICATION*);Much like creating PETSc vector and matrix objects, the first argument is an MPI communicator. An MPI [15] communicator indicates a collection of processors that will be used to evaluate the objective function, compute constraints, and provide derivative information. When only one processor is being used, the communicator MPI_COMM_SELF can be used with no understanding of MPI. Even parallel users need to be familiar with only the basic concepts of message passing and distributed-memory computing. Most applications running TAO in parallel environments can employ the communicator MPI_COMM_WORLD to indicate all processes in a given run.
The second argument is the address of a TAO_APPLICATION variable. This routine will create a new application object and set the variable, which is a pointer, to the address of the object. This application variable can now be used by the developer to define the application and by the TAO solver to solve the application.
Elsewhere in this chapter, the TAO_APPLICATION variable will be referred to as the application object.
After solving the application, the command
TaoAppDestroy(TAO_APPLICATION);will destroy the application object and free the work space associated with it.
In all of the optimization solvers, the application must provide a Vec object of appropriate dimension to represent the variables. This vector will be cloned by the solvers to create additional work space within the solver. If this vector is distributed over multiple processors, it should have a parallel distribution that allows for efficient scaling, inner products, and function evaluations. This vector can be passed to the application object using the routine
TaoAppSetInitialSolutionVec(TAO_APPLICATION,Vec);When using this routine, the application should initialize the vector with an approximate solution of the optimization problem before calling the TAO solver. If you do not know of a solution that that can be used, the routine TaoAppSetDefaultSolutionVec(TAO_APPLICATION,Vec); can be used to declare variables that will in be set to zero or some other default solution.
This vector will be used by the TAO solver to store the solution. Elsewhere in the application, this solution vector can be retieved from the application object using the routine
TaoAppGetSolutionVec(TAO_APPLICATION, Vec *);This routine takes the address of a Vec in the second argument and sets it to the solution vector used in the application.
For example, a routine that evaluates an objective function may need parameters, work vectors, and other information. This information, which may be specific to an application and necessary to evaluate the objective, can be collected in a single structure and used as one of the arguments in the routine. The address of this structure will be cast as type (void*) and passed to the routine in the final argument. There are many examples of these structures in the TAO distribution.
This technique offers several advantages. In particular, it allows for a uniform interface between TAO and the applications. The fundamental information needed by TAO appears in the arguments of the routine, while data specific to an application and its implementation is confined to an opaque pointer. The routines can access information created outside the local scope without the use of global variables. The TAO solvers and application objects will never access this structure, so the application developer has complete freedom to define it. In fact, these contexts are completely optional - a NULL pointer can be used.
TAO solvers that minimize an objective function require the application to evaluate the objective function. Some solvers may also require the application to evaluate derivatives of the objective function. Routines that perform these computations must be identified to the application object and must follow a strict calling sequence.
Routines that evaluate an objective function
,
should follow the form:
EvaluateObjective(TAO_APPLICATION,Vec,double*,void*);The first argument is the application object, the second argument is the
This routine, and the application context, should be passed to the application object using the routine
TaoAppSetObjectiveRoutine(TAO_APPLICATION, int(*)(TAO_APPLICATION,Vec,double*,void*), void*);The first argument in this routine is the application object, the second argument is a function pointer to the routine that evaluates the objective, and the third argument is the pointer an appropriate application context.
Although final argument may point to anything, it must be cast as a (void*) type. This pointer will be passed back to the developer in the fourth argument of the routine that evaluates the objective. In this routine, the pointer can be cast back to the appropriate type. Examples of these structures and there usage are provides in the distribution.
Most TAO solvers also require gradient information from the application . The gradient of the objective function can be specified in a similar manner. Routines that evaluate the gradient should have the calling sequence
EvaluateTheGradient(TAO_APPLICATION,Vec,Vec,void*);In this routine, the first argument is the application object, the second argument is the variable vector, the third argument is the gradient, and the fourth argument is the user-defined application context. Only third argument in this routine is different from the arguments in the routine that evaluates the objective function. The numbers in the gradient vector have no meaning when passed into this routine, but should represent the gradient of the objective at the specified point at the end the routine. This routine, and the user-defined pointer, can be passed to the application object using the routine:
TaoAppSetGradientRoutine(TAO_APPLICATION, int (*)(TAO_APPLICATION,Vec,Vec,void*), void *);In this routine, the first argument is the application object, the second argument is the function pointer, and the third object is the application context, cast to (void*).
Instead of evaluating the objective and its gradient in separate routines, TAO also allows the user to evaluate the function and the gradient at the same routine. In fact, some solvers are more efficient when both function and gradient information can be computed in the same routine. These routines should follow the form
EvaluateFunctionGradient(TAO_APPLICATION,Vec,double*,Vec,void*);where the first argument is the TAO solver, and the second argument points to the input vector for use in evaluating the function and gradient. The third argument should return the function value, while the fourth argument should return the gradient vector, and the fifth argument is a pointer to a user-defined context. This context and the name of the routine should be set with the call:
TaoAppSetObjectiveAndGradientRoutine(TAO_APPLICATION, int (*)(TAO_APPLICATION,Vec,double*,Vec,void*), void *);The arguments of this routine are the TAO application, a function name, and a pointer to a user-defined context.
The TAO example problems demonstrate the use of these application contexts
as well as specific instances of function, gradient, and Hessian
evaluation routines.
All of these routines should return the integer after
successful completion and a nonzero integer if the function
is undefined at that point or an error occurred.
Some optimization routines also require a Hessian matrix from the user. The routine that evaluates the Hessian should have the form:
EvaluateTheHessian(TAO_APPLICATION,Vec,Mat*,Mat*,MatStructure*,void*);The first argument of this routine is a TAO application. The second argument is the point at which the Hessian should be evaluated. The third argument is the Hessian matrix, and the sixth argument is a user-defined context. Since the Hessian matrix is usually used in solving a system of linear equations, a preconditioner for the matrix is often needed. The fourth argument is the matrix that will be used for preconditioning the linear system. In most cases, this matrix will be the same as the Hessian matrix. The fifth argument is the flag used to set the Hessian matrix and linear solver in the routine KSPSetOperators().
One can set the Hessian evaluation routine by calling
int TaoAppSetHessianRoutine(TAO_APPLICATION, int (*)(TAO_APPLICATION,Vec,Mat*,Mat*,MatStructure*,void*), void *)The first argument is the TAO application, the second argument is the function that evaluates the Hessian, and the third argument is a pointer to a user defined context, cast as a void* pointer.
For solvers that evaluate the Hessian, the matrices used to store the Hessian should be set using
TaoAppSetHessianMat(TAO_APPLICATION,Mat,Mat);The first argument is the TAO application, the second argument is the Hessian matrix, and the third argument is the preconditioning matrix. In most applications, these two matrices will be the same structure.
TaoAppDefaultComputeHessian( TAO_APPLICATION, Vec, Mat*, Mat*, MatStructure*, void*);and
TaoAppDefaultComputeHessianColor( TAO_APPLICATION, Vec, Mat*, Mat*, MatStructure*, void* );These routines can be set using TaoAppSetHessianRoutine() or through the options database. If finite differences is used with coloring, the routine
TaoAppSetColoring(TAO_APPLICATION, ISColoring);should be used to specify the coloring.
Some optimization problems also impose constraints upon the variables. The constraints may impose simple bounds upon the variables, or require that the variables satisfy a set of linear or nonlinear equations.
The simplest type of constraint upon an optimization problem puts lower or upper bounds upon the variables. Vectors that represent lower and upper bounds for each variable can be set with the command
TaoAppSetVariableBoundsRoutine(TAO_APPLICATION, int (*)(TAO_APPLICATION, Vec,Vec,void*),void *);The first vector and second vectors should contain the lower and upper bounds, respectively. When no upper or lower bound exists for a variable, the bound may be set to TAO_INFINITY or TAO_NINFINITY. After the two bound vectors have been set, they may be accessed with the with the command TaoGetApplicationVariableBounds(). Since not all solvers use bounds on variables, the user must be careful to select a type of solver that acknowledges these bounds.
Constraints in the form of nonlinear equations have the form
where
.
These constraints should be specified in a
routine, written by the user, that evaluates C(X).
The routine that evaluates the constraint equations should have the form:
int EqualityConstraints(TAO_APPLICATION,Vec,Vec,void*);The first argument of this routine is a TAO application object. The second argument is the variable vector at which the constraint function should be evaluated. The third argument is the vector of function values C, and the fourth argument is a pointer to a user-defined context. This routine and the user-defined context should be set in the TAO solver with the command
TaoAppSetConstraintRoutine(TAO_APPLICATION, int (*)(TAO_APPLICATION,Vec,Vec,void*), void*);In this function, first argument is the TAO application, the second argument is vector in which to store the function values, and the third argument is a pointer to a user-defined context that will be passed back to the user.
The Jacobian of the function C is the matrix in
such that each column contains the partial derivatives of f with respect
to one variable.
The evaluation of the Jacobian of f should be performed in a routine
of the form
int J(TAO_APPLICATION,Vec,Mat*,Mat*,MatStructure*,void*);In this function, the second argument is the variable vector at which to evaluate the Jacobian matrix, the third argument is the Jacobian matrix, and the sixth argument is a pointer to a user-defined context. This routine should be specified using
TaoAppSetJacobianRoutine(TAO_APPLICATION,Mat, int (*)(TAO_APPLICATION,Vec,Mat*,Mat*, MatStructure*,void*), void*);The first argument is the TAO application, the second argument is the matrix in which the information can be stored, the third argument is the function pointer, and the fourth argument is an optional user-defined context. The Jacobian matrix should be created in a way such that the product of it and the variable vector can be put in the constraint vector.
For solvers that evaluate the Jacobian, the matrices used to store the Jacobian should be set using
TaoAppSetJacobianMat(TAO_APPLICATION,Mat,Mat);The first argument is the TAO application, the second argument is the Jacobian matrix, and the third argument is the preconditioning matrix. In most applications, these two matrices will be the same structure.
The routine set by TaoAppSetMonitor() is called once during each iteration of the optimization solver. Hence, the user can employ this routine for any application-specific computations that should be done after the solution update. .
TaoAppSetMonitor(TAO_APPLICATION, int (*)(TAO_APPLICATION,void*),void *);
TaoGetKSP(TAO_SOLVER, KSP *);With access to the KSP object, users can customize it for their application to achieve additional performance.
Once the TAO solver and TAO application object have been created and customized, they should be matched with one another using the routine
TaoSetupApplicationSolver( TAO_APPLICATION, TAO_SOLVER);This routine will set up the TAO solver for the application. Different solvers may set up differently, but they typically create the work vectors and linear solvers needed to find a solution. These structures were not created during the creation of the solver because the size of the application was not known. After calling this routine the routine TaoAppGetTaoSolver() can be used to obtain the TAO solver object.
TaoGetGradientVec( TAO_SOLVER, Vec*);will set a pointer to a Vec to the vector object containing the gradient vector and the routine
TaoGetVariableBoundVecs( TAO_SOLVER, Vec*, Vec*);will set the pointers to the lower and upper bounds on the variables - if they exist. These vectors may be viewed at before, during, and after the solver is running.
Options for the application and solver can be be set from the command line using the routine
TaoSetOptions( TAO_APPLICATION, TAO_SOLVER);This routine will call TaoSetupApplicationSolver() if it has not been called already. This command also provides information about runtime options when the user includes the -help option on the command line.
Once the application and solver have been set up, the solver can be called using the routine
TaoSolveApplication( TAO_APPLICATION, TAO_SOLVER);This routine will call the TAO solver. If the routine TaoSetupApplicationSolver() has not already been called, this routine will call it.
After a solution has been found, the routine
TaoCopyDualsOfVariableBounds( TAO_APPLICATION, Vec, Vec );can compute the dual values of the variables bounds and copy them into the vectors passed into this routine.
Occasionally TAO users will have to interact directly with the linear algebra objects used by the solvers. Solvers within TAO use vector, matrix, index set, and linear solver objects that have no native data structures. Instead they have methods whose implementation is uses structures and routines provided by PETSc or other external software packages.
Given a PETSc Vec object X, the user can create a TaoVec object. By declaring the variables
TaoVec *xx;the routine
TaoWrapPetscVec(Vec,TaoVec **);takes the Vec x and creates and sets TaoVec *xx equal to a new TaoVec object. This object actually has the derived type TaoVecPetsc. Given a TaoVec whose underlying representation is a PETSc Vec, the command
TaoVecGetPetscVec( TaoVec *, Vec *);will retrieve the underlying vector. The routine TaoVecDestroy() will destroy the TaoVec object, but the Vec object must also be destroyed.
TaoWrapPetscMat(Mat,TaoMat **);takes the Mat H and creates and sets TaoMat *HH equal to the new TaoMat object. The second argument specifies whether the Mat object should be destroyed when the TaoVec object is destroy. This object actually has the derived type TaoMatPetsc. Given a TaoMat whose underlying representation is a PETSc Vec, the command
TaoMatGetPetscMat( TaoMat *, Mat *);will retrieve the underlying matrix. The routine TaoMatDestroy() will destroy the TaoMat object, but the Mat object must also be destroyed.
TaoWrapKSP( KSP, TaoLinearSolver **);takes a KSP object and creates a TaoLinearSolver object. The
TaoLinearSolverGetKSP( TaoLinearSolver *, KSP *);gets the underlying KSP object from the TaoLinearSolver object.
For index sets, the routine
TaoWrapPetscIS( IS, int, TaoIndexSet **);creates a TaoIndexSet object. In this routine, however, the second argument is the local size of the vectors that this object will describe. For instance, this object may describe with elements of a vector are positive. The second argument should be be local length of the vector. The IS object will be destroyed when the TaoIndexSet is destroyed. The routine
TaoIndexSetGetPetscIS( TaoIndexSet *, IS *);will return the underlying IS object.
Portable TAO makefiles follow the rules and definitions of PETSc makefiles. In Figures 5.1 we present a sample makefile.
This small makefile is suitable for maintaining a single program that uses the TAO library. The most important line in this makefile is the line starting with include:
include ${TAO_DIR}/bmake/tao_commonThis line includes other makefiles that provide the needed definitions and rules for the particular base software installations (specified by ${TAO_DIR} and ${PETSC_DIR}) and architecture (specified by ${PETSC_ARCH}), which are typically set as environmental variables prior to compiling TAO source or programs. As listed in the sample makefile, the appropriate include file is automatically completely specified; the user should not alter this statement within the makefile.
TAO applications using PETSc should be linked with the to the PETSC_SNES_LIB library as well as the TAO_LIB library. This version uses PETSc 2.3.0, and the PETSC_DIR variable should be set accordingly. Many examples of makefiles can be found in the example problems.
Most of the functionality of TAO can be obtained by people who program purely in Fortran 77 or Fortran 90. Note, however, that we recommend the use of C and/or C++ because these languages contain several extremely powerful concepts that the Fortran77/90 family does not. The TAO Fortran interface works with both F77 and F90 compilers.
Since Fortran77 does not provide type checking of routine input/output parameters, we find that many errors encountered within TAO Fortran programs result from accidentally using incorrect calling sequences. Such mistakes are immediately detected during compilation when using C/C++. Thus, using a mixture of C/C++ and Fortran often works well for programmers who wish to employ Fortran for the core numerical routines within their applications. In particular, one can effectively write TAO driver routines in C++, thereby preserving flexibility within the program, and still use Fortran when desired for underlying numerical computations.
Only a few differences exist between the C and Fortran TAO interfaces, all of which are due to differences in Fortran syntax. All Fortran routines have the same names as the corresponding C versions, and command line options are fully supported. The routine arguments follow the usual Fortran conventions; the user need not worry about passing pointers or values. The calling sequences for the Fortran version are in most cases identical to the C version, except for the error checking variable discussed in Section 5.14.2. In addition, the Fortran routine TaoInitialize(char *filename,int info) differs slightly from its C counterpart; see the manual page for details.
Currently, TAO users must employ the Fortran file suffix .F rather than .f. This convention enables use of the CPP preprocessor, which allows the use of the #include statements that define TAO objects and variables. (Familiarity with the CPP preprocessor is not needed for writing TAO Fortran code; one can simply begin by copying a TAO Fortran example and its corresponding makefile.)
The TAO directory ${TAO_DIR}/include/finclude contains the Fortran include files and should be used via statements such as the following:
#include "include/finclude/includefile.h"Since one must be very careful to include each file no more than once in a Fortran routine, application programmers must manually include each file needed for the various TAO (or other supplementary) components within their program. This approach differs from the TAO C++ interface, where the user need only include the highest level file, for example, tao.h, which then automatically includes all of the required lower level files. As shown in the various Fortran example programs in the TAO distribution, in Fortran one must explicitly list each of the include files.
In the Fortran version, each TAO routine has as its final argument an integer error variable, in contrast to the C++ convention of providing the error variable as the routine's return value. The error code is set to be nonzero if an error has been detected; otherwise, it is zero. For example, the Fortran and C++ variants of TaoSolveApplication() are given, respectively, below, where info denotes the error variable:
call TaoSolveApplication(TAO_APPLICATION taoapp, TAO_SOLVER tao, int info) info = TaoSolveApplication(TAO_APPLICATION taoapp, TAO_SOLVER tao)
Fortran programmers can use the error codes in writing their own tracebacks. For example, one could use code such as the following:
call TaoSolveApplication(taoapp, tao, info) if (info .ne. 0) then print*, 'Error in routine ...' return endifIn addition, Fortran programmers can check these error codes with the macro CHKERRQ(), which terminates all process when an error is encountered. See the PETSc users manual for details. The most common reason for crashing PETSc Fortran code is forgetting the final info argument.
Additional interface differences: taosetlinesearch - use only the first and fourth arguments. The setup, options, view, and destroy routines do not apply.
Figure 5.2 shows a sample makefile that can be used for TAO Fortran programs. You can compile a debugging version of the program rosenbrock1f with make rosenbrock1f.
Note that the TAO Fortran interface library, given by ${TAO_FORTRAN_LIB}, must precede the base TAO library, given by ${TAO_LIB}, on the link line.
The TAO library currently interfaces to the PETSc library for low-level system functionality as well as linear algebra support. The PETSc users manual discusses additional Fortran issues in these areas, including
This section discusses options and routines that apply to all TAO solvers and problem classes. In particular, we focus on monitoring routines, convergence tests, and line searches.
By default the TAO solvers run silently without displaying information about the iterations. The user can initiate monitoring with the command
int TaoSetMonitor(TAO_SOLVER solver, int (*mon)(TAO_SOLVER tao,void* mctx), void *mctx);The routine, mon indicates a user-defined monitoring routine and mctx denotes an optional user-defined context for private data for the monitor routine.
The routine set by TaoSetMonitor() is called once during each iteration of the optimization solver. Hence, the user can employ this routine for any application-specific computations that should be done after the solution update. The option -tao_monitor activates the default monitoring routine, TaoDefaultMonitor() .
One can cancel all hardwired monitoring routines for TAO at runtime with the option -tao_cancelmonitors.
As the Newton method converges so that the gradient norm is small, many of the final digits printed with the -tao_monitor option are meaningless. In addition, they vary on different machines; due to different round-off rules used by, say, the IBM RS6000 and the Sun Sparc. This situation makes testing between different machines difficult. The option -tao_smonitor activates printing fewer of the digits of the residual norm as it decreases; thus on many machines it will always print the same numbers, making cross processor testing easier.
There are many different ways to define convergence of a solver. The methods TAO uses by default are mentioned in Section 3.3. These methods include absolute and relative convergence tolerances as well as a maximum number of iterations of function evaluations. If these choices are not sufficient, the user can even specify a customized test.
Users can set their own customized convergence tests of the form
int conv(TAO_SOLVER tao, void *cctx);The second argument is a pointer to a structure defined by the user. Within this routine, the solver can be queried for the solution vector, gradient vector, or other statistic at the current iteration through routines such as TaoGetSolutionStatus() and TaoGetTolerances().
To use this convergence test within a TAO solver, use the command
int TaoSetConvergenceTest(TAO_SOLVER solver, int (*conv)(TAO_SOLVER tao, void *cctx), void *cctx);The second argument of this command is the convergence routine, and the final argument of the convergence test routine, cctx, denotes an optional user-defined context for private data. The convergence routine receives the TAO solver and this private data structure. The termination flag can be set using the routine
int TaoSetTerminationReason(TAO_SOLVER , TaoTerminationReason*);
Many solver in TAO require a line search. While these solver always offer a default line search, alternative line searches can also be used. Line searches must have the form:
int L(TAO_SOLVER tao,TaoVec *xx,TaoVec *gg,TaoVec *dx, TaoVec *ww, double *f, double *step,double *gdx,int *flg,void *lsctx);In this routine the first argument is the TAO solver, the second argument is the current solution vector, the third argument is the gradient at the current point, the fourth argument is the step direction, the fourth vector is a work vector, the fifth argument is the function value, the sixth argument is the step length, the seventh argument is the inner product of the gradient and direction vector used for the Armijo condition, the eighth argument is a flag indicating success or failure of the line search, and the last argument is a pointer to a structure that can be used to define the line search. When the routine is finished the solution vector xx, gradient vector gg, function value f, step size step, and flg should be updated to reflect the new solution.
This routine can be set with the function
int TaoSetLineSearch(TAO_SOLVER solver, int (*L)(TAO_SOLVER,TaoVec*,TaoVec*,TaoVec*,TaoVec*, double*,double*,double*,int*,void*), void *ctx);In this routine, the second argument is the function pointer to the line search routine, and the third argument is the pointer that will be passed to the line search routine.
Since one must be very careful to include each file no more than once in a Fortran routine, application programmers must manually include each file needed for the various TAO (or other supplementary) components within their program. This approach differs from the TAO C++ interface, where the user need only include the highest level file, for example, tao_solver.h, which then automatically includes all of the required lower level files. As shown in the various Fortran example programs in the TAO distribution, in Fortran one must explicitly list each of the include files.
In the Fortran version, each TAO routine has as its final argument an integer error variable, in contrast to the C++ convention of providing the error variable as the routine's return value. The error code is set to be nonzero if an error has been detected; otherwise, it is zero. For example, the Fortran and C++ variants of TaoSolve() are given, respectively, below, where info denotes the error variable:
call TaoSolve(TAO_SOLVER tao, int info) info = TaoSolve(TAO_SOLVER tao)
Fortran programmers can use the error codes in writing their own tracebacks. For example, one could use code such as the following:
call TaoSolve(tao,info) if (info .ne. 0) then print*, 'Error in routine ...' return endifIn addition, Fortran programmers can check these error codes with the macro CHKERRQ(), which terminates all process when an error is encountered. See the PETSc users manual for details. The most common reason for crashing PETSc Fortran code is forgetting the final info argument.
Additional interface differences: taosetlinesearch - use only the first and fourth arguments. The setup, options, view, and destroy routines do not apply.
Figure 2.4 shows a sample makefile that can be used for TAO programs. In this makefile, one can compile and run a debugging version of the Fortran program rosenbrock1f.F with the action make rosenbrock1f.F. The compilation command is restated below:
rosenbrock1f: rosenbrock1f.o -${FLINKER} -o rosenbrock1f rosenbrock1f.o \ ${TAO_FORTRAN_LIB} ${TAO_LIB} ${RM} rosenbrock1f.oNote that the TAO Fortran interface library, given by ${TAO_FORTRAN_LIB}, must precede the base TAO library, given by ${TAO_LIB}, on the link line.
The TAO library currently interfaces to the PETSc library for low-level system functionality as well as linear algebra support. The PETSc users manual discusses additional Fortran issues in these areas, including
New optimization solvers can be added to TAO. TAO provides tools for facilitate the implementation of a solver. The advantages of implementing a solver using TAO are several.
The first argument is always the TAO structure. This structure may be used to obtain the vectors used to store the variables and the function gradient, evaluate a function and gradient, solve a set of linear equations, perform a line search, and apply a convergence test.
The second argument is specific to this solver. This pointer will be set in the initialization routine and cast to an appropriate type in the other routines. To implement the Fletcher - Reeves conjugate gradient algorithm, for instance, the following structure may be useful.
typedef struct{ double beta; TaoVec *dx; /* step direction */ TaoVec *ww; /* work vector */ } TAO_CG;This structure contains two work vectors and a scalar. Vectors for the solution and gradient are not needed here because the TAO structure has pointers to them.
static int TaoSolve_CG_FR(TAO_SOLVER tao, void *solver){ TAO_CG *cg = (TAO_CG *) solver; TaoVec *xx,*gg; /* solution vector, gradient vector */ TaoVec *dx=cg->dx, *ww=cg->ww; int iter=0,lsflag=0,info; double gnormPrev,gdx,f,gnorm,step=0; TaoTerminateReason reason; TaoFunctionBegin; info=TaoCheckFG(tao);CHKERRQ(info); info=TaoGetSolution(tao,&xx);CHKERRQ(info); info=TaoGetGradient(tao,&gg);CHKERRQ(info); info = TaoComputeMeritFunctionGradient(tao,xx,&f,gg);CHKERRQ(info); info = gg->Norm2(&gnorm); CHKERRQ(info); info = dx->SetToZero(); CHKERRQ(info); cg->beta=0; gnormPrev = gnorm; /* Enter loop */ while (1){ /* Test for convergence */ info = TaoMonitor(tao,iter++,f,gnorm,0.0,step,&reason);CHKERRQ(info); if (reason!=TAO_CONTINUE_ITERATING) break; cg->beta=(gnorm*gnorm)/(gnormPrev*gnormPrev); info = dx->Axpby(-1.0,gg,cg->beta); CHKERRQ(info); info = dx->Dot(gg,&gdx); CHKERRQ(info); if (gdx>=0){ /* If not a descent direction, use gradient */ cg->beta=0.0; info = dx->Axpby(-1.0,gg,cg->beta); CHKERRQ(info); gdx=-gnorm*gnorm; } /* Line Search */ gnormPrev = gnorm; step=1.0; info = TaoLineSearchApply(tao,xx,gg,dx,ww,&f,&step,&lsflag); info = gg->Norm2(&gnorm);CHKERRQ(info); } TaoFunctionReturn(0); }
The first line of this routine cast the second argument to a pointer to a TAO_CG data structure. This structure contains pointers to two vectors and a scalar which will be needed in the algorithm.
After declaring an initializing several variables, the solver first checks that the function and gradient have been defined using the routine TaoCheckFG(). Next, the solver get the variable and gradient vectors which were passed to TAO by the application program. Other solvers may also want to set pointers to Hessian matrices, Jacobian matrices, or vectors containing bounds on the variables. The commands for these routines are TaoGetSolution(), TaoGetGradient(), TaoGetVariableBounds(), TaoGetHessian(), and TaoGetJacobian().
This solver lets TAO evaluate the function and gradient at the current point in the using the routine TaoComputeFunctionGradient(). Other routines may be used to evaluate the Hessian matrix or evaluate constraints. TAO may obtain this information using direct evaluation of other means, but the these details do not affect our implementation of the algorithm.
The norm of the gradient is a standard measure used by unconstrained minimization solvers to define convergence. This quantity is always nonnegative and equals zero at the solution. The solver will pass this quantity, the current function value, the current iteration number, and a measure of infeasibility to TAO with the routine
int TaoMonitor(TAO_SOLVER,int,double,double,double,double, TaoTerminateReason*);Most optimization algorithms are iterative in nature, and solvers should include this command somewhere in each iteration. This routine records this information, applies any monitoring routines and convergence tests set by default or the user.
In this routine, the second argument is the current iteration number, and the third argument is the current function value. The fourth argument is a nonnegative error measure associated with the distance between the current solution and the optimal solution. Examples of this measure are the norm of the gradient or the square root of a duality gap. The fifth measure is a nonnegative error that is nonnegative and usually represents a residual between the current function value and the optimal solution, such as the norm of the gradient. The sixth argument is a nonnegative steplength, or the multiple of the step direction added to the previous iterate. The results of the convergence test are returned in the last argument. If the termination reason is TAO_CONTINUE_ITERATING, the algorithm should continue.
After this monitoring routine, the solver computes a step direction using methods defined on the TaoVec object. These methods include adding vectors together and computing an inner product. A full list of these methods can be found in the manual pages.
Nonlinear conjugate gradient algorithms also require a line search. TAO provides several line searches and support for using them. The routine
int TaoLineSearchApply(TAO_SOLVER tao, TaoVec *xx, TaoVec *gg, TaoVec *dx, TaoVec *ww, double *f, double *step, int*flag)passes the current solution, gradient, and objective value to the solver and returns a new solution, gradient, and objective value. More details on line searches can be found in the Section 6.3 The details of this line search are should be specified elsewhere, when the line search is created.
TAO also includes support for linear solvers. Although this algorithm does not require one, linear solvers are an important part of many algorithms. Details on the use of these solvers can be found in Section 5.10.
EXTERN_C_BEGIN int TaoCreate_CG_FR(TAO_SOLVER tao) { TAO_CG *cg; int info; TaoFunctionBegin; info = TaoNew(TAO_CG,&cg); CHKERRQ(info); info = TaoSetMaximumIterates(tao,2000); CHKERRQ(info); info = TaoSetTolerances(tao,1e-4,1e-4,0,0); CHKERRQ(info); info = TaoSetMaximumFunctionEvaluations(tao,4000); CHKERRQ(info); info = TaoCreateMoreThuenteLineSearch(tao,0,0.1); CHKERRQ(info); info = TaoSetTaoSolveRoutine(tao,TaoSolve_CG_FR,(void*)cg); CHKERRQ(info); info = TaoSetTaoSetUpDownRoutines(tao,TaoSetUp_CG,TaoDestroy_CG); CHKERRQ(info); info = TaoSetTaoOptionsRoutine(tao,TaoSetOptions_CG_FR); CHKERRQ(info); info = TaoSetTaoViewRoutine(tao,TaoView_CG); CHKERRQ(info); TaoFunctionReturn(0); } EXTERN_C_ENDThe first thing this routine does after declaring some variables, is allocate memory for the TAO_CG data structure. Clones of the the variable vector assed into TAO in the TaoCreate() routine are used as the two work vectors. This routine also sets some default convergence tolerances and creates a particular line search. These defaults could be specified in the routine that solves the problem, but specifying them here gives the user the opportunity to modify these parameters.
Finally, this solvers passes to TAO the names of all the other routines used by the solver.
Note that the lines EXTERN_C_BEGIN and EXTERN_C_END surround this routine. These terms are required to preserve the name of this function without any name-mangling from the C++ compiler.
int TaoDestroy_CG(TAO_SOLVER tao, void *solver) { TAO_CG *cg = (TAO_CG *) solver; int info; TaoFunctionBegin; info = TaoVecDestroy(cg->Work);CHKERRQ(info); info = TaoVecDestroy(cg->DX);CHKERRQ(info); info = TaoLineSearchDestroy(tao);CHKERRQ(info); TaoFree(cg); TaoFunctionReturn(0); }Other algorithms may destroy matrices, linear solvers, index sets, or other objects needed by the solver. This routine is called from within the TaoDestroy() routine.
int TaoSetUp_CG(TAO_SOLVER,void*); { int info; TaoVec *xx; TaoFunctionBegin; info = TaoGetSolution(tao,&xx);CHKERRQ(info); info = xx->Clone(&cg->ww); CHKERRQ(info); info = xx->Clone(&cg->dx); CHKERRQ(info); TaoFunctionReturn(0); }The second argument can be cast to the appropriate data structure. Many solvers use this routine to allocate data structures needed by the solver but not created by the initialization routine.
This document was generated using the LaTeX2HTML translator Version 2002-2-1 (1.70)
Copyright © 1993, 1994, 1995, 1996,
Nikos Drakos,
Computer Based Learning Unit, University of Leeds.
Copyright © 1997, 1998, 1999,
Ross Moore,
Mathematics Department, Macquarie University, Sydney.
The command line arguments were:
latex2html -no_math -image_type gif -reuse 0 -split 0 -local_icons manual
The translation was initiated by Jason Sarich on 2005-05-13