Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
Accelerating the Sinkhorn Algorithm for Sparse Multi-Marginal Optimal Transport via Fast Fourier Transforms
Previous Article in Journal
Integrated Industrial Reference Architecture for Smart Healthcare in Internet of Things: A Systematic Investigation
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Implicit A-Stable Peer Triplets for ODE Constrained Optimal Control Problems

1
Department of Mathematics, Technical University of Darmstadt, Dolivostraße 15, 64293 Darmstadt, Germany
2
Department of Mathematics, Philipps-Universität Marburg, Hans-Meerwein-Straße 6, 35043 Marburg, Germany
*
Author to whom correspondence should be addressed.
Algorithms 2022, 15(9), 310; https://doi.org/10.3390/a15090310
Submission received: 15 August 2022 / Accepted: 26 August 2022 / Published: 29 August 2022
(This article belongs to the Section Analysis of Algorithms and Complexity Theory)
Figure 1
<p>Rayleigh Problem: Convergence of the maximal state errors <math display="inline"><semantics> <mrow> <mrow> <mo>∥</mo> </mrow> <mrow> <mo>(</mo> <msup> <mi>w</mi> <mi mathvariant="sans-serif">T</mi> </msup> <mo>⊗</mo> <mi>I</mi> <mo>)</mo> </mrow> <msub> <mi>Y</mi> <mi>n</mi> </msub> </mrow> </semantics></math>−<math display="inline"><semantics> <mrow> <mi>y</mi> <mrow> <mo>(</mo> <msub> <mi>t</mi> <mrow> <mi>n</mi> <mo>+</mo> <mn>1</mn> </mrow> </msub> <mo>)</mo> </mrow> <msub> <mrow> <mo>∥</mo> </mrow> <mo>∞</mo> </msub> </mrow> </semantics></math> (<b>left</b>) and adjoint errors <math display="inline"><semantics> <mrow> <mrow> <mo>∥</mo> </mrow> <mrow> <mo>(</mo> <msup> <mi>v</mi> <mi mathvariant="sans-serif">T</mi> </msup> <mo>⊗</mo> <mi>I</mi> <mo>)</mo> </mrow> <msub> <mi>P</mi> <mi>n</mi> </msub> </mrow> </semantics></math>−<math display="inline"><semantics> <mrow> <mi>p</mi> <mrow> <mo>(</mo> <msub> <mi>t</mi> <mi>n</mi> </msub> <mo>)</mo> </mrow> <msub> <mrow> <mo>∥</mo> </mrow> <mo>∞</mo> </msub> </mrow> </semantics></math> (<b>right</b>), <math display="inline"><semantics> <mrow> <mi>n</mi> <mo>=</mo> <mn>0</mn> <mo>,</mo> <mo>…</mo> <mo>,</mo> <mi>N</mi> </mrow> </semantics></math>.</p> ">
Figure 2
<p>Van der Pol Problem: Convergence of the maximal state errors <math display="inline"><semantics> <mrow> <mrow> <mo>∥</mo> </mrow> <mrow> <mo>(</mo> <msup> <mi>w</mi> <mi mathvariant="sans-serif">T</mi> </msup> <mo>⊗</mo> <mi>I</mi> <mo>)</mo> </mrow> <msub> <mi>Y</mi> <mi>n</mi> </msub> </mrow> </semantics></math>−<math display="inline"><semantics> <mrow> <mi>y</mi> <mrow> <mo>(</mo> <msub> <mi>t</mi> <mrow> <mi>n</mi> <mo>+</mo> <mn>1</mn> </mrow> </msub> <mo>)</mo> </mrow> <msub> <mrow> <mo>∥</mo> </mrow> <mo>∞</mo> </msub> </mrow> </semantics></math> (<b>left</b>) and adjoint errors <math display="inline"><semantics> <mrow> <mrow> <mo>∥</mo> </mrow> <mrow> <mo>(</mo> <msup> <mi>v</mi> <mi mathvariant="sans-serif">T</mi> </msup> <mo>⊗</mo> <mi>I</mi> <mo>)</mo> </mrow> <msub> <mi>P</mi> <mi>n</mi> </msub> </mrow> </semantics></math>−<math display="inline"><semantics> <mrow> <mi>p</mi> <mrow> <mo>(</mo> <msub> <mi>t</mi> <mi>n</mi> </msub> <mo>)</mo> </mrow> <msub> <mrow> <mo>∥</mo> </mrow> <mo>∞</mo> </msub> </mrow> </semantics></math> (<b>right</b>), <math display="inline"><semantics> <mrow> <mi>n</mi> <mo>=</mo> <mn>0</mn> <mo>,</mo> <mo>…</mo> <mo>,</mo> <mi>N</mi> </mrow> </semantics></math>.</p> ">
Figure 3
<p>Controlled Motion Problem: optimal path <math display="inline"><semantics> <mrow> <mo>(</mo> <msubsup> <mi>y</mi> <mn>1</mn> <mo>☆</mo> </msubsup> <mo>,</mo> <msubsup> <mi>y</mi> <mn>2</mn> <mo>☆</mo> </msubsup> <mo>)</mo> </mrow> </semantics></math> through the total energy field <math display="inline"><semantics> <mrow> <mi>E</mi> <mo>=</mo> <mfrac> <mn>1</mn> <mn>2</mn> </mfrac> <msubsup> <mi>y</mi> <mn>2</mn> <mn>2</mn> </msubsup> </mrow> </semantics></math>+<math display="inline"><semantics> <mrow> <mfrac> <mn>1</mn> <mn>4</mn> </mfrac> <msubsup> <mi>y</mi> <mn>1</mn> <mn>4</mn> </msubsup> </mrow> </semantics></math>−<math display="inline"><semantics> <mrow> <mfrac> <mn>1</mn> <mn>2</mn> </mfrac> <msubsup> <mi>y</mi> <mn>1</mn> <mn>2</mn> </msubsup> </mrow> </semantics></math> visualized by isolines and exhibiting a saddle point structure (<b>top</b>). Convergence of the maximal state errors <math display="inline"><semantics> <mrow> <mrow> <mo>∥</mo> </mrow> <mrow> <mo>(</mo> <msup> <mi>w</mi> <mi mathvariant="sans-serif">T</mi> </msup> <mo>⊗</mo> <mi>I</mi> <mo>)</mo> </mrow> <msub> <mi>Y</mi> <mi>n</mi> </msub> </mrow> </semantics></math>−<math display="inline"><semantics> <mrow> <mi>y</mi> <mrow> <mo>(</mo> <msub> <mi>t</mi> <mrow> <mi>n</mi> <mo>+</mo> <mn>1</mn> </mrow> </msub> <mo>)</mo> </mrow> <msub> <mrow> <mo>∥</mo> </mrow> <mo>∞</mo> </msub> </mrow> </semantics></math> (<b>bottom left</b>) and adjoint errors <math display="inline"><semantics> <mrow> <mrow> <mo>∥</mo> </mrow> <mrow> <mo>(</mo> <msup> <mi>v</mi> <mi mathvariant="sans-serif">T</mi> </msup> <mo>⊗</mo> <mi>I</mi> <mo>)</mo> </mrow> <msub> <mi>P</mi> <mi>n</mi> </msub> </mrow> </semantics></math>−<math display="inline"><semantics> <mrow> <mi>p</mi> <mrow> <mo>(</mo> <msub> <mi>t</mi> <mi>n</mi> </msub> <mo>)</mo> </mrow> <msub> <mrow> <mo>∥</mo> </mrow> <mo>∞</mo> </msub> </mrow> </semantics></math> (<b>bottom right</b>), <math display="inline"><semantics> <mrow> <mi>n</mi> <mo>=</mo> <mn>0</mn> <mo>,</mo> <mo>…</mo> <mo>,</mo> <mi>N</mi> </mrow> </semantics></math>.</p> ">
Figure 4
<p>Wave Problem: Convergence of the maximal state errors <math display="inline"><semantics> <mrow> <mrow> <mo>∥</mo> </mrow> <mrow> <mo>(</mo> <msup> <mi>w</mi> <mi mathvariant="sans-serif">T</mi> </msup> <mo>⊗</mo> <mi>I</mi> <mo>)</mo> </mrow> <msub> <mi>Y</mi> <mi>n</mi> </msub> </mrow> </semantics></math>−<math display="inline"><semantics> <mrow> <mi>y</mi> <mrow> <mo>(</mo> <msub> <mi>t</mi> <mrow> <mi>n</mi> <mo>+</mo> <mn>1</mn> </mrow> </msub> <mo>)</mo> </mrow> <msub> <mrow> <mo>∥</mo> </mrow> <mo>∞</mo> </msub> </mrow> </semantics></math> (<b>left</b>) and adjoint errors <math display="inline"><semantics> <mrow> <mrow> <mo>∥</mo> </mrow> <mrow> <mo>(</mo> <msup> <mi>v</mi> <mi mathvariant="sans-serif">T</mi> </msup> <mo>⊗</mo> <mi>I</mi> <mo>)</mo> </mrow> <msub> <mi>P</mi> <mi>n</mi> </msub> </mrow> </semantics></math>−<math display="inline"><semantics> <mrow> <mi>p</mi> <mrow> <mo>(</mo> <msub> <mi>t</mi> <mi>n</mi> </msub> <mo>)</mo> </mrow> <msub> <mrow> <mo>∥</mo> </mrow> <mo>∞</mo> </msub> </mrow> </semantics></math> (<b>right</b>), <math display="inline"><semantics> <mrow> <mi>n</mi> <mo>=</mo> <mn>0</mn> <mo>,</mo> <mo>…</mo> <mo>,</mo> <mi>N</mi> </mrow> </semantics></math>.</p> ">
Review Reports Versions Notes

Abstract

:
This paper is concerned with the construction and convergence analysis of novel implicit Peer triplets of two-step nature with four stages for nonlinear ODE constrained optimal control problems. We combine the property of superconvergence of some standard Peer method for inner grid points with carefully designed starting and end methods to achieve order four for the state variables and order three for the adjoint variables in a first-discretize-then-optimize approach together with A-stability. The notion triplets emphasize that these three different Peer methods have to satisfy additional matching conditions. Four such Peer triplets of practical interest are constructed. In addition, as a benchmark method, the well-known backward differentiation formula BDF4, which is only A ( 73.35 ) -stable, is extended to a special Peer triplet to supply an adjoint consistent method of higher order and BDF type with equidistant nodes. Within the class of Peer triplets, we found a diagonally implicit A ( 84 ) -stable method with nodes symmetric in [ 0 ,   1 ] to a common center that performs equally well. Numerical tests with four well established optimal control problems confirm the theoretical findings also concerning A-stability.

1. Introduction

The design of efficient time integrators for the numerical solution of optimal control problems constrained by systems of ordinary differential equations (ODEs) is still an active research field. Such systems typically arise from semi-discretized partial differential equations describing, e.g., the dynamics of heat and mass transfer or fluid flow in complex physical systems. Symplectic one-step Runge–Kutta methods [1,2] exploit the Hamiltonian structure of the first-order optimality system—the necessary conditions to find an optimizer—and automatically yield a consistent approximation of the adjoint equations, which can be used to compute the gradient of the objective function. The first-order symplectic Euler, second-order Störmer–Verlet and higher-order Gauss methods are prominent representatives of this class, which are all implicit for general Hamiltonian systems, see the monograph [3]. Moreover, compositions of basic integrators with different steps sizes and splitting methods have been investigated. Generalized partitioned Runge–Kutta methods which allow one to compute exact gradients with respect to the initial condition are studied in [4]. To avoid the solution of large systems of nonlinear equations, semi-explicit W-methods [5] and stabilized explicit Runge–Kutta–Chebyshev methods [6] have been proposed, too. However, as all one-step methods, also symplectic Runge–Kutta schemes join the structural suffering of order reductions, which may lead to poor results in their application, e.g., to boundary control problems such as external cooling and heating in a manufacturing process; see [7,8] for a detailed study of this behaviour.
In contrast, multistep methods including Peer two-step methods avoid order reductions and allow a simple implementation [9,10]. However, the discrete adjoint schemes of linear multistep methods are in general not consistent or show a significant decrease of their approximation order [11,12]. Recently, we have developed implicit Peer two-step methods [13] with three stages to solve ODE constrained optimal control problems of the form
m i n i m i z e   C y ( T )
s u b j e c t   t o   y ( t ) = f y ( t ) , u ( t ) , u ( t ) U a d , t ( 0 , T ] ,
y ( 0 ) = y 0 ,
with the state y ( t ) R m , the control u ( t ) R d , f : R m × R d R m , the objective function C : R m R , where the set of admissible controls U a d R d is closed and convex. Introducing for any u U a d the normal cone mapping
N U ( u ) = { w R d : w T ( v u ) 0   f o r   a l l   v U a d } ,
the first-order Karush–Kuhn–Tucker (KKT) optimality system [14,15] reads
y ( t ) = f y ( t ) , u ( t ) , t ( 0 , T ] , y ( 0 ) = y 0 ,
p ( t ) = y f y ( t ) , u ( t ) T p ( t ) , t [ 0 , T ) , p ( T ) = y C y ( T ) T ,
u f y ( t ) , u ( t ) T p ( t ) N U u ( t ) , t [ 0 , T ] .
In this paper, we assume the existence of a unique local solution ( y , p , u ) of the KKT system with sufficient regularity properties to justify the use of higher order Peer triplets, see, e.g., the smoothness assumption in Section 2 of [14].
Remark 1.
The objective function C ( y ( T ) ) in (1) is specified in the so-called Mayer form using terminal solution values only. Terms given in the Lagrange form
C l ( y , u ) : = 0 T l ( y ( t ) , u ( t ) ) d t
can be equivalently reduced to the Mayer form by introducing an additional state y m + 1 , the new state vector y ˜ : = ( y 1 , , y m , y m + 1 ) T , and an additional differential equation y m + 1 ( t ) = l ( y ( t ) , u ( t ) ) with initial values y m + 1 ( 0 ) = 0 . Then, the Lagrange term (8) simply reduces to y m + 1 ( T ) .
Following a first-discretize-and-then-optimize approach, we apply an s-stage implicit Peer two-step method to (2) and (3) with approximations Y n i y ( t n + c i h ) and U n i u ( t n + c i h ) , i = 1 , , s , on an equi-spaced time grid { t 0 , , t N + 1 } [ 0 , T ] with step size h = ( T t 0 ) / ( N + 1 ) and nodes c 1 , , c s , which are fixed for all time steps, to  get the discrete constraint nonlinear optimal control problem
m i n i m i z e   C y h ( T )
s u b j e c t   t o   A 0 Y 0 = a y 0 + h b f ( y 0 , u 0 ) + h K 0 F ( Y 0 , U 0 ) ,
A n Y n = B n Y n 1 + h K n F ( Y n , U n ) , n = 1 , , N ,
y h ( T ) = ( w T I ) Y N ,
with long vectors Y n = ( Y n i ) i = 1 s R s m , U n = ( U n i ) i = 1 s R s d , and  F ( Y n , U n ) = f ( Y n i , U n i ) i = 1 s . Further, y h ( T ) y ( T ) , u 0 u ( 0 ) , and  a , b , w R s are additional parameter vectors at both interval ends, A n , B n , K n R s × s , and  I R m × m being the identity matrix. We will use the same symbol for a coefficient matrix like A and its Kronecker product A I as a mapping from the space R s m to itself. In contrast to one-step methods, Peer two-step methods compute Y n from the previous stage vector Y n 1 . Hence, also a starting method, given in (10), for the first time interval [ t 0 , t 1 ] is required. On each subinterval, Peer methods may be defined by three coefficient matrices A n , B n , K n , where A n and K n are assumed to be nonsingular. For practical reasons, this general version will not be used, the coefficients in the inner grid points will belong to a fixed Peer method ( A n , B n , K n ) ( A , B , K ) , n = 1 , , N 1 . The last forward step has the same form as the standard steps but uses exceptional coefficients ( A N , B N , K N ) to allow for a better approximation of the end conditions.
The KKT conditions (5)–(7) for ODE-constrained optimal control problems on a time interval [ 0 , T ] lead to a boundary value problem for a system of two differential equations, see Section 2.1. The first one corresponds to the original forward ODE for the state solution y ( t ) and the second one is a linear, adjoint ODE for a Lagrange multiplier p ( t ) . It is well known that numerical methods for such problems may have to satisfy additional order conditions for the adjoint equation [2,5,14,16,17,18]. While these additional conditions are rather mild for one-step methods they may lead to severe restrictions for other types of methods such as multistep and Peer methods, especially at the right-hand boundary at the end point T. Here, the order for the approximation of the adjoint equation may be limited to one.
For Peer methods, this question was discussed first in [19] and the adjoint boundary condition at T was identified as the most critical point. In a more recent article [13], these bottlenecks could be circumvented by two measures. First, equivalent formulations of the forward method are not equivalent for the adjoint formulation and using a redundant formulation of Peer methods with three coefficient matrices ( A , B , K ) adds additional free parameters. The second measure is to consider different methods for the first and last time interval. Hence, instead of one single Peer method (which will be called the standard method) we discuss triplets of Peer methods consisting of a common standard method ( A , B , K ) for all subintervals of the grid from the interior of [ 0 , T ] , plus a starting method ( A 0 , K 0 ) for the first subinterval and an end method ( A N , B N , K N ) for the last one. These two boundary methods may have lower order than the standard method since they are used only once.
The present work extends the results from [13] which considered methods with s = 3 stages only, in two ways. We will now concentrate on methods with four stages and better stability properties such as A-stability. The purpose of an accurate solution of the adjoint equation increases the number of conditions on the parameters of the method. Requiring high order of convergence s for the state variable y ( t ) and order s 1 for the adjoint variable p ( t ) —which we combine to the pair ( s , s 1 ) from now on—a variant of the method BDF3 was identified in [13] as the most attractive standard method. However, this method is not A-stable, with a stability angle of α = 86 . In order to obtain A-stability, we will reduce the required orders by one. Still, we will show that convergence with the orders ( s , s 1 ) can be regained by a superconvergence property.
The paper is organized as follows: In Section 2.1, the boundary value problem arising from the KKT system by eliminating the control and its discretization by means of discrete adjoint Peer two-step triplets are formulated. An extensive error analysis concentrating on the superconvergence effect is presented in Section 2.2. The restrictions imposed by the starting and end method on the standard Peer two-step method is studied in Section 2.3. The following Section 2.4 describes the actual construction principles of Peer triplets. Numerical tests are undertaken in Section 3. The paper concludes with a discussion in Section 4.

2. Materials and Methods

2.1. The Boundary Value Problem

Following the usual Lagrangian approach applied in [13], the first order discrete optimality conditions now consist of the forward Equations (10)–(12), the discrete adjoint equations, acting backwards in time,
A N T P N = w p h ( T ) + h Y F ( Y N , U N ) T K N T P N ,
A n T P n = B n + 1 T P n + 1 + h Y F ( Y n , U n ) T K n T P n , N 1 n 0 ,
and the control conditions
h U F ( Y n , U n ) T K n T P n N U s U n , 0 n N ,
h u 0 f ( y 0 , u 0 ) T ( b T I ) P 0 N U u 0 .
Here, p h ( T ) = y C y h ( T ) T and the Jacobians of F are block diagonal matrices Y F ( Y n , U n ) = d i a g i Y n i f ( Y n i , U n i ) and U F ( Y n , U n ) = d i a g i U n i f ( Y n i , U n i ) . The generalized normal cone mapping N U s U n is defined by
N U s ( u ) = w R s d : w T ( v u ) 0   f o r   a l l   v U a d s R s d .
The discrete KKT conditions (10)–(16) should be good approximations to the continuous ones (5)–(7). In what follows, we assume sufficient smoothness of the optimal control problem such that a local solution ( y , u , p ) of the KKT system (1)–(3) exists. Furthermore, let the Hamiltonian H ( y , u , p ) : = p T f ( y , u ) satisfy a coercivity assumption, which is a strong form of a second-order condition. Then the first-order optimality conditions are also sufficient [14]. If ( y , p ) is sufficiently close to ( y , p ) , the control uniqueness property introduced in [14] yields the existence of a locally unique minimizer u = u ( y , p ) of the Hamiltonian over all u U a d . Substituting u in terms of ( y , p ) in (5) and (6), gives then the two-point boundary value problem
y ( t ) = g y ( t ) , p ( t ) , y ( 0 ) = y 0 ,
p ( t ) = ϕ y ( t ) , p ( t ) , p ( T ) = y C y ( T ) T ,
with the source functions defined by
g ( y , p ) : = f y , u ( y , p ) , ϕ ( y , p ) : = y f y , u ( y , p ) T p .
The same arguments apply to the discrete first-order optimality system (10)–(16). Substituting the discrete controls U n = U n ( Y n , P n ) in terms of ( Y n , P n ) and defining
Φ ( Y n , K n T P n ) : = ϕ ( Y n i , ( K n T P n ) i ) i = 1 s , G ( Y n , P n ) : = g ( Y n i , P n i ) i = 1 s ,
the approximations for the forward and adjoint differential equations read in a compact form
A 0 Y 0 = a y 0 + h b g y 0 , p h ( 0 ) + h K 0 G ( Y 0 , P 0 ) ,
A n Y n = B n Y n 1 + h K n G ( Y n , P n ) , 1 n N ,
y h ( T ) = ( w T I ) Y N ,
p h ( 0 ) = ( v T I ) P 0 ,
A n T P n = B n + 1 T P n + 1 h Φ ( Y n , K n T P n ) , 0 n N 1 ,
A N T P N = w p h ( T ) h Φ ( Y N , K N T P N ) , n = N .
Here, the value of p h ( 0 ) is determined by an interpolant p h ( 0 ) = ( v T I ) P 0 p ( 0 ) with v R s of appropriate order. In a next step, these discrete equations are now treated as a discretization of the two-point boundary value problem (18) and (19). We will derive order conditions and give bounds for the global error.

2.2. Error Analysis

2.2.1. Order Conditions

The local error of the standard Peer method and the starting method is easily analyzed by Taylor expansion of the stage residuals, if the exact ODE solutions are used as stages. Hence, defining y n ( k ) ( h c ) : = y ( k ) ( t n + h c i ) i = 1 s , k = 0 , 1 , for the forward Peer method, where c = ( c 1 , , c s ) T , local order q 1 means that
A n y n ( h c ) B n y n 1 ( h c ) h K n y ( h c ) = O ( h q 1 ) .
In all steps of the Peer triplet, requiring local order q 1 for the state variable and order q 2 for the adjoint solution leads to the following algebraic conditions from [13]. These conditions depend on the Vandermonde matrix V q = ( 𝟙 , c , c 2 , , c q 1 ) R s × q , the Pascal matrix P q = j 1 i 1 i , j = 1 q and the nilpotent matrix E ˜ q = i δ i + 1 , j i , j = 1 q which commutes with P q = exp ( E ˜ q ) . For the different steps (22)–(24) and (25)–(27), and in the same succession we write down the order conditions from [13] when K n is diagonal. The forward conditions are
A 0 V q 1 = a e 1 T + b e 2 T + K 0 V q 1 E ˜ q 1 , n = 0 ,
A n V q 1 = B n V q 1 P q 1 1 + K n V q 1 E ˜ q 1 , 1 n N ,
w T V q 1 = 𝟙 T ,
with the cardinal basis vectors e i R s , i = 1 , , s . The conditions for the adjoint methods are given by
v T V q 2 = e 1 T ,
A n T V q 2 = B n + 1 T V q 2 P q 2 K n V q 2 E ˜ q 2 , 0 n N 1 ,
A N T V q 2 = w 𝟙 T K N V q 2 E ˜ q 2 , n = N .

2.2.2. Bounds for the Global Error

In this section, the errors Y ˇ n j : = y ( t n j ) Y n j , P ˇ n j : = p ( t n j ) P n j , n = 0 , , N , j = 1 , , s , are analyzed. According to [13], the equation for the errors Y ˇ T = ( Y ˇ 0 T , , Y ˇ N T ) and P ˇ T = ( P ˇ 0 T , , P ˇ N T ) is a linear system of the form
M h Z ˇ = τ , Z ˇ = Y ˇ P ˇ , τ = τ Y τ P ,
where the matrix M h has a 2 × 2 -block structure and ( τ Y , τ P ) denote the corresponding truncation errors. Deleting all h-depending terms from M h , the block structure of the remaining matrix M 0 is given by
M 0 = M 11 I m 0 M 21 y y C N M 22 I m
with M 11 , M 21 , M 22 R s ( N + 1 ) × s ( N + 1 ) and a mean value y y C N R m × m of the symmetric Hessian matrix of C. The index ranges of all three matrices are copied from the numbering of the grid, 0 , , N , for convenience. In fact, M 21 = ( e N 𝟙 ) ( e N w ) T has rank one only with e N = ( δ N j ) j = 0 N . The diagonal blocks of M 0 are nonsingular and its inverse has the form
M 0 1 = M 11 1 I m 0 ( M 22 1 M 21 M 11 1 ) y y C N M 22 1 I m .
The diagonal blocks M 11 , M 22 have a bi-diagonal block structure with identity matrices I s in the diagonal. The individual s × s -blocks of their inverses are easily computed with M 11 1 having lower triangular block form and M 22 1 upper triangular block form, with blocks
( M 11 1 ) n k = B ¯ n B ¯ k + 1 , k n , ( M 22 1 ) n k = B ¯ n + 1 T B ¯ k T , k n ,
with the abbreviations B ¯ n : = A n 1 B n , 1 n N and B ˜ n + 1 T : = ( B n + 1 A n 1 ) T , 0 n < N . Empty products for k = n mean the identity I s .
Defining U : = h 1 ( M h M 0 ) and rewriting (35) in fixed-point form
Z ˇ = h M 0 1 U Z ˇ + M 0 1 τ ,
it has been shown in the proof of Theorem 4.1 of [13] for smooth right hand sides f and h h 0 that
Z ˇ 2 max { M 11 1 τ Y , M 22 1 τ P }
in suitable norms, where these norms are discussed in more detail in Section 2.2.3 below. Moreover, due to the lower triangular block structure of M 0 the estimate for the error in the state variable may be refined (Lemma 4.2 in [13]) to
Y ˇ M 11 1 τ Y + h L Z ˇ
with some constant L. Without additional conditions, estimates of the terms on the right-hand side of (40) in the form Z ˇ = O ( h 1 τ ) lead to the loss of one order in the global error. However, this loss may be avoided by one additional superconvergence condition on the forward and the adjoint method each, which will be considered next.
In our convergence result Theorem 1 below, we will assume the existence of the discrete solution satisfying (22)–(27), for simplicity. However, at least for quadratic objective functions C, this existence follows quite simply along the same lines used in this paragraph here.
Lemma 1.
Let the right-hand side f of (2) be smooth with bounded second derivatives and let the function C be a polynomial of degree two, at most. Let the coefficients of the standard scheme satisfy
A 𝟙 = B 𝟙 , 𝟙 T A = 𝟙 T B , | λ 2 ( A 1 B ) | < 1 ,
where λ 2 denotes the absolutely second largest eigenvalue of the matrix A 1 B . Then, there exists a unique solution Y , P of the system (22)–(27) for small enough stepsizes h.
Proof. 
Since y y C is constant, M 0 1 in (37) is a fixed matrix and similar to (39) the Equations (22)–(27) may be written as a fixed-point problem
Z = h M 0 1 U ( Z ) , Z = Y P ,
where the function U ( Z ) is also smooth, having the Jacobian U considered in (39). Due to (42) there exist norms such that B ¯ 1 , B ˜ T 1 hold. In this case, the proof of Theorem 4.1 in [13] shows the bound M 0 1 U L with a constant L depending on the derivatives of g and ϕ . Hence, by the mean value theorem, the right-hand side of (43) has h L as a Lipschitz constant and the map Z h M 0 1 U ( Z ) is a contraction for h h 0 : = 1 / ( 2 L ) , establishing the existence of a unique fixed point solution.    □

2.2.3. Superconvergence of the Standard Method

For s-stage Peer methods, global order s may be attained in many cases if other properties of the method have lower priority. For optimal stiff stability properties like A-stability, however, it may be necessary to sacrifice one order of consistency as in [20,21]. Accordingly, in this paper the order conditions for the standard method are lowered by one compared to the requirements in the recent paper [13] to local orders ( s , s 1 ) , see Table 1. Still, the higher global orders may be preserved to some extent by the concept of superconvergence which prevents the order reduction in the global error by cancellation of the leading error term.
Superconvergence is essentially based on the observation that the powers of the forward stability matrix B ¯ : = A 1 B may converge to a rank-one matrix which maps the leading error term of τ Y to zero. This is the case if the eigenvalue 1 of B ¯ is isolated. Indeed, if the eigenvalues λ j , j = 1 , , s , of the stability matrix B ¯ satisfy
1 = λ 1 > | λ 2 | | λ s | ,
then its powers B ¯ n converge to the rank-one matrix 𝟙 𝟙 T A since 𝟙 and 𝟙 T A are the right and left eigenvectors having unit inner product 𝟙 T A 𝟙 = 1 due to the preconsistency conditions A 1 B 𝟙 = 𝟙 and A T B T 𝟙 = 𝟙 of the forward and backward standard Peer method, see (30) and (33). The convergence is geometric, i.e.,
B ¯ n 𝟙 𝟙 T A = B ¯ 𝟙 𝟙 T A n = O ( γ n ) 0 , n ,
for any γ ( | λ 2 | , 1 ) . Some care has to be taken here since the error estimate (40) depends on the existence of special norms satisfying B ¯ X 1 : = X 1 1 B ¯ X 1 = 1 , resp. B ˜ T X 2 = 1 . Concentrating on the forward error M 11 1 τ Y , a first transformation of B ¯ is considered with the matrix
X = 1 β T 𝟙 s 1 I s 1 𝟙 s 1 β T , X 1 = β 1 β T 𝟙 s 1 I s 1 ,
where ( β 1 , β T ) = 𝟙 T A . Since X e 1 = 𝟙 and e 1 T X 1 = 𝟙 T A , the matrix X 1 B ¯ X is block-diagonal with the dominating eigenvalue 1 in the first diagonal entry. Due to (44) there exists an additional nonsingular matrix Ξ R ( s 1 ) × ( s 1 ) such that the lower diagonal block of X 1 B X has norm smaller one, i.e.
Ξ 1 𝟙 s 1 , I s 1 B ¯ β T I s 1 𝟙 s 1 β T Ξ = γ < 1 .
Hence, with the matrix
X 1 : = X 1 Ξ
the required norm is found, satisfying B ¯ X 1 : = X 1 1 B ¯ X 1 = max { 1 , γ } = 1 and B ¯ 𝟙 𝟙 T A X 1 = γ < 1 in (45). Using this norm in (40) and (41) for ϵ Y : = M 11 1 τ Y , it is seen with (38) that
ϵ n Y = k = 0 n 𝟙 ( 𝟙 T A τ n k Y ) + k = 0 n B ¯ 𝟙 𝟙 T A k τ n k Y = O ( h s ) , 0 n < N .
Only for ϵ N Y a slight modification is required and the factors in the second sum have to be replaced by ( B ¯ N 𝟙 𝟙 T A ) ( B ¯ 𝟙 𝟙 T A ) k 1 for k > 1 with norms still of size O ( γ k ) . Now, in all cases the loss of one order in the first sum in (49) may be avoided if the leading O ( h s ) -term of τ n k Y is canceled in the product with the left eigenvector, i.e., if 𝟙 T A τ n k Y = O ( h s + 1 ) . An analogous argument may be applied to the second term M 22 1 τ P in (40). The adjoint stability matrix B ˜ T = ( B A 1 ) T possesses the same eigenvalues as B ¯ and its leading eigenvectors are also known: B ˜ T 𝟙 = 𝟙 and ( A 𝟙 ) T B ˜ T = 𝟙 T B T = ( A 𝟙 ) T by preconsistency and an analogous construction applies.
Under the conditions corresponding to local orders ( s , s 1 ) the leading error terms in τ n Y = 1 s ! h s η s y ( s ) ( t n ) + O ( h s + 1 ) and τ n P = 1 ( s 1 ) ! η s 1 * p ( s 1 ) ( t n ) + O ( h s ) are given by
η s = c s A 1 B ( c 𝟙 ) s s A 1 K c s 1 ,
η s 1 * = c s 1 A T B T ( c + 𝟙 ) s 1 + ( s 1 ) A T K c s 2 .
Considering now ( 𝟙 T A ) η s in (49) and similarly ( A 𝟙 ) T η s 1 * the following result is obtained.
Theorem 1.
Let the Peer triplet with s > 1 stages satisfy the order conditions collected in Table 1 and let the solutions satisfy y C s + 1 [ 0 , T ] , p C s [ 0 , T ] . Let the coefficients of the standard Peer method satisfy the conditions
𝟙 T A c s B ( c 𝟙 ) s s K c s 1 = 0 ,
𝟙 T A T c s 1 B T ( c + 𝟙 ) s 1 + ( s 1 ) K c s 2 = 0 ,
and let (44) be satisfied. Assume, that a Peer solution ( Y T , P T ) T exists and that f and C have bounded second derivatives. Then, for stepsizes h h 0 the error of these solutions is bounded by
Y n j y ( t n j ) + h P n j p ( t n j ) = O ( h s ) ,
n = 0 , , N , j = 1 , , s .
Proof. 
Under condition (52) the representation (49) shows that M 11 1 τ Y = O ( h s ) . In the same way follows M 22 1 τ P = O ( h s 1 ) from condition (53). In (40) this leads to a common error Z ˇ = O ( h s 1 ) which may be refined for the state variable with (41) to Y ˇ = O ( h s ) .    □
Remark 2.
The estimate (49) shows that superconvergence may be a fragile property and may be impaired if | λ 2 | is too close to one, leading to very large values in the bound
k = 0 n B ¯ 𝟙 𝟙 T A k X 1 1 1 γ
for the second term in (49). In fact, numerical tests showed that the value γ | λ 2 | plays a crucial role. While | λ 2 | 0.9 was appropriate for the three-stage method AP3o32f which shows order 3 in the tests in Section 3, for two four-stage methods with | λ 2 | 0.9 superconvergence was not seen in any of our three test problems below. Reducing | λ 2 | farther with additional searches we found that a value below γ = 0.8 may be safe to achieve order 4 in practice. Hence, the value | λ 2 | will be one of the data listed in the properties of the Peer methods developed below.
Remark 3.
By Theorem 1, weakening the order conditions from the local order pair ( s + 1 , s ) to the present pair ( s , s 1 ) combined with fewer conditions for superconvergence preserves global order h s for the state variable. However, it also leads to a more complicated structure of the leading error. Extending the Taylor expansion for τ Y , τ P and applying the different bounds, a more detailed representation of the state error may be derived,
Y ˇ h s η s ( 1 γ ) s ! y ( s ) + | 𝟙 T A η s + 1 | ( s + 1 ) ! y ( s + 1 ) + h s L ^ p ( s ) + L ^ p ( s 1 ) + O ( h s + 1 )
with some modified constant L ^ . Obviously, the leading error depends on four different derivatives of the solutions. Since the dependence on p is rather indirect, we may concentrate on the first line in (55). Both derivatives there may not be correlated and it may be difficult to choose a reasonable combination of both error constants as objective function in the construction of efficient methods. Still, in the local error the leading term is obviously τ Y h s 1 s ! η s y ( s ) , with  η s defined in (50), and it may be propagated through non-linearity and rounding errors. Hence, in the search for methods,
e r r s : = 1 s ! η s = 1 s ! c s A 1 B ( c 𝟙 ) s s A 1 K c s 1
is used as the leading error constant.

2.2.4. Adjoint Order Conditions for General Matrices K n

The number of order conditions for the boundary methods is so large that they may not be fulfilled with the restriction to diagonal coefficient matrices K 0 , K N for s 3 . Hence, it is convenient to make the step to full matrices K 0 , K N in the boundary methods and the order conditions for the adjoint schemes have to be derived for this case. Unfortunately, for such matrices the adjoint schemes (27) and (26) for n = 0 have a rather unfamiliar form. Luckily, the adjoint differential equation p = y f ( y , u ) T p is linear. We abbreviate the initial value problem for this equation as
p ( t ) = J ( t ) p ( t ) , p ( T ) = p T ,
with J ( t ) = y f y ( t ) , u ( t ) T and for some boundary index β { 0 , N } , we consider the matrices A β = ( a i j ( β ) ) , K β = ( κ i j ( β ) ) . Starting the analysis with the simpler end step (27), we have
j = 1 s a j i ( N ) P N j = w i p h ( T ) + h J ( t N i ) j = 1 s κ j i ( N ) P N j , i = 1 , , s ,
which is some kind of half-one-leg form since it evaluates the Jacobian J and the solution p at different time points. This step must be analyzed for the linear Equation (57) only. Expressions for the higher derivatives of the solution p follow easily:
p = ( J 2 J ) p , p = ( J 3 + 2 J J + J J J ) p .
Lemma 2.
For a smooth coefficient matrix J ( t ) , the scheme (58) for the linear differential Equation (57) has local order 3 under the conditions
A N T V 3 + K N T V 3 E ˜ 3 = w 𝟙 T
and with β = N
i = 1 i j s ( c i c j ) κ i j ( β ) = 0 , j = 1 , , s .
Proof. 
Considering the residual of the scheme with the exact solution p ( t ) , Taylor expansion at t N and the Leibniz rule yield
Δ i = j = 1 s a j i ( N ) p ( t N j ) w i p ( T ) h J ( t N i ) j = 1 s κ j i ( N ) p ( t N j ) = k = 0 q 1 h k k ! j = 1 s a j i ( N ) c j k w i p ( k ) h k = 0 q 2 h k k ! c i k J ( k ) j = 1 s κ j i ( N ) = 0 q 2 h ! c j p ( ) : = δ + O ( h q )
where all derivatives are evaluated at t N . The second term can be further reformulated as
δ = k = 1 q 1 h k = 0 k 1 j = 1 s κ j i ( N ) c i c j k 1 ! ( k 1 ) ! J ( ) p ( k 1 ) = k = 1 q 1 h k = 1 k j = 1 s κ j i ( N ) c i 1 c j k ( 1 ) ! ( k ) ! J ( 1 ) p ( k ) .
Looking at the factors of h 0 , h 1 , h 2 separately leads to the order conditions
h 0 : 0 = A N T 𝟙 w , h 1 : 0 = ( j = 1 s a j i ( N ) c j w i ) p j = 1 s κ j i ( N ) J p = ( j = 1 s a j i ( N ) c j + w i j = 1 s κ j i ( N ) ) J p , i . e . 0 = A N T c + K N T 𝟙 w , h 2 : 0 = 1 2 ( j a j i ( N ) c j 2 w i ) p j = 1 s κ j i ( N ) c j J p + c i J p = 1 2 ( j = 1 s c j 2 a j i ( N ) w i + 2 j = 1 s c j κ j i ( N ) ) J 2 p 1 2 ( j = 1 s c j 2 a j i ( N ) w i + 2 j = 1 s κ j i ( N ) c i ) J p .
Cancellation of the factor of J 2 requires the condition 0 = A N T c 2 + 2 K N T c w , which combines with the h 0 , h 1 -conditions to (60). The factor of J , however, requires 0 = A N T c 2 + 2 D c K N T 𝟙 w with D c = d i a g ( c i ) . Subtracting this expression from the factor of J 2 leaves 0 = K N T c D c K N T 𝟙 , which corresponds to (61).    □
Remark 4.
Condition (60) is the standard version for diagonal K N from [13]. Hence, the half-one-leg form of (58) introduces s additional conditions (61), only, for order 3 while the boundary coefficients K β , β = 0 , N , may now contain s ( s 1 ) additional elements. In fact, a similar analysis for step (26) with n = β = 0 reveals again (61) as the only condition in addition to (33).

2.3. Existence of Boundary Methods Imposes Restrictions on the Standard Method

In the previous paper [13], the combination of forward and adjoint order conditions for the standard method ( A , B , K ) into one set of equations relating only A and K already gave insight on some background of these methods such as the advantages of using symmetric nodes. It also simplifies the actual construction of methods leading to shorter expressions during the elimination process with algebraic software tools. For ease of reference, this singular Sylvester type equation is reproduced here,
( V q 2 P q 2 ) T A ( V q 1 P q 1 ) V q 2 T A V q 1 = ( V q 2 P q 2 ) T K V q 1 P q 1 E ˜ q 1 + ( V q 2 E ˜ q 2 ) T K V q 1 .
A similar combination of the order conditions for the boundary methods, however, reveals crucial restrictions: the triplet of methods ( A 0 , K 0 ) , ( A , B , K ) , ( A N , B N , K N ) has to be discussed together since boundary methods of an appropriate order may not exist for any standard method ( A , B , K ) , only for those satisfying certain compatibility conditions required by the boundary methods. Knowing these conditions allows one to design the standard method alone without the ballast of two more methods with many additional unknowns. This decoupled construction also greatly reduces the dimension of the search space if methods are optimized by automated search routines. We start with the discussion of the end method.

2.3.1. Combined Conditions for the End Method

We remind that we now are looking for methods having local order q 1 = s and q 2 = s 1 everywhere, which we abbreviate from now on as ( q 1 , q 2 ) = ( s , q ) . In particular this means that A T V q + K T V q E ˜ q = B T V q P q for the standard method. Looking for bottlenecks in the design of these methods, we try to identify crucial necessary conditions and consider the three order conditions for the end method ( A N , B N , K N ) in combination
A N V s B N V s P s 1 K N V s E ˜ s = 0 , B N T V q P q = A T V q + K T V q E ˜ q = B T V q P q , A N T V q + K N T V q E ˜ q = w 𝟙 q T .
From these conditions the matrices A N , B N may be eliminated, revealing the first restrictions on B. Here, the singular matrix map
L q , s : R q × s R q × s , X E ˜ q T X + X E ˜ s
plays a crucial role.
Lemma 3.
A necessary condition for a boundary method ( A N , B N , K N ) to satisfy (63) is
L q , s V q T K N V s = 𝟙 q 𝟙 s T V q T B V s P s 1 .
Proof. 
The second condition, V q T B N = V q T B , in (63) leads to the necessary equation
V q T A N V s V q T K N V s E ˜ s = V q T B V s P s 1
due to the first condition. The transposed third condition
V q T A N V s + E ˜ q T V q T K N V s = 𝟙 q w T V s = 𝟙 q 𝟙 s T
multiplied by the nonsingular matrix V s may be used to eliminate A N and leads to
E ˜ q T V q T K N V s + V q T K N V s E ˜ s = 𝟙 q 𝟙 s T V q T B V s P s 1
which is the equation (65) from the statement.    □
Unfortunately, this lemma leads to several restrictions on the design of the methods due to the properties of the map L q , s . Firstly, for diagonal matrices K N the image L q , s V q T K N V s has a very restricted shape.
Lemma 4.
If K R s × s is a diagonal matrix, then V q T K V s and L q , s V q T K V s are Hankel matrices with constant entries along anti-diagonals.
Proof. 
With K = d i a g ( κ i ) , we have x i j : = e i T ( V q T K V s ) e j = k = 1 s κ k c k i + j 2 , showing Hankel form of X = ( x i j ) = : ( ξ i + j 1 ) for i = 1 , , s 1 , j = 1 , , s . Now, e i T ( E ˜ q T X + X E ˜ s ) e j = ( i 1 ) x i 1 , j + x i , j 1 ( j 1 ) = ( i + j 2 ) ξ i + j 2 , which shows again Hankel form.    □
This lemma means that an end method with diagonal K N only exists if also V q T B V s P s 1 on the right-hand side of (65) has Hankel structure. Unfortunately it was observed that for standard methods with definite K this is the case for q 2 2 only (there exist methods with an explicit stage κ 33 = 0 ).
Remark 5.
Trying to overcome this bottleneck with diagonal matrices K N , one might consider adding additional stages of the end method. However, using general end nodes ( c ^ 1 , , c ^ s ^ ) with s ^ s does not remove this obstacle. The corresponding matrix L q , s ( V ^ q T K N V ^ s ) with appropriate Vandermonde matrices V ^ still has Hankel form.
However, even with a full end matrix K N , Lemma 3 and the Fredholm alternative enforce restrictions on the standard method ( A , B , K ) due to the singularity of L q , s . This is discussed for the present situation with s = 4 , q = 3 , only. The matrix belonging to the map L q , s is I s E ˜ q T + E ˜ s T I q and its transpose is I s E ˜ q + E ˜ s I q . Hence, the adjoint of the map L q , s is given by
L q , s T : R q × s R q × s , X E ˜ q X + X E ˜ s T .
Component-wise the map acts as
L q , s T : ( x i j ) i x i + 1 , j + j x i , j + 1
with elements having indices i > q or j > s being zero. For q = 3 , s = 4 , one gets
L 3 , 4 ( X ) T = x 12 + x 21 2 x 13 + x 22 3 x 14 + x 23 x 24 x 22 + 2 x 31 2 x 23 + 2 x 32 3 x 24 + 2 x 33 2 x 34 x 32 2 x 33 3 x 34 0 .
It is seen that the kernel of L 3 , 4 T has dimension 3 and is given by
X = ξ 1 ξ 2 ξ 3 0 ξ 2 2 ξ 3 0 0 ξ 3 0 0 0 .
In (65), the Fredholm condition leads to restrictions on the matrix B from the standard scheme. However, since the matrix A should have triangular form, it is the more natural variable in the search for good methods and an equivalent reformulation of these conditions for A is of practical interest.
Lemma 5.
Assume that the standard method ( A , B , K ) has local order ( s , q ) = ( 4 , 3 ) . Then, end methods ( A N , B N , K N ) of order ( s , q ) = ( 4 , 3 ) only exist if the standard method ( A , B , K ) satisfies the following set of three conditions, either for B or for A,
𝟙 T B 𝟙 = 1 𝟙 T B c c T B 𝟙 = 1 𝟙 T B c 2 2 c T B c + ( c 2 ) T B 𝟙 = 1 1 = 𝟙 T A 𝟙 1 = 𝟙 T A c c T A 𝟙 0 = 𝟙 T A c 2 2 c T A c + ( c 2 ) T A 𝟙 .
Proof. 
Multiplying equation (65) by P s from the right and using E ˜ s P s = P s E ˜ s , an equivalent form is
L q , s ( V q T K N V s P s ) = 𝟙 q 𝟙 s T P s V q T B V s = : R ,
with 𝟙 s T P s = ( 1 , 2 , 4 , 8 , ) by the binomial formula. The Fredholm alternative requires that t r ( X T R ) = 0 for all X from (66). We now frequently use the identities t r ( v u T ) R = u T R v and V e i = c i 1 with e i R s . The kernel in (66) is spanned by three basis elements. The first, X 1 = e ¯ 1 e 1 T (with the convention e ¯ i R q ) leads to
0 = ! t r ( X 1 T R ) = e ¯ 1 T 𝟙 q 𝟙 s T e 1 e ¯ 1 T V q T B V s e 1 = 1 𝟙 s T B 𝟙 s .
The second basis element is X 2 = e ¯ 1 e 2 T e ¯ 2 e 1 T . Here t r ( X 2 T 𝟙 q 𝟙 s T P s ) = 𝟙 s T P s e 2 𝟙 s T P s e 1 = 1 and
t r ( X 2 T V q T B V s ) = e ¯ 1 T V q B V s e 2 e ¯ 2 T V q B V s e 1 = 𝟙 s T B c c T B 𝟙 s .
For the third element, X 3 = e ¯ 1 e 3 T 2 e ¯ 2 e 2 T + e ¯ 3 e 1 T , one gets t r ( X 3 T 𝟙 q 𝟙 s T P s ) = 𝟙 s T P s ( e 3 2 e 2 + e 1 ) = 4 4 + 1 = 1 .
The third condition on B is
t r ( X 3 T V q T B V s ) = 𝟙 s T B c 2 2 c T B c + ( c 2 ) T B 𝟙 s .
The versions for A follow from the order conditions. Let again 𝟙 : = 𝟙 s . The first columns of (30) and (33) show B 𝟙 = A 𝟙 and 𝟙 T B = 𝟙 T A , which gives 𝟙 T B 𝟙 = 𝟙 T A 𝟙 = 1 . The second column of (30) reads B c = A c + A 𝟙 K 𝟙 and leads to 𝟙 T A c = 𝟙 T B c + 𝟙 T A 𝟙 𝟙 T K 𝟙 showing also 𝟙 T K 𝟙 = 𝟙 T A 𝟙 = 1 . Hence, the second condition in (67) is equivalent with
1 = ! 𝟙 T ( B c ) c T ( B 𝟙 ) = 𝟙 T ( A c + A 𝟙 K 𝟙 ) c T A 𝟙 = 𝟙 T A c c T A 𝟙 .
In order to show the last equivalence in (67), we have to look ahead at the forward condition (30) for order 3, which is B c 2 = A ( c 2 + 2 c + 𝟙 ) 2 K ( c + 𝟙 ) . This leads to 𝟙 T A c 2 = 𝟙 T B c 2 = 𝟙 T A ( c 2 + 2 c + 𝟙 ) 2 𝟙 T K ( c + 𝟙 ) , which is equivalent to 2 ( 𝟙 T A c 𝟙 T K c ) = 1 . Now, this expression is required in the last reformulation which also uses the second adjoint order condition B T c = A T c A T 𝟙 + K T 𝟙 , yielding
1 = ! ( 𝟙 T B ) c 2 + ( c 2 ) T ( B 𝟙 ) 2 c T ( B T c ) = 𝟙 T A c 2 + ( c 2 ) T A 𝟙 2 c T A T c + 2 ( c T A T 𝟙 c T K T 𝟙 ) = 1 .
   □
Remark 6.
It can be shown that only the first condition 𝟙 T A 𝟙 = 1 in (67) is required if the matrix K is diagonal. This first condition is merely a normalization fixing the free common factor in the class { α · ( A , B , K ) : α 0 } of equivalent methods. The other two conditions are consequences of the order conditions on the standard method ( A , B , K ) with diagonal K. However, the proof is rather lengthy and very technical and is omitted.
The restrictions (67) on the standard method also seem to be sufficient with (61) posing no further restrictions. In fact, with (67) the construction of boundary methods was always possible in Section 2.4.

2.3.2. Combined Conditions for the Starting Method

The starting method has to satisfy only two conditions
A 0 V s = a e 1 T + b e 2 T + K 0 V s E ˜ s , A 0 T V q = B T V q P q K 0 T V q E ˜ q .
The first two columns of the first equation may be solved for a = A 0 V s e 1 and b = ( A 0 V s K 0 V s E ˜ s ) e 2 leading to the reduced conditions
( A 0 V s K 0 V s E ˜ s ) Q 3 = 0 , Q 3 : = I s e 1 e 1 T e 2 e 2 T .
The presence of the projection Q 3 leads to changes in the condition compared to the end method.
Lemma 6.
A necessary condition for the starting method ( A 0 , K 0 ) to satisfy (68) is
L q , s P q T V q T K 0 V s V q T B V s Q 3 = 0 .
Proof. 
Transposing the second condition from (68) and multiplying with V s Q 3 gives
( V q T A 0 V s + E ˜ q T V q T K 0 V s P q T V q T B V s ) Q 3 = 0 ,
and V q T A 0 V s Q 3 may now be eliminated from both equations, yielding
( E ˜ q T V q T K 0 V s + V q T K 0 V s E ˜ s ) Q 3 = P q T V q T B V s Q 3 .
Again, P q T may be moved to the left side and leads to (69) since it commutes with  E ˜ q T .    □
The situation is now similar to the one for the end method (also concerning a diagonal form of K 0 ) and we consider again the Fredholm condition. The matrix belonging to the matrix product L q , s ( ) · Q 3 is Q 3 E ˜ q T + ( E ˜ s Q 3 ) T I q and it has the transpose Q 3 E ˜ q + ( E ˜ s Q 3 ) I q . This matrix belongs to the map
X E ˜ q ( X Q 3 ) + ( X Q 3 ) E ˜ s T = L q , s T ( X Q 3 ) .
For q = 3 , s = 4 , images of this map are given by
L 3 , 4 T ( X Q 3 ) = 0 2 x 13 x 23 + 3 x 14 x 24 0 2 x 23 2 x 33 + 3 x 24 2 x 34 0 2 x 33 3 x 34 0 .
Here, the map L 3 , 4 T alone introduces no new kernel elements, the kernel of (70) coincides with that of the map X X Q 3 given by matrices of the form X = X ˜ ( I Q 3 ) , X ˜ R q × s . Since the right-hand side of (69) is V q T B V s Q 3 , the condition for solvability
t r X T V q T B V s Q 3 = t r X ˜ T V q T B V s Q 3 ( I Q 3 ) = 0
is always satisfied since Q 3 ( I Q 3 ) = 0 . Hence, no additional restrictions on the standard method are introduced by requiring the existence of starting methods.

2.4. Construction of Peer Triplets

The construction of Peer triplets requires the solution of the collected order conditions from Table 1 and additional optimization of stability and error properties. However, it has been observed that some of these conditions may be related in non-obvious ways, see e.g., Remark 6. This means that the accuracy of numerical solutions may be quite poor due to large and unknown rank deficiencies. Instead, all order conditions were solved here exactly by algebraic manipulation with rational coefficients as far as possible.
The construction of the triplets was simplified by the compatibility conditions (67) allowing the isolated construction of the standard method ( A , B , K ) without the many additional parameters of the boundary methods. Furthermore, an elimination of the matrix B from forward and adjoint conditions derived in [13], see (62), reduces the number of parameters in A , K to s ( s + 3 ) / 2 elements with s 1 additional parameters from the nodes. This is so since ( A , B , K ) is invariant under a common shift of nodes and we chose the increments d 1 = c 2 c 1 , d j = c j c 2 , j = 3 , , s as parameters. Still, due to the mentioned dependencies between conditions, for  s = 4 a six-parameter family of methods exists which has been derived explicitly (with quite bulky expressions).
However, optimization of stability properties such as A ( α ) -stability or error constants e r r s from (56) was not possible in Maple with six free parameters. Instead, the algebraic expressions were copied to Matlab scripts for some Monte-Carlo-type search routines. The resulting coefficients of the standard method were finally approximated by rational expressions and brought back to Maple for the construction of the two boundary methods.
At first glance, having the full six-parameter family of standard methods at hand may seem to be a good work base. However, the large dimension of the search space may prevent optimal results with reasonable effort. This can be seen below, where the restriction to symmetric nodes or singly-implicit methods yielded methods with smaller e r r 4 than automated global searches.
A ( α ) -stability of the method may be checked [13] by considering the eigenvalue problem for the stability matrix ( A z K ) 1 B , i.e.,
B x = λ ( A z K ) x K 1 ( A λ 1 B ) x = z x .
as an eigenvalue problem for z C where | λ 1 | = 1 runs along the unit circle. Since we focus on A-stable methods, exact verification of this property would have been preferable, of course, but an algebraic proof of A-stability seemed to be out of reach. It turned out that the algebraic criterion of the second author [22] is rather restrictive (often corresponding to norm estimates, Lemma 2.8 ibd). On the other hand, application of the exact Schur criterion [20,23] is not straight-forward and hardly feasible, since the (rational) coefficients of the stability polynomial are prohibitively large for optimized parameters (dozens of decimal places of the numerators).

2.4.1. Requirements for the Boundary Methods

Since Lemma 5 guarantees the existence of the two boundary methods ( A 0 , K 0 ) and ( A N , B N , K N ) , their construction can follow after that of the standard method. Requirements for these two members of the triplet may also be weakened since they are applied once only. This relaxation applies to the order conditions as shown in Table 1, but also to stability requirements. Still, the number of conditions at the boundaries is so large that the diagonal triangular forms of K 0 , K N and A 0 , A N respectively have to be sacrificed and replaced by some triangular block structure. Compared to the computational effort of the complete boundary value problem, the additional complexity of the two boundary steps should not be an issue. However, for non-diagonal matrices K 0 , K N and s = 4 , the four additional one-leg-conditions (61) have to be obeyed.
Weakened stability requirements mean that the last forward Peer step (23) for n = N and the two adjoint Peer steps (26) for n = N 1 and n = 0 need not be A-stable and only nearly zero stable if the corresponding stability matrices have moderate norms. This argument also applies to the two Runge–Kutta steps without solution output (22) and (27). However, the implementation of these steps should be numerically safe for stiff problems and arbitrary step sizes. These steps require the solution of two linear systems with the matrices A 0 h K 0 J 0 , A N T h J N T K N T or, rather
K 0 1 A 0 h J 0 , ( K N 1 A N ) T h J N T ,
where J 0 , J N are block diagonal matrices of Jacobians. These Jacobians are expected to have absolutely large eigenvalues in the left complex half-plane. For such eigenvalues, non-singularity of the matrices (72) is assured under the following eigenvalue condition:
μ 0 : = min j R e λ j ( K 0 1 A 0 ) > 0 , μ N : = min j R e λ j ( K N 1 A N ) > 0 .
In Table 2 the constants μ 0 , μ N are displayed for all designed Peer triplets as well as the spectral radii ϱ ( A N 1 B N ) and ϱ ( B A 0 1 ) , ϱ ( B N A 1 ) for the boundary steps of the Peer triplet. In the search for the boundary methods with their exact algebraic parameterizations, these spectral radii were minimized, and if they were close to one the values μ 0 , μ N > 0 were maximized.

2.4.2. A ( α ) -Stable Four-Stage Methods

Although our focus is on A-stable methods, we also shortly consider A ( α ) -stable methods. We like to consider BDF4 (backward differentiation formulas) as a benchmark, since triplets based on BDF3 were the most efficient ones in [13]. In order to distinguish the different methods, we denote them in the form APso q 1 q 2 aaa, where AP stands for Adjoint Peer method followed by the stage number and the smallest forward and adjoint orders q 1 and q 2 in the triplet. The trailing letters are related to properties of the method like diagonal or singly diagonal implicitness.
The Peer triplet AP4o43bdf based on BDF4 has equi-spaced nodes c i = i / 4 , i = 1 , , 4 , yielding w = e 4 . The coefficients of the full triplet are given in Appendix A.1. Obviously the method is singly-implicit and its well-known stability angle is α = 73.35 . We also monitor the norm of the zero-stability matrix A 1 B , which may be a measure for the propagation of rounding errors. Its value is A 1 B 5.80 . Since BDF4 has full global order 4, the error constant from (56) is e r r 4 = 0 . Still, the end methods were constructed with the local orders ( 4 , 3 ) according to Table 1. The matrices of the corresponding starting method have a leading 3 × 3 block and a separated last stage. We abbreviate this as block structure ( 3 + 1 ) . All characteristics of the boundary method are given in Table 2.
Motivated by the beneficial properties of Peer methods with symmetric nodes seen in [13,19], the nodes of the next triplet with the diagonally-implicit standard method were chosen symmetric to a common center, i.e.,  c 1 + c 4 = c 2 + c 3 . Unfortunately, searches for large stability angles with such nodes in the interval [ 0 , 1 ] did not find A-stable methods, but the following method AP4o43dif with flip symmetric nodes and α = 84.00 , which is an improvement of 10 degrees over BDF4. Its coefficients are given in Appendix A.2. Although there exist A-stable methods with other nodes in [ 0 , 1 ] , this method is of its own interest since its error constant e r r 4 2.5 × 10 3 is surprisingly small. The node vector of AP4o43dif includes c 4 = 1 , leading again to w = e 4 . Further properties of the standard method ( A , B , K ) are A 1 B 2.01 and the damping factor | λ 2 | = 0.26 . See Table 2 for the boundary methods.

2.4.3. A-Stable Methods

By extensive searches with the full six-parameter family of diagonally-implicit four-stage methods many A-stable methods were found even with nodes in [ 0 , 1 ] . In fact, regions with A-stable methods exist in at least three of the eight octants in ( d 1 , d 3 , d 4 ) -space. Surprisingly, however, for none of these methods the last node c 4 was the rightmost one. In addition, it may be unexpected that some of the diagonal elements of A and K have negative signs. This does not cause stability problems if a i i κ i i > 0 , i = 1 , , s , see also (72). A-stability assured, the search procedure tried to minimize a linear combination e r r s + δ A 1 B of the error constant and the norm of the stability matrix with small δ < 10 3 to account for the different magnitudes of these data. As mentioned in Remark 2, it was also necessary to include the damping factor | λ 2 | in the minimization process. One of the best A-stable standard methods with general nodes was named AP4o43dig. Its coefficients are given in Appendix A.3. It has an error constant e r r 4 0.0260 and damping factor | λ 2 | 0.798 , see Table 3. No block structure was chosen in the boundary methods in order to avoid large norms of the mixed stability matrices B A 0 1 , B N A 1 , A N 1 B N , see Table 2.
It was observed that properties of the methods may improve, if the nodes have a wider spread than the standard interval [ 0 , 1 ] . In our setting, the general vector w allows for an end evaluation y h ( T ) at any place between the nodes. Since an evaluation roughly in the middle of all nodes may have good properties, in a further search the nodes were restricted to the interval [ 0 , 2 ] . Indeed, all characteristic data of the method AP4o43die with extended nodes presented in Appendix A.4 have improved. As mentioned before, the standard method is invariant under a common node-shift and a nearly minimal error constant was obtained with the node increments d 1 = 10 11 , d 3 = 1 and d 4 = 2 3 . Then, the additional freedom in the choice of c 2 was needed for the boundary methods, since the conditions (73) could only be satisfied in a small interval around c 2 = 5 4 . The full node vector with this choice has alternating node increments since c 3 < c 1 < c 2 < c 4 . The method is A-stable, its error constant e r r 4 0.0136 is almost half as large as for the method AP4o43dig, and  A 1 B 6.1 and | λ 2 | 2 3 are smaller, too. The data of the boundary methods can be found in Table 2.
For medium-sized ODE problems, where direct solvers for the stage equations may be used, an additional helpful feature is diagonal singly-implicitness of the standard method. In our context this means that the triangular matrices K 1 A and A K 1 have a constant value θ in the main diagonal. Using the ansatz
a i i = θ κ i i , i = 1 , , s ,
for A = ( a i j ) and K = ( κ i j ) , the order conditions from Table 1 lead to a cubic equation for θ with no rational solutions, in general. In order to avoid pollution of the algebraic elimination through superfluous terms caused by rounding errors, numerical solutions for this cubic equations were not used until the very end. This means that also the order conditions for the boundary methods were solved with θ as a free parameter, only the final evaluation of the coefficients in Appendix A.5 was performed with its numerical value. In addition, in the Matlab-search for A-stable methods, the cubic equation was solved numerically and it turned out that the largest positive solution gave the best properties. Hence, this Peer triplet was named AP4o43sil. For a first candidate with nearly minimal e r r s + δ A 1 B , the damping factor γ 0.89 was again too close to one to ensure superconvergence in numerical tests, see Remark 2. However, further searches nearby minimizing the damping factor found a better standard method with | λ 2 | = 0.60 . Its nodes are c T = 1 50 , 3 5 , 1 , 41 85 , the diagonal parameter θ = 3.34552931287687520 is the largest zero of the cubic equation
112673616 θ 3 + 106686908 θ 2 2102637319 θ + 1621264295 = 0 .
Further properties of the standard method are A-stability, nodes in ( 0 , 1 ] with c 3 = 1 , norm A 1 B = 32.2 , and an error constant e r r 4 = 0.0230 . The end method ( A N , B N , K N ) has full matrices A N , K N , see Table 2.
For the sake of completeness, we also present an A-stable diagonally-implicit three-stage method, since in the previous paper [13] we could not find such methods with reasonable parameters. After relaxing the order conditions by using superconvergence, such methods exist. Applying all conditions for forward order s = 3 and adjoint order s 1 = 2 , there remains a five-parameter family depending on the node differences d 1 = c 2 c 1 , d 3 = c 3 c 2 and three elements of A or K. A-stable methods exist in all four corners of the square [ 1 2 , 1 2 ] 2 in the ( d 1 , d 3 ) -plane, the smallest errors e r r 3 were observed in the second quadrant. The method AP3o32f with ( d 1 , d 3 ) = ( 5 27 , 2 5 ) has a node vector with c 3 = 1 . The coefficients can be found in Appendix A.6. The characteristic data are e r r 3 0.017 , A 1 B 15.3 , | λ 2 | = 0.91 . The starting method is of standard form with lower triangular A 0 and diagonal K 0 .
The main properties of the newly developed Peer triplets are summarized in Table 3 for the standard methods and Table 2 for the boundary methods.

3. Results

We present numerical results for all Peer triplets listed in Table 3 and compare them with those obtained for the third-order four-stage one-step W-method Ros3wo proposed in [5] which is linearly implicit (often called semi-explicit in the literature) and also A-stable. All calculations have been carried out with Matlab-Version R2019a, using the nonlinear solver fsolve to approximate the overall coupled scheme (22)–(27) with a tolerance 10 14 . To illustrate the rates of convergence, we consider three nonlinear optimal control problems, the Rayleigh problem, the van der Pol oscillator, and a controlled motion problem. A  linear wave problem is studied to demonstrate the practical importance of A-stability for larger time steps.

3.1. The Rayleigh Problem

The first problem is taken from [24] and describes the behaviour of a tunnel-diode oscillator. With the electric current y 1 ( t ) and the transformed voltage at the generator u ( t ) , the Rayleigh problem reads
M i n i m i z e   0 2.5 u ( t ) 2 + y 1 ( t ) 2 d t
  s u b j e c t   t o   y 1 ( t ) y 1 1.4 0.14 y 1 ( t ) 2 + y 1 ( t ) = 4 u ( t ) , t ( 0 , 2.5 ] ,
y 1 ( 0 ) = y 1 ( 0 ) = 5 .
Introducing y 2 ( t ) = y 1 ( t ) and eliminating the control u ( t ) yields the following nonlinear boundary value problem (see [5] for more details):
y 1 ( t ) = y 2 ( t ) ,
y 2 ( t ) = y 1 ( t ) + y 2 ( t ) 1.4 0.14 y 2 ( t ) 2 8 p 2 ( t ) ,
y 1 ( 0 ) = 5 , y 2 ( 0 ) = 5 ,
p 1 ( t ) = p 2 ( t ) 2 y 1 ( t ) ,
p 2 ( t ) = p 1 ( t ) ( 1.4 0.42 y 2 ( t ) 2 ) p 2 ( t ) ,
p 1 ( 2.5 ) = 0 , p 2 ( 2.5 ) = 0 .
To study convergence orders of our new methods, we compute a reference solution at the discrete time points t = t n by applying the classical fourth-order RK4 with N = 1280 steps. Numerical results for the maximum state and adjoint errors are presented in Figure 1 for N + 1 = 20 , 40 , 80 , 160 , 320 . AP3o32f and Ros3wo show their expected orders ( 3 , 2 ) and ( 3 , 3 ) for state and adjoint solutions, respectively. Order three for the adjoint solutions is achieved by all new four-stage Peer methods. The smaller error constants of AP4o43bdf, AP4o43dif and Ros3wo are clearly visible. The additional superconvergence order four for the state solutions shows up for AP4o43die and AP4o43sil and nearly for AP4o43dif and AP4o43dig. AP4o43bdf does not reach its full order four here, too.

3.2. The van der Pol Oscillator

Our second problem is the following optimal control problem for the van der Pol oscillator:
M i n i m i z e   0 2 u ( t ) 2 + y ( t ) 2 + y ( t ) 2 d t
s u b j e c t   t o   ε y ( t ) 1 y ( t ) 2 y ( t ) + y ( t ) = u ( t ) , t ( 0 , 2 ] ,
y ( 0 ) = 0 , y ( 0 ) = 2 .
Introducing Lienhardt’s coordinates y 2 ( t ) = y ( t ) , y 1 ( t ) = ε y ( t ) + y ( t ) 3 / 3 y ( t ) , and eliminating the control u ( t ) , we finally get the following nonlinear boundary value problem in [ 0 , 2 ] (see again [5] for more details):
y 1 ( t ) = y 2 ( t ) p 1 ( t ) 2 ,
y 2 ( t ) = 1 ε y 1 ( t ) + y 2 ( t ) y 2 ( t ) 3 3 ,
y 1 ( 0 ) = 2 ε , y 2 ( 0 ) = 0 ,
p 1 ( t ) = 1 ε p 2 ( t ) 1 ε 2 y 1 ( t ) + y 2 ( t ) y 2 ( t ) 3 3 ,
p 2 ( t ) = p 1 ( t ) 1 ε 1 y 2 ( t ) 2 p 2 ( t )
2 ε 2 y 1 ( t ) + y 2 ( t ) y 2 ( t ) 3 3 1 y 2 ( t ) 2 2 y 2 ( t ) ,
p 1 ( 2 ) = 0 , p 2 ( 2 ) = 0 .
The van der Pol equation with small positive values of ε gives rise to very steep profiles in y ( t ) , requiring variable step sizes for an efficient numerical approximation. Since the factor ε 2 appears in the adjoint equations, the boundary value problem is even harder to solve. In order to apply constant step sizes with N + 1 = 20 , 40 , 80 , 160 , 320 , we consider the mildly stiff case with ε = 0.1 for our convergence study.
Numerical results for the maximum state and adjoint errors are presented in Figure 2, where a reference solution is computed with AP4o43bdf for N = 1279 . Order three for the adjoint solutions is achieved by all new four-stage Peer methods and also by Ros3wo. The three-stage Peer method AP3o32f drops down to order 1.3 . For AP4o43dig and AP4o43sil applied with N = 19 , Matlab’s fsolve was not able to deliver a solution. The additional superconvergence order four for the state solutions is visible for AP4o43die and nearly for AP4o43dif which performs best and beats AP4o43bdf by a factor five. AP4o43bdf does not reach its full order four here. The methods AP4o43sil and AP4o43dig show order three only, whereas AP3o32f and Ros3wo reach their theoretical order three.

3.3. A Controlled Motion Problem

This problem was studied in [1]. The motion of a damped oscillator is controlled in a double-well potential, where the control u ( t ) acts only on the velocity y 2 ( t ) . The optimal control problem reads
M i n i m i z e   α 2 y ( 6 ) y f 2 + 1 2 0 6 u ( t ) 2 d t
s u b j e c t   t o   y 1 ( t ) y 2 ( t ) = 0 ,
y 2 ( t ) y 1 ( t ) + y 1 ( t ) 3 + ν y 2 ( t ) = u ( t ) , t ( 0 , 6 ] ,
y 1 ( 0 ) = 1 , y 2 ( 0 ) = 0 ,
where ν > 0 is the damping parameter and y f the target final position. As in [1], we set ν = 1 , y f = ( 1 , 0 ) T , and  α = 10 .
Eliminating the scalar control u ( t ) yields the following nonlinear boundary value problem:
y 1 ( t ) = y 2 ( t ) ,
y 2 ( t ) = y 1 ( t ) y 1 ( t ) 3 ν y 2 ( t ) p 2 ( t ) ,
y 1 ( 0 ) = 1 , y 2 ( 0 ) = 0 ,
p 1 ( t ) = ( 3 y 1 ( t ) 2 1 ) p 2 ( t ) ,
p 2 ( t ) = p 1 ( t ) + ν p 2 ( t ) ,
p 1 ( 6 ) = α ( y 1 ( 6 ) 1 ) , p 2 ( 6 ) = α y 2 ( 6 ) .
The optimal control u = p 2 must accelerate the motion of the particle to follow an optimal path ( y 1 , y 2 ) through the total energy field E = 1 2 y 2 2 + 1 4 y 1 4 1 2 y 1 2 , shown in Figure 3 on the top, in order to reach the final target y f behind the saddle point. The cost obtained from a reference solution with N = 1279 is C ( y ( 6 ) ) = 0.77674 , which is in good agreement with the lower order approximation in [1]. Numerical results for the maximum state and adjoint errors are presented in Figure 3 for N + 1 = 10 , 20 , 40 , 80 , 160 . Worthy of mentioning is the repeated excellent performance of AP4o43bdf and AP4o43dif, but also the convincing results achieved by the third-order method Ros3wo. All theoretical orders are well observable, except for AP4o43dig, which tends to order three for the state solutions. A closer inspection reveals that this is caused by the second state y 2 , while the first one asymptotically converges with fourth order. However, the three methods AP4o43die, AP4o43dig, and AP4o43sil perform quite similar. Observe that AP3o32f has convergence problems for N = 9 .
The stagnation of the state errors for the finest step sizes is due to the limited accuracy of Matlab’s fsolve—a fact which was already reported in [17].

3.4. A Wave Problem

The fourth problem is taken from [25] and demonstrates the practical importance of A-stability. We consider the optimal control problem
M i n i m i z e   y 1 ( 1 ) + 1 2 0 1 u ( t ) 2 d t
s u b j e c t   t o   y 1 ( t ) + ( 2 π κ ) 2 y 1 ( t ) = u ( t ) , t ( 0 , 1 ] ,
y 1 ( 0 ) = y 1 ( 0 ) = 0 ,
where κ = 16 is used. Introducing y 2 ( t ) = y 1 ( t ) and eliminating the control u ( t ) yields the following linear boundary value problem:
y 1 ( t ) = y 2 ( t ) ,
y 2 ( t ) = ( 2 π κ ) 2 y 1 ( t ) p 2 ( t ) ,
y 1 ( 0 ) = 0 , y 2 ( 0 ) = 0 ,
p 1 ( t ) = ( 2 π κ ) 2 p 2 ( t ) ,
p 2 ( t ) = p 1 ( t ) ,
p 1 ( 1 ) = 1 , p 2 ( 1 ) = 0 .
The exact solutions are given by
y 1 * ( t ) = 1 2 ( 2 π κ ) 3 sin ( 2 π κ t ) t 2 ( 2 π κ ) 2 cos ( 2 π κ t ) ,
y 2 * ( t ) = t 2 ( 2 π κ ) sin ( 2 π κ t ) ,
p 1 * ( t ) = cos ( 2 π κ t ) , p 2 * ( t ) = 1 2 π κ sin ( 2 π κ t ) ,
and the optimal control is u * ( t ) = p 2 * ( t ) . The key observation here is that the eigenvalues of the Jacobian of the right-hand side in (106)–(111) are λ 1 / 2 = 2 π κ i and λ 3 / 4 = 2 π κ i , which requests appropriate step sizes for only the A ( α )-stable methods AP4o43bdf and AP4o34dif due to their stability restrictions along the imaginary axis. Indeed, a closer inspection of the stability region of the (multistep) BDF4 method near the origin reveals that | λ h b d f 4 | 0.3 is a minimum requirement to achieve acceptable approximations for problems with imaginary eigenvalues and moderate time horizon. For the four-stage AP4o43bdf with step size h = 4 h b d f 4 , this yields | λ h | 1.2 and hence h 1 / ( 32 π ) 0.02 for the wave problem considered. A similar argument applies to AP4o34dif, too. Numerical results for the maximum state and adjoint errors are plotted in Figure 4 for N + 1 = 20 , 40 , 80 , 160 , 320 . They clearly show that both methods deliver first feasible results for h = 1 / 80 and below only, but then again outperform the other Peer methods. Once again, Ros3wo performs remarkably well. The orders of convergence for the adjoint solutions are one better than the theoretical values, possibly due to the overall linear structure of the boundary value problem.

4. Discussion

We have extended our three-stage adjoint Peer two-step methods constructed in [13] to four stages to not only improve the convergence order of the methods but also their stability. Combining superconvergence of a standard Peer method with a careful design of starting and end Peer methods with appropriately enhanced structure, discrete adjoint A-stable Peer methods of order (4,3) could be found. Still, new requirements had to be dealt with for the higher order pair (4,3). The property of A-stability comes with larger error constants and a few other minor structural disadvantages. As long as A-stability is not an issue to solve the boundary value problem arising from eliminating the control from the system of KKT conditions, a Peer variant AP4o43bdf of the A ( 73.35 ) -stable BDF4 and the A ( 84 ) -stable AP4o43dif are the most attractive methods, which perform equally well depending on the problem type. The A-stable methods AP4o43dig and AP4o43die with diagonally implicit standard Peer methods are very good alternatives if eigenvalues close or on the imaginary axis are existent. We have also constructed the A-stable method AP4o43sil with a singly-diagonal main part as an additional option if large linear systems can be still solved by a direct solver and hence the property of requesting one LU-decomposition only is highly valuable. In future work, we plan to train our novel methods in a projected gradient approach to also tackle large-scale PDE constrained optimal control problems with semi-discretizations in space. In these applications, Peer triplets may have to satisfy even more severe requirements.

Author Contributions

Conceptualization, J.L. and B.A.S.; investigation, J.L. and B.A.S.; software, J.L. and B.A.S. All authors have read and agreed to the published version of the manuscript.

Funding

The first author is supported by the Deutsche Forschungsgemeinschaft (DFG, German Research Foundation) within the collaborative research center TRR154 “Mathematical modeling, simulation and optimisation using the example of gas networks” (Project-ID 239904186, TRR154/2-2018, TP B01).

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

In what follows, we will give the coefficient matrices which define the Peer triplets discussed above. We have used the symbolic option in Maple as long as possible to avoid any roundoff errors which would pollute the symbolic manipulations by a great number of superfluous terms. If possible, we provide exact rational numbers for the coefficients and give numbers with 16 digits otherwise. It is sufficient to only show pairs ( A n , K n ) and the node vector c , since all other parameters can be easily computed from the following relations:
B n = ( A n V s K n V s E ˜ s ) P s V s 1 , a = A 0 𝟙 , b = A 0 c K 0 𝟙 , w = V s T 𝟙 , v = V s T e 1 , s = 3 , 4 ,
with e 1 = ( 1 , 0 , , 0 ) T R s and the special matrices
V s = ( 𝟙 , c , c 2 , , c s 1 ) , P q = j 1 i 1 i , j = 1 s , E ˜ s = i δ i + 1 , j i , j = 1 s .

Appendix A.1. Coefficients of AP4o54bdf

c T = 1 4 , 1 2 , 3 4 , 1
A 0 = 2 1 2 265 96 17 96 11 288 7 6 47 24 25 12 21 32 227 96 1163 288 25 12 , K 0 = 1 2 77 192 3 32 67 192 17 96 155 576 19 192 17 192 0 1 4
A = 25 12 4 25 12 3 4 25 12 4 3 3 4 25 12 , K = 1 4 1 4 1 4 1 4
A N = 635 96 1235 72 35 32 67 96 43 288 4475 288 35 24 0 43 72 5 35 96 67 96 53 96 , K N = 25 32 5 3 61 192 1 192 115 64 13 48 23 64 185 288 13 96 1 192 43 576

Appendix A.2. Coefficients of AP4o43dif

c T = 3 22 , 53 132 , 97 132 , 1
A 0 = 1.1582197171362010 0.04624378638947835 0.7020381998219871 1.55936795391810600 0.02624560436867219 0.5084723270832399 3.71639374160988500 2.21790664582949000 2.30412222276248700 2.66055187655540200 1.275350214666080
K 0 = 0.10630513138050800 0.39188176473763570 0.18135555683856680 0.05671611789968874 0.06216151000641002 0.3773797141792473 0.08101130220826174 0.03462160050989925 0.1300000000000000
A = 2.713996187194519 5.753019558612675 2.063116456071261 5.300000000000000 4.392801539381829 2.202673482804081 2.313267438350870 2.523025304770754 2.619073109161321 1.275350214666080
K = d i a g 0.2212740342685062 , 0.2910929443629617 , 0.3576330213685321 , 0.13
A N = 3.321208926131899 6.825690771130220 1.096545539465253 0.5537291752348234 0.08772005153993665 5.504481844998321 1.589672639210835 0.3213889335444567 0.44690680951897520 2.000000000000000 0.493127099745582 0.8751181087792801 0.64081324202096150
K N = 0.4780945554021703 0.7500000000000000 0.3443968810148793 0.03064999258349816 0.9263584006629529 0.2474620802680682 0.34743387239297130 0.4116869618629235 0.1378269814151266 0.03853141924782626 0.06599889592052376

Appendix A.3. Coefficients of AP4o43dig

c T = 139 1159 , 11 19 , 1 , 1375 2014
A 0 = 482.1874750642102 4.750000000000000 5.916666666666667 6.500000000000000 5295.612100386801 78.73229010791468 60.16394904407432 7.222222222222222 893.8003010294580 4.061724320422766 9.736228361340601 19.67879886925837 5707.694317901957 68.66254446706468 58.05689396607474 35.61626763467349
K 0 = 49.91295086094522 0.5250000000000000 3.439024390243902 2.894736842105263 405.5730073881453 49.31516975831240 8.193548387096774 3.428571428571429 53.62032171809015 12.67084977396168 1.304859285573478 4.013292871986014 414.6382351541371 49.08870633833948 1.334271117642095 15.23643896150272
A = 2.604429828805958 6.603320924494022 11.44234275562775 0.5317173544040980 2.710438820206414 3.550000000000000 5.000000000000000 2.026117385005894 2.376616772673509 15.21524654319290
K = d i a g 0.8973222553064913 , 3.337407156628221 , 1.164566261468968 , 2.604651162790698
A N = 3.754385964912281 0.01222493887530562 1.014925373134328 0.1403508771929825 11.35280296428295 e 45.64990373363066 15.17493010383148 20.79910559459065 0.03205698176794144 2.937595714981687 3.756123160591242 2.666624105937791 7.630473981138614 48.59972438748765 18.91612789128839 18.27283236584584
K N = 0.9578456075353955 0.3387096774193548 0.1194029850746269 0.8045112781954887 11.57142857142857 96.60667975350700 20.12500000000000 102.4846028390543 3.761888534906390 33.44512959236514 5.865106921633463 34.94704625261853 15.32045224098695 134.2026156894527 26.37615509001594 141.3196101415549

Appendix A.4. Coefficients of AP4o43die

c T = 15 44 , 5 4 , 1 4 , 23 12
A 0 = 1573 27 29140496694667 1728480384000 37071572404007 69715375488000 37246788257 8450348544 98934973036237 2160600480000 59311823623513 87144219360000 51770824817 13203669600 7653678714559 1387052160000 7872573544487 38730764160000 32557703329 23473190400 18014543 144484600
K 0 = 110 9 4 5 168095353644187 920242956441600 136571614975979 36809718257664 9857504559041 1080300240000 398472962076949 11503036955520000 18235836500357 18404859128832 1423069729157 1440400320000 398472962076949 7668691303680000 136571614975979 61349530429440 2 61
A = 45808744223 19505421000 279428522 187552125 3 4 285647 15004170 22704013 125034750 11 10 1 4 11824391 41678250 832579 4167825 18014543 144484600
K = d i a g 35085281 25006950 , 2300653 8335650 , 1780019 2500695 , 2 61
A N = 46268635184890481 6231747944448000 843924159892681 239682613248000 2095498352743 2535104563200 240331128931 253510456320 10279031063 122060590080 4 3733202770769 25351045632000 8883867896017 6337761408000 192302368031 24412118016000 39222471881369 27696657530880 637146509711 2816782848000 191242568381 352097856000 175009899277 8137372672000
K N = 95205971609617 35952391987200 28298016708823 167316901171200 1296366536717 4182922529280 958020525197 864240192000 2425439590003 104573063232000 27877023310129 41829225292800 958020525197 14980163328000 2425439590003 69715375488000 1296366536717 6971537548800 22310489177 4882423603200

Appendix A.5. Coefficients of AP4o43sil

c T = 1 50 , 3 5 , 1 , 41 85
A 0 = 18.6770976012982273 1.15212718448036531 0.684527356670693701 30.2098963703001422 19.0677876392318276 7.55433120044842482 9.81986015015644262 2.15227598175855777 4.86425591259034856 57 25 695 72 8.28572795643498617 9.37534909694128499
K 0 = 11.4061014637853601 0.0776818914116313719 0.278650826939386227 59 28 13.9738118565057040 1.13881886074868390 2498819 583100 2.75133184568842863 0.477663277652266390 161 25 779 80 0.352459713213735910 2.80235150260251923
A = 3.40824065799546119 10.5240959029253065 6.21116392867196304 1.24215119761892880 3.47742608697035235 3.58958480260299081 12.1231239821473188 3.03082301205062472 1.32154050930321039 9.37534909694123776
K = d i a g 1.01874482010281778 , 1.85655642136068493 , 1.07294973886097804 , 2.80235150260250919
A N = 3.93487127199009570 75 26 7 8 4.71932036763822580 42.7928650670737006 3.60582429888170880 36.9240613682294166 71 202 2.31371287972058488 0.191639567731266976 1.90723457480523846 9.00567678814317298 43.3637675719685003 5.28918473115044182 35.0168267934241781
K N = 0.687420439097535868 247 72 24.5972883813771674 2.02081177860650990 24.7095335501766993 0.427115478996059306 1780 289 0.897376604330415086 5.61580307958561348 3.39815952896990200 356 17 1.56153637437775765 22.4504633222483055

Appendix A.6. Coefficients of AP3o32f

c T = 106 135 , 3 5 , 1
A 0 = 13474483 2809000 2765681 1404500 753641 273375 48583191 81461000 1538339 1093500 1783 580
K 0 = d i a g 13474483 7155000 , 2513302 1366875 , 11 10
A = 11 2 6493 2700 64 25 25757 78300 121 100 1783 580
K = d i a g 93 50 , 44 25 , 11 10
A N = 3 559409 391500 5418793 1458000 2257039 1691280 1733909 391500 5418793 1458000 565759 1691280
K N = d i a g 1190159 978750 , 5418793 3645000 , 2257039 4228200

References

  1. Liu, X.; Frank, J. Symplectic Runge-Kutta discretization of a regularized forward-backward sweep iteration for optimal control problems. J. Comput. Appl. Math. 2021, 383, 113133. [Google Scholar] [CrossRef]
  2. Sanz-Serna, J. Symplectic Runge–Kutta schemes for adjoint equations, automatic differentiation, optimal control, and more. SIAM Rev. 2016, 58, 3–33. [Google Scholar] [CrossRef]
  3. Hairer, E.; Wanner, G.; Lubich, C. Geometric Numerical Integration, Structure-Preserving Algorithms for Ordinary Differential Equations; Springer Series in Computational Mathematic; Springer: Berlin/Heidelberg, Germany, 1970; Volume 31. [Google Scholar]
  4. Matsuda, T.; Miyatake, Y. Generalization of partitioned Runge–Kutta methods for adjoint systems. J. Comput. Appl. Math. 2021, 388, 113308. [Google Scholar] [CrossRef]
  5. Lang, J.; Verwer, J. W-Methods in Optimal Control. Numer. Math. 2013, 124, 337–360. [Google Scholar] [CrossRef]
  6. Almuslimani, I.; Vilmart, G. Explicit stabilized integrators for stiff optimal control problems. SIAM J. Sci. Comput. 2021, 43, A721–A743. [Google Scholar] [CrossRef]
  7. Lubich, C.; Ostermann, A. Runge-Kutta approximation of quasi-linear parabolic equations. Math. Comp. 1995, 64, 601–627. [Google Scholar] [CrossRef]
  8. Ostermann, A.; Roche, M. Runge-Kutta methods for partial differential equations and fractional orders of convergence. Math. Comp. 1992, 59, 403–420. [Google Scholar] [CrossRef]
  9. Gerisch, A.; Lang, J.; Podhaisky, H.; Weiner, R. High-order linearly implicit two-step peer—Finite element methods for time-dependent PDEs. Appl. Numer. Math. 2009, 59, 624–638. [Google Scholar] [CrossRef]
  10. Gottermeier, B.; Lang, J. Adaptive Two-Step Peer Methods for Incompressible Navier-Stokes Equations. In Proceedings of the ENUMATH 2009, the 8th European Conference on Numerical Mathematics and Advanced Applications, Uppsala, Sweden, 29 June–3 July 2009; Kreiss, G., Lötstedt, P., Malqvist, A., Neytcheva, M., Eds.; Springer: Berlin/Heidelberg, Germany, 2009; pp. 387–395. [Google Scholar]
  11. Albi, G.; Herty, M.; Pareschi, L. Linear multistep methods for optimal control problems and applications to hyperbolic relaxation systems. Appl. Math. Comput. 2019, 354, 460–477. [Google Scholar] [CrossRef]
  12. Sandu, A. Reverse automatic differentiation of linear multistep methods. In Advances in Automatic Differentiation; Lecture Notes in Computational Science and Engineering; Bischof, C., Bücker, H., Hovland, P., Naumann, U., Utke, J., Eds.; Springer: Berlin/Heidelberg, Germany, 2008; Volume 64, pp. 1–12. [Google Scholar]
  13. Lang, J.; Schmitt, B. Discrete adjoint implicit peer methods in optimal control. J. Comput. Appl. Math. 2022, 416, 114596. [Google Scholar] [CrossRef]
  14. Hager, W. Runge-Kutta methods in optimal control and the transformed adjoint system. Numer. Math. 2000, 87, 247–282. [Google Scholar] [CrossRef]
  15. Troutman, J. Variational Calculus and Optimal Control; Springer: New York, NY, USA, 1996. [Google Scholar]
  16. Bonnans, F.; Laurent-Varin, J. Computation of order conditions for symplectic partitioned Runge-Kutta schemes with application to optimal control. Numer. Math. 2006, 103, 1–10. [Google Scholar] [CrossRef]
  17. Herty, M.; Pareschi, L.; Steffensen, S. Implicit-Explicit Runge-Kutta schemes for numerical discretization of optimal control problems. SIAM J. Numer. Anal. 2013, 51, 1875–1899. [Google Scholar] [CrossRef]
  18. Sandu, A. On the properties of Runge-Kutta discrete adjoints. Lect. Notes Comput. Sci. 2006, 3394, 550–557. [Google Scholar]
  19. Schröder, D.; Lang, J.; Weiner, R. Stability and consistency of discrete adjoint implicit peer methods. J. Comput. Appl. Math. 2014, 262, 73–86. [Google Scholar] [CrossRef]
  20. Montijano, J.; Podhaisky, H.; Randez, L.; Calvo, M. A family of L-stable singly implicit Peer methods for solving stiff IVPs. BIT 2019, 59, 483–502. [Google Scholar] [CrossRef]
  21. Schmitt, B.; Weiner, R. Efficient A-stable peer two-step methods. J. Comput. Appl. Math. 2017, 316, 319–329. [Google Scholar] [CrossRef]
  22. Schmitt, B. Algebraic criteria for A-stability of peer two-step methods. Technical Report. arXiv 2015, arXiv:1506.05738. [Google Scholar]
  23. Jackiewicz, Z. General Linear Methods for Ordinary Differential Equations; John Wiley&Sons: Hoboken, NJ, USA, 2009. [Google Scholar]
  24. Jacobson, D.; Mayne, D. Differential Dynamic Programming; American Elsevier Publishing: New York, NY, USA, 1970. [Google Scholar]
  25. Dontchev, A.; Hager, W.; Veliov, V. Second-order Runge-Kutta approximations in control constrained optimal control. SIAM J. Numer. Anal. 2000, 38, 202–226. [Google Scholar] [CrossRef]
Figure 1. Rayleigh Problem: Convergence of the maximal state errors ( w T I ) Y n y ( t n + 1 ) (left) and adjoint errors ( v T I ) P n p ( t n ) (right), n = 0 , , N .
Figure 1. Rayleigh Problem: Convergence of the maximal state errors ( w T I ) Y n y ( t n + 1 ) (left) and adjoint errors ( v T I ) P n p ( t n ) (right), n = 0 , , N .
Algorithms 15 00310 g001
Figure 2. Van der Pol Problem: Convergence of the maximal state errors ( w T I ) Y n y ( t n + 1 ) (left) and adjoint errors ( v T I ) P n p ( t n ) (right), n = 0 , , N .
Figure 2. Van der Pol Problem: Convergence of the maximal state errors ( w T I ) Y n y ( t n + 1 ) (left) and adjoint errors ( v T I ) P n p ( t n ) (right), n = 0 , , N .
Algorithms 15 00310 g002
Figure 3. Controlled Motion Problem: optimal path ( y 1 , y 2 ) through the total energy field E = 1 2 y 2 2 + 1 4 y 1 4 1 2 y 1 2 visualized by isolines and exhibiting a saddle point structure (top). Convergence of the maximal state errors ( w T I ) Y n y ( t n + 1 ) (bottom left) and adjoint errors ( v T I ) P n p ( t n ) (bottom right), n = 0 , , N .
Figure 3. Controlled Motion Problem: optimal path ( y 1 , y 2 ) through the total energy field E = 1 2 y 2 2 + 1 4 y 1 4 1 2 y 1 2 visualized by isolines and exhibiting a saddle point structure (top). Convergence of the maximal state errors ( w T I ) Y n y ( t n + 1 ) (bottom left) and adjoint errors ( v T I ) P n p ( t n ) (bottom right), n = 0 , , N .
Algorithms 15 00310 g003
Figure 4. Wave Problem: Convergence of the maximal state errors ( w T I ) Y n y ( t n + 1 ) (left) and adjoint errors ( v T I ) P n p ( t n ) (right), n = 0 , , N .
Figure 4. Wave Problem: Convergence of the maximal state errors ( w T I ) Y n y ( t n + 1 ) (left) and adjoint errors ( v T I ) P n p ( t n ) (right), n = 0 , , N .
Algorithms 15 00310 g004
Table 1. Combined order conditions for the Peer triplet, including the compatibility condition (67) and the condition (61) for full matrices K N .
Table 1. Combined order conditions for the Peer triplet, including the compatibility condition (67) and the condition (61) for full matrices K N .
StepsForward: q 1 = s Adjoint: q 2 = s 1
Start, n = 0 (29)(33),(61), β = 0
Standard, 1 n < N (30)(33)
Superconvergence(52)(53)
Compatibility(67)(67)
Last step(30), n = N (33), n = N 1
End point(31)(34),(61), β = N
Table 2. Properties of the boundary methods of Peer triplets.
Table 2. Properties of the boundary methods of Peer triplets.
Starting MethodEnd Method
TripletBlocks μ 0 ϱ ( BA 0 1 ) Blocks μ N ϱ ( A N 1 B N ) ϱ ( B N A 1 )
AP4o43bdf3+15.4711+33.8111.15
AP4o43dif3+16.2711+34.4011.03
AP4o43dig40.99140.891.0011
AP4o43sil3+11.88140.7211.03
AP4o43die3+13.8011+30.662.61.98
AP3o32f1+1+11.501.021+20.9411
Table 3. Properties of the standard methods of Peer triplets.
Table 3. Properties of the standard methods of Peer triplets.
sTripletNodes α A 1 B | λ 2 | err s Remarks
4AP4o43bdfBDF4 73.35 5.790.0990singly-implicit
AP4o43dif [ 0 , 1 ] 84.0 2.01 0.260.0025diag.-implicit
AP4o43dig [ 0 , 1 ] 90 24.50.7980.0260 c 3 = 1
AP4o43sil [ 0 , 1 ] 90 32.20.600.0230 c 3 = 1, sing.impl.
AP4o43die [ 0 , 2 ] 90 6.080.660.0135nodes alternate
3AP3o32f [ 0 , 1 ] 90 15.30.910.0170nodes alternate
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lang, J.; Schmitt, B.A. Implicit A-Stable Peer Triplets for ODE Constrained Optimal Control Problems. Algorithms 2022, 15, 310. https://doi.org/10.3390/a15090310

AMA Style

Lang J, Schmitt BA. Implicit A-Stable Peer Triplets for ODE Constrained Optimal Control Problems. Algorithms. 2022; 15(9):310. https://doi.org/10.3390/a15090310

Chicago/Turabian Style

Lang, Jens, and Bernhard A. Schmitt. 2022. "Implicit A-Stable Peer Triplets for ODE Constrained Optimal Control Problems" Algorithms 15, no. 9: 310. https://doi.org/10.3390/a15090310

APA Style

Lang, J., & Schmitt, B. A. (2022). Implicit A-Stable Peer Triplets for ODE Constrained Optimal Control Problems. Algorithms, 15(9), 310. https://doi.org/10.3390/a15090310

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop