Nothing Special   »   [go: up one dir, main page]

Academia.eduAcademia.edu
Application of Multilayer Feedforward Networks in the Solution of Compressible Euler Equations Final Technical Report for NGT-70353 Principal Investigator: Andrew J. Meade, Jr. Associate Professor Mechanical Engineering and Materials Science, MS-321 William Marsh Rice University Houston, Texas 77005-1892 Phone: (713) 737-5880 E-mail: meade@rice.edu February, 1998 Abstract Application of Multilayer Feedforward Networks in the Solution of Compressible Euler Equations Final Technical Report for NGT-70353 Recent work has proven that it is possible to actually \program" multilayer feedforward arti cial neural networks (ANNs). This new paradigm is not only making it possible to logically and predictably extend the capabilities of ANNs to \hard" computing, such as Computational Fluid Dynamics (CFD), but it is also revealing a di erent, quantitative way of looking at neural networks that could help advance the level of understanding we possess about these systems. Accurate modeling of linear ordinary di erential equations (ODEs) and linear partial di erential equations (PDEs) has already been successfully completed using this paradigm. It is proposed to extend this capability further by attempting to use ANNs to model the solution of the two-dimensional compressible Euler equations. The code thus created should not only have the same accuracy as a more conventional computer code, but should still retain the ability of an ANN to modify itself when exposed to experimental data, thus yielding software that could be specialized with experimental results. To accomplish these objectives, the method of optimal incremental function approximation has been developed for the adaptive solution of di erential equations using ANN architecture. Two major attractive features of this approach are that: 1) the developed method is exible enough to use any of the popular transfer functions and 2) the developed method requires minimal user interaction. The latter is especially advantagous when dealing with complicated physical or computational domains. The method of optimal incremental function approximation is formulated from the combination of various concepts utilized by computational mechanics and arti cial neural networks (e.g. function approximation and error minimization, variational principles and weighted residuals, and adaptive grid optimization). The basis functions and associated coecients of a series expansion, representing the solution, are optimally selected by a parallel direct search technique at each step of the algorithm according to appropriate criteria; the solution is built sequentially. Complicated data structures and expensive remeshing algorithms and systems solvers, that are common in conventional computational mechanics, are avoided. Computational eciency is increased by using augmented radial bases and concurrent computing. Variational principles are utilized for the de nition of the objective function to be extremized in the associated optimization problems, ensuring that the problem is well-posed. Numerical results and convergence rates are reported for steady-state problems, including the nonlinear compressible Euler equations. Acknowledgments The author would like to acknowledge the work of research assistants Alvaro Fernandez and Michael Kokkolaras. Also, Dr. Robert Shelton of the NASA Johnson Space Center is gratefully acknowledged for his guidance and assistance. Special thanks are given to Prof. Bowen Loftin for access to the University of Houston IBM SP2 parallel computer. Contents Abstract Acknowledgments List of Illustrations List of Tables ii iv viii xi 1 Tikhonov-Based Integration of Theoretical, Experimental and Computational Mechanics 1 1.1 1.2 1.3 1.4 Introduction The Tikhonov-Based Integration Method Interpolation Methods for Experimental Data Analysis Arti cial Neural Networks 1.4.1 Arti cial Neural Networks As a General Approximation Tool 1.4.2 Network Modelling As An Ill-Posed Problem A Method for Data Analysis: Regularization by A-Priori Mathematical Models 1.5.1 Analytical Approximation of Regularized Mechanical Systems 1.5.2 Numerical Approximation of Regularized Mechanical Systems Example Problem 1.6.1 Results from the Analytical Approximation 1.6.2 Results from the Numerical Approximation Nonlinear A-Priori Models Implementation Concerns : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 1.5 : : : : : : : : : : : : : : : : : : : : : : : : : : 1.6 : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 1.7 1.8 : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 1 3 4 8 8 10 12 14 17 18 19 21 22 23 vi 2 Optimal Incremental Approximation: Background 26 2.1 2.2 2.3 2.4 The Method of Weighted Residuals Variational Principle Methods Adaptive Grid Optimization Arti cial Neural Networks : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 3 Optimal Incremental Approximation: Approach 3.1 3.2 3.3 3.4 The Proposed Algorithm The Parallel Direct Search Optimization Technique The Iterative Character of the Proposed Algorithm Function Approximation 3.4.1 Implementation 3.4.2 Numerical Examples 36 : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 4 Linear Di erential Equations 4.1 Boundary Conditions 4.2 Linear Nonself-Adjoint Problems 4.2.1 Numerical Example: The Convection-Di usion Equation 4.2.2 Computational Cost : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 5 The Euler Equations The Velocity Potential Equation The Variational Integral The Computational Domain Boundary Conditions Numerical Procedure and Results 5.5.1 Numerical Example: 1 = 0 4 : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 50 52 56 65 69 : : : : : : : : : : : : : : : : : : : : M 36 38 39 42 42 44 47 : : : : : : : : : : : : : : : : : : : : : : : : : : 5.1 5.2 5.3 5.4 5.5 26 27 30 33 : : : : : : : : : : : : : : : : 70 71 74 76 80 83 vii 5.5.2 Numerical Example: 1 = 0 43 5.5.3 Computational Work M : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 6 Conclusions and Future Work 6.1 Conclusions 6.1.1 Objectives 6.1.2 Approach 6.1.3 Numerical Examples 6.1.4 Optimization and Bases 6.2 Future Work 6.2.1 Generalization of the Proposed Method 6.2.2 Utilizing the Method of Weighted Residuals 6.2.3 Increasing Computational Eciency 6.2.4 Compact Bases 94 : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : A A.1 A.2 A.3 A.4 A.5 Norms, Inner Products, and Orthogonality -splines and Gaussian Functions Members of the Method of Weighted Residuals Classi cation of Di erential Operators PDS Computational Issues A.5.1 The Formulation of the Optimization Variables A.5.2 Scaling A.6 Diagonal Dominance and Spectral Radius A.7 Numerical Integration : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : Bibliography 94 94 94 96 96 97 97 97 99 100 102 : : : : : : : : : : : : : : B 88 89 102 105 108 109 111 111 112 113 114 119 Illustrations 1.1 Extrapolation from insucient data for  = 0:01; (solid) the mechanical system F (x), (dash) the regularized approximation f~(x), (dash-dot) the numerical approximation fa(x). : : : : : : : : : 20 1.2 Interpolation of data with noise for optimal ; (solid) F (x), (dash) f~(x), (dash-dot) fa (x). : : : : : : : : : : : : : : : : : : : : : : : : : 22 3.1 Convergence rates for the function approximation problems: (a) one-dimensional case and (b) two-dimensional case. : : : : : : : : : 46 4.1 One-dimensional convection-di usion: exact and approximate solutions for (a) = = 5 and (b) = = 20 using 16 and 9 optimal augmented radial basis functions, respectively. : : : : : : : : : : : : 4.2 One-dimensional convection-di usion: (a) exact and approximate solutions for di erent grid spacing and cell Reynolds numbers and (b) exact and approximate solutions in the presence of an upwind scheme; = = 20, plots reproduced from reference [26]. : : : : : : : 4.3 One-dimensional convection-di usion: convergence rates for (a) = = 5 and (b) = = 20. : : : : : : : : : : : : : : : : : : : : : : : 4.4 One-dimensional convection-di usion: convergence rates comparison considering the Petrov-Galerkin and Raleigh-Ritz methods; = = 20. : : : : : : : : : : : : : : : : : : : : : : : : : : : 61 63 63 67 ix 4.5 One-dimensional convection-di usion: absolute value of the coecients c for = = 20. : : : : : : : : : : : : : : : : : : : : : : : 68 i 5.1 Compressible inviscid ow: (top) the physical (x-y-plane) and (bottom) computational (r--plane) domains. : : : : : : : : : : : 5.2 Compressible inviscid ow: the approximate velocity potential auxiliary function (r; ) for M1 = 0:4 using 15 optimal basis functions. : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 5.3 Compressible inviscid ow: the approximate velocity potential auxiliary function (r; ) at di erent radii (on and near the cylinder surface) for M1 = 0:4 using 15 optimal basis functions. : 5.4 Compressible inviscid ow: nondimensionalized (curve 1) uid velocity and (curve 2) sonic speed on the cylinder surface for M1 = 0:4. : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 5.5 Compressible inviscid ow: (top) pressure coecient distribution and (bottom) nondimensionalized density on the cylinder surface for M1 = 0:4. : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 5.6 Compressible inviscid ow, M1 = 0:4: (top) local Mach number distribution over the entire ow eld and (bottom) Mach-isolines in a region close to the cylinder surface. : : : : : : : : : : : : : : : 5.7 Compressible inviscid ow: the approximate velocity potential auxiliary function (r; ) for M1 = 0:43 using 15 optimal basis functions. : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 5.8 Compressible inviscid ow: the approximate velocity potential auxiliary function (r; ) at di erent radii (on and near the cylinder surface) for M1 = 0:43 using 15 optimal basis functions. : 75 : 84 : 84 : 85 : 86 : 87 : 88 : 89 x 5.9 Compressible inviscid ow: nondimensionalized (curve 1) uid velocity and (curve 2) sonic speed on the cylinder surface for M1 = 0:43. : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 5.10 Compressible inviscid ow: (a) pressure coecient distribution and (b) nondimensionalized density on the cylinder surface for M1 = 0:43. : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 5.11 Compressible inviscid ow, M1 = 0:43: (top) local Mach number distribution over the entire ow eld and (bottom) Mach-isolines in a region close to the cylinder surface. : : : : : : : : : : : : : : : : 90 : 90 : 91 A.1 B -splines: (a) a B1-spline and (b) a B3-spline. : : : : : : : : : : : : 107 A.2 Radial bases: (a) Gaussian function with center  = 1:5 and radius  = 0:5 and (b) an augmented radial basis with  =  = 10?6 ,  = 2:5 and  = ?2:5. : : : : : : : : : : : : : : : : : : : : : : : : 107 l l r r Tables 4.1 Amount of computational work required by PDS for the linear convection-di usion problem using B1-splines (= = 5). : : : : : : 66 4.2 Amount of computational work required by PDS for the linear convection-di usion problem using augmented radial basis functions (= = 5, d = 1000, and tol = 10?4 ). : : : : : : : : : : : : 66 4.3 Amount of computational work required by PDS for the linear convection-di usion problem using augmented radial basis functions (= = 20, d = 1000, and tol = 10?4 ). : : : : : : : : : : : 67 5.1 Amount of computational work required by PDS for the Euler Equation problem using augmented radial basis functions for d = 1120 and tol = 10?4 (M = 0:4). : : : : : : : : : : : : : : : : : : 93 5.2 Amount of computational work required by PDS for the Euler Equation problem using augmented radial basis functions for d = 1120 and tol = 10?4 (M = 0:43). : : : : : : : : : : : : : : : : : 93 1 Chapter 1 Tikhonov-Based Integration of Theoretical, Experimental and Computational Mechanics 1.1 Introduction The goal of engineering analysis is to obtain a comprehensive description of a physical system of interest. Three approaches to this goal exist: 1) theoretical analysis of mathematical models, 2) physical experimentation and 3) computational mechanics [18]. None of these approaches is superior and they must be used in combination. For example, ideally, experimental measurements should provide the most adequate description of a physical system. However, from an engineering standpoint, experimental studies lack the needed generality since measurements only describe the system behavior under a speci c combination of load and system parameters. Also, errors in experimental measurements can induce a bias in the evaluation of the physical system. This may conceal important features in the system behavior. Most importantly, experimental measurements of complex physical systems may require signi cant, if not prohibitive, material and labor costs; describing complex engineering systems using sparse experimental data is an important and challenging engineering problem. Alternatively, utilizing only mathematical models for engineering analysis can yield general results which can then provide the foundation for a parametric study. However, even the most sophisticated mathematical model can only approximate the physical system. In this regard, the delity of mathematical models and its applicability for describing system behavior ultimately requires model validation 2 using experimental measurements. In addition, the majority of sophisticated mathematical models do not have analytical solutions and must be simpli ed (lowdelity) or solved by high-speed digital computers combined with accurate numerical algorithms. Unfortunately, even with the use of numerical computation only a limited number of parameter combinations can be investigated and, like the experimental approach, the theoretical approach loses its generality. In fact, computational studies are often referred to as numerical experiments [81]. Though it is known in the engineering community that successful analyses rest upon the proper balance of all three approaches, it is noted that few attempts have been made in uniting experimental, theoretical and numerical methods in the literature. It is the overall objective of this report to address this by developing a method which e ectively combines all available information, from both experimental data and mathematical models, in the emulation of physical systems. To accomplish this in a practical sense, it is the speci c objective of this report to develop a method in which the combined information can be approximated by arti cial neural network (ANN) architectures. Chapter 1 of this report discusses the development of the integration method and includes background material and numerical examples. Section 1.8 discusses the issues that must be addressed if the integraton method is to be applied to engineering problems. As a result of this discussion a novel numerical approximation method (optimal sequential function approximation) is proposed and introduced in Chapters 2 and 3. The novel approximation method is then applied to the variational form of a linear nonself-adjoint di erential equation in Chapter 4 and the compressible Euler equations in Chapter 5. Conclusions are given in Chapter 6. Section 6.2 discusses generalization of the optimal sequential function approxi- 3 mation, when a variational form is not available, utilizing the method of weighted residuals. 1.2 The Tikhonov-Based Integration Method The proposed integration method can be interpreted as the combination of experimental and computational mechanics by which the theoretical response of a physical system is \ ne-tuned" through available experimental data. Note that even mathematical models of low physical delity, while failing to provide adequate engineering accuracy, may properly capture the most signi cant qualitative aspects of system behavior. The proposed method eciently \corrects" the response of low delity mathematical models by using experimental data. Alternatively, the method can be interpreted as a low computational cost procedure for determining smooth response curves which can be used to interpolate sparse and scattered experimental data by physically motivated criteria. As a result, the problem of predicting (extrapolating) the mechanical system response at unmeasured coordinates can be examined. This chapter also shows that the developed method can be eciently used to process and lter noisy measurements. The response of the mathematical models characterizes the physical system's target features, which are used to separate the serviceable part of the data from the noisy component. The technical approach of the developed method utilizes a common tool in neural network applications, the theory of Tikhonov regularization of ill-posed problems [83]. Speci cally, the problem of mathematical analysis of experimental data is treated as an ill-posed problem. Its regularization involves the introduction of additional information regarding the physical system. It is proposed to utilize apriori mathematical models of physical systems at appropriate degrees of delity for regularization. The computational environment of ANNs is utilized in this regard. 4 The numerical examples considered in this chapter demonstrate the usefulness of the method for engineering applications. 1.3 Interpolation Methods for Experimental Data Analysis The amount of experimental data is always limited by the cost of material resources, instrumentation and labor. Therefore, engineering analysis of experimental data involves an approximation of the system response surface from a limited amount of measurements. This leads to the classical problem of interpolating scattered data [28]. Speci cally, the exact response of the system F (x) is approximated by the equation f (x) a = X k  (x) (1.1) k k where are coecients corresponding to the basis functions  (x). The function f (x) is designed to approximately pass through the experimental data points f (x ) at the coordinates x = (x1 ; :::; x ), where d denotes the dimensionality of the approximation problem. It is assumed that k k a e i i f (x ) e i = F (x ) +  i ;i d;i (1.2) i where  denotes the random noise of measurements at the coordinate x . Several approaches have been proposed in the literature in the construction of f (x). The response surface methodology of experiment design utilizes a polynomial basis for f (x) which is tted to the data by the statistical tool of linear regression [18], [7], [67], [66]. This relatively simple method of function interpolation and the statistical framework for data analysis provide for the accurate approximation of i i a a 5 multi-dimensional functions in local regions of interest. Also this approach is quite useful in forming a strategy (design of experiments) for selecting data for the best local approximation of the response surface by Eq. (1.1). The nite element based method of data interpolation is based on the idea of using continuous, or continuously di erentiable, local bases placed on a triangularization of the region of approximation [28]. The radial basis function interpolation of scattered data is an increasingly popular method for data tting [28], [21], [20]. Though this approach incorporates \global" basis functions, the approximation fa (x) = X k kx ? k k ck   ! k (1.3) ; can be eciently computed for large data sets [21]. Here, k  k represents the Euclidean norm in the space Rd , and ck , k , and k are parameters to be evaluated. Further, the data interpolation procedure for certain radial bases can be seen, in a Hilbert space setting [87], as the minimization of the functional  [fa(x)] = Z fba( ) 2 b Rd G (kk) d (1.4) with the constraint fa(xk ) = fe(xk ); here  is a dummy variable and the hat denotes the Fourier transform of the corresponding quantities. The term G (kk) is a positive function such that b k k) ! 0 as kk ! 0: b G ( Note that, in this case, the function (1.5)  k k) b G ( ?1  can be interpreted as a high 6 pass lter which increases the contribution of the high frequency components of  ? the approximation fa(x) in Eq. (1.4). Therefore, the kernel Gb (  ) with the conditions of Eq. (1.5) induces a penalty for a non-smooth interpolation of data. The Hilbert space framework of scattered data interpolation was rst proposed by Duchon [19] and Meinguet [62] who derived multi-dimensional spline bases. A more general class of semi-spline radial functions [74], [29] can be obtained by using the smoothing functional, k k X 1 am Om fa(x) L2 R m 0 m Z 1 X @ X = am  f (x) dxA n i1 ;:::;im a  [fa(x)] = 2 k k ( ) i1 ;:::;im R m 2 j j (1.6) where m i1;:::;im = @x @:::@x : i1 im As a result, data interpolation with the Gaussian function kxk2 (x) = e? 22 (1.7) is associated with the kernel 1 m X Gb (  ) = m  m m! 2 2 k k k k =0 m 2 !? 1 = e? 2 kxk2 2 ; the Gaussian radial basis is used in the numerical examples of this chapter. 7 With regard to the preceding discussion, it is noted that few studies in spline curve- tting have been devoted to establishing appropriate criteria in the selection of the smoothing functional  [fa(x)]. Often this functional is selected based on the consideration of computational eciency rather than on the study of the engineering system. Several authors have argued that the choice of the smoothness criterion should rely on the knowledge of the underlying system that generated the data. A probabilistic foundation for the selection of the interpolating bases of Eq. (1.1) was proposed by Matheron [55], [56] who developed the kriging method of statistical data analysis. In this method the response function F (x) is assumed to be a realization of a homogeneous random eld consistent with the experimental measurements. The approximation function fa(x) is then selected as the best linear unbiased estimate of the random eld F (x) based on the observation data fe (xi ). Note that, in this case, the auto-correlation function describing the second order properties of the random eld becomes the basis function of Eq. (1.1). That is, fa (x) = X k k Rf (x ? xk ) ; (1.8) where Rf (x) denotes the auto-correlation function specifying the a-priori probability structure of the response function F (x). Similar results were obtained by Kimeldorf and Wahba [48] who showed that generalized splines correspond to a special form of the a-priori information with respect to the spectral density function of the random eld F (x). Speci cally, b Sf ( ) = G (k k) ; (1.9) 8 where S () is the spectral density function assigned to the response function F (x). Obviously this stochastic approach for selecting the smoothness requirements on f (x) may be fully justi ed only in the case when the response function can be realistically treated as a realization of a random eld. Nevertheless, Sacks et.al. [81] and Welch et.al. [88] insisted on the application of statistical procedures, even when the response function is deterministic, as in the case of computer experiments. The statistical framework introduced for analysis of deterministic data was proven to be useful in yielding an accurate approximation of functions based only on a few generated data points. The last data interpolation method to be discussed, and the focus of this text, is the use of ANN systems which have drawn signi cant interest in engineering application for data-driven emulation and control of engineering systems. The computational environment of ANN systems will be adapted in this report for integrating experimental data and mathematical models. f a 1.4 Arti cial Neural Networks 1.4.1 Arti cial Neural Networks As a General Approximation Tool Any approximation algorithm that uses a combination of basis functions and can be mapped into a graph-directed representation, can be treated as an arti cial neural network [74]. In this context, the response of a feedforward neural network (FFANN) can be seen as a general scheme of experimental data approximation and is de ned by the following equation [11], [41], [43], [63], f (x; a ) = X k i i  ? (k 1) :::1 X d l=1 1;l x l !! ; (1.10) 9 where xi = (x1; :::; xd) is the d-dimensional input of the ANN system, the vector represents the set of network parameters, and i is the non-linear transfer function of neurons associated with the i-th intermediate layer of the network. In Eq. (1.10), the value k ? 1 can be thought of as the number of intermediate neuron layers in a graph-directed representation of the ANN system. A single hidden layer of neurons is commonly selected in network applications unless a problemrelated consideration indicates that several intermediate layers can have a bene cial impact on the overall behavior of the network. In this regard, sigmoidal transfer functions have been thoroughly investigated in connectionist literature because of their resemblance to the response of biological neurons [11], [40], [41], [43], [89]. Another class of networks popular in applications is the radial basis network (RBN), whose response can be expressed by Eq. (1.3) [74], [30], [8], [72], [75]. Note that an RBN system can also be constructed as a two-hidden-layer FFANN with p transfer functions 1() = 2 and 2() = (  ). The training of an ANN system can be formulated as the procedure of selecting the parameter vector so that the response of the neural network is \close" to the available data. Often, the network parameters are found by minimizing the following error criterion (training objective function) = N X i=1 fe (xi ) ? fa (xi ; ) 2 : (1.11) In terms of the connectionist literature, numerical minimization of this error criterion by the steepest descent method constitutes the backpropagation algorithm [79]. The conjugate gradient method, the Levenberg-Marguardt method, and other nonlinear optimization methods have also been used for network training [80]. 10 However, it has been found that even the most advanced optimization algorithms can show poor convergence in training. Often, this phenomenon is called overtraining [37]. Further, sparse data can also cause diculties in network modelling of complex mechanical systems. These diculties become pronounced in modelling systems in which experimental data is relatively expensive to obtain [38]. In this regard, arti cially augmenting the available data by generating new training examples through introduction of small random noise has been advocated in the literature [38]. Also, experimental data is often corrupted by measurement noise. Minimization of the training criterion of Eq. (1.11), for a suciently large network, yields a response function fa(x; ) which can be over-sensitive to the noisy details of the training data resulting in poor network performance. Often, this phenomenon is called over tting [71]. 1.4.2 Network Modelling As An Ill-Posed Problem The problems of network overtraining, inadequacy of training examples, and overtting are in fact closely related. They can be viewed as a result of the network training being an ill-posed problem [83]. The ill-posed features of the ANN models can be deduced from a number of diculties that prevent ANN from becoming a readily available engineering tool for emulating mechanical systems. The network modelling of physical systems by minimizing Eq. (1.11) can only be justi ed when one can show that a network which realizes the minimum of Eq. (1.11) actually exists. This leads to concerns regarding the existence of the best network models [30], [89], [61]. Note that the non-existence of the best network approximation can be used, in some cases, as a rational explanation for the e ect of overtraining. 11 Network modelling from sparse data is an ill-posed problem, since training is not capable of producing a unique solution in regions lacking data [74]. This leads to establishing sensible criteria for network based extrapolation from examples. Moreover, even for adequate data the non-uniqueness phenomenon can result from the non-linear character of the network approximation. For example, the network response f (x; ) depends linearly on only a few parameters, shown in Eq. (1.10) as . All remaining parameters a ect the network response in a non-linear manner. In this regard, some aspects of the uniqueness problem in network modelling are discussed in references [30], [89] and [61]. As alluded to previously, over t of noisy training examples by neural networks is a major concern in network applications. This phenomenon can be associated with the instability of the standard training procedures with respect to the data noise [71]. Finally, Saarinen et.al. [80] demonstrated that overtraining was inherent to the internal architecture of multi-layered FFANN systems, yielding the ill-conditioned Jacobian and Hessian matrices. As a result of the highly parallel and redundant structure, neural networks with di erent parameter values are able to produce very similar responses. Clearly, network modelling can be improved by imposing additional constraints on the network parameters and/or by appropriately changing the training objective function of Eq. (1.11) and increasing the condition numbers of the Jacobian and Hessian matrices. This also may eliminate local minima in the objective function. The most common method of imposing ANN parameter constraints, suitable for numerical computations, is by using the theory of Tikhonov regularization of illposed problems [83]. a k;i 12 1.5 A Method for Data Analysis: Regularization by APriori Mathematical Models The popular Tikhonov regularization method [83] was originally adopted for ANN systems by Poggio and Girosi [74] and extensively used in ANN applications [29], [6]. In this case a set of constraints on the network approximation fa(x; ) is introduced through the Tikhonov regularization functional  fa(x; ) which speci es the penalty given to the oscillatory behavior of fa(x; ); the network parameters can be determined by minimizing the objective function h (fa) = N X i=1 fe (xi) ? fa (xi ; 2 h i i ) +  fa(x; ) : (1.12) Unfortunately, conventional smoothness based Tikhonov regularization of Poggio and Girosi may not be consistent with the smoothness required in emulating a physical system and so would be unable to properly approximate with noisy experimental data. Instead, this report submits that the regularization should be based on an a-priori mathematical model of some level of delity. This regularized model can then be approximated by more conventional computational methods. A large class of mechanics problems can be described by the equations L [f0 (x)] = g (x) ; x 2 and B [f0(x)] = 0 ; x 2 @ ; (1.13) where L[] is a linear di erential operator, B [] is the boundary operator, f0 is the solution function, and g is some forcing function. Let lh; i be the bilinear symmetric di erential form associated with the operator L, i.e. lh; i = hL[]; i = h; L[]i, so that the expression h; i denotes an inner product (Appendix A.1). Typically, 13 q this functional is called the energy functional, and the associated norm l(; ) is called the energy norm [50]. The solution to Eq. (1.13) can be determined by minimizing the energy form [f ] = 12 lhf; f i ? hg; f i : (1.14) Since for any f that satis es the boundary operator h l f0 ; f i = hg; f i ; (1.15) then substituting Eq. (1.15) into Eq. (1.14) yields [f ] = 1 lhf ? f0; f ? f0i ? 1 lhf0; f0i = 1 lhf ? f0; f ? f0i + constant 2 2 2 : From a physical perspective, the functional [f ] penalizes the mechanical energy of the discrepancy between the function f and the response of a mathematical model f0. With the use of f and the energy functional of Eq. (1.14), regularized modelling can be formulated as the minimization of the objective function ( )=  f N X i=1   jfe(xi) ? f (xi)j2 +  21 lhf; f i ? hg; f i (1.16) which combines information from experimental data and the a-priori mathematical model. Note that  represents the degree of reliance on the a-priori mathematical model of the physical system relative to the reliability of the experimental data. 14 1.5.1 Analytical Approximation of Regularized Mechanical Systems The formation of a regularized model of a mechanical system is identical to approximating the hypersurface f~(x) that gives the global minimum of Eq. (1.16). In this context, analytical study of f~(x) is essential for evaluating the properties of the regularized model. Because f~(x) is the global minimum of Eq. (1.16), L N 2 X f~(x ) ? f (x ) (x ? x ) ; x 2 f~ = g + i e i i h i  i=1 and h i B f~ =0 ;x 2@ (1.17) : From Eq. (1.17) it can be deduced that f~ contains two components f~(x) = f0(x) + fcor (x) (1.18) : These components can be determined by the equations f0(x) = and fcor (x) R G(x;  )g ( )d N   X = 2 fe (xi ) ? f~(xi ) G(x; xi) (1.19) i=1 where the Green's function G(x; ) satis es the boundary value problem L [G(x;  )] = (x ? ) ; x 2 and B [G(x;  )] Substituting Eq. (1.19) into Eq. (1.18) yields =0 ; x 2@ : 15 f~(x) = f0(x) + N X i=1 ci G(x; xi ) ; (1.20) where the coecients are de ned as cj   = 2 fe (xj ) ? f~(xj ) :  (1.21) The coecients can be evaluated by substituting Eq. (1.20) into Eq. (1.21), resulting in N 2 f (x ) ? f (x ) ? X ci G(xj ; xi ) cj = e j 0 j  ! ; j i=1 = 1; :::; N (1.22) with which c can be isolated. With the substitution of c into Eq. (1.20), f~ can be written as f~(x) = f0(x) + N X i=1 0 G(x; xi ) @  G + 2I !?1  fe ? f0 1 A  i : (1.23) It is noted that the matrix G of Eq. (1.23) is positive de nite since for an arbitrary function (x) we can write Z Z G(x;  )( )(x)d dx = Z (x)(x)dx = Z (x)L (x)dx = lh ; i ; where (x) is the solution to Eq. (1.13), for g(x) = (x), and satis es the boundary condition. Therefore, since the quadratic form lh; i is strictly elliptic, 16 Z Z G(x;  )( )(x)d dx  k k2H s( )>0 : (1.24) Substituting (x) = PNi=1 ci(x ? xi) into Eq. (1.24) yields Z Z G(x;  )( )(x)d dx = cT Gc > 0 : Because G is a positive de nite matrix the matrix (G + I=2) is also positive de nite for any positive value of the parameter . As a result, the vector c can be evaluated in a computationally stable manner. Since the range of the regularizing parameter  is the set of all positive numbers, it is prudent to obtain qualitative results for the limits of . Equation (1.23) indicates that as  ! 1 then f~(x) ! f0(x), as expected, since with  ! 1 experimental data becomes less relevant and f~ emulates more of the a-priori mathematical model. For  ! 0, the a-priori model of the mechanical system is assumed to be increasingly unreliable compared to the experimental data. Equation (1.23) gives f~(x) ! f0(x) + N X i=1 G(x; xi )  G?1  fe ? f0  i so that the regularized solution interpolates the data fe (xj ) exactly since f0 (xj ) + N X i=1 G(xj ; xi )  G?1  fe ? f0  i = f0 (xj ) + N X i=1  ji fe  ? f0 i = fe (xj ) As a result, the a-priori mathematical model of mechanical system does not contribute to f~ in the region of adequate data. However, the mathematical model is : 17 not eliminated from the modelling procedure; it is used by f~ to extrapolate the data to the regions lacking data. In an alternate perspective, as  ! 0 the function f~ can be thought of as a response of the a-priori model \attached" to the data points. 1.5.2 Numerical Approximation of Regularized Mechanical Systems The radial basis expansion used to approximate f~ can be described by Eq. (1.3) with () = e?2 =2 : (1.25) Since the basis expansion of Eq. (1.3) does not satisfy the boundary conditions of the mathematical model for arbitrary , the objective function of Eq. (1.16) is augmented to incorporate the boundary conditions. Speci cally, (fa) = N X i=1 1 fe (xi) ? fa (xi ; ) +  lhfa ; fa i ? hg; f i + bhfa ; q i 2 2   ; (1.26) where bh; i may be de ned as bhf; qi = @ q()B [f ()] d. Preselecting the distribution and localization properties of the radial basis functions, Eq. (1.26) is minimized with respect to the remaining unknowns, ck , resulting in R M X j lhj (x); k (x)i + 2 N X i=1 hg; k (x)i + 2 ! j (xi)k (xi) cj + bhk (x); qi = N X i=1 fe (xi )k (xi) ; for k = 1; :::; M (1.27) 18 M X cj bhj (x); pi = 0 and j (1.28) where and belong to an appropriate nite dimensional subspace. Equations (1.27) and (1.28) are known as the Galerkin method [26]. This observation provides a useful link between the proposed method of data analysis and methods in computational mechanics. p q 1.6 Example Problem Consider a one-dimensional engineering system governed by the di erential equation ? dd x ! ( ) dd ( ) = a x x F x x ; x 2 (0 1) ; ; F (0) = (1) = 0 F : For this example problem the value of = 1 is selected for the exact system. An a-priori mathematical model with 0 = 1 3 is used along with the regularizing functional a a Z1 1 [ ] = 2 0 f a0 d d f x !2 d ? x : Z1 0 xf d = x 1Z 1 2 0 a0 ! d ? d 0 2 d + constant d d f f x x x : (1.29) Regularization by the functional of Eq. (1.29) suppresses oscillations of the experimental data that are dissimilar to the oscillatory behavior of the a-priori model. 19 Consequently, the approximation of a regularized mechanical system should lter the data noise. 1.6.1 Results from the Analytical Approximation The experimental data for the extrapolation study are obtained from the physical system response at x = 0:25 and x = 0:75, which are shown as triangles in Fig. 1.1. The optimal curve f~, regularized by the a-priori mathematical model, is evaluated from Eq. (1.23) and is shown in Fig. 1.1 for the regularization parameter  = 0:01. By selecting values of  Eq. (1.23) acts as a penalty method. The optimal curve passes through the data points while the curvature in the region between data matches the curvature of the a-priori mathematical model. This can be shown by noting that the function fcor (x) is a piece-wise linear function in the regions between data points. As previously mentioned, the use of Eq. (1.29) for regularization should lter data noise. Noisy experimental data was generated for this study by adding random noise , to the response of the physical system, that was statistically independent of location and uniformly distributed between d=2, where d = 0:025, and with variance V ar[] = d2=12. The generated data points are displayed as triangles in Fig. 1.2. Unlike the extrapolation example, the value of the regularization parameter  was determined such that Eq. (1.23) and the constraint N X i=1 jfe (xi) ? f (xi )j = E 2 "N X i=1 # i 2 = N  V ar[] =  (1.30) were satis ed for f = f~. More speci cally,  minimized the magnitude of the equation residual (), where 20 0.07 f (x) 0.06 0.05 0.04 0.03 0.02 0.01 0 0 0.1 0.2 0.3 0.4 0.5 0.7 0.6 0.8 0.9 1 x Extrapolation from insucient data for  = 0:01; (solid) the mechanical system F (x), (dash) the regularized approximation f~(x), (dash-dot) the numerical approximation fa(x). Figure 1.1 () = N X k=1 N X k=1 jfe(xk ) ? fa(xk )j2 ?  = fe (xk ) ? f0 (xk ) ? ? : N X i=1 0 G(xk ; xi ) @ G  + 2I !?1  fe ? f0 1 A  2 i (1.31) Use of Eq. (1.31) requires a-priori knowledge regarding the intensity of the random noise in measurements. This is usually available in engineering applications. The optimal , which acts as a Lagrange multiplier, was determined through the dichotomous search technique [82]. The regularized curve of Fig. 1.2 (dash) shows 21 that the developed method of data analysis can successfully discriminate between noise and the physical system response. 1.6.2 Results from the Numerical Approximation The numerical approximation by basis expansion consisted of seven (M = 7) Gaussian radial basis functions where the transfer function centers were uniformly distributed in the segment [0,1] with the center of the two boundary functions located at x = 0 and x = 1, respectively. The parameters k of Eq. (1.3), describing the localization properties of the Gaussian function, were all selected to be equal to 1=6. Figure 1 illustrates the acceptable matching between the analytical solution of the regularization problem and the numerical approximation. Figure 2 shows that the performance of the basis expansion in modelling with the noisy data of Section 1.6.1 is satisfactory. In the interpolation study,  was determined such that Eqs. (1.27) and (1.28) were satis ed with Eq. (1.30) for f = fa. An attractive integration method based in Tikhonov regularization has been proposed and evaluated. It has been shown in this chapter that since physical models of engineering interest can be expressed in the form of empirical, di erential and/or integral equations of varying degrees of delity then these model equations can be incorporated in the form of a regularizing functional for a robust matching of experimental data. The integration method can be viewed as the adjustment of a low delity, and computationally inexpensive, mathematical model response by experimental data. The developed integration method can also be viewed as an ecient procedure for interpolating experimental data by utilizing physically motivated criteria. Further, it has been shown how the developed method can be used to successfully extrapolate the physical system response at unmeasured coordinates and lter noisy experimental measurements. The radial basis expan- 22 0.08 f (x) 0.07 0.06 0.05 0.04 0.03 0.02 0.01 0 −0.01 0 0.1 0.2 0.3 0.4 0.6 0.5 0.7 0.8 0.9 1 x Interpolation of data with noise for optimal ; (solid) F (x), (dash) f~(x), (dash-dot) fa (x). Figure 1.2 sion has been examined as the computational apparatus of the developed method. However, the developments of the chapter retain a sucient level of generality and can be implemented in conjunction with any of the commonly used methods including nite di erence and conventional nite elements. 1.7 Nonlinear A-Priori Models It is apparent that though the mathematical model can be low delity, and computationally inexpensive, the accuracy of the interpolation and extrapolation of experimental data is directly dependent on the accuracy of the a-priori mathematical model. The scheme must therefore be able to accommodate moderate to highly nonlinear equations. Fortunately, the integration scheme can be easily extended to 23 physical systems described by nonlinear equations by modifying the regularization functional of Eq. (1.26) so that the objective function becomes (fa) = N X i=1 fe (xi) ? fa (xi ; 2 ) +  21 lhfa ? fnum ; fa ? fnum i + bhfa; qi (1.32)   with lhfa ? fnum ; fa ? fnum i = Z S X m=1 dmfa ? dmfnum 2 dx ; dxm dxm ! (1.33) where S is the highest derivative present in the a-priori model and fnum is the numerical approximation of f0 by some conventional computation mechanics method that can be performed oine. The multidimensional form of Eq. (1.33) is straightforward. 1.8 Implementation Concerns As attractive as the Tikhonov-based integration method appears, there are problems in implementation that are shared by all ANN applications. Of primary concern, the objective function given by Eq. (1.26), and constrained by Eq. (1.31), must be globally minimized through selection of optimal values of  and . This is a nonlinear optimization problem. As mentioned previously, there are two major approaches to improving the performance of optimization (ANN training) algorithms: 1) e ective selection of the initial values of the network parameters and 2) e ective introduction of network parameter constraints through the incorporation of regularizing functionals. Both approaches have been used in this chapter through the use of the integration method and the construction of radial bases func- 24 tions, respectively. The minimum of Eq. (1.26) can even be controlled through appropriate preselection of the Gaussian radial basis parameters. However, numerical experiments have shown that accuracy is very sensitive to the parameters controlling the basis functions. In addition, the accuracy is also sensitive to the number of bases used. Speci cally, after exceeding a certain number of uniformly distributed bases, the approximation becomes more and more inaccurate for optimal basis parameters [60]. Consequently, there is a limit to the number of ANN local bases that can be used in approximating a mechanical system. Of secondary concern, is that constraining the basis function parameters and minimizing Eq. (1.26) requires user interaction. For example, since the number of ANN local bases is limited, they should be judiciously placed. However, placement of bases is dependent on the type of basis, familiarity with the a-priori model and experimental data, which in turn requires user interaction. Additionally, if Eq. (1.32) is used then the a-priori model is limited by the accuracy of fnum . This is of special concern in complicated computational domains. To address these concerns a technique (optimal sequential function approximation) has been developed. This technique minimizes Eq. (1.26) in a stable and reliable fashion through problem-dependent optimal selection of  and using local minimization criteria. This approach involves relatively little user interaction and is independent of the type of basis selected by the user. It will also be shown that this technique can approximate the solution to a nonlinear partial di erential equation directly and exceed the accuracy of fnum on a uniform grid. Background information on elements used in developing the optimal sequential function approximation method are given in Chapter 2. Details on the method are discussed in Chapter 3. To demonstrate the method, the limiting case of Eq. (1.26) for  ! 1 has been minimized using the variational form of the nonlinear compressible Euler 25 equations as the a-priori model (Chapter 5). Section 6.2.1 discusses generalization of the optimal sequential function approximation, utilizing the method of weighted residuals, when a variational form is not available. 26 Chapter 2 Optimal Incremental Approximation: Background The proposed method of optimal sequential function approximation is based on a variety of techniques borrowed from distinct areas of computational methods. In particular, motivated by similarities observed in adaptive grid optimization techniques, an algorithm originated and employed in arti cial neural networks for iterative function approximation has been reformulated with the help of variational principles and the closely related method of weighted residuals. The objective of this chapter is to review the previously mentioned ideas and illustrate their exact role in, and in uence on, the proposed method. 2.1 The Method of Weighted Residuals One of the most popular methods for the solution of di erential equations is the method of weighted residuals. The basic idea is to require the inner product of the residual of a di erential equation, with respect to some weighting function, to vanish. Speci cally, if an approximate solution function (reproduced by a series expansion with speci ed functions  ( ) and unknown expansion coecients ) is introduced into the di erential equation, it will yield a nonzero equation residual . The inner product of the residual with respect to some weighting function is then de ned as the integral of their product over the domain of interest (Appendix A.1), and is set to zero. In this manner, a system of equations of the form h i, = 1 , is formed and solved for the unknown coecients. If the set G of functions f1 2  g is orthogonal, then the matrix A of the associated i x ci R w n R; i i ;:::;n ; ;:::; n 27 system of algebraic equations A =  is diagonal and the coecients are calculated with computational eciency. In general though, a function  2 G is orthogonal with respect to most, but not all, functions  2 G , = 1 ? 1, 6= . Accordingly, the matrix A will be sparse and banded (in the best case tridiagonal). In spectral methods, the matrix A will be fully, or nearly-fully, occupied. The choice of the weighting functions determines the method of weighted residuals. Among the most popular are the collocation method, the subdomain method, the least-squares method, the method of moments, and the Galerkin methods (Appendix A.3). The method of weighted residuals and its applicability to di erent classes of problems is extensively discussed in reference [25]. It is important to emphasize that no matter how signi cant the choice of the weighting (or test) functions is, the importance of the choice of the basis (or trial) functions used in the series expansion should not be underestimated. As will be demonstrated by means of speci c applications, the proposed numerical method for the solution of di erential equations via optimal sequential approximation may involve concepts of the method of weighted residuals. c b i j j ;:::;n j i i 2.2 Variational Principle Methods Variational methods in mathematics, physics, and engineering have been extensively studied and developed. A selection of the more general and popular texts include references [64, 51, 78, 68], and [25]. The variational principle method is based on the calculus of variations. Calculus of variations is concerned with variations of functionals, where a functional is dened as some form of correspondence between a function and the set of the real numbers. Such functionals (also called variational principles) may be readily dened for many problems of practical importance in engineering. The approximate 28 solution (given in terms of speci ed basis functions and their corresponding unknown coecients) of a given problem is introduced into the associated functional, and the unknown coecients are determined by making the functional stationary. In this regard, the rst variation of some functional F (f ) is de ned by  F (f + w) ? F (f ) ; F =  lim !0   (2.1) where f and w are some functions and  2 R. The functional F (f ) is typically given by a variational integral, and is made stationary by requiring its rst variation to vanish; the associated integrand is then set equal to zero, yielding the so-called Euler-Lagrange equation [65]. As mentioned above, for many physical problems of the general form H [f0 ] ?g =0 ; (2.2) an associated variational principle F (f0), whose Euler-Lagrange equation is Eq. (2.2), can be de ned, and the Raleigh-Ritz method can be employed for the computation of the unknown solution expansion coecients. In Eq. (2.2), H [] is some di erential operator, f0 is the solution function, and g is some forcing function. It is important to note that the vanishing of the rst variation of the functional is only a sucient condition for the existence of an extremum. To determine whether a functional is extremized, i.e. minimized or maximized, the examination of its second variation 2F is necessary. In particular, if 2F can be said to be positive, then the functional possess a minimum; if 2F can be said to be negative, then the functional possess a maximum, and if the sign of the second variation is inde nite, then the functional merely possesses a saddle point. 29 From a physical point of view, variational principles manifest the tendency of nature to favor the presence of extrema in physical systems. The mathematician Euler expressed his certainty that \...there is no doubt that all the e ects of the world can be derived by the method of maxima and minima from their nal causes as well as from their ecient ones" [25](page 335). However, the derivation of \natural" variational principles for all classes of problems is far from being fully understood and successful. This fact makes the use of \contrived" variational principles necessary. In particular, variational principles are de ned based on the properties of the previously introduced di erential operator H []. If H is linear, i.e. H = L, then L is either self-adjoint or not self-adjoint (Appendix A.4). If L is self-adjoint, then a variational principle is readily de ned in terms of L, f0, and g. If L is not self-adjoint and the operator L can be transformed to a self-adjoint operator L without changing the associated EulerLagrange equation, then a variational principle is de ned in terms of L , f0, and g. If the nonself-adjoint operator L cannot be transformed to a self-adjoint operator, then a variational principle is de ned in terms of L, f0, g, L, v0, and h , where L[v0] ? h = 0 is the adjoint problem of L[f0] ? g = 0. If H is nonlinear, i.e. H = N , then the term \self-adjoint" cannot be used, because the adjoint of a nonlinear operator cannot be identical to the operator, and is substituted by the term \self-sucient" [5]. If N is self-sucient, variational principles can be de ned without extreme e ort. However, if N is not self-sucient, the use of Frechret derivatives (Appendix A.4) cannot be avoided and the tediously derived variational principles are quite complicated and not amenable to computational implementations. It is interesting to note that there always exists a Galerkin method (a member of the method of weighted residuals) that is equivalent to the variational method; 0 0 30 in fact it was Galerkin himself who noted this fact [25](page 223). Due to this signi cant property, Galerkin methods have been extensively used in nite element methods. Members of the method of weighted residuals, in particular the Galerkin method, have been primarily developed and used for the solution of di erential equations due to their simplicity and computational eciency. Less frequently, variational methods have been utilized by the Raleigh-Ritz method. For both of these cases, the Euler-Lagrange equation is solved numerically in order to determine the function which satis es it and the associated initial and/or boundary conditions. However, it is argued by Finlayson [25](page 310) and Greenspan [31] that the numerical approximation of the function that extremizes a given variational principle is to be preferred to the numerical approximation of the function that satis es the associated Euler-Lagrange equation. The proposed method for optimal sequential function approximation for the solution of di erential equations is based on the use of variational principles. In particular, the variational principle is adopted as the objective function(al) in the optimization problem(s) to be solved numerically in order to obtain the extremal function that satis es the associated Euler-Lagrange equation and the boundary conditions. A major advantage of the use of variational principles is that coupled and/or uncoupled equations with general boundary conditions can be concisely and succinctly represented by a single functional. 2.3 Adaptive Grid Optimization Numerical methods for the solution of uid mechanics problems rely on the discretization of the continuous spatial (and/or temporal) domain into a grid of nodal points. The requisite computational cost and the eciency of the solution method 31 are strongly a ected by the type of discretization. For example, slow convergence and numerical instability, resulting from poor condition numbers of the associated discretization matrices, have been reported in the literature for uniform grids. Therefore, optimization in computational mechanics has focused primarily on the grid selection for a given problem. Jensen introduced a nodal density function based on local discretization error estimates as a criterion for optimal node distribution [44], while Oliveira showed that optimal node distribution provides minimal total potential energy of the system [70]. In fact, it was proven that such distributed nodes lie along isoenergetic contours. Felippa solved the direct and computationally expensive grid optimization problem for a given nite element con guration, and then relocated the nodes according to a total energy criterion [23, 24]. This process was repeated iteratively until convergence to an optimal grid was achieved. His approach was limited to problems governed by self-adjoint, positive de nite, di erential operators. Moreover, unavailable exact derivatives and/or inaccurate derivative approximations dramatically decreased the eciency of the method. Results were reported for extremely coarse two-dimensional grids (8 to 36 nodes), due to the complexity (constraints) and high dimensionality of the associated optimization problem. Diaz, Kikuchi, and Taylor classi ed the optimal nodes distribution as a relocation technique, or r-method, and included it among the well known h-, p-, and h-p adaptive methods [16]. In h-methods, the grid spacing h is reduced; additional interpolation functions are introduced to improve the quality of the solution. In p-methods, the order of the interpolation functions is increased, while h-p methods combine the introduction of additional functions with the increase of the order of the existing and/or introduced functions. In the r-method, the location of the 32 nodes of the grid, i.e. the location of the functions, is adjusted for higher accuracy. The grid optimization algorithm of Diaz et al. was based on the minimization of an estimated upper bound on the potential energy of the nite element solution. The associated objective function involved a global interpolation estimate (given by simple summation of local element interpolation error estimates) which serves as an upper bound on the discretization error. Recently developed adaptive methods of computational mechanics substitute the principle of the system potential energy minimum with the weaker criterion of the homogeneous distribution of the local approximation error over the elements [13, 17, 15]. The common characteristic of adaptive grid techniques is that a grid is generated and a solution is found. Based on the solution, the grid is adapted and regenerated in order to repeat the described process. Multigrid methods share this a-posteriori error estimation philosophy. It is well-known that most of the computational time required by codes that utilize adaptive or multigrid techniques is spent on the necessary grid generation [69, 36, 22]. According to Oden [69], the major challenges to adaptive methods in computational uid dynamics include 1) the use of unstructured meshes and therefore elaborate and complicated data structures, 2) the necessity of explicit or iterative solution techniques due to the poor performance of direct solvers on dynamically evolving unstructured meshes, 3) stability issues of the associated numerical schemes stemming from the continuous changes in the data structures and polynomial orders, and 4) the computational overhead of the error estimation and the adaptation process. The method presented in this report features the advantages of adaptive methods without sharing their disadvantages. Speci cally, while the method is adaptive in an iterative manner and is based on the optimization of some appropriate crite- 33 rion, the dimensionality of the associated optimization problem is kept low, derivatives are not required, and most importantly, a grid is not built in the traditional sense. Unstructured mesh data structures are not required for the realization of the method due to the use of simple tensor products of low-order basis functions. Moreover, computationally expensive matrix-assembling algorithms and solvers of systems of equations are avoided. Variational principles and weighted residuals concepts are utilized to de ne well-posed optimization problems; numerical stability issues are limited to the nonlinear optimization process. Most signi cantly, the grid, de ned by the location of the optimally selected basis functions, evolves with the solution and is not based on a-posteriori error estimates; the computational overhead of the a-posteriori error estimation and respective regridding is eliminated. Lastly, the parallel processing of the optimization problem, de ned by the proposed algorithm, dramatically increases the eciency of the proposed method. 2.4 Arti cial Neural Networks Function approximation, in the framework of neural network computations, has been primarily based on the results of Cybenko [11] and Hornik, Stinchcombe, and White [39] who showed that any continuous function on a d-dimensional compact subset of Rd can be approximated arbitrarily well by a linear combination of simple one-dimensional functions i( ) : Rd+m R f f0  fa = uan (x) = c0 +  ! Xn cii(x; i)) ; i=1 g (2.3) where x Rd, i Rm, and ci R, represent the independent variables, function parameters and linear combination coecients, respectively. If the linear and nonlinear network parameters in Eq. (2.3) are selected by minimizing the square 2 2 2 34 of the L2 norm, 2  Z (f0 ? fa)2dx ; (2.4) then Barron [3] showed that if the functions i are xed,  cannot be smaller than O(n?2=d). He referred to this phenomenon as \the curse of dimensionality". Therefore, he proposed, based on the theoretical work of Jones [46, 47], the following iterative method for sequential approximation uan (x) = u ?1 (x) + cn (x; n ) ; a n n (2.5) where n; cn, and n are optimally selected at each iteration of the algorithm. As a result, the high-dimensional optimization problem associated with neural network training is reduced to a series of simpler low-dimensional problems. A general principle of statistics was utilized to show that the upper bound of the error  is p of the order C= n, where C is a positive constant. The functions i , introduced above, have traditionally been sigmoidals in neural network computations, i.e. functions with the properties  lim i() = 0 and lim  () = 1 : !1 i !?1 Alternatively, Orr introduced radial basis functions in his forward selection method of sequential network training [71]. Orr's radial basis functions are of the form (x; xc; r), where x; xc 2 Rd and r 2 R are the independent variables, the center point, and the radius, respectively. Their value decreases (if they are bounded) or increases (if they are not bounded) monotonically with increasing distance from 35 the central point. Orr's method is in essence a method of incremental function approximation. At each iteration of the algorithm, an additional basis function, which produces the largest reduction of error in the previous iteration, is chosen from a given set of functions and added to the approximation. However, the forward selection training method can be inecient in that it may require signi cant computational resources when the set of trial functions is large. A similar principle is utilized in Platt's resource allocating networks (RAN) [73]. Whenever an unusual pattern is presented to the network in an on- or o -line network training procedure, a new computational unit (according to the neural network terminology), is allocated. The computational units respond to local regions of the input space. The concept of incremental (or sequential) approximation is one of the major features of the proposed method. Following the main idea of Barron's algorithm, optimal basis functions with local support are added to the approximating series expansion. A broad class of functions can be employed. B -splines are popular and resemble radial basis functions, but Gaussian functions (see Appendix A.2), families of locally de ned orthogonal polynomials, and combinations of hyperbolic tangents can also be used. 36 Chapter 3 Optimal Incremental Approximation: Approach The proposed method of optimal incremental function approximation for the solution of di erential equations is presented in this chapter. The development of this method was motivated by the recent advances in parallel optimization techniques and by numerical concepts reviewed in the previous chapter. The main features of the proposed method are a) the iterative nature of the proposed algorithm for the approximation and b) the solution of a nonlinear optimization problem at each stage of the algorithm. The objective of the development of the method was the employment of the latter in the solution of di erential equations. 3.1 The Proposed Algorithm The developed algorithm can be summarized as follows [59]. Given a problem H [f0 (x)] ? g = 0, governed by one (or more) algebraic or di erential equation(s), the dependent variable(s) is (are) approximated by a basis expansion of the form given in Eq. (2.3). An incremental set of basis functions is built sequentially to improve the expansion-based approximation. At each stage of the algorithm, the parameters of the new basis function and the corresponding expansion coecient are optimally determined by solving a low-dimensional, nonlinear optimization problem. Assuming that the algorithm has reached step n, the approximating function is given by f0 (x)  fa (x) = uan X (x) = c  (x;  ) = c  (x;  ) + u ? (x) : n i=1 i i i n n n a n 1 (3.1) 37 The coecient c and the parameters  of the associated basis function  (x;  ) are optimally computed according to an appropriate criterion, i.e. by solving a nonlinear optimization problem, while the coecients c and the basis functions  (x;  ), i = 1; : : : ; n ? 1, are held xed. If the given problem is linear, the nonlinear optimization problem can be decomposed into two lower-dimensional problems; one for the computation of the parameters  and one for the calculation of the coecient c . In this manner, computational eciency is enhanced. A broad class of locally de ned functions can be used, e.g. low-order polynomials, B -splines, radial basis functions, and combined hyperbolic tangents. It is emphasized that there are no restrictions on the distribution of the interpolation functions over the domain of interest; overlapping is possible. Orthogonality or orthonormality properties can be attained, if desired or required, by means of the Gram-Schmidt algorithm, either during the basis-building process or at its completion. The latter requires the recalculation of the coecients and reduces the computational eciency of the proposed method since a system of equations has to be solved. The algorithm can be initialized with either an empty set or an arbitrary number of predetermined functions and coecients; the second option enables the use of coarse grid solutions obtained, for example, from a previous nite element analysis of the given problem. The sequential algorithm is halted when the relative di erence between the previous and current approximations falls below a user-de ned tolerance level. The details of the method will be presented and computational issues will be addressed with the help of examples of increasing complexity. Results for spen n n i i i n n n 38 ci c numerical examples (with emphasis on uid dynamics applications) will be presented and discussed, and the overall performance will be evaluated based on parameter studies. 3.2 The Parallel Direct Search Optimization Technique The computational eciency of the method depends on the cost of the nonlinear optimization problem solution. The dimensionality of the nonlinear problem is kept low due to the iterative nature of the algorithm. Moreover, the parallel direct search (PDS) optimization technique is used to solve the optimization problem eciently. PDS is a software package that consists of a collection of Fortran routines for solving both unconstrained and constrained nonlinear optimization problems using direct search methods. Direct search methods have the advantage of requiring information about the objective function only. Derivatives do not need to be calculated, and the condition for convergence is that the objective function has to be continuously di erentiable. The user has to provide objective function and constraint(s) evaluation routines and an input le. This fact makes PDS quite easy to use. In this manner, the combined user e ort and computer time may often be smaller than that required for gradient-based methods since the latter are theoretically more ecient but practically more complicated to implement. In addition, ingredients essential to gradient-based methods, i.e. derivatives, may be inaccurate, noisy, not available, or quite expensive to compute. PDS can achieve relatively high performance rates because of the parallel execution of the necessary objective function evaluations. Exceptional scalability is achieved by fully exploiting additional processors, yielding almost linear speed-up. The PDS optimization technique and its theoretical background are presented in references 39 [14, 84, 52], and [85]. Reference [10] reports the application of PDS to an optimal design problem. The working scheme of PDS is as follows. Given an initial guess for the optimization variables vector of length K , PDS evaluates the objective function at d vertices of a simplex de ned on the hyperplane of dimension K . The \best" vertex is determined and compared to the previous iterate. If the di erence lies within a user-de ned tolerance tol, the search is terminated; otherwise, a new iteration is performed with the current iterate as the initial vertex. The search size is controlled by the user by the input of the parameters d and tol. For unconstrained problems the restriction d  2K applies, while for constrained problems d  K 2 + K . However, d should be typically chosen large in order to ensure a ne-grained search. PDS works best when the number of the optimization variables is kept small, i.e. when the ratio of the number of necessary function evaluations to the number of available processors is low. This requirement is satis ed by the proposed method for incremental function approximation. Computational implementation issues will be addressed in the following sections and in Appendix A.5. 3.3 The Iterative Character of the Proposed Algorithm This section intends to demonstrate how the iterative nature of the proposed algorithm relates the method of optimal incremental approximation to the Gauss-Seidel method for the solution of a linear system of equations. There are two types of numerical methods for the solution of a linear system of algebraic equations of the form Az = b, where A 2 RN N and b 2 RN are known and z 2 RN is the unknown solution vector; direct methods yield the exact solution, as far as computer arithmetic accuracy allows, while iterative methods 40 approximate the exact solution up to a user speci ed level of accuracy. Iterative methods are computationally more ecient than direct methods for large systems (large ), especially when the matrix A is sparse, but their performance, i.e. convergence rate, strongly depends on the structure and properties of the matrix A. In general, iterative methods are more stable, as they will usually dampen errors with increasing number of iterations [49]. Various methods and algorithms have been developed to enhance and/or exploit the properties of the matrix A [2]. Iterative methods are classi ed as either stationary or nonstationary. In stationary iterative methods (Jacobi, Gauss-Seidel, SOR, etc.), the matrix A remains invariant, while in nonstationary methods (e.g. conjugate gradients (CG), GMRES, etc.) A varies with each iteration. The objective of the latter is to accelerate the convergence to a suciently accurate solution. The Gauss-Seidel method belongs to the class of stationary iterative methods. Diagonal dominance is required for convergence, and a low spectral radius is desired for a satisfactory rate of convergence (Appendix A.6). In the Gauss-Seidel method, the matrix A is decomposed into its lower triangular part, which contains the diagonal of the original matrix, and its strictly upper triangular part, i.e. an upper triangular matrix with a zero diagonal, N A  =  , (L + D) + U  =  z b z z b and the iterative algorithm is (L + D)k+1 =  ? U k z b z : (3.2) 41 If A is diagonal dominant, the Gauss-Seidel method converges for any starting vector 0. Typically, 0 = 0. Given a system of algebraic equations, e.g. a di erential equation discretized by means of the Bubnov-Galerkin method, the use of the Gauss-Seidel method for its solution can be interpreted as follows. The rst coecient 1 of the solution vector  is readily calculated by solving the rst equation of the system. Notice that the rst equation depends only on the rst interpolation function 1 since the Bubnov-Galerkin scheme implies that the inner product h  i (where is the di erential equation residual and the weighting function  is identical to the trial function  ) vanishes 8 , assuming that 0 = 0. The second equation is then solved for 2 with known 1, employing the functions 1 and 2. This process is repeated for all equations; it is clear that at step all components with are known and the employed functions are the functions  with  . After all equations have been solved, a new outer iteration begins to improve the accuracy of the solution vector. The proposed algorithm of Section 3.1 resembles the Gauss-Seidel method in that it proceeds in the same manner, with the exception that at each step of the algorithm the -th interpolation function is optimized instead of being xed and preselected. It can be argued that the analogy of the matrix (L + D) is not prede ned and constant, but built in an adaptive way. It therefore resembles a \nonstationary" Gauss-Seidel method. Consequently, the convergence rate should be improved, compared to that of the stationary Gauss-Seidel method. Moreover, the proposed algorithm is not just an improved technique for the solution of linear systems of algebraic equations. It is an adaptive, optimal approximation algorithm z z z z R; R n n n n z z z n zi i n i n i < n 42 that imitates grid optimization schemes and can be also employed in nonlinear problems, as will be demonstrated in Chapter 5. 3.4 Function Approximation In this section, the implementation of the proposed method for two simple nonlinear function approximation problems in one and two dimensions will be demonstrated. 3.4.1 Implementation At the n-th stage of the proposed sequential approximation algorithm, some function f0(x) is approximated by Eq. (3.1). Let en = f0(x) ? uan (x). The squared error between the function and its approximation can be de ned as   ken k2 = hen ; eni = Z e2n dx (3.3) and can be rewritten as Z Z Z (f0(x) ? uan(x))2dx = (f0(x) ? uan?1(x) ? cnn (x))2dx = Z Z 2 2 2  (en?1 ? cnn (x; n)) dx = en?1 dx + cn (n(x; n))2dx ? Z 2cn en?1n (x; n)dx = ken?1k2 + c2nkn (x; n)k2 ? 2cn hen?1; n (x; n)i : (3.4) The linear coecient cn can be determined by di erentiating  with respect to cn 43 and requiring the derivative to vanish @ @cn = 0 = 2cn kn (x; n)k2 ? 2hen?1 ; n(x; n)i ; (3.5) which yields cn = hen?1 ; n(x; n)i : kn (x; n)k2 (3.6) Note that the procedure for evaluating the coecient cn is equivalent to the Bubnov-Galerkin scheme since it provides for the orthogonality of the error en to the weighting function n (x; n). Indeed, Eq. (3.5) shows that Z en?1 n (x; n )dx ? cn Z (n(x; n))2dx = hen; n (x; n)i = 0 : (3.7) Substituting Eq. (3.6) into Eq. (3.4) yields  = ken?1 k2 ? hen?1; n (x; n)i2 kn(x; n)k2 : (3.8) To minimize the error, the parameters of the n-th basis function must be selected to maximize the last term of Eq. (3.8). This nonlinear optimization problem can be solved eciently by PDS. Clearly, the proposed method resembles some of the conventional adaptive numerical techniques from computational mechanics in that the new basis function is selected to provide the largest projection on the error of the preceding iteration. Speci cally, the new basis function is positioned at the location of the largest previous error. Moreover, the basis function can be also 44 viewed as a weighting function; the proposed method resembles the method of weighted residuals. 3.4.2 Numerical Examples The proposed method has been applied to two nonlinear function approximation examples. In particular the one-dimensional function f0(x) = x2 and the twodimensional function f0(x) = x2 + y2 have been approximated. For the purpose of evaluating the algorithm, piecewise linear transfer functions were used to construct [57] the computationally straightforward piecewise linear \hat" (B1-splines) basis functions. B1-splines are de ned as 8 >> x?(xM ?xl) >< xl xr )?x (x) = > (xM +  x r >> :0 if xM ? xl  x  xM if xM  x  xM + xr otherwise, (3.9) where the parameters xM , xl, and xr denote the location of the center of the function, its width to the left, and its width to the right, respectively. The simple product of two one-dimensional functions, which can be constructed from a sigmapi network, has been used in the two-dimensional case, i.e. (x; y) = (x)(y). The optimization variables are given by n = (xM ; xl; xr)Tn (3.10) for the one-dimensional case and by n = ( x; y )Tn = (xM ; xl; xr; yM ; yl; yr)Tn (3.11) 45 for the two-dimensional case. The initial values for the optimization variables, required by PDS, are chosen so that each new basis function is initially centered in the domain of interest and encompasses it. As an alternative, the initial guess can be chosen such that the center of the basis function is located at the position of the maximum error which has been calculated in the previous iteration in order to improve the eciency of the PDS technique. However, this does not a ect the accuracy of the results since PDS performs a ne-grained search throughout the domain. The approximations obtained by the proposed method are satisfactorily accurate. The RMS error, de ned as RMSNM = s PM j=1(f0(xj ) ? uaN (xj ))2 M (3.12) ; where is some sucient number of uniformly distributed trial points, is plotted versus the number of optimized basis functions in Fig. 3.1 for both the one-dimensional and the two-dimensional problems using = 11 and = 101 uniformly distributed evaluation points, respectively. Linear and quadratic convergence lines are included to illustrate the satisfactory convergence rate. Convergence rates with respect to two di erent numbers of used basis functions 1 and 2 are de ned as linear if M N M M N NM1 RMS N 2 RMSM = N =N 2 1 ; N (3.13) 46 (a) −1 (b) 0 10 10 linear RMS RMS linear −2 10 −1 10 quadratic quadratic −3 10 −2 0 1 10 2 10 N 10 10 0 1 10 2 10 N 10 (b) (a) Convergence rates for the function approximation problems: (a) one-dimensional case and (b) two-dimensional case. Figure 3.1 and quadratic if NM1 RMS N 2 RMSM = (N =N )2 2 1 (3.14) ; in order to be compatible with the convergence terminology used in the nite element method, where the grid spacing is used as the convergence metric (for example 2-convergence etc.). If Eq.s (3.13) and (3.14) become inequalities with larger right sides, the convergence rates are de ned as superlinear and superquadratic, respectively. When nite element methods are applied on uniform grids, the grid spacing and the number of basis functions used are directly related through the formulas = 1 ( ? 1), for one-dimensional problems, and x y = ( Nx1?1 )( Ny1?1 ), for two-dimensional problems. h h h h N = N h h 47 Chapter 4 Linear Di erential Equations Consider the general boundary value problem with homogeneous Dirichlet boundary conditions [ ( )] = ( ) in L f0 x g x ; ( ) = 0 on f0 x @ (4.1) ; where [] is a linear, self-adjoint di erential operator. Homogeneous Dirichlet boundary conditions are assumed for the sake of simplicity. For linear, self-adjoint problems associated with homogeneous Dirichlet boundary conditions, the variational principle associated with Eq. (4.1) is given by the functional L  [ a( )]  21 h a( ) a( )i ? h ( ) a( )i f x l f x ;f x g x ;f x (4.2) : where h i is the bilinear symmetric di erential form associated with the operator discussed in Section 1.5. If the functional  is numerically minimized by the function a( ), then a( ) automatically satis es the associated Euler-Lagrange equation Eq. (4.1); moreover, if is positive de nite, the solution is unique [64](pages 74-75). However, a function a( ) that satis es Eq. (4.1) does not necessarily minimize the functional . It is due to this fact that variational principles are employed by the proposed method. In this regard, the approximate function given by Eq. (3.1), is introduced l ; L f L f x x f x 48 into the functional  to yield  [uan(x)]  21 lhuan(x); uan(x)i ? huan(x); g(x)i : (4.3) This functional may be used as the objective function to be minimized in an optimization problem in order to determine the optimal parameters n and the associated coecient cn . However, in this case the properties of the di erential operator can be exploited to simplify the optimization problem. Since the di erential operator is self-adjoint, the boundary conditions are homogeneous and of the Dirichlet type, and huan (x); g(x))i = lhuan (x); f0(x)i, Eq. (4.3) can be rewritten as  [uan(x)] = 21 lhf0(x) ? uan (x); f0(x) ? uan (x)i ? 21 lhf0(x); f0(x)i : (4.4) At the n-th step of the proposed sequential method, the approximate solution is given by Eq. (3.1). Again, let en = f0(x) ? uan(x). Then, en = en?1 ? cnn (x; n), and Eq. (4.4) becomes  [uan(x)] = 21 lhen?1 ? cn n(x; n); en?1 ? cnn (x; n)i ? 21 lhf0(x); f0(x)i = 1 lhe ; e i + 1 c2 lh (x;  );  (x;  )i ? c lhe ;  (x;  )i ? n n?1 n n 2 n?1 n?1 2 n n n n n 1 lhf (x); f (x)i : (4.5) 0 2 0 As in the function approximation case, the coecient cn can be determined by 49 equating the partial derivative of  with respect to cn to zero @  [uan (x)] @cn = 0 = cn lhn(x; n); n(x; n)i ? lhen?1; n (x; n)i : (4.6) The coecient cn can then be calculated from cn = lhen?1 ; n (x; n)i lhn (x; n); n (x; n)i (4.7) and Eq. (4.5) can be rewritten as  2  [uan(x)] = 21 lhen?1; en?1i? 12 lhf0(x); f0(x)i? 12 lh(lhe(nx;?1; )n; (x;(x;n)i) )i :(4.8) n n n n The equivalence of the method to the Bubnov-Galerkin method is, as in Section 3.4.1, readily deduced by examining Eq. (4.6): 0 = cnlhn (x; n); n(x; n)i ? lhen?1; n (x; n)i = hL[cn n(x; n)]; n(x; n)i ? hL[en?1]; n(x; n)i = hRn ; n(x; n)i : (4.9) To improve the accuracy of the approximation, the last term of Eq. (4.8) must be maximized by appropriately selecting the parameters n of the basis function n (x; n) since the rst two terms of Eq. (4.8) are constant with respect to n. 50 4.1 Boundary Conditions Consider the linear, self-adjoint problem L[f0(x)] = g(x) associated with general boundary conditions B [f0(x)] = b(x). The boundary operator B may represent a mixture of Dirichlet and Neumann conditions. Assuming that an arbitrarily chosen function s(x) satis es the non-homogeneous boundary conditions but not necessarily the di erential equation, then the function q (x) = fa (x) ? s(x) will satisfy the homogeneous boundary conditions, provided that fa(x) satis es the non-homogeneous boundary conditions; since B [] is the general linear boundary operator, then B [q] = B [fa] ? B [s] = 0. If p = f0 ? s, where f0 is the exact solution of the problem, then, since L is linear L[p] = L[f0 ] ? L[s] = g ? L[s] = # : (4.10) In this manner, the non-homogeneous boundary value problem has been transformed to a homogeneous boundary value problem. The variational principle is then  [q]  21 hq; L[q]i ? hq; #i (4.11) and can be reformulated in terms of the unknown function fa Z Z  [fa] = (faL[fa] ? 2fag + faL[s] ? sL[fa])dx + (2sg ? sL[s])dx ; (4.12) where the second integral can be omitted due to its independence with respect to fa . Since the di erential operator is self-adjoint, integration by parts would force the last two terms of the rst integral to cancel each other, but the presence of 51 non-homogeneous boundary conditions will result in the appearance of non-zero boundary terms. The same is true for the integration by parts of the rst term of the rst integral, i.e. the boundary terms will only be zero for homogeneous boundary conditions. It is clear that the trial functions have to only satisfy the Dirichlet boundary conditions; the Neumann boundary conditions are automatically satis ed by the variational principle. Though our problem of primary interest is nonlinear, the previous analysis hints at how boundary conditions can be addressed and hints at problems that popular ANN bases may encounter using the optimal sequential approximation technique. A preselected sum of weighted bases can be used to assemble the function s(x). The remaining weighted bases are then constrained to satisfy homogeneous boundary conditions and approximate q(x). Unfortunately, Gaussian radial bases and local bases formed from hyperbolic transfer functions [60] cannot be used to accurately approximate q(x), by the optimal sequential approximation technique, in a straight-forward manner since the value of the basis function (x) and its derivatives approach zero as x ! 1. A basis should be used such that the derivatives of (x) are nonzero as (x) vanishes for x ! 1. The remedy is relatively straightforward. A radial basis function is modi ed such that only its value, but not its derivatives, becomes zero at some nite value of x (Fig. A.2 (b)). The augmented basis function to be used in this report is (x) = e e ??1 1 ; ()  x  2 R where 8> x?(x ?x ) >> l Mxl l < (x +x )?x  (x) = > r M xrr >: 0 if xM ? xl  x  xM if xM  x  xM + xr otherwise, (4.13) (4.14) 52 Note that the use of the augmented radial basis function with the proposed sequential method advances the method from an h-r scheme to an h-r-p adaptive scheme since, by a Taylor series expansion,  x  +  +   +  (x) = e e ??1 1 = : 1 +  + 2 +    2 ( ) 2 3 (4.15) As a result, the order of the basis function increases with the magnitude of . In addition, we can recover the B basis since Eq. (4.15) shows that (x) = (x) for  = 0. 2! 3! 2! 3! 1 4.2 Linear Nonself-Adjoint Problems Application of the method is straightforward for a linear self-adjoint ordinary differential equation with homogeneous Dirichlet boundary conditions. Therefore, as a test, a linear, second-order, nonself-adjoint ordinary di erential equation with non-homogeneous Dirichlet boundary conditions will be solved by means of the optimal incremental approximation method. The transformation of the nonselfadjoint operator into a self-adjoint operator will be demonstrated. In the case when this transformation is not possible, the shortcoming of the method will be discussed and potential alternatives to overcome this shortcoming will be proposed. Consider the general, linear, second-order, nonself-adjoint di erential equation L[f ] = 0 d f +  df + f = g ; x 2 [a; b] ; dx dx 2 0 2 0 0 (4.16) associated with general boundary conditions. The coecients , , and of the di erential equation may also be functions of the independent variable x. The operator L may be adjusted for self-adjointness by premultiplication with some 53 function of the independent variable x, say P (x) [90], [25](page 309). In order for the new operator L to be self-adjoint, it is necessary and sucient (Appendix A.4) that the following expression for two arbitrary functions z and w 0 Zb a Zb a wL [z ]dx ? 0 wP Zb a zL [w]dx = 0 ! Zb a wP L[z ]dx ? d2z +  dz + z dx ? Z b zP dx2 dx a Zb a zP L[w]dx = ! d2w +  dw + w dx dx2 dx (4.17) is a function of boundary terms that arise from integration by parts only. Integration by parts yields Zb ! d2z + wP  dz + wP z dx ? wP d x2 dx a ! Zb 2 d w d w zP dx2 + zP  dx + zP w dx = a ! Z b d(wP ) dz d z ? dx dx + wP  dx + wP z dx + (B.T.)1 ? a ! Z b d(zP ) dw d w ? dx dx + zP  dx + zP w dx ? (B.T.)2 = a Zb Zb Zb ? a P ddwx ddxz dx ? a w d(dPx ) ddxz dx + a wP  ddxz dx + Z b dw dz Zb wP z dx + (B.T.)1 + P dx + d x dx a a Z b d(P ) dw Zb dw Z b z d x ? zP  ? zP wdx ? (B.T.)2 = dx dx a a ! dx a Z b dz d(P ) dx + (B.T.) ? P ? w 1 dx dx a ! Z b dw d( P ) z dx P  ? dx dx ? (B.T.)2 ; a (4.18) where (B.T.)1 and (B.T.)2 denote the boundary terms stemming from the two integrations by parts. By examining Eq. (4.18), it is obvious that for nonzero and 54 , the necessary and sucient condition for L to be self-adjoint is 0 P ? d(P ) = 0 ; dx (4.19) which yields after integration P (x) = e ; (4.20) R where  = ?1( ? ddx )dx. Consequently, the operator L is self-adjoint, and a variational principle for Eq. (4.16) is readily de ned as [12] 0  [fa]  21 hfa; L [fa]i ? hfa; gi ; (4.21) 0 provided that the boundary conditions are homogeneous. If the boundary conditions are non-homogeneous, then the functional is de ned, according to the developments of Section 4.1, as  [q]  21 hq; L [q]i ? hq; #i (4.22) 0 and can be rewritten in terms of fa  [f a ] = Z b a Z b (faL [fa] ? 2fag + faL [s] ? sL [fa])dx + (2sg ? sL [s])dx :(4.23) 0 0 0 0 a Since the second integral is invariant with respect to fa, it can be dropped from the functional. 55 It is interesting to note that after integrating by parts all terms of the rst integral involving the di erential operator, and if the non-homogeneous boundary conditions are of the Dirichlet type, the functional of Eq. (4.23) will be apparently identical to the functional of Eq. (4.21), derived for the homogeneous boundary value problem, after this is also integrated by parts. The di erence is that in this case the function fa(x) must satisfy the non-homogeneous Dirichlet boundary condition; hence, Dirichlet boundary conditions are essential boundary conditions. In the case of Neumann boundary conditions, the functional of Eq. (4.23) will include terms that represent the Neumann boundary conditions. In this regard, the function fa(x) does not have to satisfy the Neumann boundary conditions since they are included in the functional (natural boundary conditions). It should be noted that the rst variation of the functional of Eq. (4.23) yields Eq. (4.16) as the associated Euler-Lagrange equation since the transformation of the operator can be interpreted as a simple variable transformation [25](page 309). The second variation of  indicates that the functional should be minimized. Remarks 1. If the coecients of the di erential equation are constant, then even ordered derivatives are self-adjoint while odd ordered derivatives are nonself-adjoint. 2. If  = 0, then the original operator L is self-adjoint and no transformation or adjustment is necessary. 3. If = 0, then a transformation or an adjustment is impossible. The introduction of the adjoint problem L[v] = h into the variational principle is 56 required to yield [5]  Z b a (v(L[fa] ? g) ? h fa)dx = Zb a (fa(L[v] ? h ) ? gv)dx : (4.24) Unfortunately, it is possible that the sign of the second variation of Eq. (4.24) is inde nite. In this case the functional cannot be extremized. An alternative would be to minimize the square of the rst variation of the functional, in other words to apply the least squares optimization method to the associated Euler-Lagrange equation. However, the simplicity and eciency of the proposed method would decrease dramatically. Alternatives for the generalization of the method due to the lack of variational principles are considered in Chapter 6.2.1. 4.2.1 Numerical Example: The Convection-Di usion Equation The linear ordinary di erential equation 2 ? ddxf20 +  ddfx0 = 0 with ;  > 0 and 0  x  1 (4.25) is a mathematical model of adequate delity for the steady-state balance between convection and di usion. Guymon derived in references [34] and [35] variational principles for the unsteady problem in one and two dimensions using a similar approach and utilized them for the solution by means of the nite element method. The coecients and  represent the viscosity and convection coecients, respectively, and will assumed to be constant for the sake of simplicity. Equation (4.25) is nondimensionalized and, in combination with the non-homogeneous Dirichlet 57 boundary conditions f0(0) = 0 and f0(1) = 1, possess the exact solution e x ?1 f0 (x) =  : e ?1  (4.26) The solution function, representing velocity, is characterized by the presence of a boundary layer adjacent to the boundary at x = 1 [27]. From a physical perspective dissipation is, for many ow problems, of signi cance only in the region of the boundary layer. Therefore, numerical solutions are often oscillatory when the dependent variable exhibits large gradients across the boundary layer. Speci cally, in nite di erence methods the accuracy and oscillations of the numerical solution [27] are highly dependent on the cell Reynolds number, de ned as Recell = x . In this regard, Eq. (4.25) poses a challenging problem to be solved by the proposed method. 2 Given the fact that the di erential operator L[] = ? ddx2 +  ddx is nonselfadjoint, and following the approach presented in the previous section, it can be found that the operator L can be transformed into a self-adjoint operator L , if it is premultiplied by the function P (x) = e?  x. The associated variational principle is then de ned, according to Eq. (4.23) and for g = 0, as 0  [fa] = Z1 0 (faL [fa] + faL [s] ? sL [fa])dx ; 0 0 0 (4.27) where s(x) = x is chosen to satisfy the boundary conditions. Integrating by parts the rst term of the integral yields Z1 0  fa e? x ! Z1 2fa 2  x d fa  Z 1 ?  x dfa d fa d ? ? dx2 +  dx dx = 0 fae dx dx ? 0 fae dx2 dx = 58 Z1  df fa e? x a dx + Z 1 d(fa e?  x ) dfa !1  df fae? x a = dx dx dx dx ? dx 0 0 !1 Z 1  dfa !2 Z1  Z 1 ?  x dfa   d f d f  a a ? x ? x ? x fa e dx dx + 0 e dx dx ? 0 fae dx dx ? fae dx 0 = 0 ! Z 1  dfa !2  dfa ? x ? dx ? e (4.28) e dx dx x=1 ; 0 0 while integrating by parts the second and third terms of the integral in Eq.(4.27) yields Z1 ! ! Z1  2 2 d s d s ? x ? d fa +  dfa dx = ? se d x ? +  dx2 dx dx2 dx 0 0 Z Z 1  d2 f a Z Z 2 1 xd s  1 ?  x ds  1 ?  x dfa ? fa e se d x ? fa e d x? d x + se? x 2 dx = 2 dx dx dx! dx 0 0 0 0 x 1 Z Z ? 1 1 d(fae ) ds dx ? f e?  x ds ?  ds  fa e? x dx + a dx dx dx dx  fa e? x 0 0 Z1 !1 Z 1 d(se?  x ) 0  df dfa dx ? d x + se? x a = dx dx dx 0 0 0 ! ! Z Z 1  dfa ds 1    d f d s d s d f a a ? x ? ? e? x dx dx dx ? e dx x=1 ? 0 e dx dx dx + e dx x=1 = 0 !  dfa d s ? e (4.29) dx ? dx x=1 : se? x The nal functional is obtained by combining Eq.s (4.28) and (4.29)  [fa] = Z1 0  e? x ! dfa 2 dx ? e?  ds dx dx ! x=1 : (4.30) The second term of the functional can be omitted since it does not depend on fa (x). This functional can be directly minimized in order to approximate the function f0(x) that satis es the di erential Eq. (4.16). The expansion of the ap- 59 proximate solution is given by uan (x) = x + n X i=1 ci i (x; i) = uan?1 (x) + cn n (x; n) ; (4.31) where the rst term is included in order to satisfy the boundary conditions. By introducing Eq. (4.31) into Eq. (4.30), the augmented optimization variables vector n = ( n; cn)T can be computed using PDS in one step. However, this nonlinear optimization problem can be decomposed into two lower-order problems. By di erentiating  with respect to cn, requiring the derivative to vanish, and abbreviating (x; ) as  for simplicity, the following expression is obtained 0 2 Z1 0  du d e? x n n dx = 0 : dx dx a (4.32) The derivative of Eq. (4.31) with respect to the independent variable x is given by n d dua (x) + c dn : duan(x) = 1 + X ci i = n?1 n dx dx dx dx i=1 (4.33) Substituting Eq. (4.33) into Eq. (4.32) and solving for the coecient cn yields cn = Z1  d d e? x n n dx dx dx 0 ? Z1 0 !?1  ! nX ?1 Z 1  di d  x dn n x ? ? e dx dx ? i=1 ci 0 e dx dx dx : (4.34) Combining Eq.s (4.34), (4.33), and (4.30), a new objective function is obtained  [uan] = Z1 0 e ?x 0 n?1 Z 1  dn dn !?1 @1 + X ci di + e? x dx  i=1 dx 0 dx dx 60 ? Z1 0 ?x e !2 ! ?1 Z 1  d d d dx ? X d c e? dx dx dx dx dx dx : (4.35) 0 =1 n n i x i n n i The parameters  of the n-th basis function are determined by minimizing the functional in Eq. (4.35), while the associated coecient c is readily calculated from Eq. (4.34). Two PDS input parameters control the number of required PDS iterations and function and constraint evaluations; the number of pattern points d for the search directions and the user-selected tolerance tol, which constitutes the convergence criterion of the PDS algorithm. PDS requires d  2K pattern points for unconstrained problems and d  K 2 + K pattern points for constrained problems, where K is the number of optimization variables. Numerical experimentations have shown that d should be chosen to be larger than 2K or K 2 + K , respectively, in order to ensure a detailed search and, as far as this is possible, a global extremum. The number of pattern points for the search directions should be relatively small if the objective function is convex, and large if the objective function is suspected to be nearly concave. Typically, if a small d is chosen, a small tolerance should also be chosen since the search interval is relatively coarse. Similarly, if a large d is selected for the same number of design variables, a larger tolerance should be chosen to ensure computational eciency. The approximate solutions are compared to the exact in Fig. 4.1(a) for = = 5 and Fig. 4.1(b) for = = 20 for the augmented radial bases. The relatively low number of necessary basis functions, in both gures, and the excellent agreement between exact and approximate solutions, even in the presence of a boundary layer characterized by a large gradient, con rm the potential of the method. Most interesting of all, note that the accuracy of the augmented radial bases rivals both n n 61 nite elements on a uniform grid and B1 splines by the proposed method (Fig. 4.3). Note that since the augmented radial bases are of the same form as the solution in Eq. (4.26),  is limited to 0  jj  3. This is to demonstrate the relatively low number of bases needed by the method when compared to a uniform grid and traditional radial bases. 1.2 1.2 approx. exact approx. exact 1 1 0.8 0.8 u(x) u(x) 0.6 0.6 0.4 0.4 0.2 0.2 0 0 0 0.1 0.2 0.3 0.4 0.5 x (a) 0.6 0.7 0.8 0.9 1 −0.2 0 0.1 0.2 0.3 0.4 0.5 x 0.6 0.7 0.8 0.9 1 (b) Figure 4.1 One-dimensional convection-di usion: exact and approximate solutions for (a) = = 5 and (b) = = 20 using 16 and 9 optimal augmented radial basis functions, respectively. Two plots from Fletcher [27] are reproduced in Fig. 4.2 in order to illustrate the large oscillations in the solutions obtained by traditional numerical methods (in this case by the nite di erence method). The problem and the boundary conditions associated with Fig. 4.2 are the same as those discussed in this section with the parameter ratio is = = 20. The symbol T has been chosen by Fletcher to represent the variable for which the symbol u was used in this text, while the symbol u has been used by Fletcher for the parameter introduced here as . 62 Figure 4.3 compares the RMS error convergence rates obtained for = = 5 and = = 20 to the nite element method on a uniform grid, the proposed method with B1 splines and the proposed method using the augmented radial bases. PDS was executed using a tolerance tol of 10?4 and 1000 pattern points d. The RMS error was calculated based on 101 uniformly distributed trial points. Though a large di erence in the RMS values are apparent, a small quantitative di erence in the convergence rates can be observed. The RMS error convergence rate for = = 20, tol = 10?4 , and d = 1000 is plotted in Fig. 4.3(b). It can be concluded that when the solution of the problem has a relatively large gradient in some region of the domain, the proposed method will fully exploit its adaptive nature and perform much better than traditional methods and traditional bases. It can be argued that the adjustment of the di erential operator L by the introduction of the premultiplying function P (x) is equivalent to the application of the Petrov-Galerkin method to the original problem, when the latter is solved by the traditional nite element method. Since the new, equivalent problem is L [f0 (x)] = P (x)h(x) = P (x)L[f0 (x)] (4.36) 0 or, more speci cally,  e? x ! 2 ? ddxf20 +  ddfx0 = 0 ; (4.37) the equation residual, after introducing the approximate solution fa, is  e? x ! 2 d ? dxf2a +  ddfxa = e?  x (R) = R : 0 (4.38) 63 (b) (a) One-dimensional convection-di usion: (a) exact and approximate solutions for di erent grid spacing and cell Reynolds numbers and (b) exact and approximate solutions in the presence of an upwind scheme; = = 20, plots reproduced from reference [26]. Figure 4.2 0 10 −1 10 uniform grid opt. incr. approx. B1−splines opt. incr. approx. RBFs linear −1 10 quadratic −2 10 RMS RMS linear quadratic −2 10 −3 10 −3 10 uniform grid opt. incr. approx. B1−splines opt. incr. approx. RBFs −4 10 0 1 10 10 N (a) 2 10 −4 10 0 10 1 10 N (b) One-dimensional convection-di usion: convergence rates for (a) = = 5 and (b) = = 20. Figure 4.3 2 10 64 Applying the Bubnov-Galerkin method of weighted residuals to the residual is equivalent to applying the Petrov-Galerkin method of weighted residuals to the original residual of Eq. (4.38) since R 0 R hR 0 ; k i = h R; e ?  x k i : (4.39) Moreover, the system of equations obtained by applying the Raleigh-Ritz method to the problem, i.e. the algebraic equations obtained by requiring the derivatives of the variational principle of Eq. (4.30) with respect to the coecients =1 , to vanish, is equivalent to the system of equations obtained k, by the Petrov-Galerkin. It was found that the equivalent Petrov-Galerkin and Raleigh-Ritz methods eliminated the oscillatory behavior of the solution but did not enhance the RMS error convergence rate when utilizing 1-splines on a uniform grid; the convergence rate is slightly worse, as seen in Fig. 4.4. In addition, it can be seen that the convergence rate of the proposed method is better when the coecients are computed one by one at each stage of the algorithm and not when all the coecients are recomputed by solving a system of equations at each step. An additional characterisic of the method is illustrated in Fig. 4.5, using augmented radial bases. Since the method is based on the adaptive improvement of the series expansion, the absolute value of the coecients should decrease with increasing number of expansion terms. The oscillations observed in Fig. 4.5 can be explained by the fact that the coecients do not necessarily have to decrease strictly monotonically, especially when their sign oscillates in the early stage of the proposed algorithm. N c k ;:::;N B 65 4.2.2 Computational Cost Tables 4.1, 4.2, and 4.3 summarize the amount of computational work required by PDS for = = 5 using B1 splines and augmented radial bases and = = 20 using augmented radial bases, respectively. Step 1 involves selection of optimal parameters in (x) of Eq. (4.14). Step 2 involves the selection of the remaining basis function parameters  and  . The number of required iterations and function and constraint evaluations do not uctuate signi cantly. A low number of pattern points associated with a tighter tolerance requires more iterations but fewer function and constraint evaluations than a large number of pattern points associated with a relaxed tolerance. It is interesting to note that in all cases the algorithm requires less work in its initial stage than in all the subsequent steps. l r 66 Amount of computational work required by PDS for the linear convection-di usion problem using B1-splines (= = 5). Table 4.1 Algorithm # of PDS Total # of Total # of step iterations function eval. constraint eval. d = 100 tol = 10?6 1 24 836 2272 8 18 1493 1536 16 15 1307 1365 24 16 1168 1335 d = 1000 tol = 10?4 1 7 916 5344 8 11 5721 6683 16 12 6970 8719 24 10 7401 8475 Amount of computational work required by PDS for the linear convection-di usion problem using augmented radial basis functions (= = 5, d = 1000, and tol = 10?4 ). Table 4.2 Algorithm # of PDS Total # of Total # of step iterations function eval. constraint eval. Step 1 1 7 883 5263 3 9 6776 7643 6 13 4419 7943 9 10 7320 7598 Step 2 1 6 2061 2061 3 5 3620 3620 6 5 1778 1778 9 7 3042 3042 67 Amount of computational work required by PDS for the linear convection-di usion problem using augmented radial basis functions (= = 20, d = 1000, and tol = 10?4 ). Table 4.3 Algorithm # of PDS Total # of Total # of step iterations function eval. constraint eval. Step 1 1 12 2420 10245 3 13 7819 9160 6 13 3883 4719 9 8 4903 5080 Step 2 1 6 2596 2596 3 6 4537 4537 6 7 1760 1760 9 6 3469 3469 0 10 −1 10 RMS linear quadratic −2 10 −3 10 opt. seq. approx. one−by−one opt. seq. approx. system Bubnov−Galerkin Raleigh−Ritz = Petrov−Gal. −4 10 0 10 1 10 N 2 10 One-dimensional convection-di usion: convergence rates comparison considering the Petrov-Galerkin and Raleigh-Ritz methods; = = 20. Figure 4.4 68 0 10 −1 |c| 10 −2 10 −3 10 −4 10 1 2 3 4 5 6 j 7 8 9 10 11 One-dimensional convection-di usion: absolute value of the coecients c for = = 20. Figure 4.5 i 69 Chapter 5 The Euler Equations Many kinds of ow of practical interest in uid dynamics, including compressible ow , can be assumed to be irrotational. When a ow is irrotational, the vorticity of the uid, which is proportional to its angular velocity, is equal to zero. This is the case for supersonic ows around sharp edges or cones, for two-dimensional or axisymmetric nozzle ows, and it can be assumed to be the case for ows around slender sharp-nosed bodies. However, supersonic ow around blunt bodies and viscous ow inside a boundary layer are certainly rotational [1]. Whenever a ow can be assumed to be irrotational, the velocity vector can be de ned as the gradient of the velocity potential, i.e. V = r'. In this manner, the governing equations (continuity and momentum equations) can be combined into a single nonlinear partial di erential equation with ' as the single unknown. Once ' is found, all other ow properties can be calculated. Variational principles for the problem of compressible, inviscid, unsteady ow have been derived for the one- and two-dimensional cases from Prozan [77, 76]. Manwell discusses in reference [54] a variational principle for steady, compressible, inviscid ow with shocks, but does not present numerical results. His approach is similar to that presented earlier by Greenspan and Jain in reference [33], where a variational principle, derived for the problem of steady, irrotational, inviscid, compressible ow, was extremized using nite di erences. In fact, the results of an extensive literature search indicated that Greenspan has been one of few reA ow is assumed to be compressible if the freestream Mach number M1 exceeds the numerical value of 0.3.  70 searchers to work directly with the variational principle and attempt to determine the extremal function by direct optimization [31, 33, 32]. In this regard, the proposed method for optimal sequential approximation will be applied to the problem studied by Greenspan and Jain in reference [33]. 5.1 The Velocity Potential Equation The equations governing steady, two-dimensional, homentropic, irrotational, inviscid, compressible ow past a xed body are given by the continuity and momentum equations, respectively @ (u) @ (v ) + @y = 0 @x @u 1 @p @u + v + =0 u @x @y  @x @v @v 1 @p u + v + =0: @x @y  @y (5.1) (5.2) (5.3) In these equations, u and v are the x- and y- Cartesian velocity components, respectively,  is density, and p is pressure. The velocity vector is de ned as the gradient of the velocity potential ' V = (u; v) = T '= r @' @' ; @x @y !T (5.4) and has magnitude V 2 = V 22= u2 + v2. The set of equations is completed by the equation of state for an ideal gas jj p = a ; jj (5.5) 71 where a is a constant and is the ratio of speci c heats. Equations (5.1) - (5.3) can be combined to yield the potential velocity equation [1, 33] 0 @c2 ? @' @x !21 2 A @ '2 ? 2 @' @' @x @ 2' @x @y @x@y 0 + @c2 ? @' @y !21 2 A @ '2 = 0 ; @y (5.6) where c2 = dp=d is the square of the local speed of sound. According to the equation of state, for an ideal gas c2 =a  ?1 : (5.7) The associated boundary conditions are @' @ = 0 on the body surface, and r' = (V1 ; 0)T = constant far from the body, (5.8) (5.9) where  is the direction normal to the body surface and V1 is the freestream velocity, which is recovered in a distance far from the body. 5.2 The Variational Integral Bateman proposed in reference [4] two variational principles whose variations yield the governing equations for steady, irrotational, isentropic, compressible, inviscid ow. The rst is implicitly expressed in terms of the velocity potential and has to be maximized if the stationary ow is subsonic throughout the domain of interest. 72 The second is de ned with help of the stream function , where @ @ V = (u; v )T = ; @y @x !T (5.10) ; and has to be minimized [53]. In the work presented in this report, the rst variational principle has been considered. Since pressure is a function of density, and density is in turn a function of the velocity magnitude, the variational integral is given by = Z pd = Z f (u2 + v 2)d = Z f ((@'=@x)2 + (@'=@y )2)d (5.11) and is associated with the following Euler-Lagrange equation @ 2f @ 2 ' @ 2f @ 2' @ 2f @ 2' + 2 + =0 : @u2 @x2 @u@v @x@y @v 2 @y 2 (5.12) By comparing Eq.s (5.6) and (5.12) it can be concluded that ! @ 2f @' 2 2 = c ? @x ; @u2 @ 2f @' @' = ? ; and @u@v @x @y ! @' 2 @ 2f 2 = c ? @y : @v 2 (5.13) (5.14) (5.15) The functional of Eq. (5.11) is divergent for in nite domains. This drawback has been remedied by Lush and Cherry by means of introducing appropriate terms into Eq. (5.11) that do not a ect the associated Euler-Lagrange equation [53]. The modi ed variational integral is convergent and has to be maximized. It is de ned 73 as  ['] = ! Z I @ @ (V1 ) p ? p1 + V1 1 d @x ! @' @ 2 ' ? V11  @ ds ; @ + (5.16) where 1 denotes freestream conditions,  is some unknown auxiliary function, and @ represents the closed body curve. The potential ' is given by ' = V1 x + V1  : (5.17) According to the boundary condition,  has to vanish far from the body. The body used by Greenspan and Jain in reference [33] was a unit cylinder, and in this regard, polar coordinates were utilized to derive the nal functional  [] = p1 Z 2 Z 1  =0 r =1 (F ? 1 + 2 H )rdrd ; M1 (5.18) where   ?1  2  M1 F = 1? 2 0 @2 cos() @ ? 2 sin() @ + @r r @ @ @r !2 + r12 @ @ !211 AA ?1 (5.19) and  1  @ sin()  1  @ cos() ? 1 + r2 @ r : H = 1? 2 r @r (5.20) 74 M1 is the freestream Mach number given by M1 = Vc 1 1 (5.21) ; where c1 is the freestream speed of sound. Finally, the boundary conditions are and @ @r = ? cos() at r = 1 =0 far from the body : (5.22) (5.23) 5.3 The Computational Domain The variational principle of Eq. (5.18) is de ned over an in nite domain. However, it is assumed that the radius spans from r = 1 (body surface) to r = 20 (far eld). This assumption is more than justi ed considering the fact that typical numerical grids in airfoil ow analysis, extend six to seven times the length of the bodies. In addition, the ow is completely symmetrical about the lines  = 0,  = 180,  = 90 , and  = 270 . In this regard, the computational domain can be restricted to the range 90    180 . With respect to the above assumptions, the variational integral to be numerically maximized is given by Z 2 = 1 ==2 Z r2 =20 r1 =1 (F ? 1 + 2 M1 H )rdrd : (5.24) Figure 5.1 depicts the physical (x-y-plane) and computational (r--plane) domains. The points A,B,C, and D with x-y coordinates (-20,0), (-1,0), (0,1), and (0,20), 75 D 20 y 15 10 5 0 C A B −20 20 −15 −10 x −5 0 D* A* C* c B* a r 15 10 5 0 80 b 100 120 140 θ 160 180 200 Compressible inviscid ow: (top) the physical (x-y-plane) and (bottom) computational (r--plane) domains. Figure 5.1 76 respectively, are mapped on the points A, B, C, and D, with r- coordinates (20,180 ), (1,180 ), (1,90), and (20,90), respectively. The presence of homogeneous Dirichlet boundary conditions on the lines r = 20 and  = 90 facilitates the use of basis functions that vanish at the domain boundaries. In this regard, and due to the presence of Neumann boundary conditions on the lines r = 1 and  = 180 , the computational domain has been extended, as shown in Fig. 5.1. The extended computational domain is described by the points a, b, c, and D in the r- plane. 5.4 Boundary Conditions The boundary conditions have to be adjusted for the modi ed computational domain. Speci cally, the appropriate boundary conditions are r  = ? cos() at r = 1 ; = 0 at r = 20 and  = 0 at  = 180 ;  (5.25) = 90 ; and (5.26) (5.27) where the subscripts denote partial di erentiation. It must now be determined whether the Neumann boundary conditions are naturally satis ed by the variational principle or if additional terms and/or modi cations of the integral are necessary. Assuming that the Dirichlet boundary conditions are satis ed by the trial functions, the variation of  [] with respect to a variation in , de ned as  =  +  (5.28) 77 with  =  +  and  =  +  r r r    (5.29) ; is given by @ @ ! = =0 Z Z @ (F r) @ ! =0 drd + Z Z @ ( M 2Hr) @ ! =0 drd (5.30) and must vanish, for  to be extremized. In this regard, and by noting that = F (r; ;  ;  ) ; F 6= F () H = H (r; ;  ;  ) ; H 6= H () r 6= r() ; and  6= () ; F r  r  it can be found that and with and @ (F r) @ ! =0 @ ( M 2Hr) @ @ r @ ! =0 ! @ (F r) @ r @ r @ = ! = =0 r ; = ! =0 + @ (F r) @  @  @ @ ( M 2Hr) @ r @ r @ @  @ ! =0 =  ! =0 + ; 1 @ (F r) = ? G ?1 rM 2 (cos  + r )  P @ r =0 !   1 @ (F r) 2 1 ? 1 = ?G M r  ? sin   Q @  =0 !  1 @ ( M 2Hr) 2 = M r ? r cos   R @ r =0 ! =0 @ ( M 2Hr) @  @  @ (5.31) ! =0 (5.32) (5.33) (5.34) (5.35) (5.36) 78 @ ( M 2 Hr) @  ! =0 =? M 2  1 1 + r2 sin   S ;  (5.37) where   1 1 1 2 2 2 G = 1 ? ( ? 1)M 2 cos  ? 2 sin  +  + 2  2 r r r  r  (5.38) : Substituting Eq.s (5.31)-(5.37) into Eq. (5.30) yields = Z Z (P  + Q )drd + r Z Z  (R + S )drd : r (5.39)  Integrating by parts the rst terms of each integral with respect to r and the second terms of each integral with respect to  yields Z Z @P Z Z @Q Z Z r2 =?  drd ?  drd + (P  )r1 d + (Q )21 dr ? @r @ Z Z @S Z Z Z Z @R @r  drd ? @  drd + (R) 21 d + (S) 21 dr : r r   (5.40) But @R @r = ? @S @ (5.41) and  is zero on r = r2 and  = 1 (due to the homogeneous Dirichlet boundary conditions on these boundaries) and so the functional variation reads as =? Z Z Z @P @r + ! @Q  drd ? @ Z Z Z (P ) 1 d + r (Q) 2 dr ? (R) 1 d + (S) 2 dr :  r  (5.42) 79 In addition, (R)r1 = 0 and (S )2 = 0; the nal functional variation is given therefore by =? Z Z @P @r + ! @Q  drd ? @ Z Z (P )r1 d + (Q)2 dr : (5.43) In order for  to vanish for arbitrary , all the integrands must vanish, i.e. @P @r + @Q =0 ; @ (5.44) (P )r1 = 0 ; and (5.45) (Q)2 = 0 : (5.46) Equation (5.44) is the associated Euler-Lagrange equation of the functional, and Eq.s (5.45) and (5.46) are the Neumann boundary conditions on r = r1 and  = 2, respectively. Speci cally, from Eq. (5.45) (?G ?1 rM 2 (cos  + r ))r1 1 = 0 =) (G ?1 )r1 = 0 or (cos  + r)r1 = 0 (5.47) 1 and from Eq. (5.46)  ?G or Since G 1 ?1 M 2 1  r 1 ?1  2 1 r  ? sin  =0:  2 = 0 =) (G ?1 )2 1 =0 (5.48) represents the density of the uid, and can therefore not be equal to 80 zero, the Neumann boundary conditions are given by r = ? cos  on r = r1 (5.49) and  = 0 on  = 2 ; (5.50) which are indeed the (both homogeneous and non-homogeneous) Neumann boundary conditions of the problem and are therefore naturally satis ed by the functional (5.18). 5.5 Numerical Procedure and Results The proposed algorithm for optimal incremental approximation has been utilized to numerically approximate the extremal function (r; ) that maximizes the variational principle de ned in Section 5.2 and adjusted to the computational domain discussed in Sections 5.3 and 5.4. The function (r; ) is approximated by the series expansion [58] an (r; ) = X c  (r;  ) (;  ) ; n i=1 i i r i i  i (5.51) where the basis functions i are, as previously introduced, augmented one-dimensional radial bases with parameters ir = (rM , rl, rr, rl , rr )Ti and i = (M , l, r, l , r )Ti . The ci are the corresponding expansion coecients. The tensor product i(r; ir)i(; i ) can be produced by a feedforward network of two hidden layers of radial bases [60]. The function (r; ) appears only implicitly, in the form of its partial derivatives with respect to r and , in the functional. They are accordingly approximated 81 by r    @an @r @an @ X c @ (r;  )  (;  ) @r  X = c  (r;  ) @  (; ) : = n i=1 n i=1 i r i i i i r i i  i (5.52) i  i (5.53) @ These approximations are introduced into the variational integral to be maximized at each step of the algorithm, and PDS is employed to determine the optimal parameters and expansion coecient. Due to the high complexity and nonlinearity of the functional, the optimization problem cannot be decomposed into lower-dimensional problems; all seven optimization variables, represented by the vector i = (rM , rl, rr, M , l, r, rl , rr , l , r , c)Ti , are computed in one step by solving an eleven-dimensional highly nonlinear optimization problem using PDS. To enhance computational eciency in the absence of simpli cation and decomposition possibilities, the scaling option of PDS was activated (Appendix A.5.2). The components of the initial guess vector for each step of the algorithm were the following: rM was positioned close to the cylinder surface, while rl and rr were chosen so that (r) covers the computational domain in the r-direction; M was positioned in the middle of the domain (with respect to the -direction), while l and r were also chosen so that () covers the computational domain in the -direction. Finally, the initial guess for rl ; rr ; l ; r and the coecient c were set equal to zero so that the functional value achieved in the previous step of the algorithm will be the initial value in the next step. In this manner, it is guaranteed that the functional value will be strictly monotonically improved with each iteration of the proposed algorithm. 82 Numerical results are presented for transonic ow of air about a unit cylinder and compared to results presented by Greenspan and Jain in reference [33]. The following ow properties were estimated after solving for the function . The nondimensionalized speed of the uid  V 2 V1 = 1 + 2 cos()r ? 2 sin(r )  + (r )2 + r12 ( )2 ; (5.54) the nondimensionalized local speed of sound  c 2 V1  2 = M12 + 21 ( ? 1) 1 ? VV 1 1 ! (5.55) ; the nondimensionalized pressure p p1  V 2 1 2 = 1 + 2 ( ? 1)M1 1 ? V 1 !! ?1 ; (5.56) the pressure coecient Cp = p p1 1 2 ?1 2 M1 (5.57) ; and the nondimensionalized density  1 = p p1 !1 ; (5.58) where = 1:405. Greenspan and Jain used a Newton-like algorithm in their numerical test cases with  = 9 and r = 0:1, 0.15, and 0.2, i.e. they used a nite di erence grid of 11  191 = 2101, 11  127 = 1397, and 11  96 = 1056 83 nodes, respectively. As will be shown on the illustrations provided below, the results reported by Greenspan and Jain are quite sensitive with respect to the value of r. Our numerical examples are presented for the freestream Mach numbers of 0.4 and 0.43. The results obtained by the proposed method are compared to those presented in reference [33]. The latter include results obtained previously by means of other numerical and analytical approximation methods and reported by Lush and Cherry [53] and Imai [42] for M1 = 0:4 and M1 = 0:43. Note that for illustrative purposes all variables requiring the di erentiation of  are curve t by a cubic polynomial. 5.5.1 Numerical Example: M1 = 0:4 The numerical example of M1 = 0:4 presented in this report simulates near sonic conditions. However, the ow remains subsonic throughout the domain of interest. As illustrated in Fig.s 5.2 and 5.3, the extremal function (r; ) was satisfactorily approximated after 15 two-dimensional augmented radial basis functions were determined and used in the series expansion. The values of the function  at di erent radii are compared to results presented in reference [33], which include numerical values obtained by 1) Imai and 2) Greenspan and Jain using a) r = 0:1 and b) r = 0:2. The agreement is good for r = 1. For higher radii the results obtained by the proposed algorithm deviate, although not signi cantly, from those reported by Greenspan and Jain. It should be emphasized, however, that these values have been reproduced from scanned plots due to the absence of tabulated data in reference [33]. Consequently, error has been introduced into the presented values. Moreover, the numerical results presented by Greenspan and Jain are over three decades old. Therefore, they merely constitute a comparison basis and not an accurate benchmark. 84 0 −0.2 χ(r,θ) −0.4 −0.6 −0.8 −1 −1.2 −1.4 80 20 100 15 120 10 140 5 160 180 0 r θ Compressible inviscid ow: the approximate velocity potential auxiliary function (r; ) for M1 = 0:4 using 15 optimal basis functions. Figure 5.2 r=1 r=1.6 0 0 opt. seq. approx. Imai Greenspan dr=0.1 Greenspan dr=0.2 −0.2 opt. seq. approx. Imai Greenspan dr=0.1 Greenspan dr=0.2 −0.1 −0.6 −0.3 χ −0.2 χ −0.4 −0.8 −0.4 −1 −0.5 −1.2 −0.6 −1.4 90 100 110 120 130 θ 140 150 160 170 −0.7 90 180 100 110 120 130 r=2.2 θ 140 150 160 170 180 r=2.8 0 0 opt. seq. approx. Imai Greenspan dr=0.1 Greenspan dr=0.2 −0.1 opt. seq. approx. Imai Greenspan dr=0.1 Greenspan dr=0.2 −0.05 −0.1 −0.2 −0.15 −0.3 χ χ −0.2 −0.25 −0.4 −0.3 −0.5 −0.35 −0.6 −0.4 −0.7 90 100 110 120 130 θ 140 150 160 170 180 −0.45 90 100 110 120 130 θ 140 150 160 170 180 Figure 5.3 Compressible inviscid ow: the approximate velocity potential auxiliary function (r; ) at di erent radii (on and near the cylinder surface) for M1 = 0:4 using 15 optimal basis functions. 85 The nondimensionalized uid velocity on the surface of the cylinder (r = 1) is plotted in Fig. 5.4 and compared to the results of Greenspan and Jain, Lush and Cherry, and Imai. Figure 5.4 also depicts the nondimensionalized local speed of sound on the cylinder surface. In this, the most sensitive test of the numerical approximation, it can be seen that the uid velocity barely exceeds the sonic speed at  = 90, though the actual ow should be just below sonic speed; the approximation is considered to be satisfactory overall. 3 (1) 2.5 2 (2) 1.5 1 opt. seq. approx. Greenspan dr=0.1 Lush and Cherry Imai 0.5 0 90 100 110 120 130 θ 140 150 160 170 180 Compressible inviscid ow: nondimensionalized (curve 1) uid velocity and (curve 2) sonic speed on the cylinder surface for M1 = 0:4. Figure 5.4 The pressure coecient distribution Cp and the nondimensionalized density on the cylinder surface are drawn in Fig. 5.5. Figure 5.6 presents the local Mach number distribution over the entire ow eld and some representative Mach-isolines (lines of equal Mach number) for the region close to the cylinder surface. 86 2 1 Cp 0 −1 −2 opt. seq. approx. Greenspan dr=0.1 Lush and Cherry Imai −3 −4 90 100 110 120 130 θ 140 150 160 170 180 1.15 1.1 1.05 1 ρ 0.95 0.9 0.85 0.8 opt. seq. approx. Greenspan dr=0.1 Lush and Cherry Imai 0.75 0.7 0.65 90 100 110 120 130 θ 140 150 160 170 180 Figure 5.5 Compressible inviscid ow: (top) pressure coecient distribution and (bottom) nondimensionalized density on the cylinder surface for 1 = 0 4. M : 87 1.4 1.2 1 Ma 0.8 0.6 0.4 0.2 0 20 15 180 160 10 140 120 5 100 0 r θ 6 5 80 6 4 7 7 3 7 9 6 10 11 1112 134 1 9 2 6 7 2 1 3 44 0 10 5 3 1 132 9 8 5 1 y 15 14 13 12 11 10 9 8 7 6 5 4 3 2 1 8 8 2 1143 2 1 1 0 1 01 1 7 7 8 9 11 4 -1 5 11 9 7 6 -25 5 8 6 -4 7 6 -3 -5 0.965213 0.901587 0.837961 0.774335 0.710709 0.647082 0.583456 0.51983 0.456204 0.392578 0.328952 0.265326 0.2017 0.138074 0.0744478 -4 -2 0 2 4 x Compressible inviscid ow, 1 = 0 4: (top) local Mach number distribution over the entire ow eld and (bottom) Mach-isolines in a region close to the cylinder surface. Figure 5.6 M : 88 5.5.2 Numerical Example: M1 = 0:43 Transonic conditions are successfully simulated when the freestream Mach number takes the value of 0.43. Speci cally, a supersonic bubble appears in the vicinity of  = 90 on and near the cylinder surface. The numerically approximated function (r; ) is depicted in Fig.s 5.7 and 5.8 using 15 optimal basis functions. It can be concluded by comparison with Fig.s 0 −0.2 χ(r,θ) −0.4 −0.6 −0.8 −1 −1.2 −1.4 80 20 100 15 120 10 140 5 160 180 0 r θ Figure 5.7 Compressible inviscid ow: the approximate velocity potential auxiliary function (r; ) for M1 = 0:43 using 15 optimal basis functions. 5.2 and 5.3 that the function reaches higher absolute values for the present test case. Once again, the values of the function  at di erent radii are compared to results presented in reference [33], which include numerical values obtained by 1) Imai and 2) Greenspan and Jain using a) r = 0:1 and b) r = 0:15. The agreement seems to be better for this test case. The nondimensionalized uid velocity and local speed of sound on the surface of the cylinder, presented in Fig. 5.9, illustrate the appearance of the supersonic 89 r=1 r=1.6 0 0 opt. seq. approx. Imai Greenspan dr=0.1 Greenspan dr=0.15 −0.2 opt. seq. approx. Imai Greenspan dr=0.1 Greenspan dr=0.15 −0.1 −0.2 −0.4 −0.3 χ χ −0.6 −0.4 −0.8 −0.5 −1 −0.6 −1.2 −1.4 90 −0.7 100 110 120 130 θ 140 150 160 170 −0.8 90 180 100 110 120 130 r=2.2 θ 140 150 160 170 180 r=2.8 0 0 opt. seq. approx. Imai Greenspan dr=0.1 Greenspan dr=0.15 −0.1 opt. seq. approx. Imai Greenspan dr=0.1 Greenspan dr=0.15 −0.05 −0.1 −0.2 −0.15 −0.3 χ χ −0.2 −0.25 −0.4 −0.3 −0.5 −0.35 −0.6 −0.4 −0.7 90 100 110 120 130 θ 140 150 160 170 180 −0.45 90 100 110 120 130 θ 140 150 160 170 180 Figure 5.8 Compressible inviscid ow: the approximate velocity potential auxiliary function (r; ) at di erent radii (on and near the cylinder surface) for M1 = 0:43 using 15 optimal basis functions. bubble. In the vicinity of  = 90 , the uid velocity exceeds the local speed of sound and the ow becomes supersonic. The pressure coecients and the nondimensionalized density, plotted in Fig. 5.10, are in satisfactory agreement with numerical results of Greenspan and Jain. Figure 5.11 presents the local Mach number distribution over the ow eld and representative Mach-isolines for the region close to the cylinder surface. The presence of supersonic ow regions is clear. 5.5.3 Computational Work The amount of computational work that PDS required at each step of the proposed algorithm is presented in Tables 5.1 and 5.2. Computational work is recorded in terms of the numbers of required PDS iterations and function and constraint eval- 90 3 2.5 (1) 2 (2) 1.5 1 opt. seq. approx. Greenspan dr=0.1 Greenspan dr=0.15 0.5 0 90 100 110 120 130 θ 140 150 160 170 180 Compressible inviscid ow: nondimensionalized (curve 1) uid velocity and (curve 2) sonic speed on the cylinder surface for 1 = 0 43. Figure 5.9 M : 1.2 1 1.1 0 1 −1 0.9 ρ Cp 2 −2 0.8 −3 0.7 opt. seq. approx. Greenspan dr=0.1 Greenspan dr=0.15 −4 −5 90 100 110 120 130 θ 140 150 160 170 180 opt. seq. approx. Greenspan dr=0.1 Greenspan dr=0.15 0.6 0.5 90 100 110 120 130 θ 140 150 160 170 180 (b) (a) Compressible inviscid ow: (a) pressure coecient distribution and (b) nondimensionalized density on the cylinder surface for 1 = 0 43. Figure 5.10 M : 91 1.4 1.2 1 Ma 0.8 0.6 0.4 0.2 0 20 15 180 160 10 140 120 5 100 0 r 80 θ 5 7 4 15 14 13 12 11 10 9 8 7 6 5 4 3 2 1 77 6 5 6 9 8 3 10 7 2 5 4 2 1 0 2 8 0 4 4 3 2 4 6 6 7 -2 14 13 1 2 1 1 99 -1 9 10 6 4 5 8 7 -3 12 13 14 3 y 9 1 1 21 1 8 5 1.08471 1.01318 0.941636 0.870097 0.798558 0.727019 0.65548 0.58394 0.512401 0.440862 0.369323 0.297784 0.226244 0.154705 0.0831659 -4 7 6 6 -5 -4 -2 0 2 4 x Compressible inviscid ow, 1 = 0 43: (top) local Mach number distribution over the entire ow eld and (bottom) Mach-isolines in a region close to the cylinder surface. Figure 5.11 M : 92 uations. The strictly monotonically increasing functional value is included. As can be concluded by inspection of Tables 5.1 and 5.2, the amount of computational work does not exhibit high uctuations as the algorithm advances. Even when the maximum numerical value of the functional is asymptotically approached, with increasing number of optimal bases, the amount of work necessary for the improvement of the variational integral value is of the same order as for the previous iterations. 93 Amount of computational work required by PDS for the Euler Equation problem using augmented radial basis functions for = 1120 and = 10?4 ( = 0 4). Table 5.1 d tol M : Algorithm # of PDS Total # of Total # of Functional step iterations function eval. constraint eval. value Step 1 1 21 6318 17198 0.0535 5 24 23824 24107 0.0960 13 25 22305 23105 0.0974 22 17 13823 16253 0.0975 Step 2 1 17 9032 9032 0.0731 5 75 76254 76254 0.0960 13 96 56877 56877 0.0974 22 15 4006 4006 0.0975 Amount of computational work required by PDS for the Euler Equation problem using augmented radial basis functions for = 1120 and = 10?4 ( = 0 43). Table 5.2 d tol M : Algorithm # of PDS Total # of Total # of Functional step iterations function eval. constraint eval. value Step 1 1 17 5341 14303 0.0694 6 113 125333 126196 0.1097 12 20 14432 16834 0.1155 19 27 20664 28026 0.1156 Step 2 1 10 1383 1383 0.0924 6 16 3903 3903 0.1097 12 21 5884 5884 0.1155 19 5 1208 1208 0.1156 94 Chapter 6 Conclusions and Future Work 6.1 Conclusions 6.1.1 Objectives The objective of the work presented in this report is not only to model the nonlinear compressible Euler equations by feedforward arti cial neural networks but to develop a computationally ecient method which requires minimal user interaction, is general enough to use any of the popular ANN transfer functions that can form compact bases, can be applied to a complex computational domain, and approaches the accuracy of conventional computational mechanics techniques. The bene ts of accomplishing this task, besides those previously given, will be a relatively painless feedforward arti cial neural network construction technique that is understandable and usable by engineers and address validation and veri cation concerns voiced by those applying ANN technology. 6.1.2 Approach To ful ll our objectives, iterative approximation concepts originating in arti cial neural networks are combined with error minimization ideas from function approximation theory to formulate a method; this method involves the incremental determination of the basis functions and the associated coecients used in the series expansion for the representation of the solution. The proposed method requires the solution of a nonlinear optimization problem at each step. Variational principles, associated with the given di erential equations and the boundary con- 95 ditions, are utilized to de ne the objective function(al). The major advantage in the use of variational principles is that the latter succinctly and concisely represent governing equations and associated boundary conditions in a single variational integral which is then extremized. In this manner, the well-posed formulation of the optimization problem is guaranteed. Computational eciency is achieved by utilizing PDS for the solution of the optimization problem and using B1-splines and augmented radial bases. The proposed method is adaptive in nature although a grid is neither built nor adapted in the traditional sense. It can be classi ed as an h-r adaptive technique when using the B1-splines and an h-r-p adaptive technique when using augmented radial bases. The computational overhead of the a-posteriori error estimation and adaptive process, shared by traditional adaptive and multigrid techniques, is avoided; the \grid", de ned by the location of the optimally selected basis functions, evolves with the solution. Moreover, additional disadvantages of traditional adaptive unstructured grid techniques are not present in the proposed method; complicated data structures are not required since systems of equations are neither assembled nor solved, and numerical stability issues are limited to the nonlinear optimization problem. If the optimization problem can be decomposed (e.g. in linear problems), the anity of the proposed method to that of weighted residuals (in particular to the Galerkin method), becomes apparent. The proposed method is iterative in nature and may be viewed as a nonstationary version of the Gauss-Seidel method for the solution of systems of algebraic equations. 96 6.1.3 Numerical Examples It is shown that the proposed method can be successfully applied to a one-dimensional, linear nonself-adjoint, and a nonlinear di erential equation associated with Dirichlet and Neumann types of boundary conditions. Emphasis is given to a uid dynamics application described by the compressible Euler equations. Accurate numerical results and satisfactory convergence rates are reported. Speci cally, the accuracy level of the presented numerical results rivals traditional nite elements on a uniform grid, considering the relatively low number and simplicity of the optimal basis functions used in the solution representation. Moreover, the proposed method does not face the diculties that traditional methods do in some applications. For example, the oscillations observed in the nite di erence and nite element solutions for the linear convection-di usion problem presented in Section 4.2.1. 6.1.4 Optimization and Bases Computational cost is reduced and eciency is enhanced by solving low-dimensional optimization problems using the parallel direct search technique. Moreover, the optimization problem may be further decomposed in linear applications. Parameter studies demonstrate that the computational work required by PDS for the solution of the nonlinear optimization problems, does not increase as the algorithm advances. The numbers of required PDS iterations and function and constraint evaluations do not uctuate dramatically. Readily parameterized conventional and augmented radial bases, and their tensor products, are used to ensure simplicity and computational eciency. However, a broad class of functions (e.g. hyperbolic tangents) can be utilized if they are combined to form compact bases, though added computational e ort may be required to satisfy the boundary conditions with each new basis. 97 6.2 6.2.1 Future Work Generalization of the Proposed Method The proposed method is based on the utilization of natural and contrived variational principles. Such variational formulations exist for many problems of practical interest in engineering and have been studied extensively. However, there still exist problems for which variational principles have not yet been de ned. It is demonstrated how alternative formulations of optimization problems can be adopted by the proposed method for the determination of the optimal basis functions and the associated coecients. 6.2.2 Utilizing the Method of Weighted Residuals The method of weighted residuals could act as the necessary alternative. For example, the minimization of the square of the residual of a di erential equation corresponds to the well-known least-squares weighted residual method; if the objective functional h i is di erentiated with respect to the coecients k , = 1 , of a series expansion, a system of algebraic equations is formed for the solution of the unknown vector . The functional h i could be used as the objective function to be minimized in the proposed method of sequential approximation. In this case, both the coecients k and the parameters of the basis functions k of the series expansion constitute the optimization variables; the proposed method is similar to the least-squares weighted residuals method. However, the disadvantages include: 1) the necessity of developing techniques for the accommodation of the boundary conditions, 2) the possible solution of a multi-objective optimization problem in the event that a given problem is governed by more than one di erential R; R c N c R; R c k ;:::;N 98 equations, and 3) solution uniqueness problems, especially in the case of nonlinear di erential equations. An alternative to the least-squares method that is within the class of the method of weighted residuals, can be deduced by noting that the numerator of the last term of Eq. (4.8), which is to be maximized, can be rewritten as (h ?1 ; n (x)i) l en 2 = ( h 0( ) ? l f ( h 0( ) n( )i ? h l f x ; x a x x ; x x a 2 a ?1 (x)]; n (x)i) L un 2 2 = = ?1 (x)[; n (x)i) L un (h ( ) n ( )i ? h [ g x ; ?1 (x); n (x)i) ?1 (x); n (x)i) l un (h [ 0( )] n( )i ? h [ L f a un = 2 =h ?1 ; n (x)i Rn 2 (6.1) ; which is equivalent to minimizing ?h n?1 n( )i2 (6.2) where the equation residual n?1 is equal to ?( [ an?1( )] ? ( )). From a geometric perspective, Eq. (6.1) indicates that by maximizing h n?1 ni2, the -th basis function will be selected such that n is as parallel as possible to the equation residual. For the sake of generalization, the objective function h n?1 ni2 will be used, where n is a function of n. The selection of the associated coecient n according to another geometric consideration follows; the residual n is forced to be as orthogonal as possible to 2 n . In this regard, the objective function to be minimized is h n n i , with n being the optimization variable. The equivalence of this technique to the PetrovGalerkin method of weighted residuals is apparent. In fact, for linear problems the coecient can be readily calculated by the Galerkin relation h n ni = 0 rather than minimizing h n ni2. R ; x ; R L u x g x R ; R i ; c R R ; R ; R ; c 99 6.2.3 Increasing Computational Eciency For practical applications we can take the preceding derivations and minimize both ? h n( ) ( n ? x ; R ?1 )i Rn and h n ( ) 2 x ; Rn i2 These equations have the advantage of reducing the required number of integrations, per PDS search, of the functional that contains most of the nonlinear optimization parameters. This translates into reduced CPU time and required memory. However, the parameter n is present in both of the equations, unlike Eq. (6.1). Also, with this method of weighted residuals approach the implementation of the boundary conditions may not be straightforward, especially if they are of the Neumann type, and this alternative approach may face solution uniqueness problems. These are relatively minor disadvantages that are present in most applications of conventional computational mechanics. The use of PDS is not required by the proposed method and is only utilized for convenience since it is a derivative-free nonlinear search technique that can be implemented on parallel and distributed computers. and is not required by the proposed method. With additional e ort from the user, better convergence can found with conventional calculus-based search techniques, such as steepest descent and the conjugate gradient method. For complex computational domains, PDS can be run on a serial machine to provide an initial guess to be used by the conventional calculus-based search technique. c 100 Complex computational domains can be accommodated by an equality constraint through the use of a crude unstructured grid. For example, (x) = P 1 ? N X i=1 i(x) !2 where i are the bases of the crude unstructured grid and P is a large positive penalty function. The function  is either 0, if the geometric parameters are within the computational domain , or  = P . The function  can be added to the objective function  during optimization. If the user has a desire to increase the resolution of the solution approximation within some subdomain D of the problem domain , a straightforward penalty function approach can be used within the context of the optimal incremental approximation method. For example, we can write 2 2  = h n (x); ?(x)Rni ? h n (x); ?(x) (Rn ? Rn?1)i 2 h (x); n(x)i 8 n > < m > 0 if x 2 D m where ?(x) = 10 > : m = 0 otherwise. 6.2.4 Compact Bases As mentioned previously, traditional transfer functions can be used by the proposed method with they are combined to form compact bases [60]. For example, the technique used to augment radial bases can be used with the hyperbolic tan- 101 gent,  (x)) A(x) = tanh( tanh() ; ( (x) ? 1)) B (x) = tanh(tanh( +1 ) (6.3) where  is the same as that de ned in Eq. (4.14). The bases A (x) and B (x) are similar to the augmented radial basis function for positive and negative values of , respectively. Note, that like the augmented radial basis, both A(x) = (x) and B (x) = (x) for  = 0. Future research work will be done on the previously mentioned items and will also focus on the extension of the proposed method to three-dimensional and/or unsteady problems. 102 Appendix A A.1 Norms, Inner Products, and Orthogonality In order to determine the \best approximation" of a (solution) function, it is necessary to de ne a metric which measures the distance between the exact and the approximate [9]. Any linear space of functions (or vectors) Q, i.e. a set S of elements obeying the addition and multiplication operations according to the usual rules of arithmetic, is called normed if a real-valued function exists and has the following properties (a) jj f jj > 0; if f 6= 0 (positivity) (b) jj f jj=j  j jj f jj (homogeneity) (c) jj f + g jjjj f jj + jj g jj (triangle inequality), where f , g 2 Q and  2 R. The metric for two functions f and g of this space is then given by jj f ? g jj. Since a large number of di erent norms can be de ned, it is obvious that the concepts of best approximation and convergence depend on the norm choice. For example, the de nition of the norm jj f (x) jj= amax j f (x) j xb (A.1) for a function f (x) de ned on an interval [a; b] enables the discussion of uniform convergence. 103 However, it is often useful to associate the norm with an operation de ned as the inner product. The inner product between two functions f and g is denoted by hf; gi and has the following properties (b) hf; f i > 0; if f 6= 0 hf; gi = hg; f i (symmetry) (c) hf; g + hi = hf; gi +  hf; hi (linearity), (a) (positivity) where ,  2 R. Then, the norm of the inner-product space is automatically dened as q jj f jj= + hf; f i (A.2) and the following properties hold (b) j hf; gi jjj f jj jj g jj jj f + g jjjj f jj + jj g jj (c) jj f + g jj2 + jj f ? g jj2= 2(jj f jj2 + jj g jj2) (parallelogram law) (d) hf; gi = 0 )jj f + g jj2=jj f jj2 + jj g jj2 (a) (Cauchy-Schwarz inequality) (triangle inequality) (Pythagorean law). Assuming x can be multidimensional, if f (x) is continuous on some domain  , where  = S @ with @ being the boundary of , and square-integrable, R i.e f (x)2dx < 1, the Hilbert space L2( ) is de ned, and the associated inner product and L2-norm are given by Z hf; f i = (f (x))2dx =jj f jj2L 2 : (A.3) 104 This is a quite popular inner product de nition, and is used throughout this report. Two functions f and g are said to be orthogonal if hf; gi = 0. In terms of definition (A.3), two functions f (x) and g(x) are said to be orthogonal with respect to some function w(x) on some domain  if hf; giw = Z f (x)g (x)w(x)dx = 0: (A.4) If an orthogonal set of functions has the additional property of jj fi jj= 1, 8i, it is said to be orthonormal. Therefore, the functions of the set share the cardinality property 8 >< 1 ; if i = j hfi; fj i = ij = >: 0 ; if i = 6 j; (A.5) where ij is the Kronecker delta. Finally note that a complete normed linear space is called a Banach space, while an inner product space that is complete in the associated norm is called a Hilbert space. A space S is said to be complete if every Cauchy sequence fxn : n = 1; 2;   g with xi 2 S converges to some x 2 S . A sequence is Cauchy if limn;m!1 jj xn ? xm jj= 0. Hilbert spaces are widely used in computational mechanics, especially in nite element methods [45]. A Hilbert space H k ( ) consists of functions that are, together with their k-th derivatives, square-integrable on . The associated inner product is then de ned as Z hf; f iH k ( ) = ((f (x))2 + (f (x))2 +    + (f k (x))2)dx =jj f jj2H k( 0 ); (A.6) 105 where k ( ) is the -th derivative of ( ). A Hilbert space 0k ( ) consists of functions that are, together with their -th derivatives, square-integrable on and vanish on its boundary . Hilbert spaces of the latter form are very useful for problems associated with homogeneous Dirichlet boundary conditions. f x k f x H k @ A.2 B-splines and Gaussian Functions The widely used 1-splines were introduced in Section 3.4.2, Eq. (3.9). Figure A.1(a) illustrates a 1-spline with center M = 1 0, width to the left  l = 0 5, and width to the right  r = 0 5. Another popular -splines family consists of 3-splines, also called cubic splines. Cubic splines are de ned in terms of ve parameters, compared to the three parameters that are needed for the de nition of the 1-splines, B B x x : x : : B B B 8 >> (x?xM ?xl1 ?xl2 )3 if M ?  l1 ?  l2   M ?  l1 (xl2 )3 >> >> (xl1)3+3(xl1)2(x?xM ?xl1 )+3xl13(x?xM ?xl1 )2?3(x?xM ?xl1 )3 (xl1 ) >> >> if M ?  l1   M >< 3 2 xr13(xM +xr1 ?x)2?3(xM +xr1 ?x)3 (A.7) 3 ( ) = > (xr1 ) +3(xr1 ) (xM +xr1 ?x)+3 ( x r1 ) >> >> if M   M +  r1 >> (xM +xr1 +xr2 ?x)3 >> if M +  r1   M +  r1 r2 (xr2 )3 >> : 0 otherwise. ; x x x x x x ; x B x x x x ; x x x x ; x x x x x x ; A 3-spline with center M = 1 0, rst width to the left  l1 = 0 25, second width to the left  l2 = 0 25, rst width to the right  r1 = 0 25, and second width to the right  r2 = 0 25 is plotted in Fig. A.1(b). The maximum value of a 3-spline is 4.0, while the maximum value of the 1-spline was 1.0. The latter occurs for the 3 -spline at the locations of the two new parameters, namely at M ?  l1 and B x x x : x : x : : : B B B x x 106 + xr1 The advantages o ered by the use of B3-splines include the improved adjustment for nonlinear approximations and the availability of up to third order derivatives. Finally, the most popular of the radial basis functions is the well known Gaussian function, de ned in terms of two parameters, its center  and radius  [71] xM G(x) = e? 2 ? k where  = kx  : (A.8) The Gaussian function is symmetric and can be employed for nonlinear approximations by exploiting its nonlinear nature and its simplicity due to the low number of parameters that are to be optimized. However, its symmetry decreases its exibility. A Gaussian function with center  = 1:5 and radius  = 0:5 is depicted in Fig. A.2 (a). Figure A.2 (b) is an example of the augmented radial basis function used in this report for the reason given in Section 4.1. For the numerical examples considered in this report, B1-splines and augmented radial bases were employed. This choice was made for the reasons that follow. Firstly, B1-splines are quite easy to use and implement computationally. In addition, they depend on three and six parameters for the one- and two-dimensional cases, respectively. In this manner, the size of the optimization problem is kept as low as possible. Secondly, the derivatives of B3-splines can achieve relatively large values, or \spikes", which are the source of numerical instabilities. Moreover, they depend on ve and ten parameters for the one- and two-dimensional cases, respectively. Accordingly, not only is the size of the associated optimization problems increased, but the computational representation of the optimization variables becomes more dicult, as will be shown in Section A.5. Finally, Gaussian functions depend on only two and three parameters to be optimized for the one- and two- 107 (a) (b) 1.5 5 4.5 4 3.5 1 B3 (x) B1 (x) 3 2.5 2 0.5 1.5 1 0.5 0 0 0.2 0.4 0.6 0.8 1 x 1.2 1.4 1.6 1.8 0 0 2 0.2 0.4 0.6 0.8 1.2 1.4 1.6 1.8 2 1.8 2 (b) (a) Figure A.1 1 x B -splines: (a) a B1-spline and (b) a B3-spline. 1 κ = κ = 10−6 L R κL = 2.5 κR = −2.5 1.2 0.9 0.8 1 0.7 0.8 Φ(x) G(x) 0.6 0.6 0.5 0.4 0.4 0.3 0.2 0.2 0.1 0 −0.5 0 0.5 1 1.5 x 2 2.5 3 3.5 0 0 0.2 0.4 0.6 0.8 1 x 1.2 1.4 1.6 (b) (a) Figure A.2 Radial bases: (a) Gaussian function with center  = 1:5?and 6 radius  = 0:5 and (b) an augmented radial basis with  =  = 10 ,  = 2:5 and  = ?2:5. l l r r 108 dimensional cases, respectively. However, the domain of in uence (or support) of the functions cannot be as readily controlled as for the B1-splines. Consequently, the necessary constraints for enforcing homogeneous Dirichlet boundary conditions may not be as straightforwardly to implement as for the B1-splines, as will be illustrated in Section A.5. In this regard, Gaussian functions are more useful in problems associated with Neumann boundary conditions, where the basis functions do not have to vanish at certain boundaries. A.3 Members of the Method of Weighted Residuals The members of the method of weighted residuals can be classi ed according to the choice of the weighting function as follows. 1. The subdomain method, where the domain of interest is divided into n smaller subdomains ! , i = 1; : : : ; n, and the weighting functions w (x) are given by i 8 >< 1 ; w (x) = >: 0 ; i 2! x 62 ! x i (A.9) i i : 2. The collocation method, were the weighting function is given by the Dirac delta function, w = (x ? x ), and possess the following property i Z i 8 >< 1 ;  (x ? x )dx = >: 0 ; i =x x 6= x x i i : (A.10) 109 In this manner, h ( ) ( )i = Z ( ) ( )d = j R x wi x R x wi x x R xi (A.11) : 3. The least-squares method, where the weighting function wi = @R @ ci wi is given by (A.12) ; where is the -th coecient of the series expansion. ci i 4. The method of moments, where the weighting functions are given by the monomials wi = i x ; i =0 ;:::;n ?1 (A.13) : 5. The Bubnov-Galerkin method, where the weighting functions are given by the basis (or trial) functions used in the series expansion of the approximate solution, i.e. =  . A modi cation of the form =  forms the Petrov-Galerkin method. wi wi i gi i A.4 Classi cation of Di erential Operators Di erential operators are classi ed into linear and nonlinear operators. The adjoint of an operator is found by integrating R [ ]d by parts to obtain H vH f x 110 R [ ]dx + (B.T.), where the term (B.T.) denotes the boundary terms stemming from the performed integration(s) by parts. Linear operators are classi ed into self- and nonself-adjoint operators. A linear operator is said to be self-adjoint if L[] = L[]. The adjoint of a nonlinear operator N  cannot be the same as the operator N . The term adjoint is therefore inappropriate for nonlinear operators; it is substituted by the term self-sucient. If the operator is nonlinear, the Frechret di erential must be examined. The Frechret di erential N [z] of a nonlinear operator N is de ned as [25](page 301) fH v 0 w N [w + z] ? N [w] = @ (N [w + z]) j ; N [z]  lim =0 !0  @ 0 w   (A.14) where N is the Frechret derivative with respect to the function w of the operator N operating on z. If the Frechret derivative is symmetric, i.e. 0 w Z v N [z]x = 0 w Z z N [v]x ; (A.15) 0 w then the operator N is said to be self-sucient. Equation (A.15) is the condition necessary for the existence of a functional that gives the operator N as its gradient [86]. It is the analogy of the condition that is required for a vector eld to be able to be derived from a potential, extented to function spaces [25](page 300). Note that the case of linear operators is in fact a special case of nonlinear operators since the Frechret derivative is given by the operator itself 0 L[w + z ] ? L[w] L[w] + L[z ] ? L[w] [ ]  lim = lim = L[z] : !0 ! 0   Lw z   (A.16) 111 Variational principles are readily available for problems governed by self-adjoint di erential operators. A di erential operator is said to be positive de nite if Z [ ]d fL f x > 0 ; f =0 6 (A.17) : Elliptic operators are self-adjoint and positive de nite in contrast to hyperbolic or parabolic operators. Steady-state problems typically yield elliptic operators, while time-dependent problems are typically associated with hyperbolic or parabolic operators. A.5 PDS Computational Issues A.5.1 The Formulation of the Optimization Variables The computational representation of the parameters to be optimized is neither unique nor trivial. For the 1-splines, for example, one option to de ne the PDS optimization variables in terms of the parameters to be optimized is given by Eq. (3.10). For this representation B M x = x(1)  l = x(2) and  r = x(3) ; x ; x ; (A.18) where x(1), x(2), and x(3) are the PDS variables. An alternative formulation could be the following M x = x(1) ; x L = x(2) and ; x R = x(3) ; (A.19) 112 where x = x ? x and x = x + x . The rst formulation requires the additional satisfaction of constraints since the functions to be chosen have to satisfy homogeneous boundary conditions, i.e. vanish at the boundaries. Speci cally, the constraints L M x(1) ? x(2)  x l R ;L M r and x(1) + x(3)  x ;R (A.20) need to be included in the PDS optimization problem formulation, where x and x denote the left and right boundaries of a one-dimensional domain, respectively. This is not necessary for the second representation, where the homogeneous boundary conditions can be readily satis ed by appropriate bounds, which PDS requires in all cases for the optimization variables. However, the rst representation was proven, by means of computational experimentations, to be more ecient than the second. It is obvious that with increasing number of parameters, as is the case for B3-splines, the number of choices increases exponentially, and a numerical study should be performed to determine the optimal representation. Lastly, note that an adequately large lower bound on x is necessary to avoid relatively large values in the evaluation of the derivatives of the basis functions. A typical value for the lower bound is 1 % of the width of the computational domain. ;L ;R A.5.2 Scaling An issue of signi cance in direct search optimization methods is the correct scaling among the variables. The objective of scaling is to avoid uneven searches in the 113 directions of the optimization variables that constitute the dimensionality of the search hyperplane. PDS o ers the option of enforcing diverse scaling factors for the optimization variables in a very simple and user-friendly manner. The scaling factor of each variable is given by the range of values that this variable can take. For example, if the variable x(1) is likely to take any value between -3 and 6, its scaling factor is 9. Scaling is unnecessary, obviously, if all optimization variables have approximately the same scaling factor, in which case PDS assigns a scaling factor of magnitude 1 to each variable. A scaling factor of magnitude 1 is to be selected in the absence of information. A.6 Diagonal Dominance and Spectral Radius A matrix A 2 R N X ja j> N N ii j is said to be diagonal dominant if 8i  =1;j 6=i j a j; ij j = 1; : : : ; N : (A.21) The spectral radius of the matrix A is given by (A) = max j  j; i = 1; : : : ; N ; i (A.22) where  are the eigenvalues of the matrix A, i.e. the roots of the characteristic polynomial det(A ? I) = 0, with I being the N  N identity matrix. i 114 A.7 Numerical Integration All numerical integrations were performed by means of ESSL library routines. ESSL is a numerical library provided by the IBM SP2 parallel computer, on which all codes presented in this report were executed. In particular, the double precision routines DGLNQ and DGLNQ2, which utilize the Gauss-Legendre quadrature rule for one- and two-dimensional numerical integration, respectively, were used. As mentioned in the main text of the report, the optimally chosen basis functions are allowed to overlap; a structured grid is not built to provide comfortably determined integration bounds. In addition, 1-splines are not uniformly de ned, but can take three di erent values according to their parameters and position in space. In this regard, it is necessary to develop algorithms for the ecient and accurate numerical integration of the variational integrals. In particular, two algorithms were developed. The rst algorithm can be employed when the structure of the functional allows the decomposition of the variational integral into simpler integrals, where the summations from the series expansions can be taken outside the integrals. For example, when the following algebraic manipulation can be performed B Z Z ( ) ( ( )) d d = Z Z X  ( ) ( ) X d ( )  ( )d d = d X X Z Z  ( ) ( ) d ( )  ( )d d = d X X Z  ( ) d ( ) d Z  ( ) ( )d a un a x; y @ un x; y @x n n i=1 n i=1 j =1 n n ci i x ci cj i=1 j =1 ci cj i i y i x x x y n j =1 i y j x d x j cj x x j x x x i j y x y j y x y y j y y ; (A.23) the integral of the product of two (or more) basis functions and/or its deriva- 115 tives in one dimension has to be computed. The algorithm used to determine the number m of the necessary subintegrations and the associated bounds aj and bj , j = 1; : : : ; m, for n basis functions (B1-splines), assuming that the bounds of the domain are x1 and x2, is the following Algorithm 1 Find the maximum xL among the xL;i ; i = 1; : : : ; n; Find the minimum xR among the xR;i ; i = 1; : : : ; n; If xL;max  x1 ; set xL;max = x1; If xR;min  x2 ; set xR;min = x2; Set m = 0; If xL;max  xR;min ; stop; For i = 1; : : : ; n If xL;max < xM;i < xR;min m m + 1; am = xL;max ; bm = xM;i ; Set xL;max = xM;i ; End if; End for; m m + 1; am = xL;max ; bm = xR;min: In the case that the variational integral cannot be simpli ed and/or decomposed as in Eq. (A.23), the approximate functions have to be numerically evaluated on an adequate number of Gauss points so that the integral can be calculated accu- 116 rately. Both the unstructured and relatively complicated overlapping of the basis functions and the change of their de nition over certain spatial domains prevent the numerical integration from being eciently and accurately performed, even for a relatively high number of Gauss points. A domain decomposition, according to the basis function parameters and their location, is necessary to yield dependable numerical results. An algorithm was designed for this purpose and utilized for the two-dimensional integration in the compressible, inviscid, irrotational ow application. Assuming that the optimal sequential approximation has reached step and that the variational integral has to be evaluated within a rectangular domain of interest with boundaries 1, 2, 1, and 2, the integral is decomposed into ( ? 1)  ( ? 1) subintegrals , = 1 ? 1, = 1 ?1 and evaluated with help of the following algorithm n I x nxsub n x y n nysub Ii;j y i ; : : : ; nxsub j Algorithm 2 Merge the n-dimensional arrays x L ; xM ; into a 3n-dimensional array called Merge the n-dimensional arrays y L ; yM ; into a 3n-dimensional array called Sort the 3n-dimensional arrays Set If nxsub xcopy = 0; (1)  nxsub ( x1 nxsub ) = 1; xsub nxsub End if; + 1; x xcopy and xcopy xR ; and yR ycopy and ; ycopy ; ; : : : ; nysub 117 Algorithm 2 (cont.) For i = 1; : : : ; n If (x1 < xcopy (i) < x2 ) nxsub nxsub xsub(nxsub) = 6 and (xcopy (i) = xsub(nxsub)) + 1; x1 ; End if; End for; If xcopy (n) nxsub  x2 + 1; nxsub xsub(nxsub) = x2 ; End if; Set If nysub = 0; ycopy (1) nysub  y1 nysub ysub(nysub) = + 1; y1 ; End if; For i = 1; : : : ; n If (y1 < ycopy (i) < y2 ) nysub nysub ysub(nysub) End if; End for; = + 1; y1 ; 6 and (ycopy (i) = ysub(nysub)) 118 Algorithm 2 (cont.) If ( ) ycopy n nysub y2 nysub ( + 1; ) = 2; ysub nysub y End if; Set I = 0; n For = 1 i ?1 ; : : : ; nxsub For = 1 j ; : : : ; nysub Compute I n End for; End for : I n + n Ii;j n ; Ii;j ; ?1 119 Bibliography [1] J.D. Anderson, Jr. Modern Compressible Flow. McGraw Hill, 1990. Pages 242-250. [2] R. Barrett, M. Berry, T.F. Chan, J. Demmel, J. Donato, J. Dongarra, V. Eijkhout, R. Pozo, C. Romine, and H. van der Vorst. Templates for the Solution of Linear Systems: Building Blocks for Iterative Methods. SIAM Publications, 1994. Pages 5-7. [3] A.R. Barron. \Universal approximation bounds for superpositions of a sigmoidal function". IEEE Transactions on Information Theory, 39(3):930{945, 1993. [4] H. Bateman. \Irrotational motion of a compressible inviscid uid". Proc. Nat. Acad. Sci, 16:816{825, 1930. [5] M. Becker. The Principles and Applications of Variational Methods. MIT Press, 1964. Research Monograph No. 27, pages 7-10. [6] C.M. Bishop. \Curvature-driven smoothing: A learning algorithm for feedforward networks". IEEE Transactions on Neural Networks, 4(5):882{884., 1993. [7] G.E.P. Box and N.R. Draper. Empirical Model-Building and Response Surface. Wiley, New York, 1997. [8] D.S. Broomhead and D. Lowe. \Multi-variable functional interpolation and adaptive networks". Complex Systems, 2:321{355., 1988. 120 [9] E.W. Cheney. Approximation Theory. McGraw Hill, 1966. Pages 3-6. [10] S. Cox and M. Overton. \Designing for optimal energy absorption - III Numerical minimization of the spectral abscissa". Structural Optimization, 13(1):17{ 22, February 1997. Journal of the International Society for Structural and Multidisciplinary Optimization (ISSMO), Springer Verlag. [11] G. Cybenko. \Approximation by superposition of a sigmoidal function". Mathematics of Control, Signals, and Systems, 2:303{314., 1989. [12] A.J. Davies. The Finite Element Method, a First Approach. Clarendon Press, 1980. Pages 48-58. [13] L. Demkowicz and J.T. Oden. \On a mesh optimization method based on a minimization of interpolation error". Int. J. Engng. Sci., 24(1):55{68, 1986. [14] J.E. Dennis, Jr. and V. Torczon. \Direct search methods on parallel machines". SIAM Journal of Optimization, 1(4):448{474, November 1991. [15] R.A. DeVore. \Degree of nonlinear approximation". In C. K. Chui, L. L. Schumaker, and J. D. Ward, editors, Approximation Theory VI, pages 175{ 201. Academic Press, 1989. [16] A.R. Diaz, N. Kikuchi, and J.E. Taylor. \A method of grid optimization for nite element methods". Computer Methods in Applied Mechanics and Engineering, 41:29{45, 1983. [17] A.R. Diaz, N. Kikuchi, and J.E. Taylor. \Optimal design formulations for nite element grid adaptation". In V. Komokov, editor, Lecture Notes in Mathematics: Sensitivity of Functionals with Applications in Engineering Science. Springer Verlag, Berlin, 1984. 121 [18] E.O. Doebelin. Engineering Experimentation: Planning, Execution, Reporting. McGraw Hill, 1995. [19] J. Duchon. Spline Minimizing Rotation-Invariant Semi-Norms in Sobolev Spaces, Constructive theory of Functions of Several Variables, Lecture Notes in Mathematics. Vol. 571, Springer Verlag, Berlin, 1977. [20] N. Dyn. \Interpolation and approximation by radial and related functions". In C. K. Chui, L. L. Schumaker, and J. D. Ward, editors, Approximation Theory VI, pages 211{234. Vol.1, Academic Press, 1989. [21] N. Dyn, D. Levin, and S. Rippa. \Numerical procedure for surface tting of scattered data by radial functions". SIAM J. Sci. Stat. Comput., 7(2):639{ 659, 1986. [22] P.R. Eiseman, Y.K. Choo, and R.E. Smith. \Algebraic grid generation with control points". In T.J. Chung, editor, Finite Elements in Fluids, Volume 8. Hemisphere Publ. Corp., 1992. [23] C.A. Felippa. \Optimization of nite element grids by direct energy search". Appl. Math. Modelling, 1:93{96, 1976. [24] C.A. Felippa. \Numerical experiments in nite element grid optimization by direct energy search". Appl. Math. Modelling, 1:239{244, 1977. [25] B.A. Finlayson. The Method of Weighted Residuals and Variational Principles. Academic Press, 1972. [26] C.A.J. Fletcher. Computational Galerkin Methods. Springer, New York, 1984. [27] C.A.J. Fletcher. Computational Techniques for Fluid Dynamics, volume 1. Springer Verlag, 1988. Pages 293-299. 122 [28] R. Franke. \Scattered data interpolation: Tests of some methods". Mathematics of Computation, 38:181{200, 1982. [29] F. Girosi, M. Jones, and T. Poggio. \Regularization theory and neural network architectures". Neural Computation, 7:219{269, 1995. [30] F. Girosi and T. Poggio. \Networks and the best approximation property". Biological Cybernetics, 63:169{176., 1990. [31] D. Greenspan. \On approximating extremals of functionals - I The method and examples for boundary value problems". Bull. Int. Comp. Centre, 4(2):99{120, April-June 1965. University of Rome. [32] D. Greenspan. \On approximating extremals of functionals - II Theory and generalizations related to boundary value problems for nonlinear di erential equations". Int. J. Engng Sci., 5:571{588, 1967. [33] D. Greenspan and P.C. Jain. \Application of a method for approximating extremals of functionals to compressible ow". Mathematical Analysis and Applications, 18(1):85{111, April 1967. Also available as technical report of the Mathematical Research Center of the University of Wisconsin, TR 597, January 1966. [34] G.L. Guymon. \A nite element solution of the one-dimensional di usionconvection equation". Water Resources Research, 6(1):204{210, February 1970. [35] G.L. Guymon, V.H. Scott, and L.R. Herrmann. \A general numerical solution of the two-dimensional di usion-convection equation by the nite element method". Water Resources Research, 6(6):1611{1617, December 1970. 123 [36] O. Hassan, K. Morgan, and J. Peraire. \Finite-element solution scheme for compressible viscous ow". In T.J. Chung, editor, Finite Elements in Fluids, Volume 8. Hemisphere Publ. Corp., 1992. [37] R. Hecht-Nielsen. Neurocomputing. Addison-Wesley, Inc., New York, 1990. [38] L. Holmstrom and P. Koistinen. \Using additive noise in back-propagation training". IEEE Transaction on Neural Networks, 3:24{38., 1992. [39] K. Hornik, M. Stinchcombe, and H. White. \Multi-layer feedforward networks are universal approximators". Neural Networks, 2:359{366, 1989. [40] K. Hornik, M. Stinchcombe, and H. White. \Multilayer feedforward networks are universal approximators". Neural Networks, 2:359{366., 1989. [41] K. Hornik, M. Stinchcombe, and H. White. \Universal approximation of an unknown mapping and its derivatives using multilayer feedforward networks". Neural Networks, 3:551{560., 1990. [42] I. Iami. \On the ow of a compressible uid past a circular cylinder". In Proc. Phys. - Math. Soc. Japan, pages 180{193, 1941. Ser. 3, volume 23. [43] Y. Ito. \Representation of functions by superpositions of a step or sigmoidal function and their applications to neural network theory". Neural Networks, 4:385{394., 1991. [44] P.S. Jensen. \Advances in computer methods for partial di erential methods". In R. Vivhnevestjy, editor, AICA Symposium, pages 80{85, Lehigh University, 1975. [45] C. Johnson. Numerical Solution of Partial Di erential Equations by the Finite Element Method. Cambridge University Press, 1990. Pages 33-38. 124 [46] L.K. Jones. \Constructive approximations for neural networks by sigmoidal functions". Proceedings of the IEEE, 78(10):1586{1589, 1990. [47] L.K. Jones. \A simple lemma on greedy approximation in Hilbert space and convergence rates for projection pursuit regression and neural network training". The Annals of Statistics, 20(1):608{613, 1992. [48] G.S. Kimeldorf and G. Wahba. \A correspondence between bayesian estimation on stochastic processes and smoothing by splines". Annals of Mathematical Statistics, 41(2):495{502, 1970. [49] D. Kincaid and W. Cheney. Numerical Analysis. Brooks and Cole, 1990. Pages 181-182. [50] M. Krizek and P. Neittaanmaeki. Finite Element Approximation of Variational Problems and Applications. Longman Scienti c & Technical, 1990. Page 14. [51] C. Lanczos. The Variational Principles in Mechanics. University of Toronto Press, 1952. [52] R.M. Lewis and V. Torczon. \Pattern search algorithms for bound constrained minimization". Technical Report 96-20, ICASE, NASA Langley Research Center, Hampton, Virginia, 1996. Submitted to SIAM Journal of Optimization. [53] P.E. Lush and T.M. Cherry. \The variational method in hydrodynamics". J. Mech. Appl. Math., 9:6{21, 1956. [54] A.R. Manwell. \A variational principle for steady homenergic compressible ow with nite shocks". Wave motion, 2(1):83{95, 1980. 125 [55] G. Matheron. \Principles of geostatistics". Economic Geology, 58:1246{1266., 1963. [56] G. Matheron. \The intrinsic random functions and their applications". Adv. Appl. Prob., 5:439{468., 1973. [57] A.J. Meade, Jr. and A.A. Fernandez. \Solution of non-linear ordinary di erential equations by feedforward neural networks". Mathematical and Computer Modelling, 20(9):19{44, 1994. [58] A.J. Meade, Jr., M. Kokkolaras, and B.A. Zeldin. \Optimal incremental approximation for the adaptive solution of di erential equations". International Journal for Numerical Methods in Engineering, 1997. Submitted. [59] A.J. Meade, Jr., M. Kokkolaras, and B.A. Zeldin. \Sequential function approximation for the solution of di erential equations". Communications in Numerical Methods in Engineering, 13:977{986, 1997. [60] A.J. Meade, Jr. and B.A. Zeldin. \Approximation properties of local bases assembled from neural network transfer functions". Mathematical and Computer Modelling, 1997. To appear. [61] A.J. Meade, Jr. and B.A. Zeldin. \Establishing criteria to ensure successful feedforward arti cial neural network modelling of mechanical systems". Mathematical and Computer Modelling, 1997. To appear. [62] J. Meinguet. \Multivariate interpolation at arbitrary points made simple". J. Appl. Math. Phys., 30:292{304., 1979. [63] H.N. Mhaskar. \Neural networks for optimal approximation of smooth and analytic functions". Neural Computation, 8:164{177., 1996. 126 [64] S.G. Mikhlin. Variational Methods in Mathematical Physics. MacMillan Pergamon, 1964. [65] P.M. Morse and H. Feshbach. Methods of Theoretical Physics, volume 1. McGraw-Hill, 1953. Pages 276-280. [66] R.H. Myers, A.I. Khuri, and W.H. Carter, Jr. \Response surface methodology: 1966-1988". Technometrics, 31(2):137{157, 1989. [67] R.H. Myers and D.C. Montgomery. Response Surface Methodology: Process and Product Optimization using Designed Experiments. John Wiley & Sons, Inc. New York, 1995. [68] J.T. Oden. Variational Methods in Theoretical Mechanics. Springer Verlag, 1976. [69] J.T. Oden. \Adaptive methods in computational uid dynamics". In T.J. Chung, editor, Finite Elements in Fluids, Volume 8. Hemisphere Publ. Corp., 1992. [70] E.R. Oliveira. \Optimization of nite element solutions". In Proc. Third Conf. Matrix Methods in Struct. Mech., Ohio, 1971. Air Force Flight Dynamics Laboratory, Wright-Patterson AFB. AFFDL TR 71-160. [71] M.J. Orr. \Regularization in the selection of radial basis function centres". Neural Computation, 7(3):606{623., 1995. [72] J. Park and I.W. Sanberg. \Universal approximation using radial-basisfunction networks". Neural Computation, 3:246{257., 1991. [73] J. Platt. \A resource-allocating network for function interpolation". Neural Computation, 3:213{225, 1991. 127 [74] T. Poggio and F. Girosi. \Networks for approximation and learning". Proceedings of the IEEE, 78(9):1481{1497., 1990. [75] T. Poggio and F. Girosi. \Regularization algorithms for learning that are equivalent to multilayer networks". Science, 247:978{982., 1990. [76] R.J. Prozan. \A variational principle for compressible uid mechanics - Discussion of the multi-dimensional theory". Technical Report 3614, NASA, 1982. [77] R.J. Prozan. \A variational principle for compressible uid mechanics - Discussion of the one-dimensional theory". Technical Report 3526, NASA, April 1982. [78] K. Rektorys. Variational Methods in Mathematics, Science, and Engineering. Dordrecht, 1980. 2nd edition. [79] D. Rumelhart, G.E. Hinton, and R.J. Williams. \Learning internal representations by error propagation". In D. Rumelhart, J.L. McClelland, and the PDP Research Group, editors, Parallel Distributed Processing: Exploration in the Microstructure of Cognition, pages 318{362. MIT Press, Cambridge, MA, 1986. [80] S. Saarinen, R. Bramley, and G. Cybenko. \Ill-conditioning in neural network training problems". SIAM J. Sci. Comput., 14(3):693{714., 1993. [81] J. Sacks, W.J. Welch, T.J. Mitchell, and J.P. Wynn. \Design and analysis of computer experiments". Statistical Science, 4(4):409{435, 1989. [82] W.F. Stoecker. Design of Thermal Systems. McGraw-Hill, New York, 1989. Pages 189-190. 128 [83] A.N. Tikhonov and V.Y. Arsenin. Solution of Ill-Posed Problems. W.H. Winston, Washington, D.C,, 1977. [84] V. Torczon. \PDS: Direct search methods for unconstrained optimization on either sequential or parallel machines". Technical Report TR 92206, CRPC, March 1992. [85] V. Torczon. \On the convergence of pattern search methods". SIAM Journal of Optimization, February 1997. [86] M.M. Vainberg. Variational Methods for the Study of Nonlinear Operators. Holden-Day Inc., 1964. Pages 54-58. [87] G. Wahba. Spline Models for Observational Data. SIAM, Philadelphia, 1990. [88] W.J. Welch, R.J. Buck, J. Sacks, H.P. Wynn, T.J. Mitchell, and M.D. Morris. \Screening, predicting, and computer experiments". Technometrics, 34(1):15{ 25., 1992. [89] R.C. Williamson and U. Helmke. \Existence and uniqueness results for neural network approximations". IEEE Transactions on Neural Networks, 6(1):2{13., 1995. [90] O.C. Zienkiewicz and R.L. Taylor. The Finite Element Method. McGraw-Hill, 1989. Pages 240-242.