Reliability Methods
This theory chapter explores local and global reliability methods in greater detail the overview in Uncertainty Quantification.
Local Reliability Methods
Local reliability methods include the Mean Value method and the family of most probable point (MPP) search methods. Each of these methods is gradient-based, employing local approximations and/or local optimization methods.
Mean Value
The Mean Value method (MV, also known as MVFOSM in [HM00]) is the simplest, least-expensive reliability method because it estimates the response means, response standard deviations, and all CDF/CCDF response-probability-reliability levels from a single evaluation of response functions and their gradients at the uncertain variable means. This approximation can have acceptable accuracy when the response functions are nearly linear and their distributions are approximately Gaussian, but can have poor accuracy in other situations.
The expressions for approximate response mean
where
respectively, where
With the introduction of second-order limit state information, MVSOSM calculates a second-order mean as
This is commonly combined with a first-order variance ((b) in (54)), since second-order variance involves higher order distribution moments (skewness, kurtosis) [HM00] which are often unavailable.
The first-order CDF probability
where
With the Mean Value method, it is possible to obtain importance factors
indicating the relative contribution of the input variables to the
output variance. The importance factors can be viewed as an extension of
linear sensitivity analysis combining deterministic gradient information
with input uncertainty information, i.e. input variable standard
deviations. The accuracy of the importance factors is contingent of the
validity of the linear Taylor series approximation used to approximate
the response quantities of interest. The importance factors are
determined as follows for each of
where it is evident that these importance factors correspond to the diagonal terms in (54), (b) normalized by the total response variance. In the case where the input variables are correlated resulting in off-diagonal terms for the input covariance, we can also compute a two-way importance factor as
These two-way factors differ from the Sobol’ interaction terms that are computed in variance-based decomposition (refer to Global sensitivity analysis: variance-based decomposition) due to the non-orthogonality of the Taylor series basis. Due to this non-orthogonality, two-way importance factors may be negative, and due to normalization by the total response variance, the set of importance factors will always sum to one.
MPP Search Methods
All other local reliability methods solve an equality-constrained
nonlinear optimization problem to compute a most probable point (MPP)
and then integrate about this point to compute probabilities. The MPP
search is performed in uncorrelated standard normal space (“u-space”)
since it simplifies the probability integration: the distance of the MPP
from the origin has the meaning of the number of input standard
deviations separating the mean response from a particular response
threshold. The transformation from correlated non-normal distributions
(x-space) to uncorrelated standard normal distributions (u-space) is
denoted as
where
where the original correlation matrix for non-normals in x-space has been modified to represent the corresponding “warped” correlation in z-space [DKL86].
The forward reliability analysis algorithm of computing CDF/CCDF
probability/reliability levels for specified response levels is called
the reliability index approach (RIA), and the inverse reliability
analysis algorithm of computing response levels for specified CDF/CCDF
probability/reliability levels is called the performance measure
approach (PMA) [TCP99]. The differences between the RIA
and PMA formulations appear in the objective function and equality
constraint formulations used in the MPP searches. For RIA, the MPP
search for achieving the specified response level
where
where
where
In the RIA case, the optimal MPP solution
where
where the limit state at the MPP (
Limit state approximations
There are a variety of algorithmic variations that are available for use within RIA/PMA reliability analyses. First, one may select among several different limit state approximations that can be used to reduce computational expense during the MPP searches. Local, multipoint, and global approximations of the limit state are possible. [EAP+07] investigated local first-order limit state approximations, and [EB06] investigated local second-order and multipoint approximations. These techniques include:
a single Taylor series per response/reliability/probability level in x-space centered at the uncertain variable means. The first-order approach is commonly known as the Advanced Mean Value (AMV) method:
(61)and the second-order approach has been named AMV
:(62)same as AMV/AMV
, except that the Taylor series is expanded in u-space. The first-order option has been termed the u-space AMV method:(63)where
and is nonzero in general, and the second-order option has been named the u-space AMV method:(64)an initial Taylor series approximation in x-space at the uncertain variable means, with iterative expansion updates at each MPP estimate (
) until the MPP converges. The first-order option is commonly known as AMV+:(65)and the second-order option has been named AMV
+:(66)same as AMV+/AMV
+, except that the expansions are performed in u-space. The first-order option has been termed the u-space AMV+ method.(67)and the second-order option has been named the u-space AMV
+ method:(68)a multipoint approximation in x-space. This approach involves a Taylor series approximation in intermediate variables where the powers used for the intermediate variables are selected to match information at the current and previous expansion points. Based on the two-point exponential approximation concept (TPEA, [FRB90]), the two-point adaptive nonlinearity approximation (TANA-3, [XG98]) approximates the limit state as:
(69)where
is the number of uncertain variables and:(70)and
and are the current and previous MPP estimates in x-space, respectively. Prior to the availability of two MPP estimates, x-space AMV+ is used.a multipoint approximation in u-space. The u-space TANA-3 approximates the limit state as:
(71)where:
(72)and
and are the current and previous MPP estimates in u-space, respectively. Prior to the availability of two MPP estimates, u-space AMV+ is used.the MPP search on the original response functions without the use of any approximations. Combining this option with first-order and second-order integration approaches (see next section) results in the traditional first-order and second-order reliability methods (FORM and SORM).
The Hessian matrices in AMV
negative
Safeguarding involves offseting negative
Probability integrations
The second algorithmic variation involves the integration approach for computing probabilities at the MPP, which can be selected to be first-order ((a) and (b) in (57)) or second-order integration. Second-order integration involves applying a curvature correction [Bre84, HR88, Hon99]. Breitung applies a correction based on asymptotic analysis [Bre84]:
where
where
Hessian approximations
To use a second-order Taylor series or a second-order integration when
second-order information (
In the finite difference case, numerical Hessians are commonly computed using either first-order forward differences of gradients using
to estimate the
to estimate the
Quasi-Newton approximations, on the other hand, do not reevaluate all of the second-order information for every point of interest. Rather, they accumulate approximate curvature information over time using secant updates. Since they utilize the existing gradient evaluations, they do not require any additional function evaluations for evaluating the Hessian terms. The quasi-Newton approximations of interest include the Broyden-Fletcher-Goldfarb-Shanno (BFGS) update
which yields a sequence of symmetric positive definite Hessian approximations, and the Symmetric Rank 1 (SR1) update
which yields a sequence of symmetric, potentially indefinite, Hessian
approximations.
Optimization algorithms
The next algorithmic variation involves the optimization algorithm selection for solving Eqs. (58) and (59). The Hasofer-Lind Rackwitz-Fissler (HL-RF) algorithm [HM00] is a classical approach that has been broadly applied. It is a Newton-based approach lacking line search/trust region globalization, and is generally regarded as computationally efficient but occasionally unreliable. Dakota takes the approach of employing robust, general-purpose optimization algorithms with provable convergence properties. In particular, we employ the sequential quadratic programming (SQP) and nonlinear interior-point (NIP) optimization algorithms from the NPSOL [GMSW86] and OPT++ [MOHW07] libraries, respectively.
Warm Starting of MPP Searches
The final algorithmic variation for local reliability methods involves
the use of warm starting approaches for improving computational
efficiency. [EAP+07] describes the acceleration of MPP
searches through warm starting with approximate iteration increment,
with
Global Reliability Methods
Local reliability methods, while computationally efficient, have well-known failure mechanisms. When confronted with a limit state function that is nonsmooth, local gradient-based optimizers may stall due to gradient inaccuracy and fail to converge to an MPP. Moreover, if the limit state is multimodal (multiple MPPs), then a gradient-based local method can, at best, locate only one local MPP solution. Finally, a linear (Eqs. (a) and (b) in (57)) or parabolic (Eqs. (73) – (74)) approximation to the limit state at this MPP may fail to adequately capture the contour of a highly nonlinear limit state.
A reliability analysis method that is both efficient when applied to expensive response functions and accurate for a response function of any arbitrary shape is needed. This section develops such a method based on efficient global optimization [JSW98] (EGO) to the search for multiple points on or near the limit state throughout the random variable space. By locating multiple points on the limit state, more complex limit states can be accurately modeled, resulting in a more accurate assessment of the reliability. It should be emphasized here that these multiple points exist on a single limit state. Because of its roots in efficient global optimization, this method of reliability analysis is called efficient global reliability analysis (EGRA) [BES+07]. The following two subsections describe two capabilities that are incorporated into the EGRA algorithm: importance sampling and EGO.
Importance Sampling
An alternative to MPP search methods is to directly perform the probability integration numerically by sampling the response function. Sampling methods do not rely on a simplifying approximation to the shape of the limit state, so they can be more accurate than FORM and SORM, but they can also be prohibitively expensive because they generally require a large number of response function evaluations. Importance sampling methods reduce this expense by focusing the samples in the important regions of the uncertain space. They do this by centering the sampling density function at the MPP rather than at the mean. This ensures the samples will lie the region of interest, thus increasing the efficiency of the sampling method. Adaptive importance sampling (AIS) further improves the efficiency by adaptively updating the sampling density function. Multimodal adaptive importance sampling [DM98, ZMMM02] is a variation of AIS that allows for the use of multiple sampling densities making it better suited for cases where multiple sections of the limit state are highly probable.
Note that importance sampling methods require that the location of at least one MPP be known because it is used to center the initial sampling density. However, current gradient-based, local search methods used in MPP search may fail to converge or may converge to poor solutions for highly nonlinear problems, possibly making these methods inapplicable. As the next section describes, EGO is a global optimization method that does not depend on the availability of accurate gradient information, making convergence more reliable for nonsmooth response functions. Moreover, EGO has the ability to locate multiple failure points, which would provide multiple starting points and thus a good multimodal sampling density for the initial steps of multimodal AIS. The resulting Gaussian process model is accurate in the vicinity of the limit state, thereby providing an inexpensive surrogate that can be used to provide response function samples. As will be seen, using EGO to locate multiple points along the limit state, and then using the resulting Gaussian process model to provide function evaluations in multimodal AIS for the probability integration, results in an accurate and efficient reliability analysis tool.
Efficient Global Optimization
Note
Chapter Efficient Global Optimization has been substantially revised to discuss EGO/Bayesian optimization theory.
Expected Feasibility Function
The expected improvement function provides an indication of how much the
true value of the response at a point can be expected to be less than
the current best solution. It therefore makes little sense to apply this
to the forward reliability problem where the goal is not to minimize the
response, but rather to find where it is equal to a specified threshold
value. The expected feasibility function (EFF) is introduced here to
provide an indication of how well the true value of the response is
expected to satisfy the equality constraint
where
where