Metamodel-based importance sampling for structural reliability analysis

Structural reliability methods aim at computing the probability of failure of systems with respect to some prescribed performance functions. In modern engineering such functions usually resort to running an expensive-to-evaluate computational model (…

Authors: V. Dubourg, F. Deheeger, B. Sudret

Reliability analysis consists in the assessment of the level of safety of a system. Given a probabilistic model (an n-dimensional random vector X with probability density function (PDF) f X ) and a performance model (a function g), it makes use of mathematical techniques in order to estimate the safety level of the system in the form of a failure probability. A basic reference approach is the Monte Carlo simulation technique that resorts to numerical simulation of the performance function through the probabilistic model. Failure is usually defined as the event F = g (X ) ≤ 0 , so that the failure probability is defined as follows: Introducing the failure indicator function 1 g≤0 being equal to one if g (x ) ≤ 0 and zero otherwise, the failure probability turns out to be the mathematical expectation of this indicator function with respect to the joint probability density function f X of the random vector X . This convenient definition allows one to derive the Monte Carlo estimator which reads: 1 g≤0 x (k) (2) where x (1) , . . . , x (N ) , N ∈ * is a set of samples from the random vector X . According to the central limit theorem, this estimator is asymptotically unbiased and normally distributed with variance: In practice, this variance is compared to the unbiased estimate of the failure probability in order to decide whether the simulation is accurate enough or not. To do so, one usually resorts to the coefficient of variation defined as δ p f MC = σ p f MC / p f MC . N being fixed, this coefficient of variation dramatically increases when the failure event becomes rare (p f → 0). This makes the Monte Carlo estimation technique intractable for real world engineering problems for which the performance function involves the output of an expensive-to-evaluate black-box function -e.g. a finite element code. Note that this remark is also true for too frequent events (p f → 1) as one should compute the coefficient of variation of 1 -p f MC that exhibits the same property. In order to reduce the number of simulation runs, different alternatives to the bruteforce Monte Carlo method have been proposed and might be classified as follows. One first approach consists in replacing the original experiment by a surrogate which is much faster to evaluate. Various surrogates have been used amongst which are: quadratic response surfaces (Bucher and Bourgund, 1990;Kim and Na, 1997;Das and Zheng, 2000) and the more recent metamodels such as support vector machines (Hurtado, 2004;Deheeger and Lemaire, 2007), neural networks (Papadrakakis and Lagaros, 2002) and kriging (Kaymaz, 2005;Bichon et al., 2008). Nevertheless, it is often difficult or even impossible to quantify the error made by such a substitution. The well-known first-and second-order reliability methods, based on Taylor expansions somewhat differ from these surrogate-based approaches because of their mathematical background (Breitung, 1984). But in practice they feature the same limitation: the assumptions (mostly the unicity of the so-called most probable failure point) they are built on may not hold and it is difficult to validate them. Bucher (2009); Naess et al. (2009); Nishijima et al. (2010) investigate the use of the extreme value theory in order to infer the tail of the cumulative distribution function (CDF) of the random variable g(X ) from a set of samples. This set is obtained by crude Monte Carlo simulation of the input random vector X propagated through the performance function g. Despite this method has a lot of attractive features (it is independent of both the dimension n and the shape of g), it may require a large number of samples to make the fitted extreme value distribution accurate far in the tail -i.e. for low failure probabilities. From another point of view, variance reduction techniques have been proposed in order to make Monte Carlo simulation more efficient. Importance sampling (Rubinstein and Kroese, 2008) aims at concentrating the Monte Carlo samples in the vicinity of the limitstate surface, e.g. around the most probable failure point (also known as design point) obtained by a preliminary FORM analysis (Melchers, 1989). Subset simulation (Au and Beck, 2001;Ching et al., 2005a,b;Santoso et al., 2010) computes the failure probability as a product of conditional probabilities, each of them being estimated by Markov Chain Monte Carlo simulation. All these approaches reveal robust, although they are often too much computationally demanding to be implemented in industrial cases. As a summary the current practice for evaluating probabilities of failure in case of computationally demanding performance functions relies on the substitution of the limit-state function by a metamodel for which no general error estimation is usually available. In this paper, a new hybrid approach combining importance sampling and an adaptive metamodeling technique is proposed. First, a kriging surrogate of the limit-state function is built and adaptively refined. Then, the probabilistic prediction provided by the kriging surrogate is used to build up a quasi-optimal importance sampling density. As a result the probability of failure is computed as the product of two terms, one which is estimated by sampling the surrogate limit-state function, and a correction factor computed from the original limit-state function. The paper is organized as follows. Section 2 recalls the basic of Gaussian process (kriging) metamodeling and introduces the probabilistic classification function. Section 3 presents the construction of a quasi-optimal importance sampling density derived from this function and the associated estimator of the failure probability. Section 4 introduces an adaptive refinement technique of this importance sampling density function so as to make the algorithm as parsimonious as possible. Section 5 and 6 eventually describe practical implementation details and application examples. A metamodel means to a model what the model itself means to the real-world. Loosely speaking, it is the model of the model. As opposed to the model, its construction does not rely on any physical assumption about the phenomenon of interest but rather on statistical considerations about the coherence of some scattered observations that result from a set of experiments. This set is usually referred to as a design of experiments (DOE): = {x 1 , . . . , x m }. It should be carefully selected in order to retrieve the largest amount of statistical information about the underlying functional relationship over the input space x . In the sequel one attempts to build a metamodel for the failure indicator function 1 g≤0 . In the statistical learning theory (Vapnik, 1995) this is referred to as a classification problem. In this paper, we define a margin metamodel as a metamodel that is able to provide a probabilistic prediction of the response quantity of interest whose spread (i.e. variance) depends on the amount of information brought by the DOE. Note that this spread is thus reducible by bringing more observations into the DOE. In other words, this is an epistemic (reducible) source of uncertainty that shall not be confused with the possible uncertainty in X that will be modelled by a PDF f X later on for structural reliability purposes. To the authors' knowledge, there exist only two families of such margin metamodels: the probabilistic support vector machines -SVM by Platt (1999) and Gaussian process-(or kriging-) based classification. The present paper makes use of the kriging metamodel, but the overall concept could easily be extended to -SVM. The theoretical aspects of the kriging prediction are briefly introduced in the following subsection before it is applied to the classification problem of interest. The Gaussian process (also known as kriging) metamodeling theory is detailed in Sacks et al. (1989); Welch et al. (1992); Santner et al. (2003). In essence, it assumes that the performance function g is a sample path of an underlying Gaussian stochastic process G that may be cast as follows: where: • f (x ) T β denotes the mean of the GP which corresponds to a classical linear regression model on a given functional basis f i , i = 1, . . . , p ∈ 2 x , ; • Z (x ) denotes a zero-mean stationary Gaussian process with a constant variance σ 2 G . It is fully defined by its autocovariance function which reads: where is a vector of parameters defining R. The most widely used class of autocorrelation functions is the anisotropic squared exponential model: The best linear unbiased estimation (BLUE) of G at point x is shown (Santner et al., 2003) to be the following Gaussian random variate: with moments: where y = 〈g(x 1 ), . . . , g(x m )〉 T is the vector of observations, R is their correlation matrix, r (x ) is the vector of cross-correlations between the observations and the prediction, and F is the so-called regression matrix. The terms of R, r (x ) and F are respectively defined as follows: Finally, the generalized least-squares solution β and the vector u(x ) respectively read: At this stage one can easily prove that µ G x i = g x i and σ G x i = 0 for i = 1, . . . , m, thus meaning the kriging surrogate interpolates the observations without any residual epistemic uncertainty. Given a choice for the regression and correlation models, the optimal set of parameters β * , * and σ 2 * G can then be inferred using the maximum likelihood principle applied to the single sparse observation of the GP sample path grouped into the vector y. This inference problem turns into an optimization problem that can be solved analytically for both β * and σ 2 * G assuming * is known. Thus the problem is solved in two steps: the maximum likelihood estimation of * is first solved by a global optimization algorithm which in turns allows one to evaluate the optimal β * and σ 2 * G . Implementation details can be found in Welch et al. (1992); Lophaven et al. (2002). As seen from Eq. ( 7), the kriging surrogate provides both an approximation of the limitstate function g(x ) which is denoted by µ G (x ) and an epistemic prediction uncertainty which is characterized by the kriging variance σ 2 G (x ). Authors usually make use only of the surrogate g(x ) ≡ µ G (x ) for reliability analysissee Kaymaz (2005); Bichon et al. (2008). In this paper it is proposed to use the complete probabilistic prediction -see Picheny (2009) for a similar idea. Let us introduce for this purpose the following probabilistic classification function: In this expression, the probability measure [•] refers to the Gaussian nature of the kriging predictor G(x ) ∼ (µ G (x ), σ G (x )) and shall not be confused with the probability measure (•) associated with the random vector X -see Eq. ( 1). Thanks to the Gaussian nature of the kriging predictor, the probabilistic classification function rewrites: and still holds for the points in the experimental design for which the prediction variance equals zero by switching to the limit: It shall be again emphasized that π(x ) is not the sought failure probability. It may be interpreted as the probability that the predictor G(x ) (for some prescribed deterministic x ) is negative with respect to the epistemic uncertainty. Figure 1 illustrates the concepts introduced in this section on a basic structural reliability example from Der Kiureghian and Dakessian (1998). This example involves two independent standard Gaussian random variates X 1 and X 2 , and the performance function reads: where b = 5, κ = 0.5 and e = 0.1. In Figure 1, the original limit-state function g(x ) = 0 is represented by the black solid line. The red "+" and the blue "-" represent the initial DOE from which the kriging metamodel is built, the color being related to the sign of the performance function g. Once kriging has been applied, the mean prediction of the limitstate surface, which is defined as x ∈ n : µ G (x ) = 0 , is represented by the dashed black line. It can be seen that the metamodel is not fully accurate since the green triangle x 0 (among others) is misclassified. Indeed, x 0 is in the safe domain according to the real performance function g, and in the failure domain according to the mean kriging predictor µ G . Note that the probabilistic classification function allows a smoother decision: x 0 belongs to the failure domain with a 60% probability w.r.t. the epistemic uncertainty in the random prediction G(x 0 ) ∼ (µ G (x 0 ), σ G (x 0 )). Note also that the red "+" and the blue "-" in the DOE are always surely classified due to the interpolating property of the kriging metamodel (g Figure 2 is the one-dimensional illustration of the two classification strategies at point x 0 (green triangle): • according to the deterministic decision function (the Heaviside function centered in zero), the real performance function g is positive whereas the mean kriging prediction µ G is negative; • according to the probabilistic classification function (the smoother Gaussian cumulative distribution function centered in zero and whose spread is characterized by σ G (x 0 )), the epistemic prediction G is negative with a 60% probability (π(x 0 ) = 0.6 as defined from Eq. ( 16)). Picheny (2009) proposes to use the probabilistic classification function π (see Eq. ( 16)) as a surrogate for the original indicator function 1 g≤0 , so that the failure probability is 8 0 8 -1 . 9 6 σ G ( x 0 ) rewritten from its definition in Eq. ( 1) as follows: It is argued here that this latter quantity does not equal the failure probability of interest because it sums the aleatory uncertainty in the random vector X and the epistemic uncertainty in the prediction G. This is the reason why p f will be referred to as the augmented failure probability in the sequel. As a matter of fact, even if the epistemic uncertainty in the prediction can be reduced (e.g. by enriching the DOE as proposed in Section 4), it is impossible to quantify the contribution of each source of uncertainty in p f . This remark motivates the approach introduced in this section where the probabilistic classification function is used in conjunction with the importance sampling technique in order to build a new estimator of the failure probability. According to Rubinstein and Kroese (2008), importance sampling (IS) is one of the most efficient variance reduction techniques. This technique consists in computing the mathematical expectation of the failure indicator function according to a biased PDF which favors the failure event of interest. This PDF is called the instrumental density. Let h denote such a probability density such that h dominates 1 g≤0 f X , meaning that: Given this instrumental density, the definition of the failure probability of Eq. ( 1) may be rewritten as follows: In this expression, the expectation is now computed with respect to the instrumental density h. The latter definition of the failure probability easily leads to the definition of the importance sampling estimator: where x (1) , . . . , x (N ) , N ∈ * is a set of samples drawn from the instrumental density h. According to the central limit theorem, this estimation is unbiased and its quality may be measured by means of its variance of estimation which reads: Rubinstein and Kroese (2008) show that this variance is zero (optimality of the IS estimator) when the instrumental PDF is chosen as: However this instrumental PDF is not implementable in practice because it involves the sought failure probability p f in its denominator. The art of importance sampling consists in building an instrumental density which is quasi-optimal. Different strategies have been proposed in order to build quasi-optimal instrumental PDF suited for specific estimation problems. For instance, Melchers (1989) uses a standard normal PDF centered onto the design point obtained by FORM in the space of the independent standardized random variables. This approach might be inaccurate if the design point is not unique though. Cannamela et al. (2008) use a kriging prediction of the performance function g in order to build an instrumental PDF suited for the estimation of extreme quantiles of the random variate g(X ). However this approach is not applied to the estimation of failure probabilities. In this paper it is proposed to use the probabilistic classification function in Eq. ( 16) as a surrogate for the real indicator function in the optimal instrumental PDF in Eq. ( 24). The proposed quasi-optimal PDF thus reads as follows: where p f is the augmented failure probability which has been already defined in Eq. ( 19). For the sake of illustration, this quasi-optimal instrumental PDF is compared to the optimal (although impractical) one in Figure 3 using the example of Section 2.2. Choosing the proposed quasi-optimal instrumental PDF (i.e. substituting h * for h in Eq. ( 21)) leads to the following new expression of the failure probability: ≡ p f α corr (28) where we have introduced: This means that the failure probability is equal to the product of the augmented failure probability p f and a correction factor α corr . This correction factor is defined as the expected ratio between the real indicator function 1 g≤0 and the probabilistic classification function π. Thus, if the kriging prediction is fully accurate, the correction factor is equal to one and the failure probability is identical to the augmented failure probability (optimality of the proposed estimator). On the other hand, in the more general case where the kriging prediction is not fully accurate (since it is obtained from a DOE of finite size m), the correction factor modifies the augmented failure probability accounting for the epistemic uncertainty in the prediction. The two terms in the latter definition of the failure probability may now be estimated using Monte Carlo simulation: where the first N -sample set is generated from the original PDF f X , and the second N corrsample set is generated from the quasi-optimal instrumental PDF h * . According to the central limit theorem, these two estimates are unbiased and normally distributed. Their respective variance denoted by σ 2 and σ 2 corr may be easily derived as in Eq. ( 23). Finally, the proposed estimator of the failure probability simply reads as follows: It is important to note that both terms in Eq. ( 32) are independent, since they rely upon sampling according to two independent PDFs. Based on this latter remark, it is shown in Appendix A that for reasonably small values of the coefficients of variation δ p f and δ α corr of the two estimators p f and α corr (say 1 -10%), the coefficient of variation of the proposed estimator approximately reads: The instrumental density h * in Eq. ( 25) is not straightforward to sample from. First one observes that it is defined up to an unknown but finite normalizing constant: Indeed the exact value of p f is not known (it will only be estimated by Monte Carlo simulation). It is clear that the usual inverse transform simulation techniques are not applicable for this non-parameteric multidimensional pseudo-PDF. This is the reason why one resorts to Markov chain Monte Carlo (MCMC) simulation techniques (Robert and Casella, 2004). These techniques consists in generating a first-order Markov chain = z (i) , i = 1, . . . , N whose asymptotic distribution is shown to be the target distribution (see Robert and Casella, 2004, pp. 206 -207). The present implementation makes use of the slice sampling technique (Neal, 2003) -see Appendix B. However, Robert and Casella (2004) also point out that the Markov chain samples are not independent and identically distributed: there is some correlation between the successive states of the chain. Thus, in order to ensure that the samples used in Eq. ( 30) are independent and identically distributed, we use the so-called thining procedure which consists in keeping only one sample every t states of the chain (say t = 10). The significance of the variance reduction introduced by the proposed metamodel-based importance sampling technique mostly relies on the optimality of the proposed instrumental PDF h * . Thus, it is proposed here to adaptively refine the probabilistic classification function so that the quasi-optimal instrumental PDF h * converges towards its optimal counterpart h * . There exists a relatively large literature about the adaptive refinement of a kriging prediction for accurate classification (or level-set approximation): see e.g. Oakley (2004); Lee and Jung (2008); Bichon et al. (2008); Vazquez and Bect (2009); Picheny et al. (2010); Echard et al. (2011). They all rely on the definition of a so-called in-fill criterion which is maximum or minimum in the region of interest: the region where the sign of the predicted performance function is the most uncertain. The interested reader may find the expressions for all these criteria together with a discussed comparison on two analytical examples in a recent paper by Bect et al. (2011). For instance, Echard et al. (2011) propose the so-called -criterion which is defined as follows: This function is minimum in the vicinity of the limit-state surface of the mean prediction x ∈ n : µ G (x ) = 0 . This vicinity corresponds to the grayest part of the shade area in Figure 1. Then the problem of finding the best improvement point is usually reduced to a global optimization problem. Using the -criterion from Echard et al. (2011), the problem reads as follows: and is usually solved by means of a numerical global optimization algorithm. Additionally it is argued that for the present reliability estimation problem, one should focus on the most "probabilistically significant" region -such as it is done in FORM, because this is the region where the misclassified points have the most significant impact on the optimality of the proposed estimator. In Echard et al. (2011), such a trade-off is achieved by picking the best improvement point in a discrete population X that is distributed according to f X . The global optimization is thus reduced to a discrete optimization problem: As an introduction, it is argued that the available approximation of the optimal instrumental PDF after m observations: is an interesting trade-off criterion because it is maximum (i) where the sign of the predicted performance function is supposed to be negative, and (ii) where the probability density function f X is maximum. Then, it is proposed to use a sampling alternative to the optimization step -as in Dubourg et al. (2011). This is achieved by considering the in-fill criteria as PDFs defined up to an unknown but finite normalizing constant. It proceeds as follows: 1. Sample a large population (say N = 10 4 samples) from the weighted in-fill criterion (here h * ) using a MCMC simulation technique such as slice sampling (Neal, 2003). 2. Reduce this population to its K clusters' center (K being given, see Section 6) using the K-means clustering algorithm (MacQueen, 1967). 3. Evaluate the performance function on the K clusters' center. 4. Enrich the former experimental design with these K new observations. 5. Update the kriging prediction and loop back to step 1 until some target accuracy is achieved -see Section 4.2. The proposed sampling-based refinement strategy allows one to refine the prediction from a batch of optimal points instead of a single best point. It thus solves the problem of locally optimal points. Indeed, the in-fill criteria proposed in the literature commonly features several optima thus meaning that there does not exist a single best point. In the proposed strategy all the maxima are interpreted as modes of the improper PDF and results in local concentrations of points in the first step. The K-means clustering algorithm used in the second step reduces the population generated in the first step to the most significant modes. Note also that this approach allows one to perform several computations of the performance model in a distributed manner (i.e. on a high performance computing platform). Finally a metric is required to check the optimality of the probabilistic classification function. This metric may then be used as a stopping criterion for the previously detailed refinement strategy thus making it adaptive. The metric is based on a cross-validation procedure. Cross-validation techniques are classically used for model selection (see e.g. Stone, 1974). Basically, the design of experiments = x i , i = 1, . . . , m is split into a learning subset L and a validation subset V such that L ∩ V = and L ∪ V = . The model is then built using the learning subset L (hence denoted by g L ) and validated by comparing the predicted values g L (x ) and the real values g(x ) onto the validation subset x ∈ V . The leave-one-out technique is a special case where the learning subset is defined as L = \ x i . In a regression context, Allen (1971) propose to use a leave-one-out estimate of the mean squared error referred to as the predicted residual sum of squares: The metric proposed in this paper is a leave-one-out estimate of the correction factor in Eq. ( 29): where G \x i is the i-th leave-one-out kriging prediction of the performance function built from the DOE without the i-th sample x i . This quantity should be estimated using a minimal number of samples (say m = 30) so that it is sufficiently accurate. The reason for introducing this leave-one-out estimate of α corr is the following. In the early steps of the refinement procedure, the kriging predictor is not accurate enough to evaluate the failure probability by means of Eq. ( 32) efficiently. It would thus be ineffective to waste costly evaluations of the true limit-state function to compute the true correction factor in Eq. ( 31). On the contrary, the leave-one-out estimate α corr LOO makes only use of the available observations in the experimental design, so that it is fast to evaluate. As discussed earlier in Section 3.3, a sufficient accuracy is reached when the estimate in Eq. ( 40) gets close to one because it means that the probabilistic classification function converges towards its deterministic counterpart. And consequently the instrumental PDF h * converges towards it optimal counterpart h * . It is thus proposed to use α corr LOO as an indication to stop the refinement strategy. Precisely, the iterative enrichment of is stopped if its size is greater than 30 and α corr LOO is in the order of magnitude of 1. In the case of high dimensional and/or highly nonlinear problems, the size of the DOE is limited to m max = 1 000. One may also measure the improvement brought by the latest refinement iteration in terms of variation of α corr LOO . Remark regarding numerical implementation. It may happen that a leave-one-out estimate of the kriging variance σ 2 G \x i (x i ) gets close or even equal to zero. This is problematic because the ratio in Eq. ( 40) might not be defined in such cases. Indeed, if the mean prediction µ G \x i (x i ) is positive, the probabilistic classification function in the denominator equals zero -and it may cause an exception. It is argued that this variance should never be exactly zero (the kriging variance equals zero only at the samples in the DOE, here \ x i ), so that it is proposed to bound the probabilistic classification function above a reasonably low value (say the machine precision, M ≈ 10 -16 ). The purpose of this section is to summarize the proposed metamodel-based importance sampling algorithm. The flowchart of the algorithm is given in Figure 5. It is essentially divided in three independent steps, two of which could potentially be run in parallel. First, the algorithm only requires the choice of a maximum number of performance function evaluation N max , a targeted coefficient of variation δ target for the final estimate p f metaIS and the number K of points that should be added at each refinement iteration. Then, the first step of the algorithm is to build a reasonably accurate approximation of the probabilistic classification function that will bring a significant variance reduction for the final importance sampling estimate. Note that this step is optional in the case the analysis is resumed from an existing DOE. The two other steps are independent and might be run in parallel. They consist in using Monte Carlo simulation for the estimation of the two terms α corr and p f defining the proposed estimator p f metaIS -see Section 3. In the current implementation, we targeted the same coefficient of variation δ = δ corr = δ target / 2 for the two estimators. It could also be possible to implement these two steps in two independent threads that would communicate between each other and decide whether to stop the simulation or not based on the estimate of the final coefficient of variation. It should also be pointed out that it is easier to reduce the estimation variance on the augmented failure probability than the one one the correction factor because the former only depends on the kriging surrogate. In this section, the proposed strategy is applied to two structural reliability problems. The first one involves a simple analytical limit-state function and may thus be used to validate the implementation of the proposed strategy whereas the second example involves a more sophisticated nonlinear finite element model coupled with a high dimensional stochastic model so as to prove the applicability of the strategy to industrial problems. This basic structural reliability example is taken from Rackwitz (2001). It involves n independent lognormal random variates with mean value µ = 1 and standard deviation σ = 0.2. The performance function reads as follows: where a is set equal to 3 for the present application. The dimension of the problem is successively set equal to n = {2, 50, 100} to assess the influence of the dimension on the proposed estimator. The results are given in Table 1. (2001). The first block provides the results from a crude Monte Carlo simulation (reference) whereas the second block provides the results from the proposed importance sampling scheme. For both methods a 2% coefficient of variation on the failure probability estimates is targeted. First, it can be observed that the proposed estimator is in good agreement with the reference Monte Carlo estimates. The estimates of both the augmented failure probability p f and the correction factor α corr are also provided. Note that the correction factor increases with the dimension. In low dimension (n = 2) the kriging surrogate almost exactly equals the performance function (at least in the vicinity of its limit-state) hence α corr = 1 and p f metaIS = p f (no misclassification). In larger dimensions the surrogate loses accuracy, and the correction factor gets more and more influent. The size of the DOE from which the kriging prediction is built is given as the product between the number of refinement iterations and the number K of clusters' center added per iteration. K is chosen equal to the number of input random variables (n = {2, 50, 100}) so that the cost for the definition of the instrumental PDF is equivalent to the cost induced by a gradient-based design point search algorithm (when the gradient is approximated using a finite forward differentiation technique). In all cases 6 iterations are required (see Table 1). Note that the total number of calls to the real performance function g for the proposed importance sampling scheme (N DOE + N corr ) is much less than the number of calls induced by Monte Carlo simulation (N ) for the same targeted coefficient of variation. The mechanical model for this example is inspired by Scordelis and Lo (1961). It concerns the buckling analysis of a cylindrical shell roof whose dimensions are given in Figure 5. The longitudinal edges of the roof are free while its circumferential edges are supported by rigid diaphragms (radial displacement fixed to zero). Its constitutive material is assumed to have a nonlinear elastic Ramberg-Osgood material behavior. It is subjected to a constant surface load q and the structure is considered to fail if its critical buckling load q cr is less than a prescribed service load of magnitude q 0 = 0.18 MPa, so that the associated performance function may be defined as follows: where ξ denotes the outcome of the random vector Ξ introduced in the sequel. The critical buckling load q cr is determined by means of the asymptotic numerical method (Cochelin, 1994) coupled with a 30 × 30 8-node Büchter-Ramm shell finite element mesh (Büchter et al., 1994) using the EVE finite element code. This academic software was initially developed by Cochelin (1994) and further developed by Noirfalise et al. (2008). The stochastic model, inspired from Dubourg et al. (2009), involves four independent random fields defined over the roof surface. They describe respectively: the initial shape imperfection ζ, the material Young's modulus E, the material yield stress σ y , and the shell thickness h. The random shape imperfection is modeled as a linear combination of the three most critical Euler buckling modes' shape k , k = 1, . . . , 3 , so that it reads as follows: where Ξ ζ k , k = 1, . . . , 3 are three independent Gaussian random variates with mean µ ζ = 0 and standard deviation σ ζ ≈ 9.5 mm. This empirical model was determined with the following objectives: • the shape imperfection is Gaussian with a zero mean, so that the random variables should also be Gaussian with a zero mean; • the maximum amplitude of the shape imperfection around the mean surface (µ ζ = 0) ±2 σ ζ is set equal to the half of the mean thickness µ h = 76 mm, leading to σ ζ ≈ 9.5 mm. The other three random fields are assumed independent and lognormal with constant means µ h = 76 mm, µ E = 200, 000 MPa and µ σ y = 390 MPa and coefficients of variation δ h = 5%, δ E = 3% and δ σ y = 7%. They are represented by three Karhunen-Loève expansions of three independent standard Gaussian random fields, whose sample paths are translated into lognormal sample paths -see Dubourg et al. (2009) for the mapping. These three Gaussian random fields are assumed to have the same isotropic squared exponential autocorrelation function with a correlation length = 3, 500 mm. Due to the choice of this correlation function, the Fredholm integral equation involved in the Karhunen-Loève discretization scheme has no analytical solution. A so-called wavelet-Galerkin numerical procedure was thus used as detailed in Phoon et al. (2002). Each random field is simulated by means of 30 independent standard Gaussian random variates leading to a relative mean squared discretization error of 3.70%. Finally the complete stochastic model involves 93 independent random variables. The reliability results are gathered in Table 2. The proposed importance sampling scheme leads to a failure probability in full agreement with the value obtained by subset simulation (as implemented in the FERUM toolbox v4.0 by Bourinet et al., 2009) validates the proposed strategy. For this example the augmented failure probability p f is equal to 2.06 × 10 -4 (with a coefficient of variation of 5.70%), and the correction factor α corr is equal to 0.641 (with a coefficient of variation of 12.49%). A multiple FORM analysis revealed the existence of 4 most probable failure configurations identified by means of the restarted i-HLRF algorithm from Der Kiureghian and Dakessian (1998). These 4 failure modes corresponds to the 4 extreme cases for which the "demand" random field ζ is maximal in the corner of the roof whereas the "capacity" random fields E, σ y and h are minimal -one of these configurations is illustrated in Figure 6. The combination of these 4 failure modes in a serial system then allowed to give a Multi-FORM approximation of the failure probability (third column in Table 2). In this case, the Ditlevsen's bounds (Ditlevsen and Madsen, 1996) coincide (up to the accuracy provided in Table 2). It thus proves the ability of the proposed strategy to deal with reliability problems featuring multiple design points in a reasonably high dimension. Starting from the double premise that the usual surrogate-based reliability analyses do not permit to quantify the error made by using the metamodel instead of the original limitstate function, and that the existing variance reduction techniques remain time-consuming when the performance function involves the output of an expensive-to-evaluate black box function, an hybrid strategy has been proposed. First, a probabilistic classification function based on the post-processing of a kriging prediction was introduced. This function allows a smoother classification than its deterministic counterpart (i.e. the indicator function of the failure domain) accounting for the epistemic uncertainty in the kriging prediction. The probabilistic classification is then used to formulate a quasi-optimal importance sampling density. Using elementary algebra the failure probability is recast as a product of two terms, namely the augmented failure probability p f which is evaluated by means of the meta-model only, and a correction factor α corr that is computed from evaluations of the original limit-state function. In order to decide whether the kriging surrogate is accurate enough, a leave-one-out estimate of α corr is used and the iterative refinement is stopped when it is in the order of magnitude of 1. Once the kriging surrogate has been built, the two terms of the product defining the failure probability may be evaluated in parallel. The method turned out to be efficient on several application examples, two of which are presented in this paper. It can handle problems featuring a reasonably high number of random variables and multiple design points. Further work is in progress to include the proposed algorithm within a reliability-based design optimization framework. x (t-1) M p(x (t-1) ) Step 1. u (t) =U( [0, M p(x (t-1) )] ) Step 2. x (t) =U({x | u (t) ≤ M p(x)}) M p(x) x x (t) u (t) 6 « Slice » 0

Original Paper

Loading high-quality paper...

Comments & Academic Discussion

Loading comments...

Leave a Comment