An Efficient Approach for Seismic Fragility Assessment with Application to Old Reinforced Concrete...

21
Journal of Earthquake Engineering, 14:231–251, 2010 Copyright Ó A.S. Elnashai & N.N. Ambraseys ISSN: 1363-2469 print / 1559-808X online DOI: 10.1080/13632460903086028 An Efficient Approach for Seismic Fragility Assessment with Application to Old Reinforced Concrete Bridges GIANMARCO DE FELICE and RENATO GIANNINI Department of Structures, University Roma Tre, Rome, Italy A procedure for seismic fragility assessment, suitable for application to non ductile RC structures is presented, which is based on the estimate of a response surface that gives the probability of failure of the structure as a function of the random variables that affect the response. The seismic fragility or risk is then evaluated through numerical integration. The method considers different sources of uncertainty: (i) in the seismic input, through the use of different accelerograms for the dynamic analysis; (ii) in the structural response, through the use of a refined nonlinear finite element model; and (iii) in the ultimate state capacity, taking into account the different modes of failure which may occur, for which a random mechanical capacity model is available. Aiming at reducing the number of simulation analyses, the uncertainties on the seismic input and on the mechanical parameters governing the response are treated according to the response surface methodology, while the limit- state randomness is treated explicitly during the simulations. Using the proposed procedure, the seismic safety of two reinforced concrete bridges from Italian highway network, with simply supported deck and either single stem or frame piers, is evaluated. The results are expressed in terms of fragility curves as function of spectral acceleration. The obtained results highlight the influence of material randomness on reliability and the relative importance of seismic input with respect to mechanical and epistemic uncertainty. Keywords Seismic Fragility; Simulation; Bridges; Assessment; Reinforced Concrete; Response Surface Method 1. Introduction Various approaches at different levels of complexity have been proposed in the recent past for the assessment of seismic risk of structures [Pinto et al., 2004]. As regards the structural behaviour, although fairly reliable models are available in the nonlinear field, they tend not to be included in the framework of probabilistic analysis, because of the long computation times involved. As regards the model of action, the use of samples of a stochastic process underwent severe criticism, since it was considered incapable of reproducing the whole complex reality and variability of seismic motion [Naeim and Lew, 1995; Bommer and Acevedo, 2004]. The only approach capable of utilizing models, of both structure and action, which are adequately accurate and realistic, is simulation; however, the Monte Carlo method, also in the more efficient versions recently proposed (i.e., Au and Beck, 2001, 2003), proves to be exceedingly onerous from the computa- tional point of view, especially in the case of complex structures analyzed with refined models, whose randomness depends on many variables. Received 24 June 2008; accepted 15 May 2009. Address correspondence to Gianmarco de Felice, Department of Structures, University Roma Tre, via Corrado Segre, 6 - 00146 Rome, Italy; E-mail: [email protected] 231

Transcript of An Efficient Approach for Seismic Fragility Assessment with Application to Old Reinforced Concrete...

Journal of Earthquake Engineering, 14:231–251, 2010

Copyright � A.S. Elnashai & N.N. Ambraseys

ISSN: 1363-2469 print / 1559-808X online

DOI: 10.1080/13632460903086028

An Efficient Approach for Seismic FragilityAssessment with Application

to Old Reinforced Concrete Bridges

GIANMARCO DE FELICE and RENATO GIANNINI

Department of Structures, University Roma Tre, Rome, Italy

A procedure for seismic fragility assessment, suitable for application to non ductile RC structures ispresented, which is based on the estimate of a response surface that gives the probability of failureof the structure as a function of the random variables that affect the response. The seismic fragilityor risk is then evaluated through numerical integration. The method considers different sources ofuncertainty: (i) in the seismic input, through the use of different accelerograms for the dynamicanalysis; (ii) in the structural response, through the use of a refined nonlinear finite element model;and (iii) in the ultimate state capacity, taking into account the different modes of failure which mayoccur, for which a random mechanical capacity model is available. Aiming at reducing the numberof simulation analyses, the uncertainties on the seismic input and on the mechanical parametersgoverning the response are treated according to the response surface methodology, while the limit-state randomness is treated explicitly during the simulations.

Using the proposed procedure, the seismic safety of two reinforced concrete bridges fromItalian highway network, with simply supported deck and either single stem or frame piers, isevaluated. The results are expressed in terms of fragility curves as function of spectral acceleration.The obtained results highlight the influence of material randomness on reliability and the relativeimportance of seismic input with respect to mechanical and epistemic uncertainty.

Keywords Seismic Fragility; Simulation; Bridges; Assessment; Reinforced Concrete; ResponseSurface Method

1. Introduction

Various approaches at different levels of complexity have been proposed in the recent

past for the assessment of seismic risk of structures [Pinto et al., 2004]. As regards the

structural behaviour, although fairly reliable models are available in the nonlinear field,

they tend not to be included in the framework of probabilistic analysis, because of the

long computation times involved. As regards the model of action, the use of samples of a

stochastic process underwent severe criticism, since it was considered incapable of

reproducing the whole complex reality and variability of seismic motion [Naeim and

Lew, 1995; Bommer and Acevedo, 2004]. The only approach capable of utilizing models,

of both structure and action, which are adequately accurate and realistic, is simulation;

however, the Monte Carlo method, also in the more efficient versions recently proposed

(i.e., Au and Beck, 2001, 2003), proves to be exceedingly onerous from the computa-

tional point of view, especially in the case of complex structures analyzed with refined

models, whose randomness depends on many variables.

Received 24 June 2008; accepted 15 May 2009.

Address correspondence to Gianmarco de Felice, Department of Structures, University Roma Tre, via

Corrado Segre, 6 - 00146 Rome, Italy; E-mail: [email protected]

231

Dow

nloa

ded

by [

Uni

vers

ita d

egli

Stud

i Rom

a T

re]

at 0

2:33

09

Apr

il 20

15

Since rigor and accuracy are not compatible with a practically acceptable amount of

computation, the present thrust of the research is towards methods that are admittedly of

less than general nature and contain approximate assumptions, but are capable of provid-

ing satisfactory results when used in their proper context (i.e., Pinto, 2001).

An efficient method was developed in Shome et al. [1998] to calibrate the partial

safety coefficients used in the FEMA-350 guidelines, which is based on a simple

analytical formulation for the probability of failure, whose parameters are obtained on

the basis of a limited number of analyses with different earthquakes, scaled to the

intensity of the expected value of the site response spectrum, corresponding to the

fundamental period of the structure. The IDA (Incremental Dynamic Analysis) method

[Vamvatsikos and Cornell, 2001] requires several analyses for each earthquake sample

and for each intensity level. In any case, the efficiency of these methods is achieved by

neglecting the randomness of the structure’s mechanical parameters (as, for instance,

strength and stiffness of the structural elements) and their influence on the possible failure

mechanisms that may occur. This is not completely satisfactory, especially in the case of

existing reinforced concrete structures, which is the focus of the present study, since not

infrequently the degree of uncertainty on the mechanical model becomes comparable

with that related to the hazard [Ibarra and Krawinkler 2005; Liel et al., 2009]. The

variability of the basic constituent materials, concrete and steel, with respect to their

nominal, or to the in situ measured values, is of course one of the factors, but rarely the

most significant one, if compared with the uncertainties due to imperfect knowledge of

the amount and layout of reinforcement. The influence of even minor differences in the

amount of reinforcement, curtailing of the rebars, detailing of certain regions, etc., can be

paramount in determining a good or bad performance of the structure, and their con-

sideration is therefore essential for a meaningful estimate of the probability of failure. All

of the mentioned aspects are carefully controlled by present seismic codes, so that

properly designed new structures should in principle only fail according to clearly defined

mechanisms, with a negligible probability of occurrence of the undesired ones. A simple

and global failure condition for these structures, such as the maximum inter-story drift,

may be adequate. Conversely, old structures often neglect these rules and may fail

according to different failure mechanisms that have to be considered in the assessment.

To take into account multiple failure modes, that may interest one or more structural

elements, is a rather difficult task; also, because the structural models are not able to

properly reproduce all the different mechanism that may develop. Besides, the assump-

tion that failure coincide with the lost of convergence of the numerical solver, is not

satisfactory since it strongly depends on the efficiency of the integration algorithm. For

these structures, in the absence of other well-recognized criteria, it seems careful and

reasonable to assume that the global failure coincides with that of the first load-bearing

element. Even with this assumption, the check of the collapse remains uncertain to a large

extent, since the capacity formula, such as the ultimate chord rotation of a plastic hinge

[Panagiotakos and Fardis, 2001] or the shear strength of a column [Priestley et al., 1994],

are affected by a large scatter, which is of the same order of magnitude as the scatter in

the demand due to record-to-record variability. Aiming at taking into account multiple

failure modes and structural randomness, the statistic technique of the response surface

has been used recently [Franchin et al., 2003; Schotanus et al. 2004; Liel et al., 2009] for

seismic assessment. The method consists in determining, through a limited number of

targeted experiments, a simple relation of polynomial type between the random variables

and a structural response parameter measuring the performance of the structure. Such

measurement, if expressed in terms of ground motion intensity (e.g., peak ground accel-

eration or spectral acceleration at the fundamental period of the structure), may be

232 G. de Felice and R. Giannini

Dow

nloa

ded

by [

Uni

vers

ita d

egli

Stud

i Rom

a T

re]

at 0

2:33

09

Apr

il 20

15

directly compared with the seismic hazard. The limit of the Response Surface Method

(RSM) lies in the fact that it loses efficiency if the number of random factors considered

is not sufficiently small (4–6). In fact, the number of simulations necessary in order to

construct the response surface grows exponentially with the number n of variables, so that

if n is large, the number of experiments necessary becomes comparable with that required

by the Monte Carlo method.

Another procedure, defined as Effective Fragility Analysis (EFA) was recently

proposed [de Felice et al., 2002; Pinto et al., 2004], in order to achieve a significant

reduction in the number of variables on which the structural response depends. In the

versions presented earlier, the method did not have recourse to the response surface

technique; the dependence of the probability of failure on the variables considered was

expressed by means of a Taylor series expansion truncated after the second order.

In the present work, the EFA is integrated with the RSM in order to permit a better

estimate of probabilities and to further reduce the number of simulations necessary, while

taking seismic variability into account. The method is able to incorporate any significant

sources of uncertainty, related to both the action and the structure; it makes use of a full

nonlinear dynamic analysis under samples (recorded accelerograms) of the excitation

process, and calculates the probability of exceeding any specified limit-state, conditioned

to each sample. By repeating the analyses according to a prescribed experimental plan,

when varying the mechanical properties of the structure, the response surface is finally

obtained, which gives the dependence of the probability of failure of the bridge as a

function of structural random variables. The seismic risk is then estimated by means of

convolution with the probability density function of the hazard at the site and of the

structural random variables, using as a model of structural behaviour the simple poly-

nomial relation obtained earlier. A similar approach was proposed in Schotanus et al.

[2004], where the RSM was used to estimate the relation between the seismic intensity

leading to collapse and the random variables, and the fragility obtained through FORM

analyses. In the present case, however, the randomness in the capacity model is included

in the formulation and a lower number of analyses are required, since the intensity

measure is included among the fixed effect variables. The result is expressed directly

in terms of probability of collapse and then the fragility results simply by convolution of

the response surface.

Hereafter, EFA will be briefly presented (Sec. 2), then after a rapid illustration of the

RSM (Sec. 3), we will clarify how the two methods are joined together (Sec. 4).

Eventually, the approach is applied to two existing bridges of the Italian highway network

(Sec. 5).

2. Effective Fragility Analysis

Accurate seismic reliability methods for nonlinear structures require performing a number

of dynamic analyses, and this usually takes the largest part of total computing effort

required. Reducing the number of the structural analyses is therefore a goal common to

all methods. With this objective in mind, in the present approach the vector of random

structural variables is partitioned into two sub-vectors. The first one, X, is of relatively

small dimensions and contains the variables (named external) which have a significant

influence on the response, as for example, the stiffness, masses, yield strength, etc., and in

general all variables a change of whose values would reasonably require performing a

new analysis. The second sub-vector, Y, of much larger dimensions, is composed instead

of variables of local nature (named internal), which are suitable for detecting a state of

failure of the structure, but whose own state is not such as to affect significantly the

Efficient Approach for Seismic Fragility Assessment 233

Dow

nloa

ded

by [

Uni

vers

ita d

egli

Stud

i Rom

a T

re]

at 0

2:33

09

Apr

il 20

15

global dynamic response of the structure. Strictly speaking, all the geometrical and

mechanical variables influence both the strength and the dynamic response of the

structure; for instance, the shear or bending failure of a single element may affect the

global structural response. The separation between ‘‘internal’’ and ‘‘external’’ variables is

accordingly largely conventional and depends to a large extent on the structural model

adopted for evaluating the seismic demand. In fact, those parameters that do not appear in

the model clearly cannot influence the seismic demand, even if this may not be com-

pletely realistic. Moreover, if the failure condition adopted is of fragile type, it is not

necessary to analyze the effect of this collapse on the subsequent response of the

structure, since, in any case, the conventional limit state has been exceeded. Experience

shows that the separation between internal and external variables, carried out according to

the model adopted and on the base of engineering judgement, often leads to acceptable

approximations.

The mechanical properties of the materials, concrete and steel, certainly alter, not

only the resistance, but also the dynamic response. However, if we separate randomness

of the mechanical properties into two parts, one, average for the whole structure, and the

other which expresses its local fluctuations, we may venture the hypothesis that the latter,

while affecting the resistance, has only a limited influence on the response.

The expedient of separating the random variables makes it possible to greatly reduce

the number of analyses implied by the simulation process; in fact, after fixing the external

variables and the action, the structural response proves to be deterministic and thus

calculation of the probability of failure (conditioned by the values of the external

variables) may be carried out relatively rapidly.

2.1. Unconditioned Probability Computation

Upon assignment of a value to the vector X, the structural response to the j-th sample

accelerogram becomes deterministic and can be evaluated. In the general case, failure

may occur according to different mechanisms in different elements of the structure, as,

for instance, when the strength is reached in a brittle element, or the displacement

capacity is attained for a ductile element. The failure of a structural element does not

strictly imply the overall collapse, especially for capacity-designed new structures; but in

the case of existing structures, as stated previously, once the load-bearing elements and

their possible failure mechanisms are singled out, we may assume that failure of any of

the elements/mechanisms k=1,n coincides with the global failure. From reliability view-

point, the problem can still be considered as that of a series system; the structure fails as a

result of the failure of the weakest component and therefore, denoting by gk Y ; tð Þ ¼ 0 the

limit-state condition of the generic k-th element/mechanism, the condition for survival of

the structure for a given seismic input a(t) and for a given vector of external variables x,

is that no limit state has been exceeded:

S YjaðtÞ; xð Þ : Yj \k

mint

gkðY; tÞ�>0

� �¼ Yjmin

kmin

tgkðY; tÞ>0

� �: (1)

Accordingly, the condition of failure writes:

F Y aðtÞj ; xð Þ : Y [k

mint

gk Y; tð Þ � 0

����� �

¼ Y maxk

mint

gk Y; tð Þ � 0

����� �

(2)

234 G. de Felice and R. Giannini

Dow

nloa

ded

by [

Uni

vers

ita d

egli

Stud

i Rom

a T

re]

at 0

2:33

09

Apr

il 20

15

and the probability of failure, given by PF ¼ Pr F aðtÞ; X ¼ xjf g , may be calculated from

(2), contemporaneously for all the element/mechanisms, using an efficient Monte Carlo

method. Application of the Monte Carlo method is straightforward since the computation

of the limit state functions gkðY; tÞ does not require a new structural analysis; each

simulation merely requires the computation of functions gk, which often have a fairly

simple form. However if, as often happens, many of the mechanisms are not well

correlated, convergence of the procedure could be considerably slowed down; in these

cases, the efficiency of the method can be improved having recourse to the subset

simulation as proposed by Au and Beck [2001].

The computation of the probability of failure becomes more rapid when the limit

state equations can be expressed in the form:

gk Y; tð Þ ¼ ck Yð Þ � dk tð Þ; k ¼ 1; n; (3)

where ck Yð Þ represents the capacity of the structure, not dependent on time t, while dk tð Þexpresses the demand that, for given input a(t) and external variables x, is obtained from

the results of the dynamic analysis as a deterministic function of time t. In such a case, the

failure condition of the generic k-th element/mechanism becomes:

Fk : mint

ck Yð Þ � dkðtÞ � 0f g ¼ ck Yð Þ �maxt2T

dk tð Þf g � 0 ¼ ck Yð Þ � dk max � 0 (4)

and the probability of failure PFk ¼ Pr ck Yð Þ � dk max � 0ja tð Þ;X ¼ xf g may be com-

puted using the usual methods of time-invariant reliability, for example by means of

the First Order Reliability Method (FORM).

The probability of the union PF ¼ Pr [k

Fk

� �may be obtained in an approximate way

using, for example, the Ditlevsen bounds [Ditlevsen, 1979], which require evalua-

tion of the probabilities of the intersections two by two Fi \ Fj

� �, or improved

bounds based on higher-order intersections [Zang, 1993; Song and Der Kiureghian,

2003]; the latter may be used also for evaluating the reliability of more general

systems consisting of elements in series and in parallel. Alternatively, the probability

of failure may be estimated by means of the approximate evaluation of the multi-

variate normal distribution [Tang and Melchers, 1987; Pandey, 1998; Yuan and

Pandey, 2006].

The probability of failure computed in the abovementioned way is conditioned to the

value x of external r.v. X and to the particular accelerogram a(t) used in the analysis. It is

therefore necessary to make this probability no longer conditional on the external random

variables. This is achieved, in the present article, by using the Response Surface Method,

i.e., approximating the probability of failure PF(x) by means of a second order poly-

nomial function defined in an appropriate sub-region of the X-space. If n is the dimension

of the X-space, the number of coefficients needed for a full second-order polynomial is:

(1+n+n2). This number represents the required minimum number of experiments, each

one providing a value of PF, necessary for estimating the coefficients of the polynomial

function. In order to improve the smoothness of the surface and its fit to the experimental

results, however, a larger number of experiments has to be employed, which also allows

computation of the dispersion due to the lack of fit of the model, as will be detailed

below.

Efficient Approach for Seismic Fragility Assessment 235

Dow

nloa

ded

by [

Uni

vers

ita d

egli

Stud

i Rom

a T

re]

at 0

2:33

09

Apr

il 20

15

3. Response Surface Method

The response surface method is a statistic technique for determining the parameters

of a function that approximate an unknown function, Z(x), in the neighborhood of

a point x0, on the basis of the values of Z obtained by means of appropriate

experiments.

The approximating function is generally expressed in quadratic form with respect to

the variables x:

Z ¼ �0 þXm

i¼1

�ixi þXm

i¼1

Xm

k¼1

�ikxixk þ "; (5)

where e is a zero-mean random variable which takes account of the lack of fit of the

model. The (5) may be more concisely written in the form:

Z ¼ v xð Þ qþ "; (6)

where

v xð Þ ¼ 1 x1 � � � xm x21 x1x2 � � � x2

m

(7)

and q is a vector constructed with the parameters yi and yij.

The model expressed by (6) is linear in the parameters q: as a result it gives a r.v.

whose average is v(x)q while the dispersion is expressed by means of the term e By

carrying out n experiments in connection with the values xi ði ¼ 1; :::; nÞ of the variables,

we obtain n values of response zi, which may be gathered in the vector z. Accordingly, we

have:

z ¼ Vqþ e (8)

in which V is the matrix constructed with v xið Þ :

V ¼v x1ð Þ

..

.

v xnð Þ

264

375 (9)

and e is a vector of n independent realisations of the r.v. e, which measures the

discrepancy between observations and the model.

The coefficients q may be determined on the basis of (8) by applying the criterion of

maximum likelihood or, more simply, with the method of least square. Assuming that e is

a vector of Gaussian variables, both methods give the same estimate for q:

q ¼ VTV� ��1

VTz: (10)

The variance of e may be estimated on the basis of the discrepancies between the

observed values and the forecasts of the model z ¼ Vq . Given r ¼ z� z, an estimate

of the variance of e without bias is the quadratic difference:

236 G. de Felice and R. Giannini

Dow

nloa

ded

by [

Uni

vers

ita d

egli

Stud

i Rom

a T

re]

at 0

2:33

09

Apr

il 20

15

s2" ¼

rTr

n� m; (11)

where n indicates the number of experiments and m the number of parameters gathered in

the vector q. Equation (11) shows that, in order to reduce the dispersion of error, it is

necessary to carry out a considerably larger number of experiments than that of the

parameters.

The variance of Z depends, not only on the variance of e and of the matrix V, but also

on point x where the estimate is made. The design of experiments that makes the variance

of Z dependent exclusively on the distance from the central point, is defined as rotatable.

The conditions for rotatability may be found in the specialized texts [Box and Draper,

1987].

The minimum number of experiments necessary for determining q is equal to the

number of the parameters but, as has been seen, this number must be increased if the

estimate of q is to be sufficiently reliable; accordingly, a compromise must be found

between the two contrasting requirements of accuracy and economy. The choice of a

rational design for the experiments is thus an essential point in the procedure.

A complete factorial design consists of fixing two levels for each of the variables m

and then carrying out 2m experiments corresponding to all of their possible combinations.

This design does not explore the effects of the variations of a single variable at a time and

accordingly does not permit an accurate estimate of pure quadratic terms of (6).

Therefore, it is customary to combine the factorial design with a ‘‘star’’ design, forming

what is called a central composite design (Fig. 1). Denoting by x�i and xþi the two levels

for each variable in the factorial design, the central point has coordinates x0i ¼ x�i þxþi

2and

thus, utilizing the coordinate transformation xi ¼ x0i þ�i�i, in which �i ¼ xþi�x�i2

, a

standardized space is obtained in which the points of the factorial experiments are the

apices of a hypercube centred on the origin and located in the points of coordinates

�i ¼ �1. The star part of the experiment is represented by points of coordinates

�i ¼ ��; i ¼ 1; . . . ; m; �j ¼ 0; 8j 6¼ i� �

. The value of a is established by the conditions

of rotatability of the design: in the absence of repetitions � ¼ 2m=4.

(a) Factorial design

( ) ( )1,1,1,, 321±±±=ξξξ

(b) Star design

( ) ( )( )( )α

α

αξξξ

±

±

±=

,0,0

0,,0

0,0,,,321

(c) Central point

( ) ( )0,0,0,, 321=ξξξ

FIGURE 1 Diagram of a central composite design of experiments.

Efficient Approach for Seismic Fragility Assessment 237

Dow

nloa

ded

by [

Uni

vers

ita d

egli

Stud

i Rom

a T

re]

at 0

2:33

09

Apr

il 20

15

When the experiment consists of a simulation with a mathematical model, if the

variables in x completely define the model, the response is clearly deterministic and

accordingly, repetition of the experiment is useless. When on the contrary the

experiment is carried out in the field, the result always depends on a large number

of parameters, many of which cannot be controlled, so that, if the experiment is

repeated for the same values of x, in general different results are obtained. This fact

may also be considered in the case of numerical simulations, by introducing other

parameters, not included in vector x, which vary randomly between one simulation

and another.

In order to take a random effect into account, an additive random term d should

be added into Eq. (6). One possible way of estimating this term consists of repeating

the entire project of experiments for the different samples of the random factor.

However, this procedure is unattractive because of the large number of experiments

required. A more performing method consists of subdividing (the factorial part of) the

experimental design into as many blocks as samples of the random factor are

considered. The division of the experiments into blocks implies a certain loss of

information; it could be shown that, assuming random effects to be of additive type,

block design will mask the interactions between some of the variables governing the

fixed effects. The order of effects that are masked depends on the dimensions of the

blocks and on the way in which the experiments are divided into the blocks. The

division of experiments into blocks must ensure that only interactions of the higher

order are masked, as illustrated in the specialized literature [Box and Draper, 1987].

In any case, the blocks cannot be too small; therefore, if the number of blocks

necessary for estimating random effects is large compared to the number of factorial

experiments, it would be advisable to repeat twice or more the whole programme of

experiments, using different random factors in the various repetitions. If z is the

vector of the results of experiments, V is the matrix of the polynomial terms already

seen in (9) and q is the vector of the unknown parameters, the model with random

effects could be formulated as follows:

z ¼ V qþ Bd þ e (12)

in which B is a matrix n · b which applies the b random effects contained in d with the n

experiments, according to the diagram of division into blocks. Thus, if experiment i is

included in block k, this gives Bik = 1, otherwise Bik = 0. Matrix B introduces a correlation

between the terms of vector z that does not allow determination of the coefficients qindependently from the variances of d and of e. The procedure for estimating q, ��, and

�", based on the criterion of maximum likelihood, becomes more complex, requiring an

iterative method, since the problem is non-linear. More details may be found in Searle

et al. [1992].

In conclusion, for the given vector of r.v. X, once the dependence of the response

parameter Z on X has been approximately established on the basis of the response

surface, it should be possible to express Z in the form:

Z ¼ v xð Þ qþ � þ " (13)

which clearly shows that Z is a r.v. depending not only on x, but also on d and e which

represent the random factor introduced by the random effects and by the lack-of-fit of the

model, respectively.

238 G. de Felice and R. Giannini

Dow

nloa

ded

by [

Uni

vers

ita d

egli

Stud

i Rom

a T

re]

at 0

2:33

09

Apr

il 20

15

4. Combination of EFA and RSM

As we have seen in Sec. 2, the probability of failure PF conditional on given values of the

external variables X can be computed for a particular accelerogram according to the EFA

method. In order to obtain the unconditional probability of failure, a functional relation-

ship between PF and X should be established using the RSM.

As a first step utilizing the RSM in this context, we should ensure that no values of

PF less than zero or greater than one are attained. This will be done by applying to PF a

one-to-one transformation which projects the interval [0.1] over the whole real axis. The

inverse of the standard normal distribution ��1 is used, so that PF is transformed into the

reliability index b:

� ¼ ���1 PFð Þ : (14)

The subsequent considerations are accordingly applied to b instead of PF directly.

The result of each simulation depends not only on the values assigned to the external

variables, but also on the ground motion used in the analysis. In order to account for such

a dependence, an intensity measure of the earthquake, such as a peak ground acceleration

or an elastic spectral value, should be defined and included among the random variables

responsible for the fixed effect. However, this does not cover completely the effect of

ground motion variability on randomness of the response, since different time histories of

ground motion with the same earthquake intensity may produce different outputs giving

rise to different values of PF. As suggested in Franchin et al. [2003], the effect of time

history, by parity of ground motion intensity, may play the role of the random effect in

the RSM described in the previous paragraph. As illustrated heretofore, account may be

taken of this effect without increasing the number of experiments, that is, of dynamic

analyses of the structure. The design of numerical simulations to be carried out is divided

into as many blocks as the number of earthquake samples used.

4.1. Earthquake Samples and Seismic Intensity Measure

The applicability of the RSM to EFA is conditioned by the fact that the random effect

introduced by the choice of a ground motion time history should not be too great, so as

not to cover the fixed effects due to the parameters controlled in the simulation. The

dispersion introduced by the variation of time-history from sample to sample, at parity of

intensity, depends to a large extent on the chosen intensity measure. This problem is not

very important if artificially generated accelerograms as samples of a random process are

employed, since the response scatter is quite low; however, the present trend favors the

use of recorded accelerograms, rightly considered to reflect more realistically the natural

variability of ground motion for given values of the macroseismic parameters like

magnitude and distance.

Several ground motion intensity measures, aiming at minimizing the scatter of the

nonlinear response without introducing bias in the result, were proposed in the recent past

[Kurama and Farrow, 2003; Baker and Cornell, 2005; Luco and Cornell, 2007]; but,

besides its efficiency in reducing the variability in structural response, the chosen

intensity, should also ensure hazard computability. Nowadays, the hazard is normally

provided either in terms of peak ground acceleration, velocity or displacement, or in

terms of the ordinates of the elastic response spectrum. At present, a widely popular

scaling factor is the value of the spectral acceleration at or around the first natural period

Efficient Approach for Seismic Fragility Assessment 239

Dow

nloa

ded

by [

Uni

vers

ita d

egli

Stud

i Rom

a T

re]

at 0

2:33

09

Apr

il 20

15

of the structure [Shome et al., 1998]. For elastic structures dominated by the first mode,

the criterion provides almost no dispersion; but when the structure becomes non-linear,

the fundamental period increases, and efficacy is less guaranteed, especially for structures

with a small period (0.2 � 0.4 s). In this range, the spectra vary very rapidly with period,

and even a slight change in the mechanical properties (either as an effect of nonlinear

behavior, or as a result of the variation in mechanical properties) may produce a high

scatter in the response.

In order to divide the experiments into blocks, in particular as regards its factorial

part consisting of 2m analyses, the number of random effects must be a power of 2. It

would seem reasonable that the number of natural accelerograms samples to be consid-

ered, should be either 8 or 16.

The design of experiments depends to a large extent on the number of variables

contained in the vector x [Schotanus et al., 2004]. If this is high, the total number of

experiments required is high, and accordingly also a division into 8 or 16 blocks may be

accepted; if not, it would be advisable to divide the experiment into a smaller number of

blocks and to repeat it integrally, combining this with a further series of accelerograms.

For example, if the terms in x are 6, the factorial part consists of 26 = 64 experiments,

which can be subdivided into 8 blocks, each of which consists of 8 experiments. The star

part of the design could combine 6 of the 8 samples, one for each variable, while at the

central point of the design, the analyses for all the samples could be repeated. In this way,

it would be necessary to carry out 26 þ 2� 6þ 8 ¼ 84 analyses for the whole experi-

mental plan. In the case of 4 variables only, the number of factorial experiments is 24 =

16, which clearly cannot be divided into 8 blocks without confusing the fixed effects. It is

advisable in that case to repeat the whole experimental plan, dividing each one into 4

blocks, so that the total number of analyses becomes 2� 24 þ 2� 4þ 4ð Þ ¼ 56, which is

not very different from that of the preceding case.

4.2. Computation of Unconditional Probability

Once the relation between b and x has been established in explicit form, the probability of

failure PF is obtained by inverting (14):

PF X; �; "ð Þ ¼ � � vðXÞ qþ � þ "ð Þ½ � ; (15)

where d and e are Gaussian r.v. with zero-mean and known standard deviations sd and se.Vector X collects the variables controlling the fixed effects. In order to compute uncon-

ditional probability, it is therefore necessary to carry out a multidimensional convolution

between the function (15) and the probability density functions of the variables to be

saturated (d, e and the components of X). If the variables in X include seismic intensity,

we are faced with two possible alternatives. Either the probability of failure is not

saturated with respect to seismic intensity, giving rise to the fragility curve of the

structure or, a convolution with the hazard curve at the site is performed, giving rise to

the seismic risk.

Denoting by X_

the vector, sub-set of X, of the m variables to be saturated, the

probability of failure, conditional on the remaining, X^

is:

PF X^ �

¼Z

Rm

Z 1�1

Z 1�1

PF x^; x_; �; "

�fX_ x

_ �

f� �ð Þ f" "ð Þ dx_

d� d": (16)

240 G. de Felice and R. Giannini

Dow

nloa

ded

by [

Uni

vers

ita d

egli

Stud

i Rom

a T

re]

at 0

2:33

09

Apr

il 20

15

If x^

is reduced to the parameter of seismic intensity alone, (16) gives the fragility curve of

the structure. Apart from the purpose of obtaining the seismic fragility, this approach is

also useful for understanding the behavior of the structure, in particular the dependence of

its probability of failure on the random mechanical properties; therefore, it is also

possible to determine those parameters that have a larger influence on the probability

of collapse, and hence to concentrate on these latter both for their probabilistic modeling

and for the acquisition of data.

The integral (16) should be computed numerically: given the nature of the integrated

functions, one appropriate method could be the Monte Carlo method itself.

The fact that, at the end of this complex procedure, we have reverted to the MC

method may appear circular, but we should bear in mind that now the integrand function

in (16) has a simple explicit form and its evaluation requires only few operations at a

limited computational cost.

5. Application to Highway R.C. Bridges

In this section, the method described in the previous paragraph is applied to the evalua-

tion of the seismic fragility of two reinforced concrete bridges, which are part of the

Italian highway network. The selected bridges have simply supported deck with either

single stem (Fig. 3) or frame piers (Fig. 5). Since the deck is merely supported, for the

purposes of seismic analysis, the bridge is modeled as if it consisted of independent piers,

each of which is connected at the top with a mass proportional to the weight transmitted

by the deck and to a fraction of its own weight.

In the first bridge examined (Vallone del Duca), the single stem piers are modelled as

simple columns fixed at the base; in the second (Olmeta), the piers are modeled as a

frame with five pillars connected at the top by the cap beam. Fiber beam elements have

been used, in which the concrete is modelled according to the Popovic-Mander stress-

strain curve and the steel follows the Menegotto-Pinto law.

The seismic response of the bridges has been considered affected by four external

random variables, three of which are structural properties: the total mass, M, the steel

yielding strength, fy, the concrete strength, fc; and the fourth is related to the seismic

input, representing the spectral acceleration Sa of the 5% elastic response spectrum,

centred at the first natural period of the structure. It is assumed that collapse may occur

according to two alternative mechanisms: as a result of exceeding maximum rotation

ductility or shear strength capacity in one of the pillars forming the pier (which is a single

one in the former case). The capacity threshold for both mechanisms is defined using

well-known empirical formulae as described in Sec. 5.2. The internal random variables

controlling the capacity are: the plastic hinge length, the yielding and ultimate bending

curvatures of the critical sections, and the modeling errors of the capacity formula.

5.1. Earthquake Samples

The seismic input variability is accounted for as a random effect, by using the eight

natural records reported in Table 1 whose response spectra are shown in Fig. 2. The

records have been selected randomly from PEER database (http://peer.berkeley.edu/

smcat/) among the registrations on intermediate soil (B type according to EC8 classification)

with a magnitude ranging from 6.0–7.5 and a focus-to-site distance in the range 20–40 km.

The selected time histories have been subdivided into two groups of four records each,

according to the blocking scheme. The experiment plan requires therefore repeating the

28 numerical analyses for each group with a total of 56 runs.

Efficient Approach for Seismic Fragility Assessment 241

Dow

nloa

ded

by [

Uni

vers

ita d

egli

Stud

i Rom

a T

re]

at 0

2:33

09

Apr

il 20

15

TA

BL

E1

Acc

eler

og

ram

su

sed

ind

yn

amic

anal

ysi

s

IdE

arth

qu

ake

Tim

eS

tati

on

Reg

.M

R[k

m]

PG

A[g

]T

d[s

]

IG

rou

p

1C

hal

fan

tV

alle

y1

98

6/0

7/2

11

4:4

2L

akeC

row

ley

-S

heh

orn

R.

00

96

.23

60

.16

34

0

2C

ape

Men

do

cin

o1

99

2/0

4/2

51

8:0

6S

hel

ter

Co

ve

Air

po

rt0

00

7.1

33

.80

.22

93

6

3K

oca

eli,

Tu

rkey

19

99

/08

/17

Go

yn

uk

00

07

.43

5.5

0.1

32

25

4L

om

aP

riet

a1

98

9/1

0/1

80

0:0

5G

ilro

yA

rray

#7

09

06

.92

4.2

0.3

23

40

IIG

rou

p

5N

ort

hri

dg

e1

99

4/0

1/1

71

2:3

1L

A-

Ch

alo

nR

d0

70

6.7

23

.70

.22

53

1

6N

ort

hri

dg

e1

99

4/0

1/1

71

2:3

1L

A-

NF

arin

gR

d0

90

6.7

23

.90

.24

23

0

7S

anF

ern

and

o1

97

1/0

2/0

91

4:0

0C

asta

ic-

Old

Rid

ge

Ro

ute

29

16

.62

4.9

0.2

68

30

8F

riu

li,

Ital

y1

97

6/0

5/0

61

4:0

0T

olm

ezzo

00

06

.5–

0.3

51

36

242

Dow

nloa

ded

by [

Uni

vers

ita d

egli

Stud

i Rom

a T

re]

at 0

2:33

09

Apr

il 20

15

5.2. Limit States and Capacity Formulas

The capacity formulas for failure mechanisms of RC structures are generally built upon a

relatively simple mechanical model, to which elements of empirical origin are added.

These formulas are presumed to be unbiased (i.e., to provide a correct estimate of the

mean value), but they are accompanied by a significant scatter due to modelling error. For

the purpose of the present approach, the capacity of the k-th mechanism/element is

expressed in the form:

Ck ¼ CkðY; tÞ"k; (17)

where CkðY;tÞrepresents the average capacity, which is a function of internal variables

gathered in vector Y and time t, and ek is the modeling error, assumed as a log-normal

random variable with unitary mean and coefficient of variation (c.o.v.) to be selected on

the basis of the model accuracy. In this case, even if correlation exists between the basic

variables governing two different mechanisms or the same mechanism in two different

elements, the overall correlation between the capacities Ck and Cj of the two mechanisms/

elements k and j is considerably reduced by the presence of the independent r.v. "k and "j.

Flexural rotational capacity of piers is defined as:

C� ¼ �yLs

3þ �u � �y

� �Lp; (18)

where �u; �y are, respectively, ultimate and yielding curvature of the critical section, Ls,

is the distance from the point of counter-flexure, Lp ¼ Lp 1� Lp=ð2LsÞ� �

, and Lp is the

length of the plastic hinge. In our case, the plastic hinge length has been estimated as

Lp ¼ 0:08Ls þ 0:22fyd, where the second term accounts for the strain penetration effects

due to anchored longitudinal rebar (d is their diameter, fy their yielding strength in MPa).

The choice to express the limit state in terms of chord rotation rather than in terms of

section curvature is due to the aim to obtain a more stable result, not affected by possible

localization of plastic deformations induced by the integration algorithm used by the

computational model. Alternative expressions may be introduced, if deemed more appro-

priate, with the general procedure remaining unaffected. The comparison between the

FIGURE 2 Response Spectra of the eight accelerograms used in the simulations.

Efficient Approach for Seismic Fragility Assessment 243

Dow

nloa

ded

by [

Uni

vers

ita d

egli

Stud

i Rom

a T

re]

at 0

2:33

09

Apr

il 20

15

flexural capacity model and the results of experimental tests available in the literature,

show a significant scatter that suggest selecting a modelling error ey with a c.o.v. of 50%

[Panagiotakos and Fardis, 2001].

For what concerns the shear collapse, predictive equations for shear strength capacity

of concrete columns have been proposed in Priestley et al. [1994] and Kowalsky and

Priestley [2000]. The demand is represented by the shear carried by the pier and accord-

ingly the capacity is given the form:

CVðtÞ ¼ VCðtÞ þ VS þ VNðtÞf g"V ; (19)

where the average term is the sum of three main contributions:

VC ¼ 0:8AgkðtÞffiffiffiffifc

p; VS ¼ Asw

sfyD cotð30Þ; VN ¼ NðtÞ tan�ðtÞ (20)

representing, respectively, the contribution of concrete, shear reinforcement, and normal

force. In Eq. (20), the meaning of variables is the following: Ag is the shear effective area

of concrete, fc is the concrete strength, kðtÞ ¼ k ðtÞð Þ is a coefficient accounting for the

decrease of concrete contribution with increasing ductility demand m, Asw is the trans-

versal reinforcement area, s the stirrup distance, fy is the steel yielding strength, D is the

net length of concrete in tension measured in the direction of shear stress, N(t) is the axial

load, and a(t) is the angle between the compression strut and the axis of the element. The

ductility demand ðtÞ ¼ �ðtÞ=�y can be computed as a function of the chord rotation y(t)

the shear length of the pier Ls and the above defined plastic hinge length Lp :

ðtÞ ¼ �ðtÞLpþ 1� Ls

3Lp: (21)

According to a limited number of experimental/theoretical comparisons carried out by the

shear model developers, a c.o.v. equal to 0.3 is assumed for the model error eV, which is

slightly higher than the value indicated in Priestly et al. [1994] to account for the

imperfect knowledge of shear region details.

As regards the components of the sub-vector Y, i.e., of the variables that define the

attainment of a limit-state, apart from the model errors, ey and eV, the ultimate and

yielding curvatures of critical sections �u; �y have been treated as internal random

variables in order to take into account the uncertainties on the estimate of their value;

the plastic hinge length, Lp, which has to be considered as a conventional rather than a

physical quantity, is treated as an internal random variable as well. In all the above cases

randomness has been treated assuming the parameter as a best estimate value times a log-

normal fluctuation with a given c.o.v. as reported in Table 2.

TABLE 2 Internal random variables and log-normal parameters

No Random varaiable c.o.v.

1 ey Model error for flexural capacity 0.5

2 eV Model error for shear capacity 0.3

3 ewy Yielding curvature 0.1

4 ewu Ultimate curvature 0.1

5 eLp Plastic hinge length 0.3

244 G. de Felice and R. Giannini

Dow

nloa

ded

by [

Uni

vers

ita d

egli

Stud

i Rom

a T

re]

at 0

2:33

09

Apr

il 20

15

5.3. The Vallone Del Duca Viaduct

The viaduct is along the highway A16 Napoli-Canosa, in Southern Italy, between

Benevento and Avellino. The bridge has been recently retrofitted restraining the deck

gaps and introducing seismic isolators at pier caps, but has been studied in the ‘‘as-built’’

conditions.

The bridge has 6 spans, each 32 mt long with a structural scheme of a simple

supported beam. The deck is composed by three pre-stressed beams 1.92 mt high,

connected by 5 traversal links and a slab 20 cm thick and 9.54 mt wide (Fig. 3). The

r.c. piers have a single stem with a rectangular solid section 1.40 · 2.70 mt and a total

longitudinal reinforcement of 40 rebars 28 and stirrups 10/28 cm. Each pier is founded

on plinths on 4 piles with diameter 1.50 mt.

Mechanical properties of materials have been obtained from the original design

documentation and from experimental tests carried out when the retrofitted was planned:

concrete strength, on the basis of core samplings at pier base, has been evaluated on

average about �fc ¼ 45MPa and c.o.v. equal to 0.20. The reinforcing steel is classified

Aq50, a mild steel type following old Italian regulation, for which we can assume an

average yielding strength �fy ¼ 370MPa, and an ultimate average strength �fu ¼ 545MPa

with a c:o:v: ¼ 0:08.

Two piers have been taken into account, namely pier n 3, the most slender, having

16 mt height, with a fundamental period T3 = 0.98 s, and pier n 5, the squatter, with 5.5

mt height, with T5 = 0.18 s (see Fig. 3). For each pier, the procedure previously described

has been applied, consisting in 56 nonlinear analyses according to the central composite

design plan with two repetitions. Each simulation provides an estimate of the reliability

index of the pier. The coefficients of the response surface are then evaluated through the

maximum likelihood method, together with the variances of the lack of fit of the model

and the random effect.

FIGURE 3 Vallone Del Duca viaduct (A16 highway, between Benevento and

Avellino).

Efficient Approach for Seismic Fragility Assessment 245

Dow

nloa

ded

by [

Uni

vers

ita d

egli

Stud

i Rom

a T

re]

at 0

2:33

09

Apr

il 20

15

The following values are obtained: �" ¼ 0:167 and �� ¼ 0:237 for pier 3, �" ¼ 0:210

and �� ¼ 0:028 for pier 5. The random effect is relevant for pier n3, while being almost

negligible for pier n5; such a difference is due to the fact that pier n5 reaches the

collapse in shear before the attainment of yielding of steel, and therefore, by parity of

spectral acceleration, the different seismic inputs provide almost the same estimate of the

probability of failure, since the structure behaves elastically and is driven by the first

vibration mode.

In Fig. 4, the influence of the external random variables X_

¼ fc; fy;M� �T

, on the

probability of failure of each pier is depicted in the standardized space �_

i ¼X_

i�X_

i

�X_

i

, for

spectral acceleration equal to 0.4 and 0.9 g for piers 3, and 5 respectively, corresponding

to a mean return period at the site of 975 years; a probability of failure of the order of

magnitude 10�3 is obtained for average values of the mass and the material strengths. The

figure shows a quite small influence of the external variables on the probability of failure,

except for the concrete strength in pier n5, in which failure is reached essentially in

shear, with a strong correlation of fc to failure probability; on the contrary, in the more

slender pier, failure is reached essentially in flexure and therefore the concrete strength

has a smaller influence on the reliability estimate.

The fragility curves of the two piers are obtained through convolution between the

response surface and the external random variables gathered in vector X_

, and the two

random variables d and e, according to (16), as shown in Fig. 5. The figure also displays

the effect of seismic record variability and concrete strength dispersion on seismic

fragility: the plots show the fragility curve evaluated at the average value and at ± one

standard deviation of r.v. d and fc. A completely different result is obtained for the two

piers: for the taller (pier n2) the concrete strength does not affect the fragility estimate,

which is mainly affected by the random effect, whilst for the squatter pier (n5) the

–2 –1 0 1 2

ξ0.003

0.0035

0.004

0.0045

0.005

Pf concrete strengthsteel strengthtotal mass

–2 –1 0 1 2

ξ0

0.004

0.008

0.012

0.016Pf concrete strength

steel strengthtotal mass

FIGURE 4 Vallone del Duca viaduct pier n.3 with 16 mt height (left) and pier n5 with

5.5 mt height (right): dependence of the probability of failure on external random

variables fc, fy, M in standardized space, evaluated for spectral acceleration equal to

0.4 and 0.9 g, respectively, both corresponding to a mean return period at the site of 975

years.

246 G. de Felice and R. Giannini

Dow

nloa

ded

by [

Uni

vers

ita d

egli

Stud

i Rom

a T

re]

at 0

2:33

09

Apr

il 20

15

concrete strength plays a crucial role in the fragility estimate, higher than the dispersion

induced by seismic input.

5.4. The Olmeta Viaduct

The Olmeta viaduct (Fig. 6) is along A1 highway between Bologna and Florence, in

central Italy. It was built during the early 1960s, without seismic provisions. It is

composed of two separate ways, with 12 spans for a total length of 254 mt; the deck

has five pre-stressed beams with a span 21.16 mt long and a slab 9.60 mt wide excluding

sidewalks about 70 cm wide; the piers (Fig. 7) have a framed structure with five columns

with a section 100 · 60 cm each and a spacing of 2.40 mt among them, linked together by

a cap beam with a section 100 · 110 cm. The pier height varies to a maximum of 12 mt;

the deck is simply supported by piers. Pier reinforcement consists of 12 rebars 16

longitudinally and 10 stirrups every 16 cm transversally. In 1974, the viaduct was

retrofitted: the slab was rebuilt and the pier columns jacketed by a 10 cm thick concrete

layer reinforced longitudinally by 6 rebars 20 and transversally by 10 stirrups every

0 0.5 1 1.5 2

Sa (g)0.0E+000

5.0E-002

1.0E-001

1.5E-001Pf

Pier n.3Pier n.5

0 0.5 1 1.5 2

Sa (g)0.0E+000

5.0E-002

1.0E-001

1.5E-001Pf − st.dev(fc)

− st.dev(δ) average+ st.dev(δ)+ st.dev( fc)Pier n.3

Pier n.5

FIGURE 5 Seismic fragility of Vallone del Duca viaduct piers n3 and n5: fragility

curves (left) and effect of seismic record variability and concrete strength randomness on

seismic fragility (right): the plots show the fragility curve of piers valuated at the average

value and at ± one standard deviation of r.v. d and fc.

FIGURE 6 Olmeta viaduct (A1 highway, between Florence and Bologna): plan and

front view.

Efficient Approach for Seismic Fragility Assessment 247

Dow

nloa

ded

by [

Uni

vers

ita d

egli

Stud

i Rom

a T

re]

at 0

2:33

09

Apr

il 20

15

20 cm. Aiming at evaluating the effects of the reinforcement, the reliability analysis has

been carried out in both the original ‘‘as-built’’ status and the present retrofitted state; the

results are presented for pier n2 (Fig. 7), having columns of 8.98 mt high and transversal

natural period T = 1.14 s in the as built status and period T = 0.76 s in the present

retrofitted state.

For this viaduct, in the absence of design values or in-site measures, the material

properties have been estimated according to the values provided in Verderame et al.

[2001], where the data of about 600 structures commissioned by public administrations in

Italy in the year 1960 have been analyzed. For concrete strength, an average value of 28.2

MPa is estimated with a c.o.v. = 0.34, while for reinforcing steel, an average yielding

strength �fy ¼ 325MPa, and an ultimate average strength �fu ¼ 467MPa can been assumed,

corresponding to Aq42 steel class; yield and ultimate strength have been considered as

correlated log-normal R.V. with a c:o:v: ¼ 0:07. For the mass, an average value of 350

tons is assumed according to the design documentation, with a c:o:v: ¼ 0:07.

The effective fragility analysis is then applied, according to the experimental plan

previously described, comprising 56 nonlinear analyses that have been performed in the

range of spectral acceleration between 0.055 and 0.24 g, corresponding to a mean return

period at the site between 95 and 2,475 years. The coefficients of the response surface

and the standard deviations of both error term and random effect have been estimated by

means of the maximum likelihood method. The following values have been obtained:

�" ¼ 0:358, �� ¼ 0:391 and �" ¼ 0:263, �� ¼ 0:501, for pier n2 before and after

strengthening, respectively. The results display a high dispersion induced by both random

effect and lack of fit of the response surface.

As shown in Figs. 8 and 9a, a significant enhancement in reliability is provided with

rehabilitation, since the probability of failure decreases by three orders of magnitude. As

in the previous section, the influence of the external random variables fc; fy;M, on the

probability of failure is depicted in a standardized space (Fig. 8): the concrete strength fc,

which has a much higher dispersion, plays the most relevant role in both cases, before and

after strengthening.

FIGURE 7 Olmeta viaduct: pier geometry and reinforcement details.

248 G. de Felice and R. Giannini

Dow

nloa

ded

by [

Uni

vers

ita d

egli

Stud

i Rom

a T

re]

at 0

2:33

09

Apr

il 20

15

The comparison between the effects of the dispersion induced by seismic input and

structural randomness is shown in Fig. 9b, where the fragility curve of the pier before

strengthening is evaluated at the average value and at ± one standard deviation of r.v. d and

fc; in the present case, the scatter induced by sample-to-sample earthquake record variability, has

a comparable effect on seismic fragility to that induced by the randomness of concrete strength.

6. Conclusions

A procedure for evaluating the seismic fragility, suitable for application to non ductile RC

structures is presented, which makes it possible to obtain an accurate estimate, using

-2 -1 0 1 2

ξ0

0.01

0.02

0.03Pf

concrete strengthsteel strengthtotal mass

-2 -1 0 1 2

ξ0.0E+000

5.0E-005

1.0E-004

1.5E-004

2.0E-004Pf concrete strength

steel strengthtotal mass

FIGURE 8 Olmeta viaduct pier n.2 before (left) and after (right) rehabilitation: depen-

dence of the probability of failure on external random variables fc, fy, M in standardized

space, evaluated for Sa = 0.115 g.

0 0.05 0.1 0.15 0.2 0.25

Sa (g)0.00

0.04

0.08

0.12

0.16

0.20

Pf unreinforcedreinforced

0 0.05 0.1 0.15 0.2 0.25

Sa (g)0.0E+000

4.0E-002

8.0E-002

1.2E-001

1.6E-001Pf -st.dev(fc)

- st.dev.(delta) average+ st.dev.(delta)+ st.dev.(fc)

FIGURE 9 Seismic fragility of Olmeta viaduct pier n2: (a) comparison of fragility

before and after rehabilitation; (b) effect of seismic record variability and concrete

strength randomness on seismic fragility: the plots show the fragility curve evaluated at

the average value and at ± one standard deviation of r.v. d and fc.

Efficient Approach for Seismic Fragility Assessment 249

Dow

nloa

ded

by [

Uni

vers

ita d

egli

Stud

i Rom

a T

re]

at 0

2:33

09

Apr

il 20

15

refined models for the structural analysis and natural accelerograms, at low computational

cost. This method consists of an enhanced version of the Effective Fragility Analysis

(EFA) integrated with the Response Surface Method (RSM). Random variables for both

structural model and seismic action are taken into account: the random variables referring

to the structure are divided into two groups, one of which contains all of those variables

(internal) that, while influencing the resistance, have little effect on the dynamic

response; whereas the other collects the remaining structural variables (external) that

strongly affect the dynamic response. Once the external variables have been fixed, and a

ground motion at the base of the structure selected, the conditioned probability of failure

(PF) is computed using standard reliability methods (FORM). The response surface

method is then used to establish a function approximating the dependence of PF on

external variables and on seismic intensity. The accelerogram sample-to-sample varia-

bility is accounted for by introducing a random effect within the RSM, without increasing

the number of analyses. The fragility curve of the structure is finally obtained by

convolution with the probability density functions of external random variables.

The procedure is applied to assess the seismic fragility of two existing RC bridges of

the Italian highway network with simply supported deck and either single stem or frame

piers. The procedure is capable of considering several modes of failure; in the examples

considered, flexural ductility and shear capacity have been taken into account. The results

highlight the features of the method, its capability of representing the dependence of the

fragility on material randomness as well as the relative importance of seismic input with

respect to mechanical and epistemic uncertainty.

References

Au, S.-K. and Beck, J. L. [2001] ‘‘Estimation of small probabilities in high dimensions by subset

simulation,’’ Probabilistic Engineering Mechanics 16(4), 263–277.

Au, S.-K. and Beck, J. L. [2003] ‘‘Important sampling in high dimensions,’’ Structural Safety 25(2),

139–163.

Baker, J. W. and Cornell, C. A. [2005] ‘‘A vector-valued ground motion intensity measure

consisting of spectral acceleration and epsilon,’’ Earthquake Engineering and Structural

Dynamics 34(10), 1193–1217.

Bommer, J. and Acevedo, A. [2004] ‘‘The use of real earthquake accellerograms as input to

dynamic analysis,’’ Journal of Earthquake Engineering 8(1), 43–91.

Box, G. E. P. and Draper, N. R. [1987] Empirical Model-Building and Response Surfaces, John

Wiley & Sons, New York.

de Felice, G., Giannini, R., and Pinto P. E. [2002] ‘‘Probabilistic seismic assessment of existing R/C

buildings: static pushover versus dynamic analysis,’’ Proc. 12th European Conference on

Earthquake Engineering, Elsevier, London.

Ditlevsen, O. [1979] ‘‘Narrow reliability bounds for structural systems,’’ Journal of Structural

Mechanics 7(4), 453–472.

Franchin, P., Lupoi, A., and Pinto, P. E. [2003] ‘‘Seismic fragility of reinforced concrete structures

using a response surface approach,’’ Journal of Earthquake Engineering 7(NS1), 45–77.

Kowalsky, M. J. and Priestley, M. J. N. [2000] ‘‘Improved analytical model for shear strength

of circular reinforced concrete columns in seismic regions,’’ ACI Structural Journal 97(3),

388–396.

Kurama, Y. C. and Farrow, K. T. [2003] ‘‘Ground motion scaling methods for different site

conditions and structure characteristics,’’ Earthquake Engineering and Structural Dynamics

32(15), 2425–2450.

Ibarra, L. F. and Krawinkler, H. [2005] ‘‘Global collapse of frame structures under seismic

excitations,’’ PEER Report 2005/06.

250 G. de Felice and R. Giannini

Dow

nloa

ded

by [

Uni

vers

ita d

egli

Stud

i Rom

a T

re]

at 0

2:33

09

Apr

il 20

15

Liel, A. B., Haselton, C. B., Deierlein G. G., and Baker, J. W. [2009] ‘‘Incorporating modeling

uncertainties in the assessment of seismic collapse risk of buildings,’’ Structural Safety 31, 197–211

Luco, N. and Cornell, C. A. [2007] ‘‘Structure-specific scalar intensity measures for near-source

and ordinary earthquake ground motions,’’ Earthquake Spectra 23(2), 357–392.

Naeim, F. and Lew, M. [1995] ‘‘On the use of design spectra compatible time histories,’’

Earthquake Spectra 11(1), 111–127.

Panagiotakos, T. B. and Fardis, M. N. [2001] ‘‘Deformation of reinforced concrete members at

yielding and ultimate,’’ ACI Structural Journal 98(2), 135–148.

Pandey, M. D. [1998] ‘‘An effective approximation to evaluate multinormal integrals,’’ Structural

Safety 20(1), 51–67.

Pinto, P. E. [2001] ‘‘Reliability methods in earthquake engineering,’’ Progress in Structural

Engineering and Materials 3(1), 76–85.

Pinto, P. E., Giannini, R., and Franchin, P. [2004] Seismic Reliability Analysis of Structures, IUSS

Press, Pavia, Italy.

Priestley, M. J. N., Seible, F., and Calvi, G. M. [1996] Seismic Design and Retrofit of Bridges. John

Wiley & Sons, New York.

Priestley, M. J. N., Verma, R., Xiao, Y. [1994] ‘‘Seismic shear strength of reinforced concrete

columns,’’ Journal of Structural Engineering, ASCE 120(4), 2310–2329.

Schotanus, M. I. J., Franchin, P., Lupoi, A., and Pinto P. E. [2004] ‘‘Seismic fragility analysis of 3D

structures,’’ Structural Safety 26(4), 421–441.

Searle, S. R., Casella, G., and McCulloch, C. E. [1992] Variance Components. Prentice-Hall,

Englewood Cliffs, NJ.

Shome, N., Cornell, C., Bazzurro, P., and Carballo, J. [1998] ‘‘Earthquakes, records, and nonlinear

responses,’’ Earthquake Spectra 14(3), 469–500.

Song, J. and Der Kiureghian, A. [2003] ‘‘Bounds on system reliability by linear programming,’’

Journal of Engineering Mechanics, ASCE,129(6), 627–636.

Tang, L. K. and Melchers, R. E. [1987] ‘‘Dominant mechanisms in stochastic plastic frames,’’

Reliability Engineering 18(2), 101–115.

Vamvatsikos, D. and Cornell, C. A. [2001] ‘‘Incremental dynamic analysis,’’ Earthquake

Engineering and Structural Dynamics 31(3), 491–514.

Verderame, G. M. and Manfredi, G. [2001] ‘‘Le proprieta meccaniche dei calcestruzzi impiegati

nelle strutture in cemento armato realizzate negli anni ’60,’’ Atti del X Congr. Naz. L’ingegneria

Sismica in Italia, Anidis, Potenza (in Italian).

Yuan X.-X. and Pandey, M. D. [2006] ‘‘Analysis of approximations for multinormal integration in

system reliability computation,’’ Structural Safety 28(4), 361–377.

Zang, Y. C. [1993] ‘‘High-order reliability bounds for series systems and application to structural

systems,’’ Computers & Structures 46(2), 381–386.

Efficient Approach for Seismic Fragility Assessment 251

Dow

nloa

ded

by [

Uni

vers

ita d

egli

Stud

i Rom

a T

re]

at 0

2:33

09

Apr

il 20

15