useful to describe if the data was collected on Material A or
Material B. Including this variable in the model can improve
the model’s ability to represent the data, and a model like
Equation 20 may be a significantly better fit. The x​ term is
useful in describing how the discontinuity size and material
type interact.
(20)​ y = β0​​ + β​1​​x + β2​​m + β​3​​mx​
(21)​ m = {​1​
0​ if Material A​
if Material B
The linear model including a categorical material variable
(​ could be rewritten as shown in Equation 22, where terms
involving the discontinuity size are kept separate. In this case,
the results will lead to two POD curves, one for each material.
(22)​ = ​ˆ 0​​​ +​ˆ1​​​x,​​​ ​​
​ˆ0​​​ =​​ˆ0​​​ +​​ˆ2​​​m,​ ​​
​ˆ1​​​x = (​ˆ1​​​ +​​ˆ 3​​​ m)​x​
For any linear model that can be written as ​​y​ = α​0​​ + α​1​​x​
where ​​ 0​​​ and ​​ 1​​​ are sums of parameters not involving discon-
tinuity size, the methodology in this section will apply. The
approach works for continuous variables as well as categori-
cal ones, but every level of a continuous variable will have a
different POD curve. For instance, if an additional continuous
variable, such as gain, is included in the experiment, there
would be a different POD curve for each value of gain and
each material.
Converting Equation 22 to probability is possible by
defining Equation 23 and using Equation 24. Note that if ​​
(​ˆ​​​ 1 +​​ ˆ 3​​​ m)​ =0​ then the model in Equation 20 fails the overall
model test (i.e., the model is not statistically significant), indi-
cating no relationship between the NDE response and discon-
tinuity size, which defeats the purpose of a POD study.
(23)​ ​​ˆd​​​ po = ydec​​ _​−​​ˆ0​​​
​ˆ1​​​
= y​dec​​__________​)m​​​​2ˆ​​+​​​​0ˆ​​​(​−
(​​ˆ​1​​​ +​​ˆ​3​​​m)​ ​​
​ˆd​​​ σpo = ​ˆε​​​ ​​
​ˆ1​​​​
= ​ˆ​ε​​​ _
(​​ˆ​1​​​ +​​ˆ​3​​​m)​​​
(24)​ POD(x)​ =Φ(​​ ˆ ydec​​​​ _
​ˆε​​​
)​​
=Φ(​ x (ydec​​​ ​​ˆ0​​​)​ ​​ˆ1​​​ __________
​ˆε​​​ ​ˆi​​​ 1
)​ =Φ​ (​
x ​​ˆd​​​ po _
​ˆd​​​ σpo )​​
The variance-covariance matrix ​​ (​​​ˆ)​​​​ from Equation 11 had
a size of × 3​ since the model estimated three parameters: ​​β​0​​​​, ​​
ˆ 1​​​​ and ​​​σ​ε​​​​ However, the model in Equation 20 has five param-
eters: ​​β​0​​​​ ​​β​1​​​​ ​​β​2​​​​ ​​β​3​​​​ and the implicit standard deviation term, ​​
ˆ ε​​​​ the estimated variance-covariance matrix for this model is
shown in Equation 25:
Recall that the goal is to estimate the × 2​ ​​Vpod​​ = U​​T​V(​ˆ)​U​​​
matrix. If has dimensions of × 5​ then should have
dimensions of × 2​ This matrix, in terms of ​​ is given in
Equation 26. Then, using Equations 25 and 26, the ​​ pod​​ =
U​​T​V(​ˆ)​U​ matrix can be calculated using linear algebra.
An alternative method to obtain the same result is to use
the standard definition from Equation 16 in terms of ​​ k​​​ then
use Equation 22 relating ​​ k​​​ and ​​ k​​​ .
(27) Var​[​​ˆd​​​]​​ po ​1
​ˆ1​​​​​2​​​(​
Var​[ˆ0​​​]​ ​​ +2ˆd​​​Cov​[​ˆ ​​ po 0​​​ , ​ˆ 1​​​ ]​​
+ ​ˆd​​​​​2​Var​[​ˆ po 1​​​ ]​ )​
,​
Var​[ˆd​​​]​ ​​ po = ​1
​ˆ1​​​​​2​​​(​
Var​[​​ˆ​ε​​​]​ 2​​ˆd​​​Cov​[​​ˆ1​​​,​​ˆ​ε​​​]​​​o​p
+​​​ˆd​​​​​2​Var​[​ˆ1​​​]​ po )​
,
Cov​[ˆd​​​,​ˆd​​​]​​ ​​ po po ​1
​ˆ1​​​​​2​​​(​
​ˆd​​​​Cov​[​​ˆ0​​​,​​ˆ1​​​]​ σpo Cov​[​​ˆ0​​​,​​ˆ​ε​​​]​−​​
​ˆd​​​Cov​[ˆ1​​​,​ˆ​ε​​​]​ μpo ​​ +​ˆd​​​​​ˆd​​​Var​[ˆ1​​​]​​)​​​​​o​po​p​
A U G U S T 2 0 2 5 M AT E R I A L S E V A L U AT I O N 61
The variance and covariance estimates for each param-
eter are obtained when fitting the model (see Equation 25).
However, Equation 27 is written with sums of variances and
covariances. Statistics theory [19] provides the equations (see
Appendix A: Sums of Variances and Covariances). All the
sums needed for Equation 27 are provided in Equation 28.
Also note that, since is a scalar, ​​ ar(m​β​i​​)​ = m​​ 2​ Var​(​​​β​i​​​)​​​​ and
Cov(​β​i​​,m​β​j​​)​ =mCov(​β​i​​,​β​j​​)​​.​
(28)​ Var​[​ˆ 0​​​ ]​ =Var[​ˆ 0​​​ +m​​ˆ 2​​​ ]​​ Var[​​ˆ0​​​]​ + m​​2​Var[​​ˆ2​​​]​ +2mCov[​​ˆ 0​​​ ,​​ˆ 2​​​ ]​​
Var​[​ˆ1​​​]​ =Var[​​ˆ1​​​ +m​ˆ3​​​]​​ Var[​​ˆ 1​​​ ]​ +mVar[​​ˆ3​​​]​ +2mCov[​​ˆ 1​​​ ,​​ˆ3​​​]​​
Var[​ˆ​ε​​​]​ =Var[​​ˆ​ε​​​]​​
Cov​[ˆ0​​​,​ˆ1​​​]​ ​​ =Cov[​​ˆ0​​​ +m​ˆ2​​​,​​ˆ1​​​ +m​​ˆ3​​​]​​ Cov[​​ˆ 0​​​ ,​​ˆ1​​​]​ +mCov[​​ˆ 0​​​ ,​​ˆ3​​​]​​
+mCov[​​ˆ2​​​,​​ˆ 1​​​ ]​ + m​​2​Cov[​​ˆ 2​​​ ,​​ˆ3​​​]​​
Cov​[​ˆ0​​​,​​ˆ​ε​​​]​ =Cov[​ˆ 0​​​ +m​​ˆ 2​​​ ,​​ˆε​​​]​​ =Cov[​ˆ 0​​​ ,​​ˆε​​​]​ +mCov[​ˆ 2​​​ ,​​ˆε​​​]​​​
Cov​[​ˆ1​​​,​​ˆ​ε​​​]​ =Cov[​ˆ 1​​​ +m​​ˆ 3​​​ ,​​ˆε​​​]​​ Cov[​​ˆ 1​​​ ,​​ˆε​​​]​ +mCov[​​ˆ 3​​​ ,​​ˆε​​​]​​
Then, to calculate ​​ 90/95​​​ one can follow the steps in
Equations 16–19, using the ​​ pod​​​ values from Equation 27.
3.2. Polynomial Alternatives to the Simple Linear
Model Setup
The conversion from a linear model to a POD curve is nontriv-
ial for any linear model that extends beyond the simple linear
model. This section describes how to perform POD for a linear
model defined by an invertible function of discontinuity size ,
using a change-of-variable methodology.
First, fit a linear model relating to ​​ (​​x)​​​​ where is a dif-
ferentiable and invertible function. For example, consider
a linear model where is a second-order polynomial.
Then ​​ =f​(​​x​)​​ = α​0​​ + α​1​​x + α​2​​​x​​ 2​​​ and the estimated model
would be ​y​ =​​ ˆ 0​​​ +​​ ˆ 1​​​ x + α​2​​​x​​ 2​​ The next step in POD estima-
tion requires writing ​​​y​l​​​ =​​ ˆ 0​​​ +​​ ˆ 1​​​ x​i​​ +​​ ˆ 2​​​ x​i​​​​2​​ in terms of prob-
ability (as in Equation 6). However, as shown in Equation 29,
the quadratic form does not allow for to be separated into the
form of ([​​x ˆ μ​pod​​​​]​​​/ˆ​​​)​​ σ​pod so that ​ˆ​​​​ μ​pod and ​ˆ​​​​​2​​ σ​pod are able to
be estimated separately.
(29)​ POD(a)​ =Φ(​​ ˆ ydec​​ _
​ˆε​​​
)​​​=Φ(​​ ​ˆ​1​​​x +​​ˆ2​​​​x​​2​ (​​ydec​​ ​​ˆ​0​​​)​ _____________
​ˆ​ε​​​
)​​
Define a new variable, ​​ =f(​​x)​​​​ to facilitate a change of
variables so that a separable version of Equation 29 is possible.
However, to use the variable we need to know its mean and
variance. The next section will show how to estimate them.
3.2.1. MOMENTS OF Y: EXPECTED VALUE AND VARIANCE
Consider the fitted model, ​f(​​ ˆ )​​ which describes the mean
behavior of ​​ (​​x)​​​​ with respect to Then, a good estimate of the
expected value of comes from estimating ​​f(​​ ˆ )​​ In practice,
one can define the function = β​0​z​​​​ + β​1​z​​​​z =f(x)​​ and fit a linear
model. As shown in Equation 30, the form of the expected
value is identical to the simple linear case, except is replaced
by so ​ˆ​​​ μ​pod = (​​ydec​​ ​β​0​z​​​​​)​​​/​ˆ​ ˆ β​1z​​​​​​ Since ​​
[​​ z​
]​​ = ˆ f( )​ then ​​
ˆ β​0​z​​​​​ 0​ and ​​ˆ​ β​1z​​​​​ 1​ (which are approximate only because of
potential rounding errors during estimation), so ​​​μ​pod​​​ ˆ y​dec​​​.
(30)​ E[y]​ =E[​β​0z​​​​ + β​1z​​​​z]​ = β​0z​​​​ + β​1z​​​​E[z]​​​​​
Next, the variance of is needed. A naive approach would
use the residual variance ​​​σ​z​​​​​2​​ provided during model fitting.
However, since the variable depends on the predicted values
(i.e., the mean prediction) of ​f( ˆ )​ this may underestimate the
variance of Equation 31 provides the variance of in terms
of ​​ (​​x)​​​​:​
(31)​ Var[y]​ =Var​[​β0z​​​​ + β1z​​​​f(x)​]​​​​​=β1z​​​​​​2​Var​[f(x)​]​​​​​​​​​​
Next, the variance of ​​ (​​x)​​​​ is needed. According to the Delta
method, for a differentiable function with derivative ​​ (T)​
where [x]​ =µ​ then ar​[f(X)​]​ = i=0​ fi​(​T)​​ (μ)​​​2​ Var[​x​i​​]​​ +
2​∑i j​​​ fi​(​T)​​​​fj​​(​T)​​Cov[​x​i​​,​x​j​​]​​ This requires taking derivatives of ​​ (​​x)​​​​​
with respect to each parameter (​​ i​​​ in ​​ (​​x)​​​​ and similarly, the
variance and covariance of each parameter.
Besides differentiability of the regression function (which
is assured for the models we propose), the Delta method also
assumes asymptotic normality for the statistic of interest. Since
the statistic of interest is the predicted response, when normal-
ity is not met, transformations found through methods such
as the Box-Cox transformation can help justify the use of the
Delta method for variance estimation.
For the example where =f(x)​ =​​ ˆ 0​​​ +​​ ˆ 1​​​ x +​​ ˆ 2​​​ x​​ 2​​ ,
the necessary first derivatives are given in Equation 32. All
higher-order derivatives are zero in this case, so they do not
contribute to the sum. Note, however, that in some functions,
these higher-order derivatives may need to be included. We
assume that normality is either directly met in the model or
through the use of a transform such as Box-Cox.
(32)​ (​
f(x)​ _
​​ˆ​0​​​
)​ =1, (​
f(x)​ _
​​ˆ1​​​
)​ =x,​​ (​
f(x)​ _
​​ˆ2​​​
)​ = x​​2​​
Thus, using the Delta method, the variance of ​​ (​​x)​​​​ is given
in Equation 33:
(33)​ Var​[f(x)​]=​∑fi​(​T)​​(μ)​​​2​Var[​xi​​]​​ ​​
i=0​
2
+2​∑fi​
i j​
(T)​ Cov[​xi​​,​xj​​]​​​​​​
=(​_​ f(x)
​​ˆ0​​​
)​​​
2​ Var[​α0​​]​ + (​
f(x)​ _
​​ˆ1​​​
)​​​
2​ Var[​α​1​​]​​​+(​_​ f(x)
​​ˆ2​​​
)​​​
2​ Var[​α2​​]​​​​
+2(​_​ f(x)
​​ˆ​0​​​
f(x)​ _
​​ˆ1​​​
Cov[​α1​​,​α​0​​])​​ 2(​_​ f(x)
​​ˆ0​​​
f(x)​​ _
​​ˆ​2​​​
Cov[​α2​​,​α0​​])​​​​​​
+2​ (​
f(x)​ _
​​ˆ​1​​​
f(x)​ _
​​ˆ2​​​
Cov[​α2​​,​α​1​​]​ )​​​ =Var[​α​0​​]​ + x​​2​Var[​α1​​]​ + x​​4​Var[​α​2​​]​​​
+2xCov[​α​1​​,​α0​​]​ +2 x​​2​Cov[​α2​​,​α0​​]​ +2 x​​3​Cov[​α2​​,​α​1​​]​​​​
ME
|
PODMODELING
62
M AT E R I A L S E V A L U AT I O N A U G U S T 2 0 2 5
Previous Page Next Page