combine the flexibility of machine
learning with the statistical rigor of a
Bayesian inference, enabling models to
capture both data variability and model
uncertainty.
BAYESIAN NEURAL NETWORKS
FOR UNCERTAINTY
Unlike conventional neural networks,
which produce deterministic outputs,
BNNs treat network weights as prob-
ability distributions rather than fixed
values. A BNN builds on a standard
feedforward or convolutional neural
network by replacing each fixed weight
and bias with a probability distribu-
tion. Instead of learning a single “best”
value for each parameter, a BNN learns
a posterior distribution over parame-
ters given the training data, enabling
uncertainty-aware predictions critical for
safety-critical NDE applications [75].
Typically, a normal neural network
defines a deterministic mapping:
y =f(x w) where is the vector of all
weights and biases. In a BNN, we define
a prior (w) (commonly Gaussian) over
these parameters and combine it with a
likelihood (D|w) derived from the loss
function. Bayesian inference then yields a
posterior: (w|D) ∝ p(D|w)p(w) which
is approximated via variational methods
or Monte Carlo techniques. At prediction
time, instead of a single forward pass, we
draw multiple weight samples t ~p(w|D)
and compute:
(9) p(y|x, D) ≈ 1
T
∑
t=1
T (x wt)
This ensemble of outputs provides
both a mean prediction and a variance
that quantifies both epistemic uncer-
tainty (from limited data, reflected in
weight spread) and aleatoric uncertainty
(modeled via an explicit output-noise
term, if included).
BNNs offer advantages such as inter-
pretable uncertainty metrics and the
ability to integrate physical constraints,
as seen in material property prediction
[78] and bearing remaining useful life
estimation [79]. However, BNNs face
computational challenges due to the
intractability of exact posterior inference.
Comparative studies suggest hybrid
approaches, combining BNNs with
Monte Carlo dropout or deep ensembles,
could improve robustness and scalability
in NDE workflows [80].
MONTE CARLO DROPOUT FOR MODEL
UNCERTAINTY
MC dropout is another popular
AI-driven UQ technique that helps deep
learning models estimate uncertainty
by introducing dropout at inference
time [76]. Dropout, a regularization
method, randomly deactivates a subset
of neurons during each forward pass
to prevent overfitting in deep neural
networks. During the NN training
process, each forward pass randomly
“drops” a subset of neurons at the
dropout layer, which effectively samples
from a simpler, approximate weight
distribution (w) By keeping dropout
active at test time and running several
forward passes (MC dropout), multiple
weight samples are drawn from (w) and
the outputs are averaged. This process
approximates the full Bayesian posterior
p(w|D) without needing explicit priors
or complex inference. Additionally,
uncertainty metrics such as variance or
standard deviation can be derived from
the resulting prediction distribution.
MC dropout has high adaptability
to various base NN architectures (e.g.,
CNNs [81] and ResNets [82]). It has been
widely applied for reliability assessment
in NDE tasks such as crack characteriza-
tion [83] and seismic data reconstruction
[84]. Yonekura et al. [85] also leveraged
MC dropout in a generative adversarial
network (GAN) framework for uncer-
tainty reduction in airfoil design.
DEEP ENSEMBLES FOR ROBUST UQ
The deep ensembles (DE) technique
is effective for uncertainty estimation.
Unlike single deep learning models,
deep ensembles do not require architec-
tural modifications or complex training
procedures, simplifying their implemen-
tation in practical scenarios. The tech-
nique combines outputs from multiple
independently trained deep learning
models, ensuring predictions account for
sensor noise, material inconsistencies,
and operational variations. Each network
in the ensemble learns a different
mapping from input data (e.g., ultrasonic
waveforms, thermographic images) to
outputs (e.g., discontinuity probability,
size estimate). At inference, an input is
passed through every member of the
ensemble the mean of their outputs
serves as the final prediction, while
the variance captures epistemic uncer-
tainty—that is, the model’s “disagree-
ment” about unfamiliar or ambiguous
cases.
In NDE, deep ensembles have
demonstrated success in tasks such as
crack detection and material character-
ization, where reliable uncertainty esti-
mates are critical for decision-making
[77]. Their parallelizable nature allows
for efficient deployment, though compu-
tational resource requirements remain a
limitation compared to simpler methods
like MC dropout. In practice, Pyle et al.
[86] demonstrated that DE achieved
markedly better calibration and anomaly
detection compared with MC dropout
for ultrasonic crack detection, where
DE applied spectral normalization and
residual connections, which further
sharpened their calibration and boosted
out-of-distribution detection.
Recent Advancements and Trends
in UA&UQ for NDE
With advancements in sensor technol-
ogy, AI-driven automation, and NDE 4.0,
industries are increasingly adopting
UQ. Sectors such as aerospace, nuclear
energy, oil and gas, and advanced man-
ufacturing leverage UQ to enhance dis-
continuity detection reliability, minimize
false alarms, and ensure structural
integrity.
Multi-Sensor Data Fusion
and Bayesian Inference for
Comprehensive Uncertainty
Assessment
Unlike single-sensor methods, the inte-
gration of multiple sensors in NDE
has revolutionized UQ by combining
different NDE techniques and leverag-
ing their complementary strengths to
reduce false positives and negatives,
improve spatial resolution, and increase
NDT TUTORIAL
|
UA&UQ
32
M AT E R I A L S E V A L U AT I O N • A U G U S T 2 0 2 5
learning with the statistical rigor of a
Bayesian inference, enabling models to
capture both data variability and model
uncertainty.
BAYESIAN NEURAL NETWORKS
FOR UNCERTAINTY
Unlike conventional neural networks,
which produce deterministic outputs,
BNNs treat network weights as prob-
ability distributions rather than fixed
values. A BNN builds on a standard
feedforward or convolutional neural
network by replacing each fixed weight
and bias with a probability distribu-
tion. Instead of learning a single “best”
value for each parameter, a BNN learns
a posterior distribution over parame-
ters given the training data, enabling
uncertainty-aware predictions critical for
safety-critical NDE applications [75].
Typically, a normal neural network
defines a deterministic mapping:
y =f(x w) where is the vector of all
weights and biases. In a BNN, we define
a prior (w) (commonly Gaussian) over
these parameters and combine it with a
likelihood (D|w) derived from the loss
function. Bayesian inference then yields a
posterior: (w|D) ∝ p(D|w)p(w) which
is approximated via variational methods
or Monte Carlo techniques. At prediction
time, instead of a single forward pass, we
draw multiple weight samples t ~p(w|D)
and compute:
(9) p(y|x, D) ≈ 1
T
∑
t=1
T (x wt)
This ensemble of outputs provides
both a mean prediction and a variance
that quantifies both epistemic uncer-
tainty (from limited data, reflected in
weight spread) and aleatoric uncertainty
(modeled via an explicit output-noise
term, if included).
BNNs offer advantages such as inter-
pretable uncertainty metrics and the
ability to integrate physical constraints,
as seen in material property prediction
[78] and bearing remaining useful life
estimation [79]. However, BNNs face
computational challenges due to the
intractability of exact posterior inference.
Comparative studies suggest hybrid
approaches, combining BNNs with
Monte Carlo dropout or deep ensembles,
could improve robustness and scalability
in NDE workflows [80].
MONTE CARLO DROPOUT FOR MODEL
UNCERTAINTY
MC dropout is another popular
AI-driven UQ technique that helps deep
learning models estimate uncertainty
by introducing dropout at inference
time [76]. Dropout, a regularization
method, randomly deactivates a subset
of neurons during each forward pass
to prevent overfitting in deep neural
networks. During the NN training
process, each forward pass randomly
“drops” a subset of neurons at the
dropout layer, which effectively samples
from a simpler, approximate weight
distribution (w) By keeping dropout
active at test time and running several
forward passes (MC dropout), multiple
weight samples are drawn from (w) and
the outputs are averaged. This process
approximates the full Bayesian posterior
p(w|D) without needing explicit priors
or complex inference. Additionally,
uncertainty metrics such as variance or
standard deviation can be derived from
the resulting prediction distribution.
MC dropout has high adaptability
to various base NN architectures (e.g.,
CNNs [81] and ResNets [82]). It has been
widely applied for reliability assessment
in NDE tasks such as crack characteriza-
tion [83] and seismic data reconstruction
[84]. Yonekura et al. [85] also leveraged
MC dropout in a generative adversarial
network (GAN) framework for uncer-
tainty reduction in airfoil design.
DEEP ENSEMBLES FOR ROBUST UQ
The deep ensembles (DE) technique
is effective for uncertainty estimation.
Unlike single deep learning models,
deep ensembles do not require architec-
tural modifications or complex training
procedures, simplifying their implemen-
tation in practical scenarios. The tech-
nique combines outputs from multiple
independently trained deep learning
models, ensuring predictions account for
sensor noise, material inconsistencies,
and operational variations. Each network
in the ensemble learns a different
mapping from input data (e.g., ultrasonic
waveforms, thermographic images) to
outputs (e.g., discontinuity probability,
size estimate). At inference, an input is
passed through every member of the
ensemble the mean of their outputs
serves as the final prediction, while
the variance captures epistemic uncer-
tainty—that is, the model’s “disagree-
ment” about unfamiliar or ambiguous
cases.
In NDE, deep ensembles have
demonstrated success in tasks such as
crack detection and material character-
ization, where reliable uncertainty esti-
mates are critical for decision-making
[77]. Their parallelizable nature allows
for efficient deployment, though compu-
tational resource requirements remain a
limitation compared to simpler methods
like MC dropout. In practice, Pyle et al.
[86] demonstrated that DE achieved
markedly better calibration and anomaly
detection compared with MC dropout
for ultrasonic crack detection, where
DE applied spectral normalization and
residual connections, which further
sharpened their calibration and boosted
out-of-distribution detection.
Recent Advancements and Trends
in UA&UQ for NDE
With advancements in sensor technol-
ogy, AI-driven automation, and NDE 4.0,
industries are increasingly adopting
UQ. Sectors such as aerospace, nuclear
energy, oil and gas, and advanced man-
ufacturing leverage UQ to enhance dis-
continuity detection reliability, minimize
false alarms, and ensure structural
integrity.
Multi-Sensor Data Fusion
and Bayesian Inference for
Comprehensive Uncertainty
Assessment
Unlike single-sensor methods, the inte-
gration of multiple sensors in NDE
has revolutionized UQ by combining
different NDE techniques and leverag-
ing their complementary strengths to
reduce false positives and negatives,
improve spatial resolution, and increase
NDT TUTORIAL
|
UA&UQ
32
M AT E R I A L S E V A L U AT I O N • A U G U S T 2 0 2 5