Analysis of Uncertainty and Variability in Finite Element Computational Models for Biomedical Engineering: Characterization and Propagation
- 1Simbiosys Group, Universitat Pompeu Fabra, Barcelona, Spain
- 2International Center for Numerical Methods in Engineering (CIMNE), Barcelona, Spain
- 3Catalan Institution for Research and Advanced Studies (ICREA), Barcelona, Spain
Computational modeling has become a powerful tool in biomedical engineering thanks to its potential to simulate coupled systems. However, real parameters are usually not accurately known, and variability is inherent in living organisms. To cope with this, probabilistic tools, statistical analysis and stochastic approaches have been used. This article aims to review the analysis of uncertainty and variability in the context of finite element modeling in biomedical engineering. Characterization techniques and propagation methods are presented, as well as examples of their applications in biomedical finite element simulations. Uncertainty propagation methods, both non-intrusive and intrusive, are described. Finally, pros and cons of the different approaches and their use in the scientific community are presented. This leads us to identify future directions for research and methodological development of uncertainty modeling in biomedical engineering.
In the last decades, the finite element (FE) method has gained popularity within bioengineering problems, becoming a standard procedure in the implementation of biomedical systems. Most often, simulations only take into account predefined conditions, which consider a set of fixed values, carrying out a deterministic study. However, these conditions present uncertainties, which are omnipresent in all modeling approaches. In the bioengineering field, considering uncertainties has a high impact in the output due to their clinical relevance.
Uncertainty quantification is one main aspect of uncertainty management and is the name given to a set of techniques to deal with uncertainties. It consists of several aspects: (1) uncertainty identification, the first step to identify the source of variability uncertainty, (2) uncertainty categorization, which describes the kind of uncertainty regarding its source, (3) uncertainty characterization, which deals with the statistical description of the input, (4) uncertainty propagation, which analyzes the effects of the input variability on the output, and, finally, (5) uncertainty analysis, which assesses the variability effects and their sources.
Quantifying and dealing with such uncertainties has been explored in biomedical modeling and computational simulation to support diagnosis (González Ballester et al., 2000, 2002, 2004; Kohara et al., 2011; Schmid et al., 2011), pre and post-operative decisions (Talib et al., 2005; Rajamani et al., 2007; Bode et al., 2012), and therapy treatments (Belenguer et al., 2006; Zheng et al., 2007, 2009; Bednarz et al., 2011; Cao et al., 2012; Rüegsegger et al., 2012). Uncertainties influence in both the computation of the output and its interpretation. Thus, the aim of uncertainty quantification and propagation analysis is to determine the impact of uncertainty and variability on the response of the model. While probabilistic and non-probabilistic methods can be used to obtain this response, this article focuses mainly on probabilistic analysis. Applications of probabilistic analysis in bioengineering range from biomechanical analysis of implants across a given population (Belenguer et al., 2006; Kozic et al., 2010) to the study of morphometrical changes in body structures to diagnose diseases (Linguraru et al., 2006, 2008).
In computational modeling, uncertainty and variability can have many different sources. Some of them are related to the model itself, such as the ones that come from simplifications in the model, while others are related to the input parameters. There are also numerical approximation uncertainties that are related to the approximation of the numerical solution and model-form uncertainties that result from all the assumptions, abstractions, and mathematical formulation.
Uncertainty can be categorized into aleatory and epistemic. Aleatory uncertainty is related to the intrinsic variation of the system caused by model input parameters, which can lead to an unpredictable variation in the outcomes (Roy and Oberkampf, 2010; Lin et al., 2012). It can also be referred to as “variability,” “irreducible,” “stochastic,” or “random uncertainty” and is usually characterized using probabilistic approaches due to its random nature. In contrast, epistemic uncertainty stems from the lack of knowledge of the real system behavior. It is related to the approximation of the numerical solution. The error in these assumptions can be due to a wrong model interpretation or reductionist hypotheses of the governing equations. Theoretically, it can be overcome by defining a better physical-mathematical model or considering more data by carrying out thorough measurements. It is also known as “subjective” or “reducible uncertainty.” Epistemic uncertainty is not well characterized by probabilistic approaches, as it relates to the lack of knowledge, rather than statistical information (Lin et al., 2012).
Classical probabilistic analysis predicts the output according to the uncertainty in input data, rather than from a predefined set of inputs like in deterministic studies. Uncertainty analysis using probabilistic descriptions of model inputs can be employed to describe the probabilistic distribution of the model outputs. For this, methods are needed to propagate the extensive range of uncertainties present in the model, such as those coming from the model geometry or material properties (Figure 1). These propagation methods can be divided into intrusive and non-intrusive (Lin et al., 2012). Intrusive models require reformulating the governing equation of the model, while non-intrusive methods use ensembles of simulations created by a sampling scheme. Non-intrusive methods are often preferred since commercial FE solvers can be used as black-boxes. On the other hand, intrusive approaches are usually in-house codes and need to be implemented according to specific conditions of the physical problem and constitutive model. Figure 2 gives an overview of the pipeline for the analysis of uncertainty and variability in the context of computational FE simulations.
Figure 1. General workflow analysis of input uncertainty and variability in computational models: (1) identification of all sources of uncertainty, (2) characterization of model input uncertainty, and (3) propagation of input uncertainty through the model.
Figure 2. Probabilistic design of a cemented hip prosthesis, adapted from (Kayabasi and Ekici, 2008). (A) Geometry of a hip prosthesis and variable design parameters. (B) New design parameters which reduce the probability of failure.
This article reviews the methods and approaches to analyze input aleatory uncertainties in biomedical models for FE simulations. Characterization and propagation methods are presented in Sections 2 and 3, respectively. Sampling techniques and non-intrusive methods, such as random sampling and stochastic collocation, are discussed among others. Also, intrusive methods in the stochastic finite element framework are presented. Finally, Section 4 discusses current limitations and future perspectives for research.
2. Uncertainty and Variability in Computational Models: Identification and Characterization
The first step in the analysis of uncertainty is the identification and characterization of the different sources that can affect the outcome of the system. Such sources may have their origin in model inputs (Roy and Oberkampf, 2010), which include parameters such as material properties, geometry, and boundary conditions. Different techniques can be used to model input parameter variations. The simplest way consists in defining a set of values for each parameter, where no statistical distribution function is defined (section 2.2.1). Another approach is to consider a statistical distribution of the input parameters, such as a Gaussian or lognormal distributions, among others (section 2.2.2). Finally, a more accurate approach is to create statistical models where the whole variability across the selected population is considered (section 2.2.3).
2.1. Uncertainty Source Identification
Three main uncertainty sources that affect the inputs of computational models are: geometry, material properties, and boundary conditions. Most authors have focused on simplified FE models considering exclusively uncertainties in material properties and/or loading conditions (Nicolella et al., 2006; Easley et al., 2007; Dopico-González et al., 2009; Berthaume et al., 2012). However, others are aware that model geometry plays a key role in the behavior of anatomical structures, since morphometrical variations have important effects on the output results (Easley et al., 2007; Noailly et al., 2007; Kozic et al., 2010; Mousavi et al., 2012; Niemeyer et al., 2012; Rao et al., 2013). Considering different sources simultaneously allows coping with a wider range of uncertainties and hence provides a better accuracy in the model output.
2.2. Source Variability Characterization
2.2.1. Discrete Sets of Values
The simplest approach to characterize input parameter variation is to use combinations from a predefined set of fixed values of the input data. The best way to do that is by using design of experiments (DOE) (section 3.1.1), which explore the simultaneous effect of multiple input variables (factors) on the output (response) (Yoon et al., 2005; Chen et al., 2011). DOE determines discrete values within a range for each factor, called levels. Assuming there are f factors, each one with l levels, the number of runs to be evaluated is equal to lf, increasing exponentially. Thus, depending on the amount of levels of each factor, the space of possible combinations varies. For instance, from 324 runs for the cervical cage evaluation (Yang et al., 2007) to 1024 for the pressure-relieving foot orthosis (Cheung and Zhang, 2008) for a full DOE.
The applications of these methods have not been extensively explored in bioengineering problems. Nonetheless, specific literature regarding product design and sensitivity analysis can be found. Malandrino et al. (2009) carried out a sensitivity analysis of an intervertebral disc FE model to analyze the influence of tissue material properties. Yang et al. (2007) used DOE to optimize a cervical ring cage by FE analysis and Cheung and Zhang (2008) a pressure-relieving foot orthosis. Both identified the sensitivity of the design factors and tested the effect of the structural combinations of FE models created by DOE.
Although DOE methods have been mainly used for product design optimization, they provide a simple approach to generate a group of instances that cover all combinations of parameters previously defined. Thus, the population of models created could be used also to account for the uncertainty derived by geometrical tolerances of a device or for anatomical variability, for instance, which is unfeasible to capture on a deterministic model.
2.2.2. Statistical Distribution of Parameters
For source variability characterization, the most widespread approach considers a statistical distribution of the input random parameters (Easley et al., 2007). Biomechanical analyses that consider model input uncertainty using statistical distribution are usually focused on the study of joint behavior or prostheses. For instance, one can add information about the variability of geometrical parameters to analyze failure probability of a hip prosthesis (Bah and Browne, 2009), to validate probabilistic models (Mehrez and Browne, 2012), or to propose an optimized design (Kayabasi and Ekici, 2008) (Figure 2). The effects in a hip replacement in terms of strain fields generated by the combination of applied loads, bone and implant stiffness (Viceconti et al., 2006; Dopico-González et al., 2009), and bone-implant angle (Dopico-González et al., 2010a,b) have also been evaluated.
Other issues to take into account are the effect of the alignment variability (Laz et al., 2006a), loading profiles, friction, and wear coefficient (Laz et al., 2006b; Easley et al., 2007; Pal et al., 2008). To characterize the relevant parameters to be considered in a knee implant design, Fitzpatrick et al. (2010) identified the relationship between design parameters, surgical alignment, and loading conditions. Also accounting for component alignment variability, Rohlmann et al. (2009) found a strong correlation between alignment and gap size of an artificial disc and joint forces in the lumbar spine.
Although spine, knee or hip related studies are the most common ones, due to their clinical relevance, other studies can also be found. For example, related to hand representation, Valero-Cuevas et al. (2003) accounted for 50 musculoskeletal parameters to create a realistic model of the thumb. Berthaume et al. (2012) focused on a craniofacial FE model, where uncertainties in material properties were considered to determine the stress and strain output variability.
Some authors used also anthropological parameters, such as body segments, to perform inverse dynamic studies (Holden and Stanhope, 1998; Rao et al., 2006; Langenderfer et al., 2008). In a similar way, others have investigated the knee kinematic variability due to the uncertainty in anatomical landmark locations (Morton et al., 2007).
Other works involved the study of shear strength and fatigue of bone cement (Nicolella et al., 2006; Jeffers et al., 2007) or fracture risk prediction of the bone (Laz et al., 2007). In a more detailed way, Grasa et al. (2005) created a damage model for bone cement with the hip joint contact force as a random variable. Using this probabilistic damage model, Pérez et al. (2006) predicted the failure probability of the stem–cement interface accounting for the non-linearity of the cement degradation, rather than considering it linear (Grasa et al., 2005).
All these methods represented random parameters as Gaussian or lognormal distributions, defined by experimental results or from the literature. However, these variables or features could be misrepresented due to an imprecise characterization for not considering the dependency between them. Thus, more accurate models to represent the variability found in the population under study, as those described in the next section, are needed.
2.2.3. Statistical Models
A more detailed approach to characterize source variability creates statistical models able to capture the variation of the input data. Based on Cootes’ theory (Cootes and Taylor, 1995), statistical shape models (SSM) are capable to capture the complex geometric variability of a large number of shapes within a class or population. Shapes are represented by a set of points called landmarks, forming a point distribution model. Through principal component analysis (PCA), the dimensionality of the object is reduced from a large set of correlated variables to a compact set of uncorrelated ones (Rajamani et al., 2007; Zheng et al., 2009, 2010; Baldwin et al., 2010; Bredbenner et al., 2010; Bonaretti et al., 2011; Fitzpatrick et al., 2011a). Then, the SSM provides the mean position of the points as well as the modes of variation that capture the main variability of the set of shapes. Further work adds gray-level intensity information to yield a so-called statistical appearance or density model (Cootes and Taylor, 2001). Then, gray-level value and density properties (e.g., of the bone tissue) can be related and the variability of the tissue material properties can be represented (Belenguer et al., 2006; Bryan et al., 2010; Bonaretti et al., 2011; Nicolella and Bredbenner, 2012).
This approach has become a robust and powerful tool in bioengineering to extract the shape and density variability of the morphometrical data among patients. Researchers have applied SSM to a wide range of applications. Some studies focused on relating geometry to diseases, for instance the risk of developing osteoarthritis (Bredbenner et al., 2010) or geometry to kinematics (Smoger et al., 2015), creating thus a statistical shape–function model. Others considered the study of implant behavior or even optimization of implant designs across the population (see Figure 3) (Belenguer et al., 2006; Kozic et al., 2010). For instance, Belenguer et al. (Belenguer et al., 2006) presented a framework for orthopedic implant design using both statistical shape and intensity models. They aimed at finding the implant shape to fit the maximum percentage of patients across a given population. Also considering both shape and intensity statistical model, Nicolella and Bredbenner (2012) developed a parametric representation of the proximal femur to study how the structure affects bone strength.
Figure 3. (A) Shape space defined by the PCA of a population of human femur. (B) Implant-fitting method applied for tibial orthopedic implants. Adapted from Kozic et al. (2010).
The deformation behavior of an anatomical structure due to a surgical procedure or a medical test can be considered too. Ashraf et al. (2002) created a statistical estimation tool for intra-operative prostate deformation. The main modes of variation were extracted from the deformation of the prostate during transrectal ultrasound (TRUS) probe insertion, carried out computationally with different insertion parameters. Khalaji et al. (2008) estimated in real time the tissue deformation of prostate during TRUS by statistical models and neural networks. Similarly, Mousavi et al. (2012) compared multilayer neural networks and PCA to better fit the relationship between the prostate shape and FE deformation and stress fields.
The methods described above create a FE mesh for each sample of the point distribution model. Another approach is to employ a mesh-morphing framework (Bah et al., 2009), where a reference template mesh is morphed onto the remaining surfaces of the data set. In particular, to consider the material properties of the bone, Bryan et al. (2009) assigned to each node a value corresponding to the gray intensity extracted from a computer tomography scan. The statistical model was built by morphing a template tetrahedral mesh by elastic surface matching registration and mesh morphing (Bryan et al., 2010).
Baldwin et al. (2010) developed an integrated segmentation approach based on mesh morphing (Figure 4), where a template was morphed according to the manually manipulated perimeter of anatomical structures by a graphical user interface. Once the meshes from the subjects were created, a PCA was applied to capture the variability in the control points that define each structure of the joint, including size, shape changes, and positional alignment. Using this mesh-morphing approach, Fitzpatrick et al. (2011a,b) analyzed the changes in knee kinematics and contact mechanics due to articular alignment. These studies were followed by Rao et al. (2013) and Smoger et al. (2015) who evaluated and related knee mechanics and kinematics to both shape and alignment variability (Figure 5). Malandrino et al. (2015) used a similar approach to morph a generic mesh of lumbar vertebrae to subject-specific geometries and assess the interaction among biomechanical and biophysical processes and intervertebral disc condition. Despite the promising results obtained with these approaches, Bonaretti et al. (2011) found that a better mesh quality to generate a femur mesh – regarding the homogeneity of edge ratio and minimum angle – was obtained using image-based models, rather than by using a mesh-morphing approach to conduct a strain analysis at femoral neck. However, the FE results using both approaches slightly differed, and the choice of one specific solution is also often driven by the desired type of elements. Morphing-based methods have the advantage that they deform the same template mesh, which results in isotopological meshes and one to one correspondences. This means that the resulting meshes have the same number of nodes and elements and the same connectivity. This allows for an easier treatment in further steps of the computational FE analysis, such as boundary condition definition. On the other hand, large deformations of the reference mesh may lead to FE meshes with degenerated elements. In order to avoid this, quality checks are performed and fitting constraints are included in the morphing process, based on mesh quality measures such as the aspect ratio.
Figure 4. Illustration of the internal nodes of a mesh before (A) and after (B) the morphing, adapted from Baldwin et al. (2010).
Figure 5. Knee contact pressure (in MPa) of the statistical shape and alignment model (Rao et al., 2013). Models presented correspond to the mean shape and the variation of the first two modes between ±1 SD from the mean shape.
3. Uncertainty and Variability in Computational Models: Propagation
Uncertainty quantification of input parameters needs to be propagated through the simulation model to system outputs. Propagation methods can be divided into intrusive and non-intrusive (Figure 1). Intrusive methods reformulate the deterministic standard FE method in a stochastic FE method (SFEM) by integrating the random variables. Since the structural system is usually modeled by partial differential equations (PDE), the stochastic nature of the uncertain system is analogously modeled as a random field or stochastic process by stochastic PDE (SPDE). Despite its computational complexity, SFEM has received an important attention over the last decades, since it is a powerful tool for the solution of SPDE, and the growth of computational power makes it feasible to deal with large-scale problems. Non-intrusive techniques, rather than modifying the mathematical model, use sampling methods to generate an ensemble of simulations where each model is created by a sampling scheme accounting for model input variability. Then, each model is analyzed through FE and the results are finally statistically studied. In this review, pure sampling methods that create instances of the whole computational model have been classified into sampling techniques, including design of experiments, statistical methods and adaptive sampling methods. In biomedical computational models, non-intrusive methods are often preferred for practical reasons since, unlike intrusive methods, they do not require modification of the mathematical and numerical formulation of the underlying deterministic model. In this section, sampling methods are first described. Afterward, intrusive and non-intrusive methods are presented and discussed (see Table 1 for a comparison of some of their main characteristics).
3.1. Sampling Techniques
Propagating uncertainty and variability can be easily done by means of sampling techniques such as statistical sampling methods or DOE. These methods are able to cope with a high number of computational simulations, where the results of the problem have to be obtained through FE analysis for each of the models created.
3.1.1. Design of Experiments (DOE)
DOE is a statistical technique which studies the effect of multiple variables simultaneously defined as discrete sets of values (Yoon et al., 2005; Chen et al., 2011) (see 2.2.1). It received a special attention mainly in industries, such as electronics or chemistry, but it was recently adopted into bioengineering (Montgomery, 1997; Guvenis, 2013). In particular, some authors pointed out its potential in medical imaging (Taner and Sezen, 2007), and it was defined as a cost-effective way to analyze multiple parameters in biomedical research (Li, 1998).
In a classical DOE, one variable is changed randomly while keeping the others constant. However, this implies assessing a large number of combinations, and it is not easy to neither establish interaction between parameters nor determine the optimum. On the other hand, improved methods offer an optimized approach carrying out an efficient performance analysis where the number of simulations is minimum and the optimization process can include multiple objectives (Guvenis, 2013). In particular, the most widely used statistical design method is the Taguchi method (Dar et al., 2002). It is based on factorial analysis, in which only a fraction of all possible combinations of the experimental runs of a full factorial design. Thus, an efficient DOE is carried out with fewer simulations. In order to do that, factors are assigned to a crossed array layout. Controllable factors are part of the inner array, and uncontrollable ones are part of the outer array. Usually, the controllable factors are the ones that improve product characteristics (Yoon et al., 2005) and uncontrollable factors those that cause variations in a production process (Khuri and Mukhopadhyay, 2010). Once the value or level of the controllable factors is established, each run from the controllable factor array is tested across each run from the uncontrollable one. In other words, a series of experiments is carried out to obtain the optimal combination of parameters. This combination will have the greatest effect on the final performance with the minimum variation of the design (Yang et al., 2007; Cheung and Zhang, 2008). Finally, a statistical analysis can be performed to determine the contribution and significance of the main effects of each factor to conduct a sensitivity analysis or to assess the uncertain outcome for a given combination of parameters.
The strengths of the Taguchi method are (1) consistency and robustness of performance considering noise factors and (2) the reduction of time and cost of production (Yang et al., 2007). However, this method has also received criticisms due to both the large number of runs that are required and the difficulty to estimate interaction among control factors (Khuri and Mukhopadhyay, 2010). The Taguchi method has been successfully applied in bioengineering applications. For instance, studies of lumbar or cervical intervertebral discs (Espino et al., 2003; Ng et al., 2004; Malandrino et al., 2009, 2015), designs of a pressure-relieving foot orthosis (Cheung and Zhang, 2008), cervical ring cage (Yang et al., 2007), or even a finger probe for non-invasive hemoglobin monitor (Yoon et al., 2005).
Other approach related to DOE is the response surface method, also known as surrogate model or meta-model. However, this method is exclusively focused on design optimization. It consists to find the optimal parameters by iteratively fitting a mathematical model to the results obtained experimentally (Guvenis, 2013). A series of DOE are carried out to determine the optimum and the relationship between input and output results. For each settings of input variables, the response is measured. The fitting of these responses is known as the response surface (Khuri and Mukhopadhyay, 2010).
3.1.2. Statistical Sampling Methods
Monte Carlo method is a statistical sampling technique in which a random value of each input variable is generated according to a prescribed probability density function. This method is very useful to compute full statistics, and it is considered an exact method to deal with uncertainty since it does not require any assumptions or approximations. However, the simulations computed must converge from a statistical point of view, which is one of the main drawbacks of the method. In computational bioengineering, Monte Carlo method has been applied to a large number of studies, such as modeling fatigue damage evolution in bone (Pidaparti et al., 2001) or the creation of a new hip prosthesis design (Kayabasi and Ekici, 2008). Specifically, Monte Carlo simulation has been used to evaluate artificial joint replacements (Kayabasi and Ekici, 2008; Bah and Browne, 2009; Rohlmann et al., 2009; Dopico-González et al., 2010b; Fitzpatrick et al., 2010, 2011a), to analyze the structural behavior of bone or biomaterials (Pidaparti et al., 2001; Jeffers et al., 2007), and to study anatomical structures (Valero-Cuevas et al., 2003; Belenguer et al., 2006; Bryan et al., 2009; Lin et al., 2009).
Latin Hypercube sampling can be classified as a stratified sampling method and provides an improvement of the convergence ratio over Monte Carlo simulation. It divides the sampling space into subsets – or equally probable intervals – and selects values from each of them randomly, rather than generating samples independently from the previously generated ones. This independence is one of its main advantages over Monte Carlo, since a fewer number of runs are needed to accurately approximate a random distribution (Laz and Browne, 2010; Lin et al., 2012). Latin Hypercube sampling is often used in uncertainty analysis, and, more specifically, it has been applied to study the design of mechanics of joint replacement (Chang et al., 2001; Bah and Browne, 2009; Dopico-González et al., 2009), the behavior of anatomical structures such as the craniofacial structures (Berthaume et al., 2012), the lumbar spine (Niemeyer et al., 2012) or the femur (Bonaretti et al., 2011), or even to analyze multivariable sensitivity in patelofemoral mechanics (Fitzpatrick et al., 2011a).
Most probable point-based (MPP) methods also present a better computational efficiency than Monte Carlo simulation since they reduce the analysis time and have a good accuracy–efficiency ratio (Easley et al., 2007). MPP methods represent the combination of parameters which predicts a certain performance within a specific probability level (Easley et al., 2007; Laz and Browne, 2010). Thus, they determine the most probable point of an objective function by applying an optimization scheme such as first-order Taylor series approximation. MPP methods are the basis of well-known reliability methods like first- and second-order reliability methods (FORM/SORM) (Zhao and Ono, 1999). These have been commonly applied in the structural analysis field to study the reliability of a system to estimate the mechanical failure probability. Although FORM has been rarely used in bioengineering, it has been employed to validate a model of a hip replacement as a time-efficient alternative to Monte Carlo (Mehrez and Browne, 2012).
The mean-value method obtains a mean-based response function and employs the MPP method. However, for higher orders, the advanced mean value (AMV) is used, which incorporates higher order terms to obtain a better representation of the response (Easley et al., 2007). Some authors have demonstrated in orthopedic and bone mechanics studies that the AMV is more efficient than the Monte Carlo simulation, reducing significantly the number of trials [e.g., from 1000 trials to 10 trials, approximately, in (Easley et al., 2007; Laz et al., 2007)].
Nicolella et al. (2006) found that a combination of Monte Carlo and MPP methods is more efficient than the individual use of either, although it requires a high number of runs to compute the entire response. They applied this combined probabilistic scheme to obtain the most influencial parameters in the design of a femoral hip prosthesis. AMV has been applied together with Monte Carlo in other applications such as kinematics and dynamics studies of the human body (Morton et al., 2007; Langenderfer et al., 2008).
3.1.3. Adaptive Sampling Methods
Adaptive sampling methods perform a sampling strategy that chooses the next sample based on the results obtained by previous samples. These approaches usually focus on optimization since they are able to adjust the input parameters improving the results of the final model design. We have broadly divided them into two types: level sets and genetic algorithms.
Level sets have their origin in the image processing field. Specifically, they were employed as a segmentation technique based on active contours. Active contours, also known as snakes, are a technique based on evolving curves or surfaces that aims to detect boundaries of objects in an image, constrained by predefined geometrical conditions (Caselles et al., 1997; Chan and Vese, 2001). Level sets describe an interface Γ as a zero level of a continuous function ϕ defined on the image domain Ω. Thus, the continuous function ϕ is positive inside the domain, negative outside, and zero on the interface Γ (see Figure 6). This approach does not follow the interface itself, but it locates the set Γ(t) where the function ϕ vanishes (Osher and Fedkiw, 2001). Then, this makes it easier to follow interfaces with topology changes, such as holes or shape splits. For this reason, level sets are not only used as a segmentation method but also in other applications, such as topological optimization (Chen et al., 2010; Dijk et al., 2013) or tracking of moving objects (Paragios and Deriche, 2004). Level sets are widely used in medical image segmentation. In terms of physical interpretation as active contours, they are elastic bodies that react and move in a natural way to applied forces and constraints (McKeighen, 1996; Osher and Fedkiw, 2001).
Figure 6. Examples of level-set functions ϕ and their corresponding material domain Ω before and after the design update, adapted from Dijk et al. (2013).
Some authors adapted this approach of segmentation and tracking into a sampling method (Kozic et al., 2010). The authors used a level set-based optimization scheme in a shape space created by a PCA, thus creating a statistical shape model of human tibia. Rather than using level set as a shape representation method, they analyzed the statistical shape space to create a partition of the population, based on an implant-fitting criterion. This criterion was based on reducing the geometrical fitting error between the orthopedic implant, used for fracture fixation, and the surface of the proximal tibia (see Figure 3). Their scheme allowed them to find iteratively new samples in the shape space which meet certain geometry characteristics to finally obtain an implant design which satisfied the maximum percentage of the target population, according to a fitting criterion.
Another approach is based on the so-called genetic algorithms (GA). They are adaptive heuristic search algorithms rooted in the mechanism of evolution and natural selection found in genetics (Srinivas and Patnaik, 1994; Kumar et al., 2010). GA are based on a selection process which leads to the survival of the fittest individuals.
First, a set of initial instances, called population, are generated. Then, the population will be improved in successive iterations in the so-called generations. The set of parameters of a population are represented by a chromosome which will be evaluated assigning a fitness value. This value describes the measured solution according to an objective function and the results obtained. Thus, depending on their fitness value, some chromosomes are selected for the new generation. The higher the fitness value, the higher the probability of survival in the subsequent generation (Srinivas and Patnaik, 1994; Lodygowski et al., 2009; Kumar et al., 2010).
GA are particularly suited to find optimal solutions in complex problems. They have been applied in a wide range of fields such as biology, engineering, or computer science. Specifically, in bioengineering applications some examples are the optimization of prostate implants (Yu and Schell, 1996; Yang et al., 1998) or cochlear implant fitting (Choi et al., 2004; BaSkent et al., 2007; Legrand et al., 2007). Chang et al. (2015) used the idea of artificial neural network together with GA to optimize the screw orientation of an anterior cervical discectomy, obtaining an effective reduction of time and effort to find the optimal solution.
Despite the wide range of propagation methods presented so far, all of them are considered sampling schemes that eventually cope with a high number of computational models that need to be evaluated to obtain a statistically valid global response. In the next section, we present methods that deal directly with the uncertainty presented in the input random variables or fields. Both intrusive and non-intrusive methods can be used to propagate this input uncertainty. Some of the limitations of intrusive methods are the background required to implement them and their particularity (each modification of the physical problem involves defining again the SPDE), thus the implementation needs to be adapted to each physical problem addressed. Non-intrusive methods offer a good balance between generality and computational cost, being a common approach in engineering fields such as structural or aerodynamic studies, when uncertainty is considered.
3.2. Intrusive Methods
Intrusive approaches require to reformulate the deterministic standard FE method into a SFEM. Since the structural system is usually modeled by PDE, the stochastic nature of the uncertain system is analogously modeled as a stochastic process by SPDE. SFEM has received an important attention over the last decades as a powerful tool for the solutions of SPDE, thanks to the growth of the computational power, which makes it feasible and efficient to deal with these large-scale problems.
Intrusive methods consist in describing the uncertain input variable by a stochastic process or random field. Often, stochastic process denotes uncertainty on time, while random field refers to uncertainty on a domain in higher dimensions (e.g., space). They could be defined as families of random variables represented by their distribution and statistical moments. They may have a correlation function, or correlation length, that define the variability of the random field describing the dependency of the random variables in space (Keese, 2003). For continuous random variables, Gaussian random fields are often the most suitable and simplest, where all distributions are jointly Gaussian. However, non-Gaussian ones have also been employed lately due to the wide range of engineering problems with non-Gaussian characteristics (Sachdeva et al., 2007; Stefanou, 2009).
For numerical computation, random fields need to be discretized. This leads to treat both dimensions separately by defining two different meshes which are not required to be the same. The one to discretize the spatial dimension relies on the geometry, and it is usually created by standard techniques like FE. The second one is the “stochastic” mesh, used to discretize the stochastic dimension by approximating the continuous random field by a finite combination of random variables (Keese, 2003; Matthies and Keese, 2005; Stefanou, 2009).
Usually the FE mesh is coarser than the “stochastic” one. In contrast to the outcome seek through traditional mesh convergence studies, it has been even suggested that the later should be as fine as possible according to the computational resources available. To discretize the stochastic dimensions, different series representation techniques have been proposed, such as the mid-point method, the shape function method, or the optimal linear estimation (Keese, 2003; Sachdeva et al., 2007; Stefanou, 2009). However, the most common approach is the Karhunen–Loeve decomposition. Given the domain D and the sample space Ω, this approach, based on model reduction techniques, expands any random field, , in a sum of products of functions of the stochastic and deterministic parameters in a Fourier-type series. In this way, a random variable is defined within the random field k at a certain point x ∈ D, thus a real number k(x, ω) is obtained for each realization ω ∈ Ω (Betz et al., 2004; Eiermann et al., 2007; Sachdeva et al., 2007).
where ξi is a series of uncorrelated random variables, λi and ki (x) are the eigenvalues and eigenfunctions of the eigenproblem [equation (2)], being Cov(x1, x2) the autocovariance function of the random field.
This approach is suitable for representation only when a few terms are needed to capture the maximum fluctuation of the random field that is defined by its correlation length. An upgrade of this technique is the proper generalized decomposition (Nadal et al., 2015), suitable to be used in any engineering and biomedical engineering problem. Another approach is the spectral representation method which similarly expands the random field as a sum of trigonometric functions with random phase angles and amplitudes (Eiermann et al., 2007; Stefanou, 2009).
Once the probability density function of the input random field is represented, it can be reformulated to introduce a finite dimensional subspace of the stochastic parameter space, Wh (see section 3.2.2) (Eiermann et al., 2007). This space, previously discretized, is created by a combination of the basis of the stochastic space. Several approaches have been proposed to construct this finite dimensional subspace. However, the most widely employed is the use of global polynomials in each random variable. This results on a space called polynomial chaos (PC) expansion.
3.2.1. Polynomial Chaos Method
The PC method has been extensively used when dealing with uncertainty quantification in structural analysis and fluid dynamic studies. The method represents each random field k as an expansion of orthogonal polynomials, Φi(ε(ω)), being ε(ω) and ω the random variable and the random results, respectively.
being npc and ppc the dimensionality and order of expansion, respectively. This representation, treats independently the stochastic part, the polynomial chaos functions Φi(ε(ω)), and the deterministic ones, the coefficients ki (x,t) (Loeven et al., 2008). When all parameters are independent Gaussian random variables, a basis of orthogonal polynomials sequence, called Hermite polynomials, is defined, creating the PC expansion space (Li and Zhang, 2007). Some issues, such as the convergence of the polynomial expansion, the high computational cost, and the complex implementation, have led to other approaches, such as non-intrusive PC method, to be able to use deterministic solvers for uncertainty studies (see section 3.3)
The solution of the stochastic response of the system already discretized is directly computed via high-dimensional integration over a probability space using different approaches (Fonseca et al., 2002; Eiermann et al., 2007; Sachdeva et al., 2007; Geneser et al., 2008; Stefanou, 2009). The most commonly employed is the direct numerical integration, which computes the solution in a straightforward way via Monte Carlo: N instances of the stochastic system matrix are sampled by Monte Carlo according to a probability measure (Matthies, 2008; Stefanou, 2009). Then, the system is solved N number of times and the population of response variability vector is computed by statistics (Stefanou, 2009). There exist a number of different numerical methods for solving the system of linear random equations, such as Monte Carlo (as previously mentioned), stochastic Galerkin approach, perturbation method, and Neumann series expansion, among others.
3.2.2. Stochastic Galerkin Method
The Stochastic Galerkin method is a further development of the PC method used to approximate the stochastic response of the system via Galerkin methods (Pons-Prats, 2011). In the standard FE, the classical Galerkin method is a numerical technique to solve partial differential equations using a spatial discretization and weighted residual formulation to approximate the continuous problem to a discrete one. Analogously, the stochastic Galerkin method discretizes the random field using a PC expansion, achieving an efficient representation of the response with arbitrary probability density functions (Geneser et al., 2008).
This approach computes the stochastic process by a weighted sum of orthogonal polynomials of random variables, i.e., projecting the expansion of the random variables to the polynomials basis. Therefore, by using a stochastic projection Galerkin scheme, the solution lies in the stochastic subspace of order m. Hence, the approximation of the solution process can be defined as u(θ) = ξ0ψ0(θ) + ξ1ψ1(θ) + Λ ξnψm(θ) = Ψ(θ)ξ, being a set of basis vectors spanning the stochastic subspace Wh and is a vector of undetermined coefficients (Sachdeva et al., 2007). By using this Galerkin projection scheme, the error of the expansion is minimized employing more basis vectors.
Once the approximation of the stochastic response is computed via stochastic Galerkin projection, the set of equations obtained for the expansion coefficients are deterministic, thus conventional quadrature methods can be used to evaluate the integral, and, therefore, the process is computationally cheaper than direct numerical integration (Xiu, 2007; Matthies, 2008).
3.2.3. Perturbation Method
The perturbation method is another numerical technique to compute the solution of the stochastic process. It represents the output response of the system by expanding the output quantity around its mean value using Taylor’s series. Thus, it approximates the solution by a Taylor expansion with random coefficients (Lee and Lim, 1998). These coefficients are unknown deterministic vectors of the input uncertainty source (Kaminski, 2005; Stefanou, 2009), expressing in terms of lower-order polynomial function the stiffness matrix, load vector, and nodal displacement – in the particular case of a structural analysis – with respect to the random variables (Lee and Lim, 1998; Fonseca et al., 2002). The main drawback of this method compared to Monte Carlo approach, for instance, is its high intrusive character. There exists a high dependency between the evaluation of the sensitivity derivatives from the expansion and the code implementation.
Instrusive methods have shown satisfactory results in only few applications in the bioengineering field. Regarding model reduction techniques, Karhunen–Loeve decomposition was used to simulate in real time the deformation of non-linear and anisotropic organic tissues, more specifically the human cornea, providing an alternative to FE analysis for the case study (Niroomandi et al., 2008). Geneser et al. (2008) presented a novel methodology which combines generalized PC expansion with stochastic Galerkin to study the sensitivity of organ tissue conductivity (see Figure 7). Shi and Liu (2003) estimated cardiac kinematic functions by SPDE. Finally, other authors compared different approaches, such as deterministic, probabilistic, and SFEM, in the study of a mid-cervical spine and concluded that SFEM could provide a high value to minimize uncertainty influences to carry out further reliability analysis of biological systems (Jang and Ekwaro-Osire, 2010).
Figure 7. Effect of tissue conductivity uncertainty of three organs on torso potential. (A) Mean and (B) standard deviation of the electrical potential resulting from varying the lung conductivity. Standard deviation of the potential distribution on torso for 50% variation in (C) muscle and (D) fat conductivity. All units are in millivolts. (Geneser et al., 2008).
The intrusive nature of these methods – high computational cost and extensive mathematical manipulation – still remains their main drawback. Some methods, such as non-intrusive PC expansion or the stochastic collocation method, were developed to address their limitations.
3.3. Non-Intrusive Methods
As mentioned, intrusive methods need to modify the governing equations, which could be difficult or even unfeasible. This led to the development of similar approaches, so as to facilitate the implementation of the analysis.
3.3.1. Polynomial Chaos Method (Non-Intrusive Approach)
One of these approaches is the non-intrusive PC expansion. It estimates the coefficients for known orthogonal polynomial functions based on a response metric of interest according to a set of simulations. Thus, it creates a meta-model by encompassing the results space (Huberts et al., 2015). The response coefficients of the expansion can be computed by (1) spectral projection, which employs inner products and polynomial orthogonality properties, or (2) linear regression, called also point collocation, which extracts the coefficients that best match a set of output spaces by linear least square (Xiu, 2007; Eldred and Burkardt, 2009; Huberts et al., 2015). In order to estimate its coefficients, spectral projection can use a simple sampling approach – known as non-intrusive spectral projection (Loeven et al., 2008; Abgrall et al., 2010). The idea is to sample a discrete parameter space where the deterministic simulation is computed obtaining a set of responses (Abgrall et al., 2010), and the variable solution is reconstructed as a polynomial expansion. Huberts et al. (2015) reported a better performance of the least-squares regression approach, compared to the spectral projection, in a cardiovascular pulse wave propagation model (Huberts et al., 2015). PC expansion can be also employed together with surrogate models, previously introduced as response surface method, to reduce the computational cost, for instance, creating the surrogate model in the form of the Hermite polynomials (Isukapalli et al., 1998). Thus, the solution can be computed by fast-converging polynomial approximation based on least-squares linear regression (Du and Chen, 2000; Berveiller et al., 2006; Baroth et al., 2007; Eck et al., 2015).
3.3.2. Probabilistic Collocation
The second technique to compute the coefficients of the expansion is the probabilistic collocation method. This approach combines the idea of PC expansions and the collocation approach. It represents the behavior of the uncertain parameters by a set of weighted values computed by a numerical integration, such as Gauss-quadrature, of the stochastic space based on Lagrange polynomials (Loeven and Bijl, 2008; Eldred and Burkardt, 2009; Pons-Prats, 2011). As mentioned previously, it uses the concept of collocation points to obtain the coefficients of the expansion that better represent the uncertain response (Eldred and Burkardt, 2009) by doing the Galerkin projection (Loeven and Bijl, 2008). Then, the approximation of the distribution function is integrated from each collocation point (Loeven and Bijl, 2008; Loeven et al., 2008; Eldred and Burkardt, 2009). The stochastic representation of a function is described as follows:
being Li(ε(ω)) the Lagrangian polynomial which leads to the calculation of the weight, wi, for each fi(x). It has been proved that this probabilistic collocation method is more efficient than the Galerkin PC expansion approach (section 3.2) and that the computational cost is reduced considerably compared to the direct integration method using Monte Carlo (Li and Zhang, 2007).
3.3.3. Stochastic Collocation Method
Closely related to PC expansion, stochastic collocation is another non-intrusive method where the simulations are performed at specific collocation points in the uncertainty space, combining the fast convergence of the PC approach and the decoupled nature of sampling techniques such as Monte Carlo (Xiu, 2007; Loeven et al., 2008; Eldred and Burkardt, 2009; Pons-Prats, 2011; Sankaran and Marsden, 2011). The solution of the stochastic response is constructed employing Lagrange interpolation to obtain the expansion polynomial which is obtained using a collocation grid of points rather than by random sampling (Loeven et al., 2008; Eldred and Burkardt, 2009).
As described by Loeven et al. (2008), the solution in a stochastic domain α, which is defined according to a standard domain of orthogonal polynomials (Abgrall et al., 2010), is u(x,t,α). From the α domain, Np collocation points αi are taken. The solution u(x,t,α) is approximated by the following expansion
where ui (x,t) are the values of u(x,t,α) at the given collocation points αi, being hi(α) the interpolating polynomials of degree Np − 1 (Loeven et al., 2008).
This method provides a set of deterministic equations that are computed using standard solvers. That means that for each collocation point, a deterministic problem is solved. The main difference, compared to PC expansion is that it forms multidimensional interpolants for known coefficients, rather than estimating the coefficients for known basis functions of the orthogonal polynomial (Eldred and Burkardt, 2009).
These non-intrusive approaches have been employed also in the biomedical engineering fields, mainly in topics such as drug infusion (Preston et al., 2009), hemodynamics (Xiu and Sherwin, 2007; Sankaran and Marsden, 2011), or cardiovascular simulations (Eck et al., 2015, 2016). Both Sankaran and Marsden (2011) and Eck et al. (2015) employed the approach of stochastic collocation to carry out sensitivity analysis on blood flow simulations.
Concretely, Eck et al. (2015) evaluated how arterial stiffness influenced on backward-propagating pressure waves, using a previously developed model that employs PC expansion in order to set the parameters of the model to create a patient-specific predictive model (Huberts et al., 2015). Later, Eck et al. (2016) provided a guideline for the uncertainty analysis in cardiovascular applications. Sankaran and Marsden (2011) proposed an adaptive collocation algorithm to reduce the computational cost of the stochastic collocation scheme and quantify the confidence of the stochastic response in abdominal aortic aneurysm and carotid artery bifurcation simulations.
This article reviews the analysis of uncertainty and variability in the context of bioengineering computational models. Characterization techniques and propagation methods have been presented, as well as examples of their applications in biomedical simulations.
As shown, a great number of uncertainty and variability factors are found in bioengineering computational models. Among them, the most studied are the ones related to anatomical shape and implant design parameters. The techniques involving statistical shape models are the most commonly used due to their ability to capture the huge variability among patients and then obtaining realistic results according to the population.
The most common techniques of uncertainty propagation employed in this field are the ones within the sampling techniques category. These methods have the advantage that they do not require modifications to existing computational tools, hence the solver of FE can be used as a “black box,” and their use and implementation is straightforward. Even so, their computational cost is rather high due to the number of samples that need to be analyzed. Table 2 summarizes the studies within biomedical engineering field presented, where geometrical variability and non-intrusive approaches are the ones most considered.
Table 2. Uncertainty studies according to the source of uncertainty and its characterization in the context of biomedical computational analysis.
Many methods, intrusive and non-intrusive, firmly settled into some engineering fields such as structural or fluid dynamic analysis, have become regular tools to study uncertainty and variability in computational models. However, intrusive methods have been rarely used in biomedical computational models. Intrusive approaches require a high knowledge of the computational problem since an adaptation of the deterministic case is needed. Despite their advantages, these approaches are not commonly used due to their high computational cost, issues related with the convergence of the polynomial expansions, and the high mathematical manipulation. In consequence, this led to study further non-intrusive methods, such as the non-intrusive PC method or probabilistic collocation (see Table 1 for a comparison of the main techniques discussed). These non-intrusive methods started to be used in the last decades, showing promising results in biomedical engineering, mainly in topics related with cardiovascular applications.
Furthermore, the different approaches discussed have been compared to each other in extremely few cases. It could be interesting to contrast the results obtained by a non-intrusive approach to an intrusive one in different applications to determine the feasibility of certain methods in biomedical fields. Another consideration to point out is that, in biomedical engineering, epistemic uncertainties are present too. Moreover, further uncertainties can be found in the data acquisition process. There is a lack of studies which consider both uncertainties, aleatory and epistemic, in the FE simulations. Considering both types of uncertainties would provide more accurate results according to data origin, methodology of processing, and the variability and uncertainty found in the physical problem definition.
Finding the best suitable method is a trade-off among several factors. Namely, the cost of implementing the method for each application – leading to the selection of an intrusive or a non-intrusive method, the computational cost associated with the specific analysis – leading to the selection of a low number of samples, or of an intrusive method – and many other considerations which are problem specific. The analysis with uncertainties is becoming more and more relevant in any engineering field, and more specifically in biomedical engineering. While the cost is still higher than a standard deterministic analysis, the benefits of better fitting the real world applications are a clear advantage.
From this review, it is clear that there is still work that can be done in bioengineering applications with respect to uncertainty quantification. Some of them would be, for instance: (1) carefully analyze and model the input parameters uncertainty, dealing not exclusively with geometrical parameters or material properties, but having also into account boundary conditions and altogether. It is a common problem in many engineering fields, but it takes a relevant importance when focusing on biomedical engineering problems due to its clinical relevance. (2) As mentioned, compare the efficiency and applicability of each method, both intrusive and non-intrusive. The simplicity of this fact could seem meaningless, but it can provide really useful information to choose the approach that better fit the study requirements, regarding computational cost and accuracy of the solution. Furthermore, the analysis between approaches should be done on the application under study, since the reference is the equivalent result on the deterministic approach, and the efficiency is directly related to each application. (3) Finally, the analysis of the output behavior is one of the key factors to take real benefit of uncertainty quantification. Finding the best statistical moments and parameters to represent the propagated variability, while obtaining a good understanding of the interaction between input–output variability efficiency is the final aim of uncertainty propagation methods.
The authors are aware that uncertainty quantification is a vast field, and some methods are missing in this review. Only those which have been already used in bioengineering applications have been studied, leaving other methods, such as fuzzy FE methods or multi-level Monte Carlo, for future reviews and applications.
All authors have contributed equally to this work and approved it for publication.
Conflict of Interest Statement
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
This work is partly supported by the Spanish Ministry of Economy and Competitiveness under the Maria de Maeztu Units of Excellence Programme (MDM-2015-0502) and by the European Union Seventh Frame Programme (FP7/2007-2013), Grant agreement 304857, HEAR-EU project.
Abgrall, R., Congedo, P. M., Galéra, S., and Geraci, G. (2010). “Semi-intrusive and non-intrusive stochastic methods for aerospace applications,” in 4th European Conference for Aerospace Sciences (Saint Petersburg, Russia), 1–8.
Ashraf, M., Davatzikos, C., and Taylor, R. (2002). “A combined statistical and biomechanical model for estimation of intra-operative prostate deformation,” in Medical Image Computing and Computer-Assisted Intervention MICCAI 2002, LNCS, Springer, Vol. 2489, 452–460.
Bah, M. T., Nair, P. B., and Browne, M. (2009). Mesh morphing for finite element analysis of implant positioning in cementless total hip replacements. Med. Eng. Phys. 31, 1235–1243. doi:10.1016/j.medengphy.2009.08.001
Baldwin, M. A., Langenderfer, J. E., Rullkoetter, P. J., and Laz, P. J. (2010). Development of subject-specific and statistical shape models of the knee using an efficient segmentation and mesh-morphing approach. Comput. Meth. Prog. Bio. 97, 232–240. doi:10.1016/j.cmpb.2009.07.005
Baroth, J., Bressolette, P., Chauvière, C., and Fogli, M. (2007). An efficient SFE method using Lagrange polynomials: application to nonlinear mechanical problems with uncertain parameters. Comput. Methods Appl. Mech. Eng. 196, 45–48. doi:10.1016/j.cma.2007.04.017
BaSkent, D., Eiler, C. L., and Edwards, B. (2007). Using genetic algorithms with subjective input from human subjects: implications for fitting hearing aids and cochlear implants. Ear. Hearing. 28, 370–380. doi:10.1097/AUD.0b013e318047935e
Bednarz, B., Lu, H. M., Engelsman, M., and Paganetti, H. (2011). Uncertainties and correction methods when modeling passive scattering proton therapy treatment heads with Monte Carlo. Phys. Med. Biol. 56, 2837–2854. doi:10.1088/0031-9155/56/9/013
Belenguer, L., Büchler, P., Rueckert, D., Nolte, L. P., and González Ballester, M. A. (2006). “Statistical finite element model for bone shape and biomechanical properties,” in Medical Image Computing and Computer-Assisted Intervention MICCAI 2006, LNCS, Springer, Vol. 4190, 405–411.
Berthaume, M. A., Dechow, P. C., Iriarte-Diaz, J., Ross, C. F., Strait, D. S., Wang, Q., et al. (2012). Probabilistic finite element analysis of a craniofacial finite element model. J. Theor. Biol. 300, 242–253. doi:10.1016/j.jtbi.2012.01.031
Betz, W., Papaioannou, I., and Straub, D. (2004). Numerical methods for the discretization of random fields by means of the Karhunen–Loéve expansion. Comput. Methods Appl. Mech. Eng. 271, 109–129. doi:10.1016/j.cma.2013.12.010
Bode, A. S., Huberts, W., Bosboom, E. M., Kroon, W., van der Linden, W. P., Planken, R. N., et al. (2012). Patient-specific computational modeling of upper extremity arteriovenous fistula creation: its feasibility to support clinical decision-making. PLoS ONE 7:e34491. doi:10.1371/journal.pone.0034491
Bonaretti, S., Seiler, C., Boichon, C., Philippe, B., and Reyes, M. (2011). “Mesh-based vs. image-based statistical model of appearance of the human femur: a preliminary comparison study for the creation of finite element meshes,” in Mesh Processing in Medical Image Analysis Workshop (Toronto, Canada: MICCAI).
Bredbenner, T. L., Eliason, T. D., Potter, R. S., Mason, R. L., Havill, L. M., and Nicolella, D. P. (2010). Statistical shape modeling describes variation in tibia and femur surface geometry between Control and Incidence groups from the osteoarthritis initiative database. J. Biomech. 43, 1780–1786. doi:10.1016/j.jbiomech.2010.02.015
Bryan, R., Mohan, P. S., Hopkins, A., Galloway, F., Taylor, M., and Nair, P. B. (2010). Statistical modelling of the whole human femur incorporating geometric and material properties. Med. Eng. Phys. 32, 57–65. doi:10.1016/j.medengphy.2009.10.008
Bryan, R., Nair, P. B., and Taylor, M. (2009). Use of a statistical model of the whole femur in a large scale, multi-model study of femoral neck fracture risk. J. Biomech. 42, 2171–2176. doi:10.1016/j.jbiomech.2009.05.038
Cao, W., Lim, G. J., Lee, A., Li, Y., Liu, W., Ronald Zhu, X., et al. (2012). Uncertainty incorporated beam angle optimization for IMPT treatment planning. Med. Phys. 39, 5248–5256. doi:10.1118/1.4737870
Chang, P. B., Williams, B. J., Bhalla, K. S., Belknap, T. W., Santner, T. J., Notz, W. I., et al. (2001). Design and analysis of robust total joint replacements: finite element model experiments with environmental variables. J. Biomech. Eng. 123, 239–246. doi:10.1115/1.1372701
Chang, T. K., Hsu, C. C., and Chen, K. T. (2015). Optimal screw orientation for the fixation of cervical degenerative disc disease using nonlinear C3-T2 multi-level spinal models and neuro-genetic algorithms. Acta Bioeng. Biomech. 17, 59–66.
Chen, D. C., You, C. S., Nian, F. L., and Guo, M. W. (2011). “Using the Taguchi method and finite element method to analyze a robust new design for titanium alloy prick hole extrusion,” in 11th International Conference on the Mechanical Behavior of Materials, Vol. 10. Milano, Italy, 82–87.
Cheung, J. T. M., and Zhang, M. (2008). Parametric design of pressure-relieving foot orthosis using statistics-based finite element method. Med. Eng. Phys. 30, 269–277. doi:10.1016/j.medengphy.2007.05.002
Choi, C. T. M., Lai, W. D., and Chen, Y. B. (2004). Optimization of cochlear implant electrode array using genetic algorithms and computational neuroscience models. IEEE Trans. Magn. 40, 639–642. doi:10.1109/TMAG.2004.824912
Dopico-González, C., New, A. M., and Browne, M. (2010a). A computational tool for the probabilistic finite element analysis of an uncemented total hip replacement considering variability in bone-implant version angle. Comput. Methods Biomech. Biomed. Engin. 13, 1–9. doi:10.1080/10255840902911536
Dopico-González, C., New, A. M., and Browne, M. (2010b). Probabilistic finite element analysis of the uncemented hip replacement-effect of femur characteristics and implant design geometry. J. Biomech. 43, 512–520. doi:10.1016/j.jbiomech.2009.09.039
Easley, S. K., Pal, S., Tomaszewski, P. R., Petrella, A. J., Rullkoetter, P. J., and Laz, P. J. (2007). Finite element-based probabilistic analysis tool for orthopaedic applications. Comput. Methods Programs Biomed. 85, 32–40. doi:10.1016/j.cmpb.2006.09.013
Eck, V. G., Donders, W. P., Sturdy, J., Feinberg, J., Delhaas, T., Hellevik, L. R., et al. (2016). A guide to uncertainty quantification and sensitivity analysis for cardiovascular applications. Int. J. Numer. Method. Biomed. Eng. 32. doi:10.1002/cnm.2755
Eck, V. G., Feinberg, J., Langtangen, H. P., and Hellevik, L. R. (2015). Stochastic sensitivity analysis for timing and amplitude of pressure waves in the arterial system. Int. J. Numer. Method. Biomed. Eng. 31, e02711. doi:10.1002/cnm.2711
Eldred, M., and Burkardt, J. (2009). “Comparison of non-intrusive polynomial chaos and stochastic collocation methods for uncertainty quantification,” in Proceedings of the 47th AIAA Aerospace, Orlando, Florida, 1–20.
Espino, D. M., Meakin, J. R., Hukins, D. W. L., and Reid, J. E. (2003). Stochastic finite element analysis of biological systems: comparison of a simple intervertebral disc model with experimental results. Comput. Methods Biomech. Biomed. Engin. 6, 243–248. doi:10.1080/10255840310001606071
Fitzpatrick, C. K., Baldwin, M. A., Rullkoetter, P. J., and Laz, P. J. (2011a). Combined probabilistic and principal component analysis approach for multivariate sensitivity evaluation and application to implanted patellofemoral mechanics. J. Biomech. 44, 13–21. doi:10.1016/j.jbiomech.2010.08.016
Fitzpatrick, C. K., Baldwin, M. A., Laz, P. J., FitzPatrick, D. P., Lerner, A. L., and Rullkoetter, P. J. (2011b). Development of a statistical shape model of the patellofemoral joint for investigating relationships between shape and function. J. Biomech. 44, 2446–2452. doi:10.1016/j.jbiomech.2011.06.025
Fitzpatrick, C. K., Clary, C. W., Laz, P. J., and Rullkoetter, P. J. (2010). Relative contributions of design, alignment, and loading variability in knee replacement mechanics. J. Orthop. Res. 30, 2015–2024. doi:10.1002/jor.22169
Fonseca, J., Mares, C., Friswell, M., and Mottershead, J. (2002). “Review of parameter uncertainty propagation methods in structural dynamic analysis,” in International Conference on Noise and Vibration Engineering, Vol. 4. Leuven, Belgium, 1853–1860.
Geneser, S. E., Kirby, R. M., and MacLeod, R. S. (2008). Application of stochastic finite element methods to study the sensitivity of ECG forward modeling to organ conductivity. IEEE T Bio. Med. Eng. 55, 31–40. doi:10.1109/TBME.2007.900563
González Ballester, M. A., Pennec, X., George Linguraru, M., and Ayache, N. (2004). Generalized image models and their application as statistical models of images. Med. Image Anal. 8, 361–369. doi:10.1016/j.media.2004.06.012
González Ballester, M. A., Zisserman, A., and Brady, M. (2000). Segmentation and measurement of brain structures in MRI including confidence bounds. Med. Image Anal. 4, 189–200. doi:10.1016/S1361-8415(00)00013-X
Grasa, J., Pérez, M. A., Bea, J., García-Aznar, J., and Doblaré, M. (2005). A probabilistic damage model for acrylic cements. Application to the life prediction of cemented hip implants. Int. J. Fatigue 27, 891–904. doi:10.1016/j.ijfatigue.2004.12.009
Huberts, W., Donders, W. P., Delhaas, T., and van de Vosse, F. N. (2015). Applicability of the polynomial chaos expansion method for personalization of a cardiovascular pulse wave propagation model. Int. J. Numer. Method. Biomed. Eng. 31, e02720. doi:10.1002/cnm.2720
Isukapalli, S. S., Roy, A., and Georgopoulos, P. G. (1998). Stochastic response surface methods (SRSMs) for uncertainty propagation: application to environmental and biological systems. Risk Anal. 18, 351–363. doi:10.1111/j.1539-6924.1998.tb01301.x
Jang, T. H., and Ekwaro-Osire, S. (2010). “Stochastic finite element approach in mid-cervical spine (C4-C6) for increasing analysis reliability,” in 12th International Conference on Integrated Design and Process Technology (Dallas, Texas). Jun 6–11.
Jeffers, J. R. T., Browne, M., Lennon, A. B., Prendergast, P. J., and Taylor, M. (2007). Cement mantle fatigue failure in total hip replacement: experimental and computational testing. J. Biomech. 40, 1525–1533. doi:10.1016/j.jbiomech.2006.07.029
Keese, A. (2003). A Review of Recent Developments in the Numerical Solution of Stochastic Partial Differential Equations (Stochastic Finite Elements). Internal Report. Brunswick, Germany: Technical University Braunschweig.
Khuri, A. I., and Mukhopadhyay, S. (2010). “Response surface methodology,” in Wiley Interdisciplinary Reviews: Computational Statistics, Vol. 2, eds Wegman E. J., Said Y. H., and Scott D. W. (Hoboken, New Jersey: Wiley), 128–149.
Kohara, S., Tateyama, T., Foruzan, A. H., Furukawa, A., Kanasaki, S., Wakamiya, M., et al. (2011). “Preliminary study on statistical shape model applied to diagnosis of liver cirrhosis,” in 18th IEEE International Conference on Image Processing (Brussels, Belguim), 2921–2924.
Kozic, N., Weber, S., Büchler, P., Lutz, C., Reimers, N., González Ballester, M. A., et al. (2010). Optimisation of orthopaedic implant design using statistical shape space analysis based on level sets. Med. Image Anal. 14, 265–275. doi:10.1016/j.media.2010.02.008
Langenderfer, J. E., Laz, P. J., Petrella, A. J., and Rullkoetter, P. J. (2008). An efficient probabilistic methodology for incorporating uncertainty in body segment parameters and anatomical landmarks in joint loadings estimated from inverse dynamics. J. Biomech. Eng. 130, 014502. doi:10.1115/1.2838037
Laz, P. J., Pal, S., Halloran, J. P., Petrella, A. J., and Rullkoetter, P. J. (2006a). Probabilistic finite element prediction of knee wear simulator mechanics. J. Biomech. 39, 2303–2310. doi:10.1016/j.jbiomech.2005.07.029
Laz, P. J., Pal, S., Fields, A., Petrella, A. J., and Rullkoetter, P. J. (2006b). Effects of knee simulator loading and alignment variability on predicted implant mechanics: a probabilistic study. J. Orthop. Res. 24, 2212–2221. doi:10.1002/jor.20254
Laz, P. J., Stowe, J. Q., Baldwin, M. A., Petrella, A. J., and Rullkoetter, P. J. (2007). Incorporating uncertainty in mechanical properties for finite element-based evaluation of bone mechanics. J. Biomech. 40, 2831–2836. doi:10.1016/j.jbiomech.2007.03.013
Legrand, P., Bourgeois-Republique, C., Péan, V., Cohen, E. H., Levy-Vehel, J., Frachet, B., et al. (2007). Interactive evolution for cochlear implants fitting. Genet. Program. Evol. M. 8, 319–354. doi:10.1007/s10710-007-9048-4
Lin, C. F., Gross, M., Ji, C., Padua, D., Weinhold, P., Garrett, W. E., et al. (2009). A stochastic biomechanical model for risk and risk factors of non-contact anterior cruciate ligament injuries. J. Biomech. 42, 418–423. doi:10.1016/j.jbiomech.2008.12.005
Linguraru, M. G., Ayache, N., Bardinet, E., González Ballester, M. A., Galanaud, D., Haïk, S., et al. (2006). Differentiation of sCJD and vCJD forms by automated analysis of basal ganglia intensity distribution in multisequence MRI of the brain-definition and evaluation of new MRI-based ratios. IEEE Trans. Med. Imaging 25, 1052–1067. doi:10.1109/TMI.2006.876133
Linguraru, M. G., Vercauteren, T., Reyes Aguirre, M., González Ballester, M. A., and Ayache, N. (2008). Segmentation propagation from deformable atlases for brain mapping and analysis. Brain Res. 1, 269–287.
Loeven, G. J. A., and Bijl, H. (2008). “Airfoil analysis with uncertain geometry using the probabilistic collocation method,” in Proc. of the AIAA 48th AIAA/ASME/ASCE/AHS/ASC Structures, Structural Dynamics, and Materials Conference. AIAA-2008-2070. (Illinois, USA: Schaumburg).
Loeven, G. J. A., Witteveen, J. A. S., and Bijl, H. (2008). “Probabilistic collocation: an efficient non-intrusive approach for arbitrarily distributed parametric uncertainties,” in 45th AIAA Aerospace Sciences Meeting and Exhibit, 2007–2317.
Malandrino, A., Planell, J. A., and Lacroix, D. (2009). Statistical factorial analysis on the poroelastic material properties sensitivity of the lumbar intervertebral disc under compression, flexion and axial rotation. J. Biomech. 42, 2780–2788. doi:10.1016/j.jbiomech.2009.07.039
Malandrino, A., Pozo, J. M., Castro-Mateos, I., Frangi, A. F., van Rijsbergen, M. M., Ito, K., et al. (2015). On the relative relevance of subject-specific geometries and degeneration-specific mechanical properties for the study of cell death in human intervertebral disk models. Front. Bioeng. Biotechnol. 3:1–15. doi:10.3389/fbioe.2015.00005
Matthies, H., and Keese, A. (2005). Galerkin methods for linear and nonlinear elliptic stochastic partial differential equations. Comput. Methods Appl. Mech. Eng. 194, 1295–1331. doi:10.1016/j.cma.2004.05.027
Mehrez, L., and Browne, M. (2012). A numerically validated probabilistic model of a simplified total hip replacement construct. Comput. Methods Biomech. Biomed. Engin. 15, 845–858. doi:10.1080/10255842.2011.564163
Morton, N. A., Maletsky, L. P., Pal, S., and Laz, P. J. (2007). Effect of variability in anatomical landmark location on knee kinematic description. J. Orthop. Res. 25, 1221–1230. doi:10.1002/jor.20396
Mousavi, S. R., Khalaji, I., Sadeghi Naini, A., Raahemifar, K., and Samani, A. (2012). Statistical finite element method for real-time tissue mechanics analysis. Comput. Methods Biomech. Biomed. Engin. 15, 595–608. doi:10.1080/10255842.2010.550889
Nadal, E., Chinesta, F., Díez, P., Fuenmayor, F. J., and Denia, F. D. (2015). Real time parameter identification and solution reconstruction from experimental data using the proper generalized decomposition. Comput. Methods Appl. Mech. Eng. 296, 113–128. doi:10.1016/j.cma.2015.07.020
Ng, H. W., Teo, E. C., and Lee, V. S. (2004). Statistical factorial analysis on the material property sensitivity of the mechanical responses of the C4-C6 under compression, anterior and posterior shear. J. Biomech. 37, 771–777. doi:10.1016/j.jbiomech.2003.09.025
Nicolella, D. P., and Bredbenner, T. L. (2012). Development of a parametric finite element model of the proximal femur using statistical shape and density modelling. Comput. Methods Biomech. Biomed. Engin. 15, 101–110. doi:10.1080/10255842.2010.515984
Nicolella, D. P., Thacker, B. H., Katoozian, H., and Davy, D. T. (2006). The effect of three-dimensional shape optimization on the probabilistic response of a cemented femoral hip prosthesis. J. Biomech. 39, 1265–1278. doi:10.1016/j.jbiomech.2005.03.010
Niemeyer, F., Wilke, H. J. J., and Schmidt, H. (2012). Geometry strongly influences the response of numerical models of the lumbar spine–a probabilistic finite element analysis. J. Biomech. 45, 1414–1423. doi:10.1016/j.jbiomech.2012.02.021
Niroomandi, S., Alfaro, I., Cueto, E., and Chinesta, F. (2008). Real-time deformable models of non-linear tissues by model reduction techniques. Comput. Meth. Prog. Bio. 91, 223–231. doi:10.1016/j.cmpb.2008.04.008
Noailly, J., Wilke, H. J., Planell, J. A., and Lacroix, D. (2007). How does the geometry affect the internal biomechanics of a lumbar spine bi-segment finite element model? Consequences on the validation process. J. Biomech. 40, 2414–2425. doi:10.1016/j.jbiomech.2006.11.021
Paragios, N., and Deriche, R. (2004). Geodesic active contours and level sets for the detection and tracking of moving objects. IEEE Trans. Pattern Anal. Mach. Intell. 22, 266–280. doi:10.1109/34.841758
Pérez, M. A., Grasa, J., García-Aznar, J. M., Bea, J. A., and Doblaré, M. (2006). Probabilistic analysis of the influence of the bonding degree of the stem cement interface in the performance of cemented hip prostheses. J. Biomech. 39, 1859–1872. doi:10.1016/j.jbiomech.2005.05.025
Preston, J. S., Tasdizen, T., Terry, C. M., Cheung, A. K., and Kirby, R. M. (2009). Using the stochastic collocation method for the uncertainty quantification of drug concentration due to depot shape variability. IEEE Trans. Biomed. Eng. 56, 609–620. doi:10.1109/TBME.2008.2009882
Rajamani, K. T., Styner, M. A., Talib, H., Zheng, G., Nolte, L. P., and González Ballester, M. A. (2007). Statistical deformable bone models for robust 3D surface extrapolation from sparse data. Med. Image Anal. 11, 99–109. doi:10.1016/j.media.2006.05.001
Rao, C., Fitzpatrick, C. K., Rullkoetter, P. J., Maletsky, L. P., Kim, R. H., and Laz, P. J. (2013). A statistical finite element model of the knee accounting for shape and alignment variability. Med. Eng. Phys. 35, 1450–1456. doi:10.1016/j.medengphy.2013.03.021
Rao, G., Amarantini, D., Berton, E., and Favier, D. (2006). Influence of body segments’ parameters estimation models on inverse dynamics solutions during gait. J. Biomech. 39, 1531–1536. doi:10.1016/j.jbiomech.2005.04.014
Rohlmann, A., Mann, A., Zander, T., and Bergmann, G. (2009). Effect of an artificial disc on lumbar spine biomechanics: a probabilistic finite element study. Eur. Spine J. 18, 89–97. doi:10.1007/s00586-008-0836-1
Roy, C. J., and Oberkampf, W. L. (2010). “A complete framework for verification, validation, and uncertainty quantification in scientific computing,” in 48th Aerospace Sciences Meeting, Orlando, Florida, 1–15.
Rüegsegger, M. B., Bach Cuadra, M., Pica, A., Amstutz, C. A., Rudolph, T., Aebersold, D., et al. (2012). Statistical modeling of the eye for multimodal treatment planning for external beam radiation therapy of intraocular tumors. Int. J. Radiat. Oncol. Biol. Phys. 84, 541–547. doi:10.1016/j.ijrobp.2012.05.040
Sachdeva, S. K., Nair, P. B., and Keane, A. J. (2007). On using deterministic FEA software to solve problems in stochastic structural mechanics. Comput. Struct. 85, 277–290. doi:10.1016/j.compstruc.2006.10.008
Sankaran, S., and Marsden, A. L. (2011). A stochastic collocation method for uncertainty quantification and propagation in cardiovascular simulations. J. Biomech. Eng. 133, 031001. doi:10.1115/1.4003259
Schmid, J., Kim, J., and Magnenat-Thalmann, N. (2011). Robust statistical shape models for MRI bone segmentation in presence of small field of view. Med. Image Anal. 15, 155–168. doi:10.1016/j.media.2010.09.001
Shi, P., and Liu, H. (2003). Stochastic finite element framework for simultaneous estimation of cardiac kinematic functions and material parameters. Med. Image Anal. 7, 445–464. doi:10.1016/S1361-8415(03)00066-5
Smoger, L. M., Fitzpatrick, C. K., Clary, C. W., Cyr, A. J., Maletsky, L. P., Rullkoetter, P. J., et al. (2015). Statistical modeling to characterize relationships between knee anatomy and kinematics. J. Orthop. Res. 33, 1620–1630. doi:10.1002/jor.22948
Talib, H., Rajamani, K., Kowal, J., Nolte, L. P., Styner, M., and González Ballester, M. A. (2005). A comparison study assessing the feasibility of ultrasound-initialized deformable bone models. Comput. Aided Surg. 10, 293–299. doi:10.1080/10929080500379390
Valero-Cuevas, F. J., Johanson, M. E., and Towles, J. D. (2003). Towards a realistic biomechanical model of the thumb: the choice of kinematic description may be more critical than the solution method or the variability/uncertainty of musculoskeletal parameters. J. Biomech. 36, 1019–1030. doi:10.1016/S0021-9290(03)00061-7
Viceconti, M., Brusi, G., Pancanti, A., and Cristofolini, L. (2006). Primary stability of an anatomical cementless hip stem: a statistical analysis. J. Biomech. 39, 1169–1179. doi:10.1016/j.jbiomech.2005.03.024
Yang, G., Reinstein, L. E., Pai, S., Xu, Z., and Carroll, D. L. (1998). A new genetic algorithm technique in optimization of permanent [sup 125]I prostate implants. Med. Phys. 25, 2308. doi:10.1118/1.598460
Zheng, G., Dong, X., Rajamani, K. T., Zhang, X., Styner, M., Thoranaghatte, R. U., et al. (2007). Accurate and robust reconstruction of a surface model of the proximal femur from sparse-point data and a dense-point distribution model for surgical navigation. IEEE Trans. Biomed. Eng. 54, 2109–2122. doi:10.1109/TBME.2007.895736
Zheng, G., Gollmer, S., Schumann, S., Dong, X., Feilkas, T., and González Ballester, M. A. (2009). A 2D/3D correspondence building method for reconstruction of a patient-specific 3D bone surface model using point distribution models and calibrated X-ray images. Med. Image Anal. 13, 883–899. doi:10.1016/j.media.2008.12.003
Zheng, G., Schumann, S., and González Ballester, M. A. (2010). An integrated approach for reconstructing a surface model of the proximal femur from sparse input data and a multi-resolution point distribution model: an in vitro study. Int. J. Comput. Assist. Radiol. Surg. 5, 99–107. doi:10.1007/s11548-009-0386-y
Keywords: uncertainty quantification, finite element models, random variables, intrusive and non-intrusive methods, sampling techniques, computational modeling
Citation: Mangado N, Piella G, Noailly J, Pons-Prats J and González Ballester MÁ (2016) Analysis of Uncertainty and Variability in Finite Element Computational Models for Biomedical Engineering: Characterization and Propagation. Front. Bioeng. Biotechnol. 4:85. doi: 10.3389/fbioe.2016.00085
Received: 09 September 2016; Accepted: 19 October 2016;
Published: 07 November 2016
Edited by:Tien Tuan Dao, University of Technology of Compiègne, France
Reviewed by:Henrique De Amorim Almeida, Polytechnic Institute of Leiria, Portugal
André P. G. Castro, University of Sheffield, UK
Emiliano Votta, Politecnico di Milano, Italy
Copyright: © 2016 Mangado, Piella, Noailly, Pons-Prats and González Ballester. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Nerea Mangado, firstname.lastname@example.org