Refine Your Search

Search Results

Viewing 1 to 13 of 13
Journal Article

Value of Information for Comparing Dependent Repairable Assemblies and Systems

2018-04-03
2018-01-1103
This article presents an approach for comparing alternative repairable systems and calculating the value of information obtained by testing a specified number of such systems. More specifically, an approach is presented to determine the value of information that comes from field testing a specified number of systems in order to appropriately estimate the reliability metric associated with each of the respective repairable systems. Here the reliability of a repairable system will be measured by its failure rate. In support of the decision-making effort, the failure rate is translated into an expected utility based on a utility curve that represents the risk tolerance of the decision-maker. The algorithm calculates the change of the expected value of the decision with the sample size. The change in the value of the decision represents the value of information obtained from testing.
Technical Paper

Random Vibration Analysis Using Quasi-Random Bootstrapping

2018-04-03
2018-01-1104
Reliability analysis of engineering structures such as bridges, airplanes, and cars require calculation of small failure probabilities. These probabilities can be calculated using standard Monte Carlo simulation, but this method is impractical for most real-life systems because of its high computational cost. Many studies have focused on reducing the computational cost of a reliability assessment. These include bootstrapping, Separable Monte Carlo, Importance Sampling, and the Combined Approximations. The computational cost can also be reduced using an efficient method for deterministic analysis such as the mode superposition, mode acceleration, and the combined acceleration method. This paper presents and demonstrates a method that uses a combination of Sobol quasi-random sequences and bootstrapping to reduce the number of function calls. The study demonstrates that the use of quasi-random numbers in conjunction bootstrapping reduces dramatically computational cost.
Journal Article

Assessing the Value of Information for Multiple, Correlated Design Alternatives

2017-03-28
2017-01-0208
Design optimization occurs through a series of decisions that are a standard part of the product development process. Decisions are made anywhere from concept selection to the design of the assembly and manufacturing processes. The effectiveness of these decisions is based on the information available to the decision maker. Decision analysis provides a structured approach for quantifying the value of information that may be provided to the decision maker. This paper presents a process for determining the value of information that can be gained by evaluating linearly correlated design alternatives. A unique approach to the application of Bayesian Inference is used to provide simulated estimates in the expected utility with increasing observations sizes. The results provide insight into the optimum observation size that maximizes the expected utility when assessing correlated decision alternatives.
Technical Paper

Inverse Modeling: Theory and Engineering Examples

2016-04-05
2016-01-0267
Over the last two decades inverse problems have become increasingly popular due to their widespread applications. This popularity continuously demands designers to find alternative methods, to solve the inverse problems, which are efficient and accurate. It is important to use effective techniques that are both accurate and computationally efficient. This paper presents a method for solving inverse problems through Artificial Neural Network (ANN) theory. The paper also presents a method to apply Grey Wolf optimizer (GWO) algorithm to inverse problems. GWO is a recent optimization method producing superior results. Both methods are then compared to traditional methods such as Particle Swarm Optimization (PSO) and Markov Chain Monte Carlo (MCMC). Four typical engineering design problems are used to compare the four methods. The results show that the GWO outperforms other methods both in terms of efficiency and accuracy.
Journal Article

Bootstrapping and Separable Monte Carlo Simulation Methods Tailored for Efficient Assessment of Probability of Failure of Structural Systems

2015-04-14
2015-01-0420
There is randomness in both the applied loads and the strength of systems. Therefore, to account for the uncertainty, the safety of the system must be quantified using its reliability. Monte Carlo Simulation (MCS) is widely used for probabilistic analysis because of its robustness. However, the high computational cost limits the accuracy of MCS. Smarslok et al. [2010] developed an improved sampling technique for reliability assessment called Separable Monte Carlo (SMC) that can significantly increase the accuracy of estimation without increasing the cost of sampling. However, this method was applied to time-invariant problems involving two random variables. This paper extends SMC to problems with multiple random variables and develops a novel method for estimation of the standard deviation of the probability of failure of a structure. The method is demonstrated and validated on reliability assessment of an offshore wind turbine under turbulent wind loads.
Technical Paper

Combined Approximation for Efficient Reliability Analysis of Linear Dynamic Systems

2015-04-14
2015-01-0424
The Combined Approximation (CA) method is an efficient reanalysis method that aims at reducing the cost of optimization problems. The CA uses results of a single exact analysis, and it is suitable for different types of structures and design variables. The second author utilized CA to calculate the frequency response function of a system at a frequency of interest by using the results at a frequency in the vicinity of that frequency. He showed that the CA yields accurate results for small frequency perturbations. This work demonstrates a methodology that utilizes CA to reduce the cost of Monte Carlo simulation (MCs) of linear systems under random dynamic loads. The main idea is to divide the power spectral density function (PSD) of the input load into several frequency bins before calculating the load realizations.
Technical Paper

Multi-Level Decoupled Optimization of Wind Turbine Structures

2015-04-14
2015-01-0434
This paper proposes a multi-level decoupled method for optimizing the structural design of a wind turbine blade. The proposed method reduces the design space by employing a two-level optimization process. At the high-level, the structural properties of each section are approximated by an exponential function of the distance of that section from the blade root. High-level design variables are the coefficients of this approximating function. Target values for the structural properties of the blade are determined at that level. At the low-level, sections are divided into small decoupled groups. For each section, the low-level optimizer finds the thickness of laminate layers with a minimum mass, whose structural properties meet the targets determined by the high-level optimizer. In the proposed method, each low-level optimizer only considers a small number of design variables for a particular section, while traditional, single-level methods consider all design variables simultaneously.
Technical Paper

Reliability Analysis of Composite Inflatable Space Structures Considering Fracture Failure

2014-04-01
2014-01-0715
Inflatable space structures can have lower launching cost and larger habitat volume than their conventional rigid counterparts. These structures are made of composite laminates, and they are flexible when folded and partially inflated. They contain light-activated resins, and can be cured with the sun light after being inflated in space. A spacecraft can burst due to cracks caused by meteor showers or debris. Therefore, it is critical to identify the important fracture failure modes, and assess their probability. This information will help a designer minimize the risk of failure and keep the mass and cost low. This paper presents a probabilistic approach for finding the required thickness of an inflatable habitat shell for a prescribed reliability level, and demonstrates the superiority of probabilistic design to its deterministic counterpart.
Journal Article

Managing the Computational Cost of Monte Carlo Simulation with Importance Sampling by Considering the Value of Information

2013-04-08
2013-01-0943
Importance Sampling is a popular method for reliability assessment. Although it is significantly more efficient than standard Monte Carlo simulation if a suitable sampling distribution is used, in many design problems it is too expensive. The authors have previously proposed a method to manage the computational cost in standard Monte Carlo simulation that views design as a choice among alternatives with uncertain reliabilities. Information from simulation has value only if it helps the designer make a better choice among the alternatives. This paper extends their method to Importance Sampling. First, the designer estimates the prior probability density functions of the reliabilities of the alternative designs and calculates the expected utility of the choice of the best design. Subsequently, the designer estimates the likelihood function of the probability of failure by performing an initial simulation with Importance Sampling.
Technical Paper

Modeling Dependence and Assessing the Effect of Uncertainty in Dependence in Probabilistic Analysis and Decision Under Uncertainty

2010-04-12
2010-01-0697
A complete probabilistic model of uncertainty in probabilistic analysis and design problems is the joint probability distribution of the random variables. Often, it is impractical to estimate this joint probability distribution because the mechanism of the dependence of the variables is not completely understood. This paper proposes modeling dependence by using copulas and demonstrates their representational power. It also compares this representation with a Monte-Carlo simulation using dispersive sampling.
Journal Article

An RBDO Method for Multiple Failure Region Problems using Probabilistic Reanalysis and Approximate Metamodels

2009-04-20
2009-01-0204
A Reliability-Based Design Optimization (RBDO) method for multiple failure regions is presented. The method uses a Probabilistic Re-Analysis (PRRA) approach in conjunction with an approximate global metamodel with local refinements. The latter serves as an indicator to determine the failure and safe regions. PRRA calculates very efficiently the system reliability of a design by performing a single Monte Carlo (MC) simulation. Although PRRA is based on MC simulation, it calculates “smooth” sensitivity derivatives, allowing therefore, the use of a gradient-based optimizer. An “accurate-on-demand” metamodel is used in the PRRA that allows us to handle problems with multiple disjoint failure regions and potentially multiple most-probable points (MPP). The multiple failure regions are identified by using a clustering technique. A maximin “space-filling” sampling technique is used to construct the metamodel. A vibration absorber example highlights the potential of the proposed method.
Technical Paper

An Efficient Re-Analysis Methodology for Vibration of Large-Scale Structures

2007-05-15
2007-01-2326
Finite element analysis is a well-established methodology in structural dynamics. However, optimization and/or probabilistic studies can be prohibitively expensive because they require repeated FE analyses of large models. Various reanalysis methods have been proposed in order to calculate efficiently the dynamic response of a structure after a baseline design has been modified, without recalculating the new response. The parametric reduced-order modeling (PROM) and the combined approximation (CA) methods are two re-analysis methods, which can handle large model parameter changes in a relatively efficient manner. Although both methods are promising by themselves, they can not handle large FE models with large numbers of DOF (e.g. 100,000) with a large number of design parameters (e.g. 50), which are common in practice. In this paper, the advantages and disadvantages of the PROM and CA methods are first discussed in detail.
Technical Paper

System Reliability-Based Design using a Single-Loop Method

2007-04-16
2007-01-0555
An efficient approach for series system reliability-based design optimization (RBDO) is presented. The key idea is to apportion optimally the system reliability among the failure modes by considering the target values of the failure probabilities of the modes as design variables. Critical failure modes that contribute the most to the overall system reliability are identified. This paper proposes a computationally efficient, system RBDO approach using a single-loop method where the searches for the optimum design and for the most probable failure points proceed simultaneously. Specifically, at each iteration the optimizer uses approximated most probable failure points from the previous iteration to search for the optimum. A second-order Ditlevsen upper bound is used for the joint failure probability of failure modes. Also, an easy to implement active strategy set is employed to improve algorithmic stability.
X