16 resultados para Earthquake hazard analysis

em University of Queensland eSpace - Australia


Relevância:

90.00% 90.00%

Publicador:

Resumo:

Formal methods have significant benefits for developing safety critical systems, in that they allow for correctness proofs, model checking safety and liveness properties, deadlock checking, etc. However, formal methods do not scale very well and demand specialist skills, when developing real-world systems. For these reasons, development and analysis of large-scale safety critical systems will require effective integration of formal and informal methods. In this paper, we use such an integrative approach to automate Failure Modes and Effects Analysis (FMEA), a widely used system safety analysis technique, using a high-level graphical modelling notation (Behavior Trees) and model checking. We inject component failure modes into the Behavior Trees and translate the resulting Behavior Trees to SAL code. This enables us to model check if the system in the presence of these faults satisfies its safety properties, specified by temporal logic formulas. The benefit of this process is tool support that automates the tedious and error-prone aspects of FMEA.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The Accelerating Moment Release (AMR) preceding earthquakes with magnitude above 5 in Australia that occurred during the last 20 years was analyzed to test the Critical Point Hypothesis. Twelve earthquakes in the catalog were chosen based on a criterion for the number of nearby events. Results show that seven sequences with numerous events recorded leading up to the main earthquake exhibited accelerating moment release. Two occurred near in time and space to other earthquakes preceded by AM R. The remaining three sequences had very few events in the catalog so the lack of AMR detected in the analysis may be related to catalog incompleteness. Spatio-temporal scanning of AMR parameters shows that 80% of the areas in which AMR occurred experienced large events. In areas of similar background seismicity with no large events, 10 out of 12 cases exhibit no AMR, and two others are false alarms where AMR was observed but no large event followed. The relationship between AMR and Load-Unload Response Ratio (LURR) was studied. Both methods predict similar critical region sizes, however, the critical point time using AMR is slightly earlier than the time of the critical point LURR anomaly.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Earthquakes have been recognized as resulting from stick-slip frictional instabilities along the faults between deformable rocks. A three-dimensional finite-element code for modeling the nonlinear frictional contact behaviors between deformable bodies with the node-to-point contact element strategy has been developed and applied here to investigate the fault geometry influence on the nucleation and development process of the stick-slip instability along an intra-plate fault through a typical fault bend model, which has a pre-cut fault that is artificially bent by an angle of 5.6degrees at the fault center. The numerical results demonstrate that the geometry of the fault significantly affects nucleation, termination and restart of the stick-slip instability along the intra-plate fault, and all these instability phenomena can be well simulated using the current finite-element algorithm.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We examine the event statistics obtained from two differing simplified models for earthquake faults. The first model is a reproduction of the Block-Slider model of Carlson et al. (1991), a model often employed in seismicity studies. The second model is an elastodynamic fault model based upon the Lattice Solid Model (LSM) of Mora and Place (1994). We performed simulations in which the fault length was varied in each model and generated synthetic catalogs of event sizes and times. From these catalogs, we constructed interval event size distributions and inter-event time distributions. The larger, localised events in the Block-Slider model displayed the same scaling behaviour as events in the LSM however the distribution of inter-event times was markedly different. The analysis of both event size and inter-event time statistics is an effective method for comparative studies of differing simplified models for earthquake faults.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A statistical fractal automaton model is described which displays two modes of dynamical behaviour. The first mode, termed recurrent criticality, is characterised by quasi-periodic, characteristic events that are preceded by accelerating precursory activity. The second mode is more reminiscent of SOC automata in which large events are not preceded by an acceleration in activity. Extending upon previous studies of statistical fractal automata, a redistribution law is introduced which incorporates two model parameters: a dissipation factor and a stress transfer ratio. Results from a parameter space investigation indicate that a straight line through parameter space marks a transition from recurrent criticality to unpredictable dynamics. Recurrent criticality only occurs for models within one corner of the parameter space. The location of the transition displays a simple dependence upon the fractal correlation dimension of the cell strength distribution. Analysis of stress field evolution indicates that recurrent criticality occurs in models with significant long-range stress correlations. A constant rate of activity is associated with a decorrelated stress field.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The aim of this study was to apply multifailure survival methods to analyze time to multiple occurrences of basal cell carcinoma (BCC). Data from 4.5 years of follow-up in a randomized controlled trial, the Nambour Skin Cancer Prevention Trial (1992-1996), to evaluate skin cancer prevention were used to assess the influence of sunscreen application on the time to first BCC and the time to subsequent BCCs. Three different approaches of time to ordered multiple events were applied and compared: the Andersen-Gill, Wei-Lin-Weissfeld, and Prentice-Williams-Peterson models. Robust variance estimation approaches were used for all multifailure survival models. Sunscreen treatment was not associated with time to first occurrence of a BCC (hazard ratio = 1.04, 95% confidence interval: 0.79, 1.45). Time to subsequent BCC tumors using the Andersen-Gill model resulted in a lower estimated hazard among the daily sunscreen application group, although statistical significance was not reached (hazard ratio = 0.82, 95% confidence interval: 0.59, 1.15). Similarly, both the Wei-Lin-Weissfeld marginal-hazards and the Prentice-Williams-Peterson gap-time models revealed trends toward a lower risk of subsequent BCC tumors among the sunscreen intervention group. These results demonstrate the importance of conducting multiple-event analysis for recurring events, as risk factors for a single event may differ from those where repeated events are considered.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Purpose - In many scientific and engineering fields, large-scale heat transfer problems with temperature-dependent pore-fluid densities are commonly encountered. For example, heat transfer from the mantle into the upper crust of the Earth is a typical problem of them. The main purpose of this paper is to develop and present a new combined methodology to solve large-scale heat transfer problems with temperature-dependent pore-fluid densities in the lithosphere and crust scales. Design/methodology/approach - The theoretical approach is used to determine the thickness and the related thermal boundary conditions of the continental crust on the lithospheric scale, so that some important information can be provided accurately for establishing a numerical model of the crustal scale. The numerical approach is then used to simulate the detailed structures and complicated geometries of the continental crust on the crustal scale. The main advantage in using the proposed combination method of the theoretical and numerical approaches is that if the thermal distribution in the crust is of the primary interest, the use of a reasonable numerical model on the crustal scale can result in a significant reduction in computer efforts. Findings - From the ore body formation and mineralization points of view, the present analytical and numerical solutions have demonstrated that the conductive-and-advective lithosphere with variable pore-fluid density is the most favorite lithosphere because it may result in the thinnest lithosphere so that the temperature at the near surface of the crust can be hot enough to generate the shallow ore deposits there. The upward throughflow (i.e. mantle mass flux) can have a significant effect on the thermal structure within the lithosphere. In addition, the emplacement of hot materials from the mantle may further reduce the thickness of the lithosphere. Originality/value - The present analytical solutions can be used to: validate numerical methods for solving large-scale heat transfer problems; provide correct thermal boundary conditions for numerically solving ore body formation and mineralization problems on the crustal scale; and investigate the fundamental issues related to thermal distributions within the lithosphere. The proposed finite element analysis can be effectively used to consider the geometrical and material complexities of large-scale heat transfer problems with temperature-dependent fluid densities.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

(Magill, M., Quinzii, M., 2002. Capital market equilibrium with moral hazard. Journal of Mathematical Economics 38, 149-190) showed that, in a stockmarket economy with private information, the moral hazard problem may be resolved provided that a spanning overlap condition is satisfed. This result depends on the assumption that the technology is given by a stochastic production function with a single scalar input. The object of the present paper is to extend the analysis of Magill and Quinzii to the case of multiple inputs. We show that their main result extends to this general case if and only if, for each firm, the number of linearly independent combinations of securities having payoffs correlated with, but not dependent on, the firms output is equal to the number of degrees of freedom in the firm's production technology.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We present AUSLEM (AUStralian Land Erodibility Model), a land erodibility modelling system that utilizes a rule-set of surficial and climatic thresholds applied through a Geographic Information System (GIs) modelling framework to predict landscape susceptibility to wind erosion. AUSLEM is distinctive in that it quantitatively assesses landscape susceptibility to wind erosion at a 5 x 5 km. spatial resolution on a monthly time-step across Australia. The system was implemented for representative wet (1984), dry (1994), and average rainfall (1997) years with corresponding low, high and moderate dust storm day frequencies. Results demonstrate that AUSLEM can identify landscape erodibility, and provide an interpretation of the physical nature and distribution of erodible landscapes in Australia. Further, results offer an assessment of the dynamic tendencies of erodibility in space and time in response to the El Nino Southern Oscillation (ENSO) and seasonal synoptic scale climate variability. A comparative analysis of AUSLEM output with independent national and international wind erosion, atmospheric aerosol and dust event records indicates a high level of model competency. (c) 2006 Elsevier B.V. All rights reserved.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Fuzzy signal detection analysis can be a useful complementary technique to traditional signal detection theory analysis methods, particularly in applied settings. For example, traffic situations are better conceived as being on a continuum from no potential for hazard to high potential, rather than either having potential or not having potential. This study examined the relative contribution of sensitivity and response bias to explaining differences in the hazard perception performance of novices and experienced drivers, and the effect of a training manipulation. Novice drivers and experienced drivers were compared (N = 64). Half the novices received training, while the experienced drivers and half the novices remained untrained. Participants completed a hazard perception test and rated potential for hazard in occluded scenes. The response latency of participants to the hazard perception test replicated previous findings of experienced/novice differences and trained/untrained differences. Fuzzy signal detection analysis of both the hazard perception task and the occluded rating task suggested that response bias may be more central to hazard perception test performance than sensitivity, with trained and experienced drivers responding faster and with a more liberal bias than untrained novices. Implications for driver training and the hazard perception test are discussed.