931 resultados para Interval discrete log problem
Resumo:
The critical excavation depth of a jointed rock slope is an important problem in rock engineering. This paper studies the critical excavation depth for two idealized jointed rock slopes by employing a face-to-face discrete element method (DEM). The DEM is based on the discontinuity analysis which can consider anisotropic and discontinuous deformations due to joints and their orientations. It uses four lump-points at each surface of rock blocks to describe their interactions. The relationship between the critical excavation depth D-s and the natural slope angle alpha, the joint inclination angle theta as well as the strength parameters of the joints c(r) ,phi(r) is analyzed, and the critical excavation depth obtained with this DEM and the limit equilibrium method (LEM) is compared. Furthermore, effects of joints on the failure modes are compared between DEM simulations and experimental observations. It is found that the DEM predicts a lower critical excavation depth than the LEM if the joint structures in the rock mass are not ignored.
Resumo:
The effects of complex boundary conditions on flows are represented by a volume force in the immersed boundary methods. The problem with this representation is that the volume force exhibits non-physical oscillations in moving boundary simulations. A smoothing technique for discrete delta functions has been developed in this paper to suppress the non-physical oscillations in the volume forces. We have found that the non-physical oscillations are mainly due to the fact that the derivatives of the regular discrete delta functions do not satisfy certain moment conditions. It has been shown that the smoothed discrete delta functions constructed in this paper have one-order higher derivative than the regular ones. Moreover, not only the smoothed discrete delta functions satisfy the first two discrete moment conditions, but also their derivatives satisfy one-order higher moment condition than the regular ones. The smoothed discrete delta functions are tested by three test cases: a one-dimensional heat equation with a moving singular force, a two-dimensional flow past an oscillating cylinder, and the vortex-induced vibration of a cylinder. The numerical examples in these cases demonstrate that the smoothed discrete delta functions can effectively suppress the non-physical oscillations in the volume forces and improve the accuracy of the immersed boundary method with direct forcing in moving boundary simulations.
Resumo:
The first thesis topic is a perturbation method for resonantly coupled nonlinear oscillators. By successive near-identity transformations of the original equations, one obtains new equations with simple structure that describe the long time evolution of the motion. This technique is related to two-timing in that secular terms are suppressed in the transformation equations. The method has some important advantages. Appropriate time scalings are generated naturally by the method, and don't need to be guessed as in two-timing. Furthermore, by continuing the procedure to higher order, one extends (formally) the time scale of valid approximation. Examples illustrate these claims. Using this method, we investigate resonance in conservative, non-conservative and time dependent problems. Each example is chosen to highlight a certain aspect of the method.
The second thesis topic concerns the coupling of nonlinear chemical oscillators. The first problem is the propagation of chemical waves of an oscillating reaction in a diffusive medium. Using two-timing, we derive a nonlinear equation that determines how spatial variations in the phase of the oscillations evolves in time. This result is the key to understanding the propagation of chemical waves. In particular, we use it to account for certain experimental observations on the Belusov-Zhabotinskii reaction.
Next, we analyse the interaction between a pair of coupled chemical oscillators. This time, we derive an equation for the phase shift, which measures how much the oscillators are out of phase. This result is the key to understanding M. Marek's and I. Stuchl's results on coupled reactor systems. In particular, our model accounts for synchronization and its bifurcation into rhythm splitting.
Finally, we analyse large systems of coupled chemical oscillators. Using a continuum approximation, we demonstrate mechanisms that cause auto-synchronization in such systems.
Resumo:
This dissertation is concerned with the development of a new discrete element method (DEM) based on Non-Uniform Rational Basis Splines (NURBS). With NURBS, the new DEM is able to capture sphericity and angularity, the two particle morphological measures used in characterizing real grain geometries. By taking advantage of the parametric nature of NURBS, the Lipschitzian dividing rectangle (DIRECT) global optimization procedure is employed as a solution procedure to the closest-point projection problem, which enables the contact treatment of non-convex particles. A contact dynamics (CD) approach to the NURBS-based discrete method is also formulated. By combining particle shape flexibility, properties of implicit time-integration, and non-penetrating constraints, we target applications in which the classical DEM either performs poorly or simply fails, i.e., in granular systems composed of rigid or highly stiff angular particles and subjected to quasistatic or dynamic flow conditions. The CD implementation is made simple by adopting a variational framework, which enables the resulting discrete problem to be readily solved using off-the-shelf mathematical programming solvers. The capabilities of the NURBS-based DEM are demonstrated through 2D numerical examples that highlight the effects of particle morphology on the macroscopic response of granular assemblies under quasistatic and dynamic flow conditions, and a 3D characterization of material response in the shear band of a real triaxial specimen.
Resumo:
This article investigates the convergence properties of iterative processes involving sequences of self-mappings of metric or Banach spaces. Such sequences are built from a set of primary self-mappings which are either expansive or non-expansive self-mappings and some of the non-expansive ones can be contractive including the case of strict contractions. The sequences are built subject to switching laws which select each active self-mapping on a certain activation interval in such a way that essential properties of boundedness and convergence of distances and iterated sequences are guaranteed. Applications to the important problem of stability of dynamic switched systems are also given.
Resumo:
A extração de regras de associação (ARM - Association Rule Mining) de dados quantitativos tem sido pesquisa de grande interesse na área de mineração de dados. Com o crescente aumento das bases de dados, há um grande investimento na área de pesquisa na criação de algoritmos para melhorar o desempenho relacionado a quantidade de regras, sua relevância e a performance computacional. O algoritmo APRIORI, tradicionalmente usado na extração de regras de associação, foi criado originalmente para trabalhar com atributos categóricos. Geralmente, para usá-lo com atributos contínuos, ou quantitativos, é necessário transformar os atributos contínuos, discretizando-os e, portanto, criando categorias a partir dos intervalos discretos. Os métodos mais tradicionais de discretização produzem intervalos com fronteiras sharp, que podem subestimar ou superestimar elementos próximos dos limites das partições, e portanto levar a uma representação imprecisa de semântica. Uma maneira de tratar este problema é criar partições soft, com limites suavizados. Neste trabalho é utilizada uma partição fuzzy das variáveis contínuas, que baseia-se na teoria dos conjuntos fuzzy e transforma os atributos quantitativos em partições de termos linguísticos. Os algoritmos de mineração de regras de associação fuzzy (FARM - Fuzzy Association Rule Mining) trabalham com este princípio e, neste trabalho, o algoritmo FUZZYAPRIORI, que pertence a esta categoria, é utilizado. As regras extraídas são expressas em termos linguísticos, o que é mais natural e interpretável pelo raciocício humano. Os algoritmos APRIORI tradicional e FUZZYAPRIORI são comparado, através de classificadores associativos, baseados em regras extraídas por estes algoritmos. Estes classificadores foram aplicados em uma base de dados relativa a registros de conexões TCP/IP que destina-se à criação de um Sistema de Detecção de Intrusos.
Resumo:
An asymptotic recovery design procedure is proposed for square, discrete-time, linear, time-invariant multivariable systems, which allows a state-feedback design to be approximately recovered by a dynamic output feedback scheme. Both the case of negligible processing time (compared to the sampling interval) and of significant processing time are discussed. In the former case, it is possible to obtain perfect. © 1985 IEEE.
Resumo:
This paper discusses the application of Discrete Event Simulation (DES) in modelling the complex relationship between patient types, case-mix and operating theatre allocation in a large National Health Service (NHS) Trust in London. The simulation model that was constructed described the main features of nine theatres, focusing on operational processes and patient throughput times. The model was used to test three scenarios of case-mix and to demonstrate the potential of using simulation modelling as a cost effective method for understanding the issues of healthcare operations management and the role of simulation techniques in problem solving. The results indicated that removing all day cases will reduce patient throughput by 23.3% and the utilization of the orthopaedic theatre in particular by 6.5%. This represents a case example of how DES can be used by healthcare managers to inform decision making. © 2008 IEEE.
Resumo:
The paper presents centrifuge test data of the problem of tunnelling effects on buried pipelines and compares them to predictions made using DEM simulations. The paper focuses on the examination of pipeline bending moments, their distribution along the pipe, and their development with tunnel volume loss. Centrifuge results are obtained by PIV analysis and compared to results obtained using the DEM model. The DEM model was built to replicate the centrifuge model as closely as possible and included numerical features formulated specially for this task, such as structural elements to replicate the tunnel and pipeline. Results are extremely encouraging, with deviations between DEM and centrifuge test bending moment results being very small. © 2010 Taylor & Francis Group, London.
Resumo:
We study the problem of finding a local minimum of a multilinear function E over the discrete set {0,1}n. The search is achieved by a gradient-like system in [0,1]n with cost function E. Under mild restrictions on the metric, the stable attractors of the gradient-like system are shown to produce solutions of the problem, even when they are not in the vicinity of the discrete set {0,1}n. Moreover, the gradient-like system connects with interior point methods for linear programming and with the analog neural network studied by Vidyasagar (IEEE Trans. Automat. Control 40 (8) (1995) 1359), in the same context. © 2004 Elsevier B.V. All rights reserved.
Resumo:
A small strain two-dimensional discrete dislocation plasticity framework coupled to vacancy diffusion is developed wherein the motion of edge dislocations is by a combination of glide and climb. The dislocations are modelled as line defects in a linear elastic medium and the mechanical boundary value problem is solved by the superposition of the infinite medium elastic fields of the dislocations and a complimentary non-singular solution that enforces the boundary conditions. Similarly, the climbing dislocations are modelled as line sources/sinks of vacancies and the vacancy diffusion boundary value problem is also solved by a superposition of the fields of the line sources/sinks in an infinite medium and a complementary non-singular solution that enforces the boundary conditions. The vacancy concentration field along with the stress field provides the climb rate of the dislocations. Other short-range interactions of the dislocations are incorporated via a set of constitutive rules. We first employ this formulation to investigate the climb of a single edge dislocation in an infinite medium and illustrate the existence of diffusion-limited and sink-limited climb regimes. Next, results are presented for the pure bending and uniaxial tension of single crystals oriented for single slip. These calculations show that plasticity size effects are reduced when dislocation climb is permitted. Finally, we contrast predictions of this coupled framework with an ad hoc model in which dislocation climb is modelled by a drag-type relation based on a quasi steady-state solution. © 2013 Elsevier Ltd. All rights reserved.
Resumo:
Heart disease is one of the main factor causing death in the developed countries. Over several decades, variety of electronic and computer technology have been developed to assist clinical practices for cardiac performance monitoring and heart disease diagnosis. Among these methods, Ballistocardiography (BCG) has an interesting feature that no electrodes are needed to be attached to the body during the measurement. Thus, it is provides a potential application to asses the patients heart condition in the home. In this paper, a comparison is made for two neural networks based BCG signal classification models. One system uses a principal component analysis (PCA) method, and the other a discrete wavelet transform, to reduce the input dimensionality. It is indicated that the combined wavelet transform and neural network has a more reliable performance than the combined PCA and neural network system. Moreover, the wavelet transform requires no prior knowledge of the statistical distribution of data samples and the computation complexity and training time are reduced.
Resumo:
This paper addresses the problem of nonlinear multivariate root finding. In an earlier paper we described a system called Newton which finds roots of systems of nonlinear equations using refinements of interval methods. The refinements are inspired by AI constraint propagation techniques. Newton is competative with continuation methods on most benchmarks and can handle a variety of cases that are infeasible for continuation methods. This paper presents three "cuts" which we believe capture the essential theoretical ideas behind the success of Newton. This paper describes the cuts in a concise and abstract manner which, we believe, makes the theoretical content of our work more apparent. Any implementation will need to adopt some heuristic control mechanism. Heuristic control of the cuts is only briefly discussed here.
Resumo:
The class of all Exponential-Polynomial-Trigonometric (EPT) functions is classical and equal to the Euler-d’Alembert class of solutions of linear differential equations with constant coefficients. The class of non-negative EPT functions defined on [0;1) was discussed in Hanzon and Holland (2010) of which EPT probability density functions are an important subclass. EPT functions can be represented as ceAxb, where A is a square matrix, b a column vector and c a row vector where the triple (A; b; c) is the minimal realization of the EPT function. The minimal triple is only unique up to a basis transformation. Here the class of 2-EPT probability density functions on R is defined and shown to be closed under a variety of operations. The class is also generalised to include mixtures with the pointmass at zero. This class coincides with the class of probability density functions with rational characteristic functions. It is illustrated that the Variance Gamma density is a 2-EPT density under a parameter restriction. A discrete 2-EPT process is a process which has stochastically independent 2-EPT random variables as increments. It is shown that the distribution of the minimum and maximum of such a process is an EPT density mixed with a pointmass at zero. The Laplace Transform of these distributions correspond to the discrete time Wiener-Hopf factors of the discrete time 2-EPT process. A distribution of daily log-returns, observed over the period 1931-2011 from a prominent US index, is approximated with a 2-EPT density function. Without the non-negativity condition, it is illustrated how this problem is transformed into a discrete time rational approximation problem. The rational approximation software RARL2 is used to carry out this approximation. The non-negativity constraint is then imposed via a convex optimisation procedure after the unconstrained approximation. Sufficient and necessary conditions are derived to characterise infinitely divisible EPT and 2-EPT functions. Infinitely divisible 2-EPT density functions generate 2-EPT Lévy processes. An assets log returns can be modelled as a 2-EPT Lévy process. Closed form pricing formulae are then derived for European Options with specific times to maturity. Formulae for discretely monitored Lookback Options and 2-Period Bermudan Options are also provided. Certain Greeks, including Delta and Gamma, of these options are also computed analytically. MATLAB scripts are provided for calculations involving 2-EPT functions. Numerical option pricing examples illustrate the effectiveness of the 2-EPT approach to financial modelling.
Resumo:
BACKGROUND: Dropouts and missing data are nearly-ubiquitous in obesity randomized controlled trails, threatening validity and generalizability of conclusions. Herein, we meta-analytically evaluate the extent of missing data, the frequency with which various analytic methods are employed to accommodate dropouts, and the performance of multiple statistical methods. METHODOLOGY/PRINCIPAL FINDINGS: We searched PubMed and Cochrane databases (2000-2006) for articles published in English and manually searched bibliographic references. Articles of pharmaceutical randomized controlled trials with weight loss or weight gain prevention as major endpoints were included. Two authors independently reviewed each publication for inclusion. 121 articles met the inclusion criteria. Two authors independently extracted treatment, sample size, drop-out rates, study duration, and statistical method used to handle missing data from all articles and resolved disagreements by consensus. In the meta-analysis, drop-out rates were substantial with the survival (non-dropout) rates being approximated by an exponential decay curve (e(-lambdat)) where lambda was estimated to be .0088 (95% bootstrap confidence interval: .0076 to .0100) and t represents time in weeks. The estimated drop-out rate at 1 year was 37%. Most studies used last observation carried forward as the primary analytic method to handle missing data. We also obtained 12 raw obesity randomized controlled trial datasets for empirical analyses. Analyses of raw randomized controlled trial data suggested that both mixed models and multiple imputation performed well, but that multiple imputation may be more robust when missing data are extensive. CONCLUSION/SIGNIFICANCE: Our analysis offers an equation for predictions of dropout rates useful for future study planning. Our raw data analyses suggests that multiple imputation is better than other methods for handling missing data in obesity randomized controlled trials, followed closely by mixed models. We suggest these methods supplant last observation carried forward as the primary method of analysis.