84 resultados para deterministic fractals


Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper describes an approach to pointwise construction of general aggregation operators, based on monotone Lipschitz approximation. The aggregation operators are constructed from a set of desired values at certain points, or from empirically collected data. It establishes tight upper and lower bounds on Lipschitz aggregation operators with a number of different properties, as well as the optimal aggregation operator, consistent with the given values. We consider conjunctive, disjunctive and idempotent n-ary aggregation operators; p-stable aggregation operators; various choices of the neutral element and annihilator; diagonal, opposite diagonal and marginal sections; bipolar and double aggregation operators. In all cases we provide either explicit formulas or deterministic numerical procedures to determine the bounds. The findings of this paper are useful for construction of aggregation operators with specified properties, especially using interpolation schemata.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The inherent variability in incoming material and process conditions in sheet metal forming makes quality control and the maintenance of consistency extremely difficult. A single FEM simulation is successful at predicting the formability for a given system, however lacks the ability to capture the variability in an actual production process due to the numerical deterministic nature. This paper investigates a probabilistic analytical model where the variation of five input parameters and their relationship to the sensitivity of springback in a stamping process is examined. A range of sheet tensions are investigated, simulating different operating windows in an attempt to highlight robust regions where the distribution of springback is small. A series of FEM simulations were also performed, to compare with the findings from the analytical model using AutoForm Sigma v4.04 and to validate the analytical model assumptions.

Results show that an increase in sheet tension not only decreases springback, but more importantly reduces the sensitivity of the process to variation. A relative sensitivity analysis has been performed where the most influential parameters and the changes in sensitivity at various sheet tensions have been investigated. Variation in the material parameters, yield stress and n-value were the most influential causes of springback variation, when compared to process input parameters such as friction, which had a small effect. The probabilistic model presented allows manufacturers to develop a more comprehensive assessment of the success of their forming processes by capturing the effects of inherent variation.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The eigenvector associated with the smallest eigenvalue of the autocorrelation matrix of input signals is called minor component. Minor component analysis (MCA) is a statistical approach for extracting minor component from input signals and has been applied in many fields of signal processing and data analysis. In this letter, we propose a neural networks learning algorithm for estimating adaptively minor component from input signals. Dynamics of the proposed algorithm are analyzed via a deterministic discrete time (DDT) method. Some sufficient conditions are obtained to guarantee convergence of the proposed algorithm.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The stability of minor component analysis (MCA) learning algorithms is an important problem in many signal processing applications. In this paper, we propose an effective MCA learning algorithm that can offer better stability. The dynamics of the proposed algorithm are analyzed via a corresponding deterministic discrete time (DDT) system. It is proven that if the learning rate satisfies some mild conditions, almost all trajectories of the DDT system starting from points in an invariant set are bounded, and will converge to the minor component of the autocorrelation matrix of the input data. Simulation results will be furnished to illustrate the theoretical results achieved.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Minor component analysis (MCA) is an important statistical tool for signal processing and data analysis. Neural networks can be used to extract online minor component from input data. Compared with traditional algebraic  approaches, a neural network method has a lower computational complexity. Stability of neural networks learning algorithms is crucial to practical applications. In this paper, we propose a stable MCA neural networks learning algorithm, which has a more satisfactory numerical stability than some existing MCA algorithms. Dynamical behaviors of the proposed algorithm are analyzed via deterministic discrete time (DDT) method and the conditions are obtained to guarantee convergence. Simulations are carried out to illustrate the theoretical results achieved.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

There has been an increasing interest in face recognition in recent years. Many recognition methods have been developed so far, some very encouraging. A key remaining issue is the existence of variations in the input face image. Today, methods exist that can handle specific image variations. But we are yet to see methods that can be used more effectively in unconstrained situations. This paper presents a method that can handle partial translation, rotation, or scale variations in the input face image. The principal is to automatically identify objects within images using their partial self-similarities. The paper presents two recognition methods which can be used to recognise objects within images. A face recognition system is then presented that is insensitive to limited translation, rotation, or scale variations in the input face image. The performance of the system is evaluated through four experiments. The results show that the system achieves higher recognition rates than those of a number of existing approaches.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

One of the key challenges in geographic routing is how to deal with dead-ends, where greedy routing fails to find a neighbor node which is closer to the destination. Most existing geographic routing algorithms just switch to the deterministic face routing or limits its face searching range. In this paper, we demonstrate that we can improve routing performance by considering local connectivity status at each node before making routing decision. We present a protocol, Density Ripple Exchange (DRE), that maintains local density information at each node, and a new geographic routing algorithm, Geographic Ripple Routing (GRR), that achieves better routing performance in both hop stretch and transmission stretch than existing geographic routing algorithms by exploiting available connectivity information. Our simulations demonstrate that we increased the performance for GRR over Greedy Perimeter Stateless Routing (GPSR) by about 15%. The cost of this improved performance is a small amount of additional local connectivity information required for our algorithm.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Over the last couple of months a large number of distributed denial of service (DDoS) attacks have occurred across the world, especially targeting those who provide Web services. IP traceback, a counter measure against DDoS, is the ability to trace IP packets back to the true source/s of the attack. In this paper, an IP traceback scheme using a machine learning technique called intelligent decision prototype (IDP), is proposed. IDP can be used on both probabilistic packet marking (PPM) and deterministic packet marking (DPM) traceback schemes to identify DDoS attacks. This will greatly reduce the packets that are marked and in effect make the system more efficient and effective at tracing the source of an attack compared with other methods. IDP can be applied to many security systems such as data mining, forensic analysis, intrusion detection systems (IDS) and DDoS defense systems.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The sheet forming industry is plagued by inherent variations in its many input variables, making quality control and improvements a major hurdle. This is particularly poignant for Advanced High Strength Steels (AHSS), which exhibit a large degree of property variability. Current FE-based simulation packages are successful at predicting the manufacturability of a particular sheet metal components, however, due to their numerical deterministic nature are inherently unable to predict the performance of a real-life production process. Though they are now beginning to incorporate the stochastic nature of production in their codes. This work investigates the accuracy and precision of a current stochastic simulation package, AutoForm Sigma v4.1, by developing an experimental data set where all main sources of variation are captured through precise measurements and standard tensile tests. Using a Dual Phase 600Mpa grade steel a series of semi-cylindrical channels are formed at two Blank Holder Pressure levels where the response metric is the variation in springback determined by the flange angle. The process is replicated in AutoForm Sigma and an assessment of accuracy and precision of the predictions are performed. Results indicate a very good correspondence to the experimental trials, with mean springback response predicted to within 1 ° of the flange angle and the interquartile spread of results to within 0.22°.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Most ecological and evolutionary processes are thought to critically depend on dispersal and individual movement but there is little empirical information on the movement strategies used by animals to find resources. In particular, it is unclear whether behavioural variation exists at all scales, or whether behavioural decisions are primarily made at small spatial scales and thus broad-scale patterns of movement simply reflect underlying resource distributions. We evaluated animal movement responses to variable resource distributions using the grey teal (Anas gracilis) in agricultural and desert landscapes in Australia as a model system. Birds in the two landscapes differed in the fractal dimension of their movement paths, with teal in the desert landscape moving less tortuously overall than their counterparts in the agricultural landscape. However, the most striking result was the high levels of individual variability in movement strategies, with different animals exhibiting different responses to the same resources. Teal in the agricultural basin moved with both high and low tortuosity, while teal in the desert basin primarily moved using low levels of tortuosity. These results call into question the idea that broad-scale movement patterns simply reflect underlying resource distributions, and suggest that movement responses in some animals may be behaviourally complex regardless of the spatial scale over which movement occurs.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The development and use of cocycles for analysis of non-autonomous behaviour is a technique that has been known for several years. Initially developed as an extension to semi-group theory for studying rion-autonornous behaviour, it was extensively used in analysing random dynamical systems [2, 9, 10, 12]. Many of the results regarding asymptotic behaviour developed for random dynamical systems, including the concept of cocycle attractors were successfully transferred and reinterpreted for deterministic non-autonomous systems primarily by P. Kloeden and B. Schmalfuss [20, 21, 28, 29]. The theory concerning cocycle attractors was later developed in various contexts specific to particular classes of dynamical systems [6, 7, 13], although a comprehensive understanding of cocycle attractors (redefined as pullback attractors within this thesis) and their role in the stability of non-autonomous dynamical systems was still at this stage incomplete. It was this purpose that motivated Chapters 1-3 to define and formalise the concept of stability within non-autonomous dynamical systems. The approach taken incorporates the elements of classical asymptotic theory, and refines the notion of pullback attraction with further development towards a study of pull-back stability arid pullback asymptotic stability. In a comprehensive manner, it clearly establishes both pullback and forward (classical) stability theory as fundamentally unique and essential components of non-autonomous stability. Many of the introductory theorems and examples highlight the key properties arid differences between pullback and forward stability. The theory also cohesively retains all the properties of classical asymptotic stability theory in an autonomous environment. These chapters are intended as a fundamental framework from which further research in the various fields of non-autonomous dynamical systems may be extended. A preliminary version of a Lyapunov-like theory that characterises pullback attraction is created as a tool for examining non-autonomous behaviour in Chapter 5. The nature of its usefulness however is at this stage restricted to the converse theorem of asymptotic stability. Chapter 7 introduces the theory of Loci Dynamics. A transformation is made to an alternative dynamical system where forward asymptotic (classical asymptotic) behaviour characterises pullback attraction to a particular point in the original dynamical system. This has the advantage in that certain conventional techniques for a forward analysis may be applied. The remainder of the thesis, Chapters 4, 6 and Section 7.3, investigates the effects of perturbations and discretisations on non-autonomous dynamical systems known to possess structures that exhibit some form of stability or attraction. Chapter 4 investigates autonomous systems with semi-group attractors, that have been non-autonomously perturbed, whilst Chapter 6 observes the effects of discretisation on non-autonomous dynamical systems that exhibit properties of forward asymptotic stability. Chapter 7 explores the same problem of discretisation, but for pullback asymptotically stable systems. The theory of Loci Dynamics is used to analyse the nature of the discretisation, but establishment of results directly analogous to those discovered in Chapter 6 is shown to be unachievable. Instead a case by case analysis is provided for specific classes of dynamical systems, for which the results generate a numerical approximation of the pullback attraction in the original continuous dynamical system. The nature of the results regarding discretisation provide a non-autonomous extension to the work initiated by A. Stuart and J. Humphries [34, 35] for the numerical approximation of semi-group attractors within autonomous systems. . Of particular importance is the effect on the system's asymptotic behaviour over non-finite intervals of discretisation.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This thesis reports on a quantitative exposure assessment and on an analysis of the attributes of the data used in the estimations, in particular distinguishing between its uncertainty and variability. A retrospective assessment of exposure to benzene was carried out for a case control study of leukaemia in the Australian petroleum industry. The study used the mean of personal task-based measurements (Base Estimates) in a deterministic algorithm and applied factors to model back to places, times etc for which no exposure measurements were available. Mean daily exposures were estimated, on an individual subject basis, by summing the task-based exposures. These mean exposures were multiplied by the years spent on each job to provide exposure estimates in ppm-years. These were summed to provide a Cumulative Estimate for each subject. Validation was completed for the model and key inputs. Exposures were low, most jobs were below TWA of 5 ppm benzene. Exposures in terminals were generally higher than at refineries. Cumulative Estimates ranged from 0.005 to 50.9 ppm-years, with 84 percent less than 10 ppm-years. Exposure probability distributions were developed for tanker drivers using Monte Carlo simulation of the exposure estimation algorithm. The outcome was a lognormal distribution of exposure for each driver. These provide the basis for alternative risk assessment metrics e.g. the frequency of short but intense exposures which provided only a minimal contribution to the long-term average exposure but may increase risk of leukaemia. The effect of different inputs to the model were examined and their significance assessed using Monte Carlo simulation. The Base Estimates were the most important determinant of exposure in the model. The sources of variability in the measured data were examined, including the effect of having censored data and the between and within-worker variability. The sources of uncertainty in the exposure estimates were analysed and consequential improvements in exposure assessment identified. Monte Carlo sampling was also used to examine the uncertainties and variability associated with the tanker drivers' exposure assessment, to derive an estimate of the range and to put confidence intervals on the daily mean exposures. The identified uncertainty was less than the variability associated with the estimates. The traditional approach to exposure estimation typically derives only point estimates of mean exposure. The approach developed here allows a range of exposure estimates to be made and provides a more flexible and improved basis for risk assessment.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Shared clusters represent an excellent platform for the execution of parallel applications given their low price/performance ratio and the presence of cluster infrastructure in many organisations. The focus of recent research efforts are on parallelism management, transport and efficient access to resources, and making clusters easy to use. In this thesis, we examine reliable parallel computing on clusters. The aim of this research is to demonstrate the feasibility of developing an operating system facility providing transport fault tolerance using existing, enhanced and newly built operating system services for supporting parallel applications. In particular, we use existing process duplication and process migration services, and synthesise a group communications facility for use in a transparent checkpointing facility. This research is carried out using the methods of experimental computer science. To provide a foundation for the synthesis of the group communications and checkpointing facilities, we survey and review related work in both fields. For group communications, we examine the V Distributed System, the x-kernel and Psync, the ISIS Toolkit, and Horus. We identify a need for services that consider the placement of processes on computers in the cluster. For Checkpointing, we examine Manetho, KeyKOS, libckpt, and Diskless Checkpointing. We observe the use of remote computer memories for storing checkpoints, and the use of copy-on-write mechanisms to reduce the time to create a checkpoint of a process. We propose a group communications facility providing two sets of services: user-oriented services and system-oriented services. User-oriented services provide transparency and target application. System-oriented services supplement the user-oriented services for supporting other operating systems services and do not provide transparency. Additional flexibility is achieved by providing delivery and ordering semantics independently. An operating system facility providing transparent checkpointing is synthesised using coordinated checkpointing. To ensure a consistent set of checkpoints are generated by the facility, instead of blindly blocking the processes of a parallel application, only non-deterministic events are blocked. This allows the processes of the parallel application to continue execution during the checkpoint operation. Checkpoints are created by adapting process duplication mechanisms, and checkpoint data is transferred to remote computer memories and disk for storage using the mechanisms of process migration. The services of the group communications facility are used to coordinate the checkpoint operation, and to transport checkpoint data to remote computer memories and disk. Both the group communications facility and the checkpointing facility have been implemented in the GENESIS cluster operating system and provide proof-of-concept. GENESIS uses a microkernel and client-server based operating system architecture, and is demonstrated to provide an appropriate environment for the development of these facilities. We design a number of experiments to test the performance of both the group communications facility and checkpointing facility, and to provide proof-of-performance. We present our approach to testing, the challenges raised in testing the facilities, and how we overcome them. For group communications, we examine the performance of a number of delivery semantics. Good speed-ups are observed and system-oriented group communication services are shown to provide significant performance advantages over user-oriented semantics in the presence of packet loss. For checkpointing, we examine the scalability of the facility given different levels of resource usage and a variable number of computers. Low overheads are observed for checkpointing a parallel application. It is made clear by this research that the microkernel and client-server based cluster operating system provide an ideal environment for the development of a high performance group communications facility and a transparent checkpointing facility for generating a platform for reliable parallel computing on clusters.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The fact that an occurrence of a unit root in real output is inconsistent with the notion that business cycles are stationary fluctuations around a deterministic trend makes this an important topic for empirical investigation. We examine this issue for 24 Chinese provinces using the recently developed Lagrange multiplier panel unit root test which allows for a structural break. Our main finding is that real gross domestic product (GDP) and real GDP per capita for Chinese provinces are stationary fluctuations around a deterministic trend.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Recently, many unified learning algorithms have been developed to solve the task of principal component analysis (PCA) and minor component analysis (MCA). These unified algorithms can be used to extract principal component and if altered simply by the sign, it can also serve as a minor component extractor. This is of practical significance in the implementations of algorithms. Convergence of the existing unified algorithms is guaranteed only under the condition that the learning rates of algorithms approach zero, which is impractical in many practical applications. In this paper, we propose a unified PCA & MCA algorithm with a constant learning rate, and derive the sufficient conditions to guarantee convergence via analyzing the discrete-time dynamics of the proposed algorithm. The achieved theoretical results lay a solid foundation for the applications of our proposed algorithm.