948 resultados para soft computing methods
Resumo:
A variety of current and future wired and wireless networking technologies can be transformed into a seamless communication environments through application of context-based vertical handovers. Such seamless communication environments are needed for future pervasive/ubiquitous systems. Pervasive systems are context aware and need to adapt to context changes, including network disconnections and changes in network Quality of Service (QoS). Vertical handover is one of many possible adaptation methods. It allows users to roam freely between heterogeneous networks while maintaining the continuity of their applications. This paper proposes a vertical handover mechanism suitable for multimedia applications in pervasive systems. The paper focuses on the handover decision making process which uses context information regarding user devices, user location, network environment and requested QoS. (C) 2004 Elsevier B.V. All rights reserved.
Resumo:
This paper gives a review of recent progress in the design of numerical methods for computing the trajectories (sample paths) of solutions to stochastic differential equations. We give a brief survey of the area focusing on a number of application areas where approximations to strong solutions are important, with a particular focus on computational biology applications, and give the necessary analytical tools for understanding some of the important concepts associated with stochastic processes. We present the stochastic Taylor series expansion as the fundamental mechanism for constructing effective numerical methods, give general results that relate local and global order of convergence and mention the Magnus expansion as a mechanism for designing methods that preserve the underlying structure of the problem. We also present various classes of explicit and implicit methods for strong solutions, based on the underlying structure of the problem. Finally, we discuss implementation issues relating to maintaining the Brownian path, efficient simulation of stochastic integrals and variable-step-size implementations based on various types of control.
Resumo:
What is the computational power of a quantum computer? We show that determining the output of a quantum computation is equivalent to counting the number of solutions to an easily computed set of polynomials defined over the finite field Z(2). This connection allows simple proofs to be given for two known relationships between quantum and classical complexity classes, namely BQP subset of P-#P and BQP subset of PP.
Resumo:
The uneven distribution of women and men in IT employment is often depicted as reflecting adistinction between 'hard' and 'soft' tasks, skills and attributes. This article uses detailed occupational data on professional computing jobs in Australia to examine the extent to which patterns of gender segregation are consistent with such dichotomies. Additionally, we draw on qualitative interview data from aset oforganisational case studies for insights into the ways in which segregation patterns are reproduced and/or reshaped at"the workplace level. While perceptions ofgendered dichotomies were evident among many of our interviewees, overall our analysis shows considerably more complexity, with segregation patterns not necessarily aligned with clear-cut dichotomies and career directions often directly influenced by the organisation ofworking time in particular occupational streams.
Resumo:
Systems biology is based on computational modelling and simulation of large networks of interacting components. Models may be intended to capture processes, mechanisms, components and interactions at different levels of fidelity. Input data are often large and geographically disperse, and may require the computation to be moved to the data, not vice versa. In addition, complex system-level problems require collaboration across institutions and disciplines. Grid computing can offer robust, scaleable solutions for distributed data, compute and expertise. We illustrate some of the range of computational and data requirements in systems biology with three case studies: one requiring large computation but small data (orthologue mapping in comparative genomics), a second involving complex terabyte data (the Visible Cell project) and a third that is both computationally and data-intensive (simulations at multiple temporal and spatial scales). Authentication, authorisation and audit systems are currently not well scalable and may present bottlenecks for distributed collaboration particularly where outcomes may be commercialised. Challenges remain in providing lightweight standards to facilitate the penetration of robust, scalable grid-type computing into diverse user communities to meet the evolving demands of systems biology.
Resumo:
This paper describes methods used to support collaboration and communication between practitioners, designers and engineers when designing ubiquitous computing systems. We tested methods such as “Wizard of Oz” and design games in a real domain, the dental surgery, in an attempt to create a system that is: affordable; minimally disruptive of the natural flow of work; and improves human-computer interaction. In doing so we found that such activities allowed the practitioners to be on a ‘level playing ground’ with designers and engineers. The findings we present suggest that dentists are willing to engage in detailed exploration and constructive critique of technical design possibilities if the design ideas and prototypes are presented in the context of their work practice and are of a resolution and relevance that allow them to jointly explore and question with the design time. This paper is an extension of a short paper submitted to the Participatory Design Conference, 2004.
Resumo:
Many variables that are of interest in social science research are nominal variables with two or more categories, such as employment status, occupation, political preference, or self-reported health status. With longitudinal survey data it is possible to analyse the transitions of individuals between different employment states or occupations (for example). In the statistical literature, models for analysing categorical dependent variables with repeated observations belong to the family of models known as generalized linear mixed models (GLMMs). The specific GLMM for a dependent variable with three or more categories is the multinomial logit random effects model. For these models, the marginal distribution of the response does not have a closed form solution and hence numerical integration must be used to obtain maximum likelihood estimates for the model parameters. Techniques for implementing the numerical integration are available but are computationally intensive requiring a large amount of computer processing time that increases with the number of clusters (or individuals) in the data and are not always readily accessible to the practitioner in standard software. For the purposes of analysing categorical response data from a longitudinal social survey, there is clearly a need to evaluate the existing procedures for estimating multinomial logit random effects model in terms of accuracy, efficiency and computing time. The computational time will have significant implications as to the preferred approach by researchers. In this paper we evaluate statistical software procedures that utilise adaptive Gaussian quadrature and MCMC methods, with specific application to modeling employment status of women using a GLMM, over three waves of the HILDA survey.
Resumo:
Pervasive systems need to be context aware and need to adapt to context changes, including network disconnections and changes in network Quality of Service (QoS). Vertical handover (handover between heterogeneous networks) is one of possible adaptation methods. It allows users to roam freely between heterogeneous networks while maintaining continuity of their applications. This paper proposes a vertical handover approach suitable for multimedia applications in pervasive systems. It describes the adaptability decision making process which uses vertical handovers to support users mobility and provision of QoS suitable for users’ applications. The process evaluates context information regarding user devices, User location, network environment, and user perceived QoS of applications.
Resumo:
While developments in distributed object computing environments, such as the Common Object Request Broker Architecture (CORBA) [17] and the Telecommunication Intelligent Network Architecture (TINA) [16], have enabled interoperability between domains in large open distributed systems, managing the resources within such systems has become an increasingly complex task. This challenge has been considered for several years within the distributed systems management research community and policy-based management has recently emerged as a promising solution. Large evolving enterprises present a significant challenge for policy-based management partly due to the requirement to support both mutual transparency and individual autonomy between domains [2], but also because the fluidity and complexity of interactions occurring within such environments requires an ability to cope with the coexistence of multiple, potentially inconsistent policies. This paper discusses the need of providing both dynamic (run-time) and static (compile-time) conflict detection and resolution for policies in such systems and builds on our earlier conflict detection work [7, 8] to introduce the methods for conflict resolution in large open distributed systems.
Resumo:
Purpose: To determine the critical fitting characteristics of modern soft contact lens fits and from this to devise a simplified recording scheme. Methods: Ten subjects (aged 28.1 ± 7.4 years) wore eight different modern soft contact lenses. Video was captured and analysed of blink (central and up-gaze), excursion lag (up, down, right and left gaze) and push-up movement, centration and coverage. Results: Lens centration was on average close to the corneal centre. Movement on blink was significantly smaller in up-gaze than in primary-gaze (p<0.001). Lag was greatest in down-gaze and least in up-gaze (p<0.001). Push-up test recovery speed was 1.32±0.73mm/s. Overall lens movement was determined best by assessing horizontal lag, movement on blink in up-gaze and push-up recovery speed. Steeper lens base-curves did not have a significant effect on lens fit characteristics. Contact lens material did influence lens fit characteristics, particularly silicone-hydrogels which generally had lower centration and a faster push-up speed of recovery than HEMA lenses (p<0.05). Conclusion: Lag on vertical gaze, and movement on blink in primary gaze generally provide little extra information on overall lens movement compared to horizontal lag, movement on blink in up gaze and push-up recovery speed. They can therefore be excluded from a simplified recording scheme. A simplified and comprehensive soft contact lens fit recording system could consist of a cross-hairs indicating the centre of the cornea; a circle to indicate the lens centration; a mark on the relevant position of the circle to indicate any limbal incursion; a grade (‘B’) below for movement with blink in up-gaze, a grade (‘L’) to the side for horizontal lag and a grade above (‘P’) for the assessed push-up recovery speed.
Resumo:
Purpose: To examine the objective clinical performance of ‘comfort-enhanced’ daily disposable contact lenses over a 16-h day. Methods: Four contact lenses (Hilafilcon B, Etafilcon A Plus, Nelfilcon A and Nelfilcon A Plus) were evaluated in an investigator masked, open label trial at the end of a week’s bilateral wear. Pre-lens noninvasive tear break-up time (PL-NITBUT), tear prism height, bulbar hyperaemia and ocular surface temperature (OST) were measured with the lens in situ at 8, 12 and 16 h of wear. Results: There was no difference between how many hours the lenses types were worn each day (F = 0.90, p = 0.44). The PL-NITBUT decreased with the duration of daily lens wear (F = 32.0, p < 0.001) and was more stable with Nelfilcon A Plus (F = 6.00, p = 0.002) than with the other lenses evaluated. Bulbar blood vessels increased in coverage (F = 11.5, p < 0.001) but not overall redness (F = 0.0, p = 0.99) with the duration of daily lens wear, but there was no difference between the lenses (p > 0.05). The tear prism height decreased with the duration of daily wear (F = 27.0, p < 0.001) and differed between lenses (F = 2.9, p = 0.04). The OST decreased with the duration of lens wear (F = 119.7, p < 0.001) and was reduced by daily disposable lens wear (F = 7.88, p < 0.001), but did not differ between lenses (F = 0.88, p = 0.45). Conclusions: Objective measures of tear film indicated a difference between the lenses evaluated for PLNITBUT and tear prism height, but not for wearing time or bulbar conjunctival hyperaemia. Therefore clinical benefits of daily disposable ‘comfort enhancing’ contact lenses can be measured, but challenges remain in producing contact lenses that do not compromise anterior eye physiology over the whole day. 2010 British Contact Lens Association. Published by Elsevier Ltd. All rights reserved.
Resumo:
In this thesis various mathematical methods of studying the transient and dynamic stabiIity of practical power systems are presented. Certain long established methods are reviewed and refinements of some proposed. New methods are presented which remove some of the difficulties encountered in applying the powerful stability theories based on the concepts of Liapunov. Chapter 1 is concerned with numerical solution of the transient stability problem. Following a review and comparison of synchronous machine models the superiority of a particular model from the point of view of combined computing time and accuracy is demonstrated. A digital computer program incorporating all the synchronous machine models discussed, and an induction machine model, is described and results of a practical multi-machine transient stability study are presented. Chapter 2 reviews certain concepts and theorems due to Liapunov. In Chapter 3 transient stability regions of single, two and multi~machine systems are investigated through the use of energy type Liapunov functions. The treatment removes several mathematical difficulties encountered in earlier applications of the method. In Chapter 4 a simple criterion for the steady state stability of a multi-machine system is developed and compared with established criteria and a state space approach. In Chapters 5, 6 and 7 dynamic stability and small signal dynamic response are studied through a state space representation of the system. In Chapter 5 the state space equations are derived for single machine systems. An example is provided in which the dynamic stability limit curves are plotted for various synchronous machine representations. In Chapter 6 the state space approach is extended to multi~machine systems. To draw conclusions concerning dynamic stability or dynamic response the system eigenvalues must be properly interpreted, and a discussion concerning correct interpretation is included. Chapter 7 presents a discussion of the optimisation of power system small sjgnal performance through the use of Liapunov functions.
Resumo:
This thesis describes research on End-User Computing (EUC) in small business in an environment where no Information System (IS) support and expertise are available. The research aims to identify the factors that contribute to EUC Sophistication and understand the extent small firms are capable of developing their own applications. The intention is to assist small firms to adopt EUC, encourage better utilisation of their IT resources and gain the benefits associated with computerisation. The factors examined are derived inductively from previous studies where a model is developed to map these factors with the degree of sophistication associated with IT and EUC. This study attempts to combine the predictive power of quantitative research through surveys with the explanatory power of qualitative research through action-oriented case study. Following critical examination of the literature, a survey of IT Adoption and EUC was conducted. Instruments were then developed to measure EUC and IT Sophistication indexes based on sophistication constructs adapted from previous studies using data from the survey. This is followed by an in-depth action case study involving two small firms to investigate the EUC phenomenon in its real life context. The accumulated findings from these mixed research strategies are used to form the final model of EUC Sophistication in small business. Results of the study suggest both EUC Sophistication and the Presence of EUC in small business are affected by Management Support and Behaviour towards EUC. Additionally EUC Sophistication is also affected by the presence of an EUC Champion. Results are also consistent with respect to the independence between IT Sophistication and EUC Sophistication. The main research contributions include an accumulated knowledge of EUC in small business, the Model of EUC Sophistication, an instrument to measure EUC Sophistication Index for small firms, and a contribution to research methods in IS.
Resumo:
This thesis introduces a flexible visual data exploration framework which combines advanced projection algorithms from the machine learning domain with visual representation techniques developed in the information visualisation domain to help a user to explore and understand effectively large multi-dimensional datasets. The advantage of such a framework to other techniques currently available to the domain experts is that the user is directly involved in the data mining process and advanced machine learning algorithms are employed for better projection. A hierarchical visualisation model guided by a domain expert allows them to obtain an informed segmentation of the input space. Two other components of this thesis exploit properties of these principled probabilistic projection algorithms to develop a guided mixture of local experts algorithm which provides robust prediction and a model to estimate feature saliency simultaneously with the training of a projection algorithm.Local models are useful since a single global model cannot capture the full variability of a heterogeneous data space such as the chemical space. Probabilistic hierarchical visualisation techniques provide an effective soft segmentation of an input space by a visualisation hierarchy whose leaf nodes represent different regions of the input space. We use this soft segmentation to develop a guided mixture of local experts (GME) algorithm which is appropriate for the heterogeneous datasets found in chemoinformatics problems. Moreover, in this approach the domain experts are more involved in the model development process which is suitable for an intuition and domain knowledge driven task such as drug discovery. We also derive a generative topographic mapping (GTM) based data visualisation approach which estimates feature saliency simultaneously with the training of a visualisation model.