977 resultados para QA
Resumo:
Multilevel algorithms are a successful class of optimization techniques which addresses the mesh partitioning problem. They usually combine a graph contraction algorithm together with a local optimization method which refines the partition at each graph level. In this paper we present an enhancement of the technique which uses imbalance to achieve higher quality partitions. We also present a formulation of the Kernighan-Lin partition optimization algorithm which incorporates load-balancing. The resulting algorithm is tested against a different but related state-of-the-art partitioner and shown to provide improved results.
Resumo:
A simulation of the motion of molten aluminium inside an electrolytic cell is presented. Since the driving term of the aluminium motion is the Lorentz (j × B) body force acting within the fluid,this problem involves the solution of the magneto-hydro-dynamic equations. Different solver modules for the magnetic field computation and for the fluid motion simulation are coupled together. The interactions of all these are presented and discussed.
Resumo:
This paper describes the architecture of the case based reasoning (CBR) component of Smartfire, a fire field modelling tool for use by members of the Fire Safety Engineering community who are not expert in modelling techniques. The CBR system captures the qualitative reasoning of an experienced modeller in the assessment of room geometries so as to set up the important initial parameters of the problem. The system relies on two important reasoning principles obtained from the expert: 1) there is a natural hierarchical retrieval mechanism which may be employed; and 2) much of the reasoning on a qualitative level is linear in nature, although the computational solution of the problem is non-linear. The paper describes the qualitative representation of geometric room information on which the system is based, and the principles on which the CBR system operates.
Resumo:
Given M(r; f) =maxjzj=r (jf(z)j) , curves belonging to the set of points M = fz : jf(z)j = M(jzj; f)g were de�ned by Hardy to be maximum curves. Clunie asked the question as to whether the set M could also contain isolated points. This paper shows that maximum curves consist of analytic arcs and determines a necessary condition for such curves to intersect. Given two entire functions f1(z) and f2(z), if the maximum curve of f1(z) is the real axis, conditions are found so that the real axis is also a maximum curve for the product function f1(z)f2(z). By means of these results an entire function of in�nite order is constructed for which the set M has an in�nite number of isolated points. A polynomial is also constructed with an isolated point.
Resumo:
Sometimes, technological solutions to practical problems are devised that conspicuously take into account the constraints to which a given culture is subjecting the particular task or the manner in which it is carried out. The culture may be a professional culture (e.g., the practice of law), or an ethnic-cum-professional culture (e.g., dance in given ethnic cultures from South-East Asia), or, again, a denominational culture prescribing an orthopraxy impinging on everyday life through, for example, prescribed abstinence from given categories of workday activities, or dietary laws. Massimo Negrotti's Theory of the artificial is a convenient framework for discussing some of these techniques. We discuss a few examples, but focus on the contrast of two that are taken from the same cultural background, namely, technological applications in compliance with Jewish Law orthopraxy. •Soya-, mycoprotein- or otherwise derived meat surrogates are an example ofnaturoid; they emulate the flavours and olfactory properties, as well as the texture and the outer and inner appearance, of the meat product (its kind, cut, form) they set out to emulate (including amenability to cooking in the usual manner for the model), while satisfying cultural dietary prohibitions. •In contrast, the Sabbath Notebook, a writing surrogate we describe in this paper, is atechnoid: it emulates a technique (writing to store alphanumeric information), while satisfying the prohibition of writing at particular times of the liturgical calendar (the Sabbath and the major holidays).
Resumo:
In judicial decision making, the doctrine of chances takes explicitly into account the odds. There is more to forensic statistics, as well as various probabilistic approaches which taken together form the object of an enduring controversy in the scholarship of legal evidence. In this paper, we reconsider the circumstances of the Jama murder and inquiry (dealt with in Part I of this paper: "The Jama Model. On Legal Narratives and Interpretation Patterns"), to illustrate yet another kind of probability or improbability. What is improbable about the Jama story, is actually a given, which contributes in terms of dramatic underlining. In literary theory, concepts of narratives being probable or improbable date back from the eighteenth century, when both prescientific and scientific probability was infiltrating several domains, including law. An understanding of such a backdrop throughout the history of ideas is, I claim, necessary for AI researchers who may be tempted to apply statistical methods to legal evidence. The debate for or against probability (and especially bayesian probability) in accounts of evidence has been flouishing among legal scholars. Nowadays both the the Bayesians (e.g. Peter Tillers) and Bayesioskeptics (e.g. Ron Allen) among those legal scholars whoare involved in the controversy are willing to give AI researchers a chance to prove itself and strive towards models of plausibility that would go beyond probability as narrowly meant. This debate within law, in turn, has illustrious precedents: take Voltaire, he was critical of the application or probability even to litigation in civil cases; take Boole, he was a starry-eyed believer in probability applications to judicial decision making (Rosoni 1995). Not unlike Boole, the founding father of computing, nowadays computer scientists approaching the field may happen to do so without full awareness of the pitfalls. Hence, the usefulness of the conceptual landscape I sketch here.
Resumo:
Review of: Noel Starkey (ed), Connectionist Natural Language Processing: Readings from 'Connection Science'
Resumo:
The key problems in discussing duality and monotonicity for continuous-time Markov chains are to find conditions for existence and uniqueness and then to construct corresponding processes in terms of infinitesimal characteristics, i.e., q-matrices. Such problems are solved in this paper under the assumption that the given q-matrix is conservative. Some general properties of stochastically monotone Q-process ( Q is not necessarily conservative) are also discussed.
Resumo:
This paper describes the approach to the modelling of experiential knowledge in an industrial application of Case-Based Reasoning (CBR). The CBR involves retrieval techniques in conjunction with a relational database. The database is especially designed as a repository of experiential knowledge, and includes qualitative search indices. The system is intended to help design engineers and material engineers in the submarine cable industry. It consists of three parts: a materials database; a database of experiential knowledge; and a CBR system used to retrieve similar past designs based upon component and material qualitative descriptions. The system is currently undergoing user testing at the Alcatel Submarine Networks site in Greenwich.
Resumo:
The FIRE Detection and Suppression Simulation (FIREDASS) project was concerned with the development of water misting systems as a possible replacement for halon based fire suppression systems currently used in aircraft cargo holds and ship engine rooms. As part of this program of work, a computational model was developed to assist engineers optimize the design of water mist suppression systems. The model is based on Computational Fluid Dynamics (CFD) and comprised of the following components: fire model; mist model; two-phase radiation model; suppression model; detector/activation model. In this paper the FIREDASS software package is described and the theory behind the fire and radiation sub-models is detailed. The fire model uses prescribed release rates for heat and gaseous combustion products to represent the fire load. Typical release rates have been determined through experimentation. The radiation model is a six-flux model coupled to the gas (and mist) phase. As part of the FIREDASS project, a detailed series of fire experiments were conducted in order to validate the fire model. Model predictions are compared with data from these experiments and good agreement is found.
Resumo:
This paper describes the architecture of the knowledge based system (KBS) component of Smartfire, a fire field modelling tool for use by members of the fire safety engineering community who are not expert in modelling techniques. The KBS captures the qualitative reasoning of an experienced modeller in the assessment of room geometries, so as to set up the important initial parameters of the problem. Fire modelling expertise is an example of geometric and spatial reasoning, which raises representational problems. The approach taken in this project is a qualitative representation of geometric room information based on Forbus’ concept of a metric diagram. This takes the form of a coarse grid, partitioning the domain in each of the three spatial dimensions. Inference over the representation is performed using a case-based reasoning (CBR) component. The CBR component stores example partitions with key set-up parameters; this paper concentrates on the key parameter of grid cell distribution.
Resumo:
We study the special case of the m machine flow shop problem in which the processing time of each operation of job j is equal to pj; this variant of the flow shop problem is known as the proportionate flow shop problem. We show that for any number of machines and for any regular performance criterion we can restrict our search for an optimal schedule to permutation schedules. Moreover, we show that the problem of minimizing total weighted completion time is solvable in O(n2) time. © 1998 John Wiley & Sons, Ltd.
Resumo:
This paper considers a special class of flow-shop problems, known as the proportionate flow shop. In such a shop, each job flows through the machines in the same order and has equal processing times on the machines. The processing times of different jobs may be different. It is assumed that all operations of a job may be compressed by the same amount which will incur an additional cost. The objective is to minimize the makespan of the schedule together with a compression cost function which is non-decreasing with respect to the amount of compression. For a bicriterion problem of minimizing the makespan and a linear cost function, an O(n log n) algorithm is developed to construct the Pareto optimal set. For a single criterion problem, an O(n2) algorithm is developed to minimize the sum of the makespan and compression cost. Copyright © 1999 John Wiley & Sons, Ltd.