938 resultados para PDE-based parallel preconditioner
Resumo:
Salient pole brushless alternators coupled to IC engines are extensively used as stand-by power supply units for meeting in- dustrial power demands. Design of such generators demands high power to weight ratio, high e ciency and low cost per KVA out- put. Moreover, the performance characteristics of such machines like voltage regulation and short circuit ratio (SCR) are critical when these machines are put into parallel operation and alterna- tors for critical applications like defence and aerospace demand very low harmonic content in the output voltage. While designing such alternators, accurate prediction of machine characteristics, including total harmonic distortion (THD) is essential to mini- mize development cost and time. Total harmonic distortion in the output voltage of alternators should be as low as possible especially when powering very sophis- ticated and critical applications. The output voltage waveform of a practical AC generator is replica of the space distribution of the ux density in the air gap and several factors such as shape of the rotor pole face, core saturation, slotting and style of coil disposition make the realization of a sinusoidal air gap ux wave impossible. These ux harmonics introduce undesirable e ects on the alternator performance like high neutral current due to triplen harmonics, voltage distortion, noise, vibration, excessive heating and also extra losses resulting in poor e ciency, which in turn necessitate de-rating of the machine especially when connected to non-linear loads. As an important control unit of brushless alternator, the excitation system and its dynamic performance has a direct impact on alternator's stability and reliability. The thesis explores design and implementation of an excitation i system utilizing third harmonic ux in the air gap of brushless al- ternators, using an additional auxiliary winding, wound for 1=3rd pole pitch, embedded into the stator slots and electrically iso- lated from the main winding. In the third harmonic excitation system, the combined e ect of two auxiliary windings, one with 2=3rd pitch and another third harmonic winding with 1=3rd pitch, are used to ensure good voltage regulation without an electronic automatic voltage regulator (AVR) and also reduces the total harmonic content in the output voltage, cost e ectively. The design of the third harmonic winding by analytic methods demands accurate calculation of third harmonic ux density in the air gap of the machine. However, precise estimation of the amplitude of third harmonic ux in the air gap of a machine by conventional design procedures is di cult due to complex geome- try of the machine and non-linear characteristics of the magnetic materials. As such, prediction of the eld parameters by conven- tional design methods is unreliable and hence virtual prototyping of the machine is done to enable accurate design of the third har- monic excitation system. In the design and development cycle of electrical machines, it is recognized that the use of analytical and experimental methods followed by expensive and in exible prototyping is time consum- ing and no longer cost e ective. Due to advancements in com- putational capabilities over recent years, nite element method (FEM) based virtual prototyping has become an attractive al- ternative to well established semi-analytical and empirical design methods as well as to the still popular trial and error approach followed by the costly and time consuming prototyping. Hence, by virtually prototyping the alternator using FEM, the important performance characteristics of the machine are predicted. Design of third harmonic excitation system is done with the help of results obtained from virtual prototype of the machine. Third harmonic excitation (THE) system is implemented in a 45 KVA ii experimental machine and experiments are conducted to validate the simulation results. Simulation and experimental results show that by utilizing third harmonic ux in the air gap of the ma- chine for excitation purposes during loaded conditions, triplen harmonic content in the output phase voltage is signi cantly re- duced. The prototype machine with third harmonic excitation system designed and developed based on FEM analysis proved to be economical due to its simplicity and has the added advan- tage of reduced harmonics in the output phase voltage.
Resumo:
Mikrooptische Filter sind heutzutage in vielen Bereichen in der Telekommunikation unersetzlich. Wichtige Einsatzgebiete sind aber auch spektroskopische Systeme in der Medizin-, Prozess- und Umwelttechnik. Diese Arbeit befasst sich mit der Technologieentwicklung und Herstellung von luftspaltbasierenden, vertikal auf einem Substrat angeordneten, oberflächenmikromechanisch hergestellten Fabry-Perot-Filtern. Es werden zwei verschiedene Filtervarianten, basierend auf zwei verschiedenen Materialsystemen, ausführlich untersucht. Zum einen handelt es sich dabei um die Weiterentwicklung von kontinuierlich mikromechanisch durchstimmbaren InP / Luftspaltfiltern; zum anderen werden neuartige, kostengünstige Siliziumnitrid / Luftspaltfilter wissenschaftlich behandelt. Der Inhalt der Arbeit ist so gegliedert, dass nach einer Einleitung mit Vergleichen zu Arbeiten und Ergebnissen anderer Forschergruppen weltweit, zunächst einige theoretische Grundlagen zur Berechnung der spektralen Reflektivität und Transmission von beliebigen optischen Schichtanordnungen aufgezeigt werden. Auß erdem wird ein kurzer theoretischer Ü berblick zu wichtigen Eigenschaften von Fabry-Perot-Filtern sowie der Möglichkeit einer mikromechanischen Durchstimmbarkeit gegeben. Daran anschließ end folgt ein Kapitel, welches sich den grundlegenden technologischen Aspekten der Herstellung von luftspaltbasierenden Filtern widmet. Es wird ein Zusammenhang zu wichtigen Referenzarbeiten hergestellt, auf denen diverse Weiterentwicklungen dieser Arbeit basieren. Die beiden folgenden Kapitel erläutern dann ausführlich das Design, die Herstellung und die Charakterisierung der beiden oben erwähnten Filtervarianten. Abgesehen von der vorangehenden Epitaxie von InP / GaInAs Schichten, ist die Herstellung der InP / Luftspaltfilter komplett im Institut durchgeführt worden. Die Herstellungsschritte sind ausführlich in der Arbeit erläutert, wobei ein Schwerpunktthema das trockenchemische Ä tzen von InP sowie GaInAs, welches als Opferschichtmaterial für die Herstellung der Luftspalte genutzt wurde, behandelt. Im Verlauf der wissenschaftlichen Arbeit konnten sehr wichtige technische Verbesserungen entwickelt und eingesetzt werden, welche zu einer effizienteren technologischen Herstellung der Filter führten und in der vorliegenden Niederschrift ausführlich dokumentiert sind. Die hergestellten, für einen Einsatz in der optischen Telekommunikation entworfenen, elektrostatisch aktuierbaren Filter sind aus zwei luftspaltbasierenden Braggspiegeln aufgebaut, welche wiederum jeweils 3 InP-Schichten von (je nach Design) 357nm bzw. 367nm Dicke aufweisen. Die Filter bestehen aus im definierten Abstand parallel übereinander angeordneten Membranen, die über Verbindungsbrücken unterschiedlicher Anzahl und Länge an Haltepfosten befestigt sind. Da die mit 357nm bzw. 367nm vergleichsweise sehr dünnen Schichten freitragende Konstrukte mit bis zu 140 nm Länge bilden, aber trotzdem Positionsgenauigkeiten im nm-Bereich einhalten müssen, handelt es sich hierbei um sehr anspruchsvolle mikromechanische Bauelemente. Um den Einfluss der zahlreichen geometrischen Strukturparameter studieren zu können, wurden verschiedene laterale Filterdesigns implementiert. Mit den realisierten Filter konnte ein enorm weiter spektraler Abstimmbereich erzielt werden. Je nach lateralem Design wurden internationale Bestwerte für durchstimmbare Fabry-Perot-Filter von mehr als 140nm erreicht. Die Abstimmung konnte dabei kontinuierlich mit einer angelegten Spannung von nur wenigen Volt durchgeführt werden. Im Vergleich zu früher berichteten Ergebnissen konnten damit sowohl die Wellenlängenabstimmung als auch die dafür benötigte Abstimmungsspannung signifikant verbessert werden. Durch den hohen Brechungsindexkontrast und die geringe Schichtdicke zeigen die Filter ein vorteilhaftes, extrem weites Stopband in der Größ enordnung um 550nm. Die gewählten, sehr kurzen Kavitätslängen ermöglichen einen freien Spektralbereich des Filters welcher ebenfalls in diesen Größ enordnungen liegt, so dass ein weiter spektraler Einsatzbereich ermöglicht wird. Während der Arbeit zeigte sich, dass Verspannungen in den freitragenden InPSchichten die Funktionsweise der mikrooptischen Filter stark beeinflussen bzw. behindern. Insbesondere eine Unterätzung der Haltepfosten und die daraus resultierende Verbiegung der Ecken an denen sich die Verbindungsbrücken befinden, führte zu enormen vertikalen Membranverschiebungen, welche die Filtereigenschaften verändern. Um optimale Ergebnisse zu erreichen, muss eine weitere Verbesserung der Epitaxie erfolgen. Jedoch konnten durch den zusätzlichen Einsatz einer speziellen Schutzmaske die Unterätzung der Haltepfosten und damit starke vertikale Verformungen reduziert werden. Die aus der Verspannung resultierenden Verformungen und die Reaktion einzelner freistehender InP Schichten auf eine angelegte Gleich- oder Wechselspannung wurde detailliert untersucht. Mittels Weisslichtinterferometrie wurden lateral identische Strukturen verglichen, die aus unterschiedlich dicken InP-Schichten (357nm bzw. 1065nm) bestehen. Einen weiteren Hauptteil der Arbeit stellen Siliziumnitrid / Luftspaltfilter dar, welche auf einem neuen, im Rahmen dieser Dissertation entwickelten, technologischen Ansatz basieren. Die Filter bestehen aus zwei Braggspiegeln, die jeweils aus fünf 590nm dicken, freistehenden Siliziumnitridschichten aufgebaut sind und einem Abstand von 390nm untereinander aufweisen. Die Filter wurden auf Glassubstraten hergestellt. Der Herstellungsprozess ist jedoch auch mit vielen anderen Materialien oder Prozessen kompatibel, so dass z.B. eine Integration mit anderen Bauelemente relativ leicht möglich ist. Die Prozesse dieser ebenfalls oberflächenmikromechanisch hergestellten Filter wurden konsequent auf niedrige Herstellungskosten optimiert. Als Opferschichtmaterial wurde hier amorph abgeschiedenes Silizium verwendet. Der Herstellungsprozess beinhaltet die Abscheidung verspannungsoptimierter Schichten (Silizium und Siliziumnitrid) mittels PECVD, die laterale Strukturierung per reaktiven Ionenätzen mit den Gasen SF6 / CHF3 / Ar sowie Fotolack als Maske, die nasschemische Unterätzung der Opferschichten mittels KOH und das Kritisch-Punkt-Trocken der Proben. Die Ergebnisse der optischen Charakterisierung der Filter zeigen eine hohe Ü bereinstimmung zwischen den experimentell ermittelten Daten und den korrespondierenden theoretischen Modellrechnungen. Weisslichtinterferometermessungen der freigeätzten Strukturen zeigen ebene Filterschichten und bestätigen die hohe vertikale Positioniergenauigkeit, die mit diesem technologischen Ansatz erreicht werden kann.
Resumo:
This work demonstrates how partial evaluation can be put to practical use in the domain of high-performance numerical computation. I have developed a technique for performing partial evaluation by using placeholders to propagate intermediate results. For an important class of numerical programs, a compiler based on this technique improves performance by an order of magnitude over conventional compilation techniques. I show that by eliminating inherently sequential data-structure references, partial evaluation exposes the low-level parallelism inherent in a computation. I have implemented several parallel scheduling and analysis programs that study the tradeoffs involved in the design of an architecture that can effectively utilize this parallelism. I present these results using the 9- body gravitational attraction problem as an example.
Resumo:
We describe the key role played by partial evaluation in the Supercomputing Toolkit, a parallel computing system for scientific applications that effectively exploits the vast amount of parallelism exposed by partial evaluation. The Supercomputing Toolkit parallel processor and its associated partial evaluation-based compiler have been used extensively by scientists at MIT, and have made possible recent results in astrophysics showing that the motion of the planets in our solar system is chaotically unstable.
Resumo:
This paper presents the research and development of a 3-legged micro Parallel Kinematic Manipulator (PKM) for positioning in micro-machining and assembly operations. The structural characteristics associated with parallel manipulators are evaluated and the PKMs with translational and rotational movements are identified. Based on these identifications, a hybrid 3-UPU (Universal Joint-Prismatic Joint-Universal Joint) parallel manipulator is designed and fabricated. The principles of the operation and modeling of this micro PKM is largely similar to a normal size Stewart Platform (SP). A modular design methodology is introduced for the construction of this micro PKM. Calibration results of this hybrid 3-UPU PKM are discussed in this paper.
Resumo:
In this paper we face the problem of positioning a camera attached to the end-effector of a robotic manipulator so that it gets parallel to a planar object. Such problem has been treated for a long time in visual servoing. Our approach is based on linking to the camera several laser pointers so that its configuration is aimed to produce a suitable set of visual features. The aim of using structured light is not only for easing the image processing and to allow low-textured objects to be treated, but also for producing a control scheme with nice properties like decoupling, stability, well conditioning and good camera trajectory
Resumo:
Partnership is increasingly espoused as the best relationship between members of the sustainable development aid chain, and implies a respect for the position of all and a desire to avoid a situation where one group dominates another. It also implies a form of relationship that is not just 'better' for the sake of it but that is more able to help achieve sustainable development. However, given the inevitable inequalities in power between donors that have the resources and field partners that do not it can be hard to put this ideal into practice. This paper explores the function of partnership within a group of closely related institutions that comprise the Catholic Church development chain. The research focussed on three Catholic Church based donors (one from the USA and two from Europe) and their partners in Abuja Ecclesiastical Province, Nigeria. Relationships between and within various strata of the Church in Nigeria were also examined. Relationships were 'patchy' at all levels. One of the donors had a significant operational presence in Nigeria and this was regarded by some respondents as a parallel structure that seriously undermined local bodies. However, while problems existed, there was a sense of inter-dependence arising from a shared sense of values and Catholic Social Teaching, which allowed partners to work through their stresses and conflicts. It is the innate sustainability of the aid chain itself founded upon a set of shared values that provided the space and time for problems to be addressed. Copyright (C) 2008 John Wiley & Sons, Ltd and ERP Environment.
Resumo:
Clustering is defined as the grouping of similar items in a set, and is an important process within the field of data mining. As the amount of data for various applications continues to increase, in terms of its size and dimensionality, it is necessary to have efficient clustering methods. A popular clustering algorithm is K-Means, which adopts a greedy approach to produce a set of K-clusters with associated centres of mass, and uses a squared error distortion measure to determine convergence. Methods for improving the efficiency of K-Means have been largely explored in two main directions. The amount of computation can be significantly reduced by adopting a more efficient data structure, notably a multi-dimensional binary search tree (KD-Tree) to store either centroids or data points. A second direction is parallel processing, where data and computation loads are distributed over many processing nodes. However, little work has been done to provide a parallel formulation of the efficient sequential techniques based on KD-Trees. Such approaches are expected to have an irregular distribution of computation load and can suffer from load imbalance. This issue has so far limited the adoption of these efficient K-Means techniques in parallel computational environments. In this work, we provide a parallel formulation for the KD-Tree based K-Means algorithm and address its load balancing issues.
Resumo:
One among the most influential and popular data mining methods is the k-Means algorithm for cluster analysis. Techniques for improving the efficiency of k-Means have been largely explored in two main directions. The amount of computation can be significantly reduced by adopting geometrical constraints and an efficient data structure, notably a multidimensional binary search tree (KD-Tree). These techniques allow to reduce the number of distance computations the algorithm performs at each iteration. A second direction is parallel processing, where data and computation loads are distributed over many processing nodes. However, little work has been done to provide a parallel formulation of the efficient sequential techniques based on KD-Trees. Such approaches are expected to have an irregular distribution of computation load and can suffer from load imbalance. This issue has so far limited the adoption of these efficient k-Means variants in parallel computing environments. In this work, we provide a parallel formulation of the KD-Tree based k-Means algorithm for distributed memory systems and address its load balancing issue. Three solutions have been developed and tested. Two approaches are based on a static partitioning of the data set and a third solution incorporates a dynamic load balancing policy.
Resumo:
Statistical approaches have been applied to examine amino acid pairing preferences within parallel beta-sheets. The main chain hydrogen bonding pattern in parallel beta-sheets means that, for each residue pair, only one of the residues is involved in main chain hydrogen bonding with the strand containing the partner residue. We call this the hydrogen bonded (HB) residue and the partner residue the non-hydrogen bonded (nHB) residue, and differentiate between the favorability of a pair and that of its reverse pair, e.g. Asn(HB)-Thr(nHB)versus Thr(HB)-Asn(nHB). Significantly (p < or = 0.000001) favoured pairings were rationalised using stereochemical arguments. For instance, Asn(HB)-Thr(nHB) and Arg(HB)-Thr(nHB) were favoured pairs, where the residues adopted favoured chi1 rotamer positions that allowed side-chain interactions to occur. In contrast, Thr(HB)-Asn(nHB) and Thr(HB)-Arg(nHB) were not significantly favoured, and could only form side-chain interactions if the residues involved adopted less favourable chi1 conformations. The favourability of hydrophobic pairs e.g. Ile(HB)-Ile(nHB), Val(HB)-Val(nHB) and Leu(HB)-Ile(nHB) was explained by the residues adopting their most preferred chi1 and chi2 conformations, which enabled them to form nested arrangements. Cysteine-cysteine pairs are significantly favoured, although these do not form intrasheet disulphide bridges. Interactions between positively and negatively charged residues were asymmetrically preferred: those with the negatively charged residue at the HB position were more favoured. This trend was accounted for by the presence of general electrostatic interactions, which, based on analysis of distances between charged atoms, were likely to be stronger when the negatively charged residue is the HB partner. The Arg(HB)-Asp(nHB) interaction was an exception to this trend and its favorability was rationalised by the formation of specific side-chain interactions. This research provides rules that could be applied to protein structure prediction, comparative modelling and protein engineering and design. The methods used to analyse the pairing preferences are automated and detailed results are available (http://www.rubic.rdg.ac.uk/betapairprefsparallel/).
Resumo:
Statistical approaches have been applied to examine amino acid pairing preferences within parallel beta-sheets. The main chain hydrogen bonding pattern in parallel beta-sheets means that, for each residue pair, only one of the residues is involved in main chain hydrogen bonding with the strand containing the partner residue. We call this the hydrogen bonded (HB) residue and the partner residue the non-hydrogen bonded (nHB) residue, and differentiate between the favourability of a pair and that of its reverse pair, e.g. Asn(HB)-Thr(nHB) versus Thr(HB)-Asn(nHB). Significantly (p <= 0.000001) favoured pairings were rationalised using stereochemical arguments. For instance, Asn(HB)-Thr(nHB) and Arg(HB)-Thr(nHB) were favoured pairs, where the residues adopted favoured chi(1) rotamer positions that allowed side-chain interactions to occur. In contrast, Thr(HB)-Asn(nHB) and Thr(HB)-Arg(nHB) were not significantly favoured, and could only form side-chain interactions if the residues involved adopted less favourable chi(1) conformations. The favourability of hydrophobic pairs e.g. Ile(HB)-Ile(nHB), Val(HB)-Val(nHB) and Leu(HB)-Ile(nHB) was explained by the residues adopting their most preferred chi(1) and chi(2) conformations, which enabled them to form nested arrangements. Cysteine-cysteine pairs are significantly favoured, although these do not form intrasheet disulphide bridges. Interactions between positively and negatively charged residues were asymmetrically preferred: those with the negatively charged residue at the HB position were more favoured. This trend was accounted for by the presence of general electrostatic interactions, which, based on analysis of distances between charged atoms, were likely to be stronger when the negatively charged residue is the HB partner. The Arg(HB)-Asp(nHB) interaction was an exception to this trend and its favourability was rationalised by the formation of specific side-chain interactions. This research provides rules that could be applied to protein structure prediction, comparative modelling and protein engineering and design. The methods used to analyse the pairing preferences are automated and detailed results are available (http:// www.rubic.rdg.ac.uk/betapairprefsparallel/). (c) 2005 Elsevier Ltd. All rights reserved.
Resumo:
A number of strategies are emerging for the high throughput (HTP) expression of recombinant proteins to enable structural and functional study. Here we describe a workable HTP strategy based on parallel protein expression in E. coli and insect cells. Using this system we provide comparative expression data for five proteins derived from the Autographa californica polyhedrosis virus genome that vary in amino acid composition and in molecular weight. Although the proteins are part of a set of factors known to be required for viral late gene expression, the precise function of three of the five, late expression factors (lefs) 6, 7 and 10, is unknown. Rapid expression and characterisation has allowed the determination of their ability to bind DNA and shown a cellular location consistent with their properties. Our data point to the utility of a parallel expression strategy to rapidly obtain workable protein expression levels from many open reading frames (ORFs).
Resumo:
WO3-based materials as sensors for the monitor of environmental gases such as NO2 (NO + NO2) have been rapidly developed for various potential applications (stationary and mobile uses). It has been reported that these materials are highly sensitive to NOx with the sensitivity further enhanced by adding precious group metals (PGM such as Pt, Pd, Au, etc.). However, there has been limited work in revealing the sensing mechanism for these gases over the WO3-based sensors. In particular, the role of promoter is not yet clear though speculations on their catalytic, electronic and structural effects have been made in the past. In parallel to these PGM promoters here we report,for the first time, that Ag promotion can also enhance WO3 sensitivity significantly. In addition, this promotion decreases the optimum sensor temperature of 300 degreesC for Most WO3-based sensors, to below 200 degreesC. Characterizations (XRD, TEM, and impedance measurement) reveal that there is no significant bulk structure change nor particle size alteration in the WO3 phases during the NO exposure. However, it is found that the Ag doping creates a high concentration of oxygen vacancies in form of coordinated crystallographic shear (CS) planes onto the underneath WO3. It is thus proposed that the Ag particle facilitates the oxidative conversion of NO to NO2 followed by a subsequent NO2 adsorption on the defective WO, sites created at the Ag-WO3 interface; hence, accounting for the high molecular sensitivity. (C) 2002 Elsevier Science B.V. All rights reserved.
Resumo:
Uncertainty contributes a major part in the accuracy of a decision-making process while its inconsistency is always difficult to be solved by existing decision-making tools. Entropy has been proved to be useful to evaluate the inconsistency of uncertainty among different respondents. The study demonstrates an entropy-based financial decision support system called e-FDSS. This integrated system provides decision support to evaluate attributes (funding options and multiple risks) available in projects. Fuzzy logic theory is included in the system to deal with the qualitative aspect of these options and risks. An adaptive genetic algorithm (AGA) is also employed to solve the decision algorithm in the system in order to provide optimal and consistent rates to these attributes. Seven simplified and parallel projects from a Hong Kong construction small and medium enterprise (SME) were assessed to evaluate the system. The result shows that the system calculates risk adjusted discount rates (RADR) of projects in an objective way. These rates discount project cash flow impartially. Inconsistency of uncertainty is also successfully evaluated by the use of the entropy method. Finally, the system identifies the favourable funding options that are managed by a scheme called SME Loan Guarantee Scheme (SGS). Based on these results, resource allocation could then be optimized and the best time to start a new project could also be identified throughout the overall project life cycle.
Resumo:
Background: Medication errors are an important cause of morbidity and mortality in primary care. The aims of this study are to determine the effectiveness, cost effectiveness and acceptability of a pharmacist-led information-technology-based complex intervention compared with simple feedback in reducing proportions of patients at risk from potentially hazardous prescribing and medicines management in general (family) practice. Methods: Research subject group: "At-risk" patients registered with computerised general practices in two geographical regions in England. Design: Parallel group pragmatic cluster randomised trial. Interventions: Practices will be randomised to either: (i) Computer-generated feedback; or (ii) Pharmacist-led intervention comprising of computer-generated feedback, educational outreach and dedicated support. Primary outcome measures: The proportion of patients in each practice at six and 12 months post intervention: - with a computer-recorded history of peptic ulcer being prescribed non-selective non-steroidal anti-inflammatory drugs - with a computer-recorded diagnosis of asthma being prescribed beta-blockers - aged 75 years and older receiving long-term prescriptions for angiotensin converting enzyme inhibitors or loop diuretics without a recorded assessment of renal function and electrolytes in the preceding 15 months. Secondary outcome measures; These relate to a number of other examples of potentially hazardous prescribing and medicines management. Economic analysis: An economic evaluation will be done of the cost per error avoided, from the perspective of the UK National Health Service (NHS), comparing the pharmacist-led intervention with simple feedback. Qualitative analysis: A qualitative study will be conducted to explore the views and experiences of health care professionals and NHS managers concerning the interventions, and investigate possible reasons why the interventions prove effective, or conversely prove ineffective. Sample size: 34 practices in each of the two treatment arms would provide at least 80% power (two-tailed alpha of 0.05) to demonstrate a 50% reduction in error rates for each of the three primary outcome measures in the pharmacist-led intervention arm compared with a 11% reduction in the simple feedback arm. Discussion: At the time of submission of this article, 72 general practices have been recruited (36 in each arm of the trial) and the interventions have been delivered. Analysis has not yet been undertaken.