935 resultados para Polygonal faults


Relevância:

10.00% 10.00%

Publicador:

Resumo:

Le présent travail de recherche se propose d’analyser les dispositifs de gouvernance nodale de la sécurité locale en France, alors que le paradigme a vu le jour et s’est développé dans les pays anglo-saxons fortement décentralisés. C’est qu’en France les dispositifs de gouvernance nodale s’apparentent bien plus à un dialogue entre central et local qu’entre secteur public et privé. La recherche identifie ainsi les caractéristiques de la gouvernance nodale au cœur des dispositifs partenariaux de la sécurité locale, supportés par le Contrat Local de Sécurité (CLS), le Conseil Local de Sécurité et de Prévention de la Délinquance (CLSPD) ou encore le Groupe Local de Traitement de la Délinquance (GLTD). La recherche identifie ainsi les stratégies de décentrage de l’État et de transfert de la production de sécurité vers une diversité d’acteurs locaux, dont les maires et les services municipaux. Une diversité de politiques publiques locales de sécurité de pertinences différentes voit alors le jour. Le premier enseignement de cette recherche est l’importance du rôle joué par le node super-structurel, que nous appelons super-node et qui regroupe le maire ou l’élu local à la sécurité, le responsable de la police d’État, celui de la police municipale et le représentant de l’État. Il apparaît que dans le dispositif de gouvernance nodale, ce groupe informel génère la dynamique collective qui permet de regrouper, tant les producteurs que les consommateurs de sécurité locale gravitant au sein du réseau local de sécurité. La quarantaine d’entrevues qualitatives permet également d’identifier que la Justice, productrice de sécurité comme peut l’être aussi la sécurité privée ou la médiation sociale, apparaît plus distante que ce que pouvait laisser penser l’étude des textes réglementaires organisant le partenariat. Les bailleurs sociaux, les transporteurs et l’Éducation nationale apparaissent clairement comme des acteurs importants, mais périphériques de sécurité, en intégrant cette « famille élargie » de la sécurité locale. Le deuxième enseignement est relatif au fonctionnement même du dispositif nodal ainsi que du super-node, la recherche permettant d’identifier les ressources mutualisées par l’ensemble des nodes. Cela permet également d’identifier les mécanismes de répartition des tâches entre les différents acteurs et plus particulièrement entre les deux organisations policières d’État et municipale, travaillant autant en compétition, qu’en complémentarité. Cette recherche explore également le rôle joué par l’information dans le fonctionnement du super-node ainsi que l’importance de la confiance dans les relations interpersonnelles des représentants des nodes au sein du super-node. Enfin, l’étude permet également de mettre en perspective les limites du dispositif actuel de gouvernance nodale : le défaut avéré d’outils performants permettant d’informer convenablement le super-node quant aux phénomènes de violence ainsi que d’évaluer l’efficience du dispositif. Cela permet également de questionner l’autonomie des dispositifs de gouvernance nodale, la confiance pouvant ouvrir à la déviance et la collégialité au défaut de la traçabilité de la responsabilité. La fracture avec la société civile apparaît clairement et ne facilite pas le contrôle sur un mode de production de sécurité qui se développe en parallèle des dispositifs traditionnels de démocratie locale.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The purpose of the present study is to understand the surface deformation associated with the Killari and Wadakkancheri earthquake and to examine if there are any evidence of occurrence of paleo-earthquakes in this region or its vicinity. The present study is an attempt to characterize active tectonic structures from two areas within penisular India: the sites of 1993 Killari (Latur) (Mb 6.3) and 1994 Wadakkancheri (M 4.3) earthquakes in the Precambrian shield. The main objectives of the study are to isolate structures related to active tectonism, constraint the style of near – surface deformation and identify previous events by interpreting the deformational features. The study indicates the existence of a NW-SE trending pre-existing fault, passing through the epicentral area of the 1993 Killari earthquake. It presents the salient features obtained during the field investigations in and around the rupture zone. Details of mapping of the scrap, trenching, and shallow drilling are discussed here. It presents the geologic and tectonic settings of the Wadakkancheri area and the local seismicity; interpretation of remote sensing data and a detailed geomorphic analysis. Quantitative geomorphic analysis around the epicenter of the Wadakkancheri earthquake indicates suitable neotectonic rejuvenation. Evaluation of remote sensing data shows distinct linear features including the presence of potentially active WNW-ESE trending fault within the Precambrian shear zone. The study concludes that the earthquakes in the shield area are mostly associated with discrete faults that are developed in association with the preexisting shear zones or structurally weak zones

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Sharing of information with those in need of it has always been an idealistic goal of networked environments. With the proliferation of computer networks, information is so widely distributed among systems, that it is imperative to have well-organized schemes for retrieval and also discovery. This thesis attempts to investigate the problems associated with such schemes and suggests a software architecture, which is aimed towards achieving a meaningful discovery. Usage of information elements as a modelling base for efficient information discovery in distributed systems is demonstrated with the aid of a novel conceptual entity called infotron.The investigations are focused on distributed systems and their associated problems. The study was directed towards identifying suitable software architecture and incorporating the same in an environment where information growth is phenomenal and a proper mechanism for carrying out information discovery becomes feasible. An empirical study undertaken with the aid of an election database of constituencies distributed geographically, provided the insights required. This is manifested in the Election Counting and Reporting Software (ECRS) System. ECRS system is a software system, which is essentially distributed in nature designed to prepare reports to district administrators about the election counting process and to generate other miscellaneous statutory reports.Most of the distributed systems of the nature of ECRS normally will possess a "fragile architecture" which would make them amenable to collapse, with the occurrence of minor faults. This is resolved with the help of the penta-tier architecture proposed, that contained five different technologies at different tiers of the architecture.The results of experiment conducted and its analysis show that such an architecture would help to maintain different components of the software intact in an impermeable manner from any internal or external faults. The architecture thus evolved needed a mechanism to support information processing and discovery. This necessitated the introduction of the noveI concept of infotrons. Further, when a computing machine has to perform any meaningful extraction of information, it is guided by what is termed an infotron dictionary.The other empirical study was to find out which of the two prominent markup languages namely HTML and XML, is best suited for the incorporation of infotrons. A comparative study of 200 documents in HTML and XML was undertaken. The result was in favor ofXML.The concept of infotron and that of infotron dictionary, which were developed, was applied to implement an Information Discovery System (IDS). IDS is essentially, a system, that starts with the infotron(s) supplied as clue(s), and results in brewing the information required to satisfy the need of the information discoverer by utilizing the documents available at its disposal (as information space). The various components of the system and their interaction follows the penta-tier architectural model and therefore can be considered fault-tolerant. IDS is generic in nature and therefore the characteristics and the specifications were drawn up accordingly. Many subsystems interacted with multiple infotron dictionaries that were maintained in the system.In order to demonstrate the working of the IDS and to discover the information without modification of a typical Library Information System (LIS), an Information Discovery in Library Information System (lDLIS) application was developed. IDLIS is essentially a wrapper for the LIS, which maintains all the databases of the library. The purpose was to demonstrate that the functionality of a legacy system could be enhanced with the augmentation of IDS leading to information discovery service. IDLIS demonstrates IDS in action. IDLIS proves that any legacy system could be augmented with IDS effectively to provide the additional functionality of information discovery service.Possible applications of IDS and scope for further research in the field are covered.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Drainage basins are durable geomorphic features that provide insights into the long term evolution of the landscape. River basin geometry develop response to the nature and distribution of uplift and subsidence, the spatial arrangement of lineaments (faults and joints), the relative resistance of different rock types and to climatically influenced hydrological parameters . For developing a drainage basin evolution history, it is necessary to understand physiography, drainage patterns, geomorphic features and its structural control and erosion status. The present study records evidences for active tectonic activities which were found to be responsible for the present day geomorphic set up of the study area since the Western Ghat evolution. A model was developed to explain the evolution of Chaliar River drainage basin based on detailed interpretation of morphometry and genesis of landforms with special emphasis on tectonic geomorphic indices and markers.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Embedded systems are usually designed for a single or a specified set of tasks. This specificity means the system design as well as its hardware/software development can be highly optimized. Embedded software must meet the requirements such as high reliability operation on resource-constrained platforms, real time constraints and rapid development. This necessitates the adoption of static machine codes analysis tools running on a host machine for the validation and optimization of embedded system codes, which can help meet all of these goals. This could significantly augment the software quality and is still a challenging field.Embedded systems are usually designed for a single or a specified set of tasks. This specificity means the system design as well as its hardware/software development can be highly optimized. Embedded software must meet the requirements such as high reliability operation on resource-constrained platforms, real time constraints and rapid development. This necessitates the adoption of static machine codes analysis tools running on a host machine for the validation and optimization of embedded system codes, which can help meet all of these goals. This could significantly augment the software quality and is still a challenging field.Embedded systems are usually designed for a single or a specified set of tasks. This specificity means the system design as well as its hardware/software development can be highly optimized. Embedded software must meet the requirements such as high reliability operation on resource-constrained platforms, real time constraints and rapid development. This necessitates the adoption of static machine codes analysis tools running on a host machine for the validation and optimization of embedded system codes, which can help meet all of these goals. This could significantly augment the software quality and is still a challenging field.Embedded systems are usually designed for a single or a specified set of tasks. This specificity means the system design as well as its hardware/software development can be highly optimized. Embedded software must meet the requirements such as high reliability operation on resource-constrained platforms, real time constraints and rapid development. This necessitates the adoption of static machine codes analysis tools running on a host machine for the validation and optimization of embedded system codes, which can help meet all of these goals. This could significantly augment the software quality and is still a challenging field.This dissertation contributes to an architecture oriented code validation, error localization and optimization technique assisting the embedded system designer in software debugging, to make it more effective at early detection of software bugs that are otherwise hard to detect, using the static analysis of machine codes. The focus of this work is to develop methods that automatically localize faults as well as optimize the code and thus improve the debugging process as well as quality of the code.Validation is done with the help of rules of inferences formulated for the target processor. The rules govern the occurrence of illegitimate/out of place instructions and code sequences for executing the computational and integrated peripheral functions. The stipulated rules are encoded in propositional logic formulae and their compliance is tested individually in all possible execution paths of the application programs. An incorrect sequence of machine code pattern is identified using slicing techniques on the control flow graph generated from the machine code.An algorithm to assist the compiler to eliminate the redundant bank switching codes and decide on optimum data allocation to banked memory resulting in minimum number of bank switching codes in embedded system software is proposed. A relation matrix and a state transition diagram formed for the active memory bank state transition corresponding to each bank selection instruction is used for the detection of redundant codes. Instances of code redundancy based on the stipulated rules for the target processor are identified.This validation and optimization tool can be integrated to the system development environment. It is a novel approach independent of compiler/assembler, applicable to a wide range of processors once appropriate rules are formulated. Program states are identified mainly with machine code pattern, which drastically reduces the state space creation contributing to an improved state-of-the-art model checking. Though the technique described is general, the implementation is architecture oriented, and hence the feasibility study is conducted on PIC16F87X microcontrollers. The proposed tool will be very useful in steering novices towards correct use of difficult microcontroller features in developing embedded systems.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

With the recent progress and rapid increase in mobile terminals, the design of antennas for small mobile terminals is acquiring great importance. In view of this situation, several design concepts are already been addressed by the scientists and engineers. Compactness and efficiency are the major criteria for mobile terminal antennas. The challenging task of the microwave scientists and engineers is to device compact printed radiating systems having broadband behavior, together with good efficiency. Printed antenna technology has received popularity among antenna scientists after the introduction of microstrip antenna in 1970s. The successors in this kind such as printed monopoles and planar inverted F are also equally important. Scientists and Engineers are trying to explore this technology as a viable coast effective solution for forthcoming microwave revolution. The transmission line perspectives of antennas are very interesting. The concept behind any electromagnetic radiator is simple. Any electromagnetic system with a discontinuity is radiating electromagnetic energy. The size, shape and the orientation of the discontinuities controls the radiation characteristics of the system such as radiation pattern, gain, polarization etc. It can be either resonant or non resonant structure. Microstrip antennas are suitable for wireless applications due to their low cost, high gain and ease of fabrication. But the major disadvantage of micro strip antennas is their inherent narrow bandwidth. A lot of techniques are introduced by the researchers all over the world to enhance the bandwidth of micro strip patch antennas. The thesis addresses an attempt to enhance the bandwidth of micro strip patch antennas by incorporating impedance matching strip as a part of the micro strip patch antenna. The first part of the thesis deals with the broadband operation of the tilted square slot and polygonal slot loaded square micro strip patch antennas. The resonant mechanisms are clearly mentioned using the simulation and experimental studies. The bandwidth of the polygonal slotted broadband patch antenna is again enhanced by implementing an Lstrip feed mechanism. In the second major part of the thesis, a novel gain enhancement technique for single band and broadband square micro strip patch antennas is achieved by implementing offset stacked configurations.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Biometrics deals with the physiological and behavioral characteristics of an individual to establish identity. Fingerprint based authentication is the most advanced biometric authentication technology. The minutiae based fingerprint identification method offer reasonable identification rate. The feature minutiae map consists of about 70-100 minutia points and matching accuracy is dropping down while the size of database is growing up. Hence it is inevitable to make the size of the fingerprint feature code to be as smaller as possible so that identification may be much easier. In this research, a novel global singularity based fingerprint representation is proposed. Fingerprint baseline, which is the line between distal and intermediate phalangeal joint line in the fingerprint, is taken as the reference line. A polygon is formed with the singularities and the fingerprint baseline. The feature vectors are the polygonal angle, sides, area, type and the ridge counts in between the singularities. 100% recognition rate is achieved in this method. The method is compared with the conventional minutiae based recognition method in terms of computation time, receiver operator characteristics (ROC) and the feature vector length. Speech is a behavioural biometric modality and can be used for identification of a speaker. In this work, MFCC of text dependant speeches are computed and clustered using k-means algorithm. A backpropagation based Artificial Neural Network is trained to identify the clustered speech code. The performance of the neural network classifier is compared with the VQ based Euclidean minimum classifier. Biometric systems that use a single modality are usually affected by problems like noisy sensor data, non-universality and/or lack of distinctiveness of the biometric trait, unacceptable error rates, and spoof attacks. Multifinger feature level fusion based fingerprint recognition is developed and the performances are measured in terms of the ROC curve. Score level fusion of fingerprint and speech based recognition system is done and 100% accuracy is achieved for a considerable range of matching threshold

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Fingerprint based authentication systems are one of the cost-effective biometric authentication techniques employed for personal identification. As the data base population increases, fast identification/recognition algorithms are required with high accuracy. Accuracy can be increased using multimodal evidences collected by multiple biometric traits. In this work, consecutive fingerprint images are taken, global singularities are located using directional field strength and their local orientation vector is formulated with respect to the base line of the finger. Feature level fusion is carried out and a 32 element feature template is obtained. A matching score is formulated for the identification and 100% accuracy was obtained for a database of 300 persons. The polygonal feature vector helps to reduce the size of the feature database from the present 70-100 minutiae features to just 32 features and also a lower matching threshold can be fixed compared to single finger based identification

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Mit aktiven Magnetlagern ist es möglich, rotierende Körper durch magnetische Felder berührungsfrei zu lagern. Systembedingt sind bei aktiv magnetgelagerten Maschinen wesentliche Signale ohne zusätzlichen Aufwand an Messtechnik für Diagnoseaufgaben verfügbar. In der Arbeit wird ein Konzept entwickelt, das durch Verwendung der systeminhärenten Signale eine Diagnose magnetgelagerter rotierender Maschinen ermöglicht und somit neben einer kontinuierlichen Anlagenüberwachung eine schnelle Bewertung des Anlagenzustandes gestattet. Fehler können rechtzeitig und ursächlich in Art und Größe erkannt und entsprechende Gegenmaßnahmen eingeleitet werden. Anhand der erfassten Signale geschieht die Gewinnung von Merkmalen mit signal- und modellgestützten Verfahren. Für den Magnetlagerregelkreis erfolgen Untersuchungen zum Einsatz modellgestützter Parameteridentifikationsverfahren, deren Verwendbarkeit wird bei der Diagnose am Regler und Leistungsverstärker nachgewiesen. Unter Nutzung von Simulationsmodellen sowie durch Experimente an Versuchsständen werden die Merkmalsverläufe im normalen Referenzzustand und bei auftretenden Fehlern aufgenommen und die Ergebnisse in einer Wissensbasis abgelegt. Diese dient als Grundlage zur Festlegung von Grenzwerten und Regeln für die Überwachung des Systems und zur Erstellung wissensbasierter Diagnosemodelle. Bei der Überwachung werden die Merkmalsausprägungen auf das Überschreiten von Grenzwerten überprüft, Informationen über erkannte Fehler und Betriebszustände gebildet sowie gegebenenfalls Alarmmeldungen ausgegeben. Sich langsam anbahnende Fehler können durch die Berechnung der Merkmalstrends mit Hilfe der Regressionsanalyse erkannt werden. Über die bisher bei aktiven Magnetlagern übliche Überwachung von Grenzwerten hinaus erfolgt bei der Fehlerdiagnose eine Verknüpfung der extrahierten Merkmale zur Identifizierung und Lokalisierung auftretender Fehler. Die Diagnose geschieht mittels regelbasierter Fuzzy-Logik, dies gestattet die Einbeziehung von linguistischen Aussagen in Form von Expertenwissen sowie die Berücksichtigung von Unbestimmtheiten und ermöglicht damit eine Diagnose komplexer Systeme. Für Aktor-, Sensor- und Reglerfehler im Magnetlagerregelkreis sowie Fehler durch externe Kräfte und Unwuchten werden Diagnosemodelle erstellt und verifiziert. Es erfolgt der Nachweis, dass das entwickelte Diagnosekonzept mit beherrschbarem Rechenaufwand korrekte Diagnoseaussagen liefert. Durch Kaskadierung von Fuzzy-Logik-Modulen wird die Transparenz des Regelwerks gewahrt und die Abarbeitung der Regeln optimiert. Endresultat ist ein neuartiges hybrides Diagnosekonzept, welches signal- und modellgestützte Verfahren der Merkmalsgewinnung mit wissensbasierten Methoden der Fehlerdiagnose kombiniert. Das entwickelte Diagnosekonzept ist für die Anpassung an unterschiedliche Anforderungen und Anwendungen bei rotierenden Maschinen konzipiert.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Kern der vorliegenden Arbeit ist die Erforschung von Methoden, Techniken und Werkzeugen zur Fehlersuche in modellbasierten Softwareentwicklungsprozessen. Hierzu wird zuerst ein von mir mitentwickelter, neuartiger und modellbasierter Softwareentwicklungsprozess, der sogenannte Fujaba Process, vorgestellt. Dieser Prozess wird von Usecase Szenarien getrieben, die durch spezielle Kollaborationsdiagramme formalisiert werden. Auch die weiteren Artefakte des Prozess bishin zur fertigen Applikation werden durch UML Diagrammarten modelliert. Es ist keine Programmierung im Quelltext nötig. Werkzeugunterstützung für den vorgestellte Prozess wird von dem Fujaba CASE Tool bereitgestellt. Große Teile der Werkzeugunterstützung für den Fujaba Process, darunter die Toolunterstützung für das Testen und Debuggen, wurden im Rahmen dieser Arbeit entwickelt. Im ersten Teil der Arbeit wird der Fujaba Process im Detail erklärt und unsere Erfahrungen mit dem Einsatz des Prozesses in Industrieprojekten sowie in der Lehre dargestellt. Der zweite Teil beschreibt die im Rahmen dieser Arbeit entwickelte Testgenerierung, die zu einem wichtigen Teil des Fujaba Process geworden ist. Hierbei werden aus den formalisierten Usecase Szenarien ausführbare Testfälle generiert. Es wird das zugrunde liegende Konzept, die konkrete technische Umsetzung und die Erfahrungen aus der Praxis mit der entwickelten Testgenerierung dargestellt. Der letzte Teil beschäftigt sich mit dem Debuggen im Fujaba Process. Es werden verschiedene im Rahmen dieser Arbeit entwickelte Konzepte und Techniken vorgestellt, die die Fehlersuche während der Applikationsentwicklung vereinfachen. Hierbei wurde darauf geachtet, dass das Debuggen, wie alle anderen Schritte im Fujaba Process, ausschließlich auf Modellebene passiert. Unter anderem werden Techniken zur schrittweisen Ausführung von Modellen, ein Objekt Browser und ein Debugger, der die rückwärtige Ausführung von Programmen erlaubt (back-in-time debugging), vorgestellt. Alle beschriebenen Konzepte wurden in dieser Arbeit als Plugins für die Eclipse Version von Fujaba, Fujaba4Eclipse, implementiert und erprobt. Bei der Implementierung der Plugins wurde auf eine enge Integration mit Fujaba zum einen und mit Eclipse auf der anderen Seite geachtet. Zusammenfassend wird also ein Entwicklungsprozess vorgestellt, die Möglichkeit in diesem mit automatischen Tests Fehler zu identifizieren und diese Fehler dann mittels spezieller Debuggingtechniken im Programm zu lokalisieren und schließlich zu beheben. Dabei läuft der komplette Prozess auf Modellebene ab. Für die Test- und Debuggingtechniken wurden in dieser Arbeit Plugins für Fujaba4Eclipse entwickelt, die den Entwickler bestmöglich bei der zugehörigen Tätigkeit unterstützen.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The main focus and concerns of this PhD thesis is the growth of III-V semiconductor nanostructures (Quantum dots (QDs) and quantum dashes) on silicon substrates using molecular beam epitaxy (MBE) technique. The investigation of influence of the major growth parameters on their basic properties (density, geometry, composition, size etc.) and the systematic characterization of their structural and optical properties are the core of the research work. The monolithic integration of III-V optoelectronic devices with silicon electronic circuits could bring enormous prospect for the existing semiconductor technology. Our challenging approach is to combine the superior passive optical properties of silicon with the superior optical emission properties of III-V material by reducing the amount of III-V materials to the very limit of the active region. Different heteroepitaxial integration approaches have been investigated to overcome the materials issues between III-V and Si. However, this include the self-assembled growth of InAs and InGaAs QDs in silicon and GaAx matrices directly on flat silicon substrate, sitecontrolled growth of (GaAs/In0,15Ga0,85As/GaAs) QDs on pre-patterned Si substrate and the direct growth of GaP on Si using migration enhanced epitaxy (MEE) and MBE growth modes. An efficient ex-situ-buffered HF (BHF) and in-situ surface cleaning sequence based on atomic hydrogen (AH) cleaning at 500 °C combined with thermal oxide desorption within a temperature range of 700-900 °C has been established. The removal of oxide desorption was confirmed by semicircular streaky reflection high energy electron diffraction (RHEED) patterns indicating a 2D smooth surface construction prior to the MBE growth. The evolution of size, density and shape of the QDs are ex-situ characterized by atomic-force microscopy (AFM) and transmission electron microscopy (TEM). The InAs QDs density is strongly increased from 108 to 1011 cm-2 at V/III ratios in the range of 15-35 (beam equivalent pressure values). InAs QD formations are not observed at temperatures of 500 °C and above. Growth experiments on (111) substrates show orientation dependent QD formation behaviour. A significant shape and size transition with elongated InAs quantum dots and dashes has been observed on (111) orientation and at higher Indium-growth rate of 0.3 ML/s. The 2D strain mapping derived from high-resolution TEM of InAs QDs embedded in silicon matrix confirmed semi-coherent and fully relaxed QDs embedded in defectfree silicon matrix. The strain relaxation is released by dislocation loops exclusively localized along the InAs/Si interfaces and partial dislocations with stacking faults inside the InAs clusters. The site controlled growth of GaAs/In0,15Ga0,85As/GaAs nanostructures has been demonstrated for the first time with 1 μm spacing and very low nominal deposition thicknesses, directly on pre-patterned Si without the use of SiO2 mask. Thin planar GaP layer was successfully grown through migration enhanced epitaxy (MEE) to initiate a planar GaP wetting layer at the polar/non-polar interface, which work as a virtual GaP substrate, for the GaP-MBE subsequently growth on the GaP-MEE layer with total thickness of 50 nm. The best root mean square (RMS) roughness value was as good as 1.3 nm. However, these results are highly encouraging for the realization of III-V optical devices on silicon for potential applications.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Colloidal self assembly is an efficient method for making 3-D ordered nanostructures suitable for materials such as photonic crystals and macroscopic solids for catalysis and sensor applications. Colloidal crystals grown by convective methods exhibit defects on two different scales. Macro defects such as cracks and void bands originate from the dynamics of meniscus motion during colloidal crystal growth while micro defects like vacancies, dislocation and stacking faults are indigenous to the colloidal crystalline structure. This paper analyses the crystallography and energetics of the microscopic defects from the point of view of classical thermodynamics and discusses the strategy for the control of the macroscopic defects through optimization of the liquid-vapor interface.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Este proyecto de grado pretende evaluar el comportamiento productivo de la empresa Plaspucol ubicada en la ciudad de Bogotá, identificando las deficiencias en su proceso y generando mecanismo de mejoramiento a través de un previo análisis. Para ello es necesario partir de un marco teórico del plástico como lo es su historia, su evolución, clasificación y su posicionamiento e influencia económica a nivel mundial y vista a su vez desde el ámbito nacional. Para analizar dichas situaciones se usaron herramientas aprendidas en lo largo de nuestra formación profesional como los diagramas de recorrido, diagramas de flujo, diagrama hombre-máquina, diagrama de balanceo, muestreo y un moderno simulador llamado Promodel con el cual se diagnostica a la empresa identificando puntos débiles y cuellos de botella en la producción en la situación actual y se crea una situación futura con propuestas de mejora empleando éste simulador.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

One of the techniques used to detect faults in dynamic systems is analytical redundancy. An important difficulty in applying this technique to real systems is dealing with the uncertainties associated with the system itself and with the measurements. In this paper, this uncertainty is taken into account by the use of intervals for the parameters of the model and for the measurements. The method that is proposed in this paper checks the consistency between the system's behavior, obtained from the measurements, and the model's behavior; if they are inconsistent, then there is a fault. The problem of detecting faults is stated as a quantified real constraint satisfaction problem, which can be solved using the modal interval analysis (MIA). MIA is used because it provides powerful tools to extend the calculations over real functions to intervals. To improve the results of the detection of the faults, the simultaneous use of several sliding time windows is proposed. The result of implementing this method is semiqualitative tracking (SQualTrack), a fault-detection tool that is robust in the sense that it does not generate false alarms, i.e., if there are false alarms, they indicate either that the interval model does not represent the system adequately or that the interval measurements do not represent the true values of the variables adequately. SQualTrack is currently being used to detect faults in real processes. Some of these applications using real data have been developed within the European project advanced decision support system for chemical/petrochemical manufacturing processes and are also described in this paper

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Process supervision is the activity focused on monitoring the process operation in order to deduce conditions to maintain the normality including when faults are present Depending on the number/distribution/heterogeneity of variables, behaviour situations, sub-processes, etc. from processes, human operators and engineers do not easily manipulate the information. This leads to the necessity of automation of supervision activities. Nevertheless, the difficulty to deal with the information complicates the design and development of software applications. We present an approach called "integrated supervision systems". It proposes multiple supervisors coordination to supervise multiple sub-processes whose interactions permit one to supervise the global process