882 resultados para Stone Tools
Resumo:
The need to effectively manage the documentation covering the entire production process, from the concept phase right through to market realise, constitutes a key issue in the creation of a successful and highly competitive product. For almost forty years the most commonly used strategies to achieve this have followed Product Lifecycle Management (PLM) guidelines. Translated into information management systems at the end of the '90s, this methodology is now widely used by companies operating all over the world in many different sectors. PLM systems and editor programs are the two principal types of software applications used by companies for their process aotomation. Editor programs allow to store in documents the information related to the production chain, while the PLM system stores and shares this information so that it can be used within the company and made it available to partners. Different software tools, which capture and store documents and information automatically in the PLM system, have been developed in recent years. One of them is the ''DirectPLM'' application, which has been developed by the Italian company ''Focus PLM''. It is designed to ensure interoperability between many editors and the Aras Innovator PLM system. In this dissertation we present ''DirectPLM2'', a new version of the previous software application DirectPLM. It has been designed and developed as prototype during the internship by Focus PLM. Its new implementation separates the abstract logic of business from the real commands implementation, previously strongly dependent on Aras Innovator. Thanks to its new design, Focus PLM can easily develop different versions of DirectPLM2, each one devised for a specific PLM system. In fact, the company can focus the development effort only on a specific set of software components which provides specialized functions interacting with that particular PLM system. This allows shorter Time-To-Market and gives the company a significant competitive advantage.
Resumo:
This thesis is focused on Smart Grid applications in medium voltage distribution networks. For the development of new applications it appears useful the availability of simulation tools able to model dynamic behavior of both the power system and the communication network. Such a co-simulation environment would allow the assessment of the feasibility of using a given network technology to support communication-based Smart Grid control schemes on an existing segment of the electrical grid and to determine the range of control schemes that different communications technologies can support. For this reason, is presented a co-simulation platform that has been built by linking the Electromagnetic Transients Program Simulator (EMTP v3.0) with a Telecommunication Network Simulator (OPNET-Riverbed v18.0). The simulator is used to design and analyze a coordinate use of Distributed Energy Resources (DERs) for the voltage/var control (VVC) in distribution network. This thesis is focused control structure based on the use of phase measurement units (PMUs). In order to limit the required reinforcements of the communication infrastructures currently adopted by Distribution Network Operators (DNOs), the study is focused on leader-less MAS schemes that do not assign special coordinating rules to specific agents. Leader-less MAS are expected to produce more uniform communication traffic than centralized approaches that include a moderator agent. Moreover, leader-less MAS are expected to be less affected by limitations and constraint of some communication links. The developed co-simulator has allowed the definition of specific countermeasures against the limitations of the communication network, with particular reference to the latency and loss and information, for both the case of wired and wireless communication networks. Moreover, the co-simulation platform has bee also coupled with a mobility simulator in order to study specific countermeasures against the negative effects on the medium voltage/current distribution network caused by the concurrent connection of electric vehicles.
Resumo:
Data sets describing the state of the earth's atmosphere are of great importance in the atmospheric sciences. Over the last decades, the quality and sheer amount of the available data increased significantly, resulting in a rising demand for new tools capable of handling and analysing these large, multidimensional sets of atmospheric data. The interdisciplinary work presented in this thesis covers the development and the application of practical software tools and efficient algorithms from the field of computer science, aiming at the goal of enabling atmospheric scientists to analyse and to gain new insights from these large data sets. For this purpose, our tools combine novel techniques with well-established methods from different areas such as scientific visualization and data segmentation. In this thesis, three practical tools are presented. Two of these tools are software systems (Insight and IWAL) for different types of processing and interactive visualization of data, the third tool is an efficient algorithm for data segmentation implemented as part of Insight.Insight is a toolkit for the interactive, three-dimensional visualization and processing of large sets of atmospheric data, originally developed as a testing environment for the novel segmentation algorithm. It provides a dynamic system for combining at runtime data from different sources, a variety of different data processing algorithms, and several visualization techniques. Its modular architecture and flexible scripting support led to additional applications of the software, from which two examples are presented: the usage of Insight as a WMS (web map service) server, and the automatic production of a sequence of images for the visualization of cyclone simulations. The core application of Insight is the provision of the novel segmentation algorithm for the efficient detection and tracking of 3D features in large sets of atmospheric data, as well as for the precise localization of the occurring genesis, lysis, merging and splitting events. Data segmentation usually leads to a significant reduction of the size of the considered data. This enables a practical visualization of the data, statistical analyses of the features and their events, and the manual or automatic detection of interesting situations for subsequent detailed investigation. The concepts of the novel algorithm, its technical realization, and several extensions for avoiding under- and over-segmentation are discussed. As example applications, this thesis covers the setup and the results of the segmentation of upper-tropospheric jet streams and cyclones as full 3D objects. Finally, IWAL is presented, which is a web application for providing an easy interactive access to meteorological data visualizations, primarily aimed at students. As a web application, the needs to retrieve all input data sets and to install and handle complex visualization tools on a local machine are avoided. The main challenge in the provision of customizable visualizations to large numbers of simultaneous users was to find an acceptable trade-off between the available visualization options and the performance of the application. Besides the implementational details, benchmarks and the results of a user survey are presented.
Resumo:
Autism Spectrum Disorders (ASDs) describe a set of neurodevelopmental disorders. ASD represents a significant public health problem. Currently, ASDs are not diagnosed before the 2nd year of life but an early identification of ASDs would be crucial as interventions are much more effective than specific therapies starting in later childhood. To this aim, cheap an contact-less automatic approaches recently aroused great clinical interest. Among them, the cry and the movements of the newborn, both involving the central nervous system, are proposed as possible indicators of neurological disorders. This PhD work is a first step towards solving this challenging problem. An integrated system is presented enabling the recording of audio (crying) and video (movements) data of the newborn, their automatic analysis with innovative techniques for the extraction of clinically relevant parameters and their classification with data mining techniques. New robust algorithms were developed for the selection of the voiced parts of the cry signal, the estimation of acoustic parameters based on the wavelet transform and the analysis of the infant’s general movements (GMs) through a new body model for segmentation and 2D reconstruction. In addition to a thorough literature review this thesis presents the state of the art on these topics that shows that no studies exist concerning normative ranges for newborn infant cry in the first 6 months of life nor the correlation between cry and movements. Through the new automatic methods a population of control infants (“low-risk”, LR) was compared to a group of “high-risk” (HR) infants, i.e. siblings of children already diagnosed with ASD. A subset of LR infants clinically diagnosed as newborns with Typical Development (TD) and one affected by ASD were compared. The results show that the selected acoustic parameters allow good differentiation between the two groups. This result provides new perspectives both diagnostic and therapeutic.
Resumo:
The postharvest phase has been considered an environment very suitable for successful application of biological control agents (BCAs). However, the tri-interaction between fungal pathogen, host (fruit) and antagonist is influenced by several parameters such as temperature, oxidative stresses, oxygen composition, water activity, etc. that could be determining for the success of biocontrol. Knowledge of the modes of action of BCAs is essential in order to enhance their viability and increase their potentialities in disease control. The thesis focused on the possibility to explain the modes of action of a biological control agent (BCA): Aureobasidium pullulans, in particular the strains L1 and L8, control effective against fruit postharvest fungal pathogen. In particular in this work were studied the different modes of action of BCA, such as: i) the ability to produce volatile organic compounds (VOCs), identified by SPME- gas chromatography-mass spectrometry (GC-MS) and tested by in vitro and in vivo assays against Penicillium spp., Botrytis cinerea, Colletotrichum acutatum; ii) the ability to produce lytic enzymes (exo and endo chitinase and β-1,3-glucanase) tested against Monilinia laxa, causal agent of brown rot of stone fruits. L1 and L8 lytic enzymes were also evaluated through their relative genes by molecular tools; iii) the competition for space and nutrients, such as sugars (sucrose, glucose and fructose) and iron; the latter induced the production of siderophores, molecules with high affinity for iron chelation. A molecular investigation was carried out to better understand the gene regulation strictly correlated to the production of these chelating molucules. The competition for space against M. laxa was verified by electron microscopy techniques; iv) a depth bibliographical analysis on BCAs mechanisms of action and their possible combination with physical and chemical treatments was conducted.
Resumo:
Die qualitative und quantitative Analyse von Biomolekülen hat in den letzten Jahren und Jahrzehnten immer mehr an Bedeutung gewonnen. Durch das Aufkommen und die kontinuierliche Weiterentwicklung neuer Separations- und Detektionsmethoden und deren Verbindung miteinander zu leistungsfähigen Einheiten, erlangte man Schritt für Schritt neue Erkenntnisse bei ihrer Untersuchung. Die Elementmassenspektrometrie als nachweisstarke Detektionsmethode wird von vielen wissenschaftlichen Arbeitsgruppen bei der Trennung und Quantifizierung von Proteinen und Metalloproteinen mittels Detektion der in den Biomolekülen vorkommenden Metalle und Heteroatome angewendet. Heteroatome (z.B. Schwefel, Phosphor) haben im Plasma des ICP-MS (inductively coupled plasma - mass spectrometer) schlechte Ionisationseigenschaften und dementsprechend deutlich höhere Nachweisgrenzen als Metalle. Ein Ansatz, schlecht oder nicht detektierbare Verbindungen (also solche, die keine Metalle oder Heteroatome enthalten) mit dem ICP-MS sichtbar zu machen, ist die Markierung der selbigen mit Metallionen oder -cluster. rnIn dieser Arbeit ist es gelungen, der Analyse ganz unterschiedlicher Substanzklassen, zum einen metallische Nanopartikel und zum anderen Proteine, neue Impulse zu geben und zukünftiges Potential bei der Anwendung gekoppelter Techniken zur Separation und Detektion aufzuzeigen. Durch die Verwendung einer alten, aber neu konzipierten Trenntechnik, der Gelelektrophorese (GE), und deren Kopplung an einen modernen Detektor, dem ICP-MS, kann die für die Proteinanalytik weit verbreitete Gelelektrophorese ihr enormes Potential bei der Trennung verschiedenster Verbindungsklassen mit der exzellenten Nachweisstärke und Elementspezifität des ICP-MS verbinden und dadurch mit deutlich weniger Arbeitsaufwand als bisher qualitative und auch quantitative Ergebnisse produzieren. Bisher war dies nur mit großem präparativem Aufwand unter Verwendung der laser ablation möglich. Bei der Analyse von Nanopartikeln konnte aufgezeigt werden, dass durch die GE-ICP-MS-Kopplung aufgrund der guten Trenneigenschaften der GE vorhandene Spezies bzw. Fraktionen voneinander separiert werden und mit Hilfe des ICP-MS Informationen auf atomarem Niveau gewonnen werden können. Es war möglich, das atomare Verhältnis der Metallatome im Kern und der Schwefelatome in der Ligandenhülle eines Nanopartikels zu bestimmen und damit die Größe des Partikels abzuschätzen. Auch konnte die Anzahl der Goldatome in einem dem Schmid-Cluster ähnlichen Nanopartikel bestimmt werden, was vorher nur mit Hilfe von MALDI-TOF möglich war. Bei der Analyse von Biomolekülen konnte auf einfache Weise der Phosphorylierungsgrad verschiedener Proteine bestimmt werden. Auch bei kleinen Molekülen erzielt die Gelelektrophorese ausgezeichnete Trennergebnisse, wie z. B. bei der Analyse verschiedener Brom- und Iodspezies.rnDie stöchiometrische Kopplung eines Proteins an einen Nanopartikel, ohne eine der beiden Verbindungen in einem größeren Maße zu verändern, stellte jedoch eine Herausforderung dar, die im Rahmen dieser Arbeit nicht vollständig gelöst werden konnte. Verschiedene Ansätze zur Kopplung der beiden Substanzen wurden erprobt, jedoch führte keine zu dem gewünschten Ergebnis einer stöchiometrisch vollständigen und spezifischen Modifikation eines Proteins mit einem Nanopartikel. Durch das Potential der GE-ICP-MS-Kopplung bei der Analyse beider Substanz-klassen und dem Beweis der Praktikabilität und Zuverlässigkeit der Methode ist jedoch der Grundstein für weitere Forschungen auf diesem Gebiet gelegt worden. Ist eine geeignete chemische Kopplung der beiden Substanzklassen gefunden und beherrscht, steht auf analytischer Seite eine leistungsstarke Kombination aus Trennung und Detektion zur Verfügung, um die Quantifizierung von Proteinen entscheidend zu verbessern.rn
Resumo:
In this thesis, we develop high precision tools for the simulation of slepton pair production processes at hadron colliders and apply them to phenomenological studies at the LHC. Our approach is based on the POWHEG method for the matching of next-to-leading order results in perturbation theory to parton showers. We calculate matrix elements for slepton pair production and for the production of a slepton pair in association with a jet perturbatively at next-to-leading order in supersymmetric quantum chromodynamics. Both processes are subsequently implemented in the POWHEG BOX, a publicly available software tool that contains general parts of the POWHEG matching scheme. We investigate phenomenological consequences of our calculations in several setups that respect experimental exclusion limits for supersymmetric particles and provide precise predictions for slepton signatures at the LHC. The inclusion of QCD emissions in the partonic matrix elements allows for an accurate description of hard jets. Interfacing our codes to the multi-purpose Monte-Carlo event generator PYTHIA, we simulate parton showers and slepton decays in fully exclusive events. Advanced kinematical variables and specific search strategies are examined as means for slepton discovery in experimentally challenging setups.
Resumo:
In haemodynamically stable patients with acute symptomatic pulmonary embolism (PE), studies have not evaluated the usefulness of combining the measurement of cardiac troponin, transthoracic echocardiogram (TTE), and lower extremity complete compression ultrasound (CCUS) testing for predicting the risk of PE-related death.
Resumo:
Background Through this paper, we present the initial steps for the creation of an integrated platform for the provision of a series of eHealth tools and services to both citizens and travelers in isolated areas of thesoutheast Mediterranean, and on board ships travelling across it. The platform was created through an INTERREG IIIB ARCHIMED project called INTERMED. Methods The support of primary healthcare, home care and the continuous education of physicians are the three major issues that the proposed platform is trying to facilitate. The proposed system is based on state-of-the-art telemedicine systems and is able to provide the following healthcare services: i) Telecollaboration and teleconsultation services between remotely located healthcare providers, ii) telemedicine services in emergencies, iii) home telecare services for "at risk" citizens such as the elderly and patients with chronic diseases, and iv) eLearning services for the continuous training through seminars of both healthcare personnel (physicians, nurses etc) and persons supporting "at risk" citizens. These systems support data transmission over simple phone lines, internet connections, integrated services digital network/digital subscriber lines, satellite links, mobile networks (GPRS/3G), and wireless local area networks. The data corresponds, among others, to voice, vital biosignals, still medical images, video, and data used by eLearning applications. The proposed platform comprises several systems, each supporting different services. These were integrated using a common data storage and exchange scheme in order to achieve system interoperability in terms of software, language and national characteristics. Results The platform has been installed and evaluated in different rural and urban sites in Greece, Cyprus and Italy. The evaluation was mainly related to technical issues and user satisfaction. The selected sites are, among others, rural health centers, ambulances, homes of "at-risk" citizens, and a ferry. Conclusions The results proved the functionality and utilization of the platform in various rural places in Greece, Cyprus and Italy. However, further actions are needed to enable the local healthcare systems and the different population groups to be familiarized with, and use in their everyday lives, mature technological solutions for the provision of healthcare services.
Resumo:
Domain-specific languages (DSLs) are increasingly used as embedded languages within general-purpose host languages. DSLs provide a compact, dedicated syntax for specifying parts of an application related to specialized domains. Unfortunately, such language extensions typically do not integrate well with the development tools of the host language. Editors, compilers and debuggers are either unaware of the extensions, or must be adapted at a non-trivial cost. We present a novel approach to embed DSLs into an existing host language by leveraging the underlying representation of the host language used by these tools. Helvetia is an extensible system that intercepts the compilation pipeline of the Smalltalk host language to seamlessly integrate language extensions. We validate our approach by case studies that demonstrate three fundamentally different ways to extend or adapt the host language syntax and semantics.
Resumo:
Written text is an important component in the process of knowledge acquisition and communication. Poorly written text fails to deliver clear ideas to the reader no matter how revolutionary and ground-breaking these ideas are. Providing text with good writing style is essential to transfer ideas smoothly. While we have sophisticated tools to check for stylistic problems in program code, we do not apply the same techniques for written text. In this paper we present TextLint, a rule-based tool to check for common style errors in natural language. TextLint provides a structural model of written text and an extensible rule-based checking mechanism.