928 resultados para Traduzione automatica, traduzione automatica del parlato, Machine Translation, Speech-to-Speech Translation, Google Translate, Microsoft Translator, valutazione della qualità in traduzione, translation quality assessment


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Statistical machine translation (SMT) is an approach to Machine Translation (MT) that uses statistical models whose parameter estimation is based on the analysis of existing human translations (contained in bilingual corpora). From a translation student’s standpoint, this dissertation aims to explain how a phrase-based SMT system works, to determine the role of the statistical models it uses in the translation process and to assess the quality of the translations provided that system is trained with in-domain goodquality corpora. To that end, a phrase-based SMT system based on Moses has been trained and subsequently used for the English to Spanish translation of two texts related in topic to the training data. Finally, the quality of this output texts produced by the system has been assessed through a quantitative evaluation carried out with three different automatic evaluation measures and a qualitative evaluation based on the Multidimensional Quality Metrics (MQM).

Relevância:

100.00% 100.00%

Publicador:

Resumo:

For more than forty years, research has been on going in the use of the computer in the processing of natural language. During this period methods have evolved, with various parsing techniques and grammars coming to prominence. Problems still exist, not least in the field of Machine Translation. However, one of the successes in this field is the translation of sublanguage. The present work reports Deterministic Parsing, a relatively new parsing technique, and its application to the sublanguage of an aircraft maintenance manual for Machine Translation. The aim has been to investigate the practicability of using Deterministic Parsers in the analysis stage of a Machine Translation system. Machine Translation, Sublanguage and parsing are described in general terms with a review of Deterministic parsing systems, pertinent to this research, being presented in detail. The interaction between machine Translation, Sublanguage and Parsing, including Deterministic parsing, is also highlighted. Two types of Deterministic Parser have been investigated, a Marcus-type parser, based on the basic design of the original Deterministic parser (Marcus, 1980) and an LR-type Deterministic Parser for natural language, based on the LR parsing algorithm. In total, four Deterministic Parsers have been built and are described in the thesis. Two of the Deterministic Parsers are prototypes from which the remaining two parsers to be used on sublanguage have been developed. This thesis reports the results of parsing by the prototypes, a Marcus-type parser and an LR-type parser which have a similar grammatical and linguistic range to the original Marcus parser. The Marcus-type parser uses a grammar of production rules, whereas the LR-type parser employs a Definite Clause Grammar(DGC).

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Word Sense Disambiguation, the process of identifying the meaning of a word in a sentence when the word has multiple meanings, is a critical problem of machine translation. It is generally very difficult to select the correct meaning of a word in a sentence, especially when the syntactical difference between the source and target language is big, e.g., English-Korean machine translation. To achieve a high level of accuracy of noun sense selection in machine translation, we introduced a statistical method based on co-occurrence relation of words in sentences and applied it to the English-Korean machine translator RyongNamSan. ACM Computing Classification System (1998): I.2.7.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

BACKGROUND: Clinical Trials (CTs) are essential for bridging the gap between experimental research on new drugs and their clinical application. Just like CTs for traditional drugs and biologics have helped accelerate the translation of biomedical findings into medical practice, CTs for nanodrugs and nanodevices could advance novel nanomaterials as agents for diagnosis and therapy. Although there is publicly available information about nanomedicine-related CTs, the online archiving of this information is carried out without adhering to criteria that discriminate between studies involving nanomaterials or nanotechnology-based processes (nano), and CTs that do not involve nanotechnology (non-nano). Finding out whether nanodrugs and nanodevices were involved in a study from CT summaries alone is a challenging task. At the time of writing, CTs archived in the well-known online registry ClinicalTrials.gov are not easily told apart as to whether they are nano or non-nano CTs-even when performed by domain experts, due to the lack of both a common definition for nanotechnology and of standards for reporting nanomedical experiments and results. METHODS: We propose a supervised learning approach for classifying CT summaries from ClinicalTrials.gov according to whether they fall into the nano or the non-nano categories. Our method involves several stages: i) extraction and manual annotation of CTs as nano vs. non-nano, ii) pre-processing and automatic classification, and iii) performance evaluation using several state-of-the-art classifiers under different transformations of the original dataset. RESULTS AND CONCLUSIONS: The performance of the best automated classifier closely matches that of experts (AUC over 0.95), suggesting that it is feasible to automatically detect the presence of nanotechnology products in CT summaries with a high degree of accuracy. This can significantly speed up the process of finding whether reports on ClinicalTrials.gov might be relevant to a particular nanoparticle or nanodevice, which is essential to discover any precedents for nanotoxicity events or advantages for targeted drug therapy.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Hospitals attached to the Spanish Ministry of Health are currently using the International Classification of Diseases 9 Clinical Modification (ICD9-CM) to classify health discharge records. Nowadays, this work is manually done by experts. This paper tackles the automatic classification of real Discharge Records in Spanish following the ICD9-CM standard. The challenge is that the Discharge Records are written in spontaneous language. We explore several machine learning techniques to deal with the classification problem. Random Forest resulted in the most competitive one, achieving an F-measure of 0.876.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Le macchine automatiche per il confezionamento sono sistemi complessi composti da molte parti soggette a usura. Con ritmi di produzione di migliaia di pezzi l’ora, l’eventualità di un guasto e del conseguente fermo macchina, può avere ripercussioni economiche considerevoli per un’azienda. Sempre più risorse, per tale motivo, vengono dedicate allo sviluppo di tecniche per incrementare l’affidabilità dei macchinari, e che riducano al contempo, il numero e la durata degli interventi manutentivi. In tal senso, le tecniche di manutenzione predittiva permettono di prevedere, con un certo grado di accuratezza, il tempo di vita residuo dei componenti delle macchine; consentendo l’ottimizzazione della programmazione degli interventi di manutenzione. Appositi sensori installati sui macchinari permettono l’acquisizione di dati, la cui analisi è alla base di tali tecniche. L'elaborato descrive il progetto e la realizzazione di un’infrastruttura software, nell’ambiente di sviluppo Matlab, per l’elaborazione automatizzata di dati di vibrazione. Il sistema proposto, attraverso il monitoraggio dei parametri rms e kurtosis, consiste in strumento di ausilio per la diagnostica di cuscinetti. Il progetto è stato realizzato sulla base di veri dati storici messi a disposizione da un'azienda del settore. L’elaborato affronta inizialmente il problema dell’accuratezza e affidabilità dei campioni, proponendo soluzioni pratiche per la valutazione della qualità e selezione automatica dei dati. Segue la descrizione del processo di estrapolazione dei parametri sopraccitati, nel caso di cuscinetti in moto non stazionario, con profilo di velocità periodico. Nella terza parte è discussa la metodologia con cui istruire il sistema affinché possa, attraverso la redazione di report sullo stato dei componenti, fungere da strumento di diagnosi. L’ultimo capitolo descrive gli strumenti del toolbox per l’analisi dati che è stato realizzato allo scopo di aumentare potenzialità del sistema di diagnostica.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Questo progetto di tesi è parte di un programma più ampio chiamato TIME (Tecnologia Integrata per Mobilità Elettrica) sviluppato tra diversi gruppi di ricerca afferenti al settore meccanico, termofluidodinamico e informatico. TIME si pone l'obiettivo di migliorare la qualità dei componenti di un sistema powertrain presenti oggi sul mercato progettando un sistema general purpose adatto ad essere installato su veicoli di prima fornitura ma soprattutto su retrofit, quindi permettendo il ricondizionamento di veicoli con motore a combustione esistenti ma troppo datati. Lo studio svolto si pone l'obiettivo di identificare tutti gli aspetti di innovazione tecnologica che possono essere installati all'interno del sistema di interazione uomo-macchina. All'interno di questo progetto sarà effettuata una pianificazione di tutto il lavoro del gruppo di ricerca CIRI-ICT, partendo dallo studio normativo ed ergonomico delle interfacce dei veicoli analizzando tutti gli elementi di innovazione che potranno far parte del sistema TIME e quindi programmare tutte le attività previste al fine di raggiungere gli obiettivi prefissati, documentando opportunamente tutto il processo. Nello specifico saranno analizzate e definite le tecniche da utilizzare per poi procedere alla progettazione e implementazione di un primo sistema sperimentale di Machine Learning e Gamification con lo scopo di predire lo stato della batteria in base allo stile di guida dell'utente e incentivare quest'ultimo tramite sistemi di Gamification installati sul cruscotto ad una guida più consapevole dei consumi. Questo sistema sarà testato su dati simulati con l'obiettivo di avere un prodotto configurabile da installare sul veicolo.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

A significant proportion of the cost of software development is due to software testing and maintenance. This is in part the result of the inevitable imperfections due to human error, lack of quality during the design and coding of software, and the increasing need to reduce faults to improve customer satisfaction in a competitive marketplace. Given the cost and importance of removing errors improvements in fault detection and removal can be of significant benefit. The earlier in the development process faults can be found, the less it costs to correct them and the less likely other faults are to develop. This research aims to make the testing process more efficient and effective by identifying those software modules most likely to contain faults, allowing testing efforts to be carefully targeted. This is done with the use of machine learning algorithms which use examples of fault prone and not fault prone modules to develop predictive models of quality. In order to learn the numerical mapping between module and classification, a module is represented in terms of software metrics. A difficulty in this sort of problem is sourcing software engineering data of adequate quality. In this work, data is obtained from two sources, the NASA Metrics Data Program, and the open source Eclipse project. Feature selection before learning is applied, and in this area a number of different feature selection methods are applied to find which work best. Two machine learning algorithms are applied to the data - Naive Bayes and the Support Vector Machine - and predictive results are compared to those of previous efforts and found to be superior on selected data sets and comparable on others. In addition, a new classification method is proposed, Rank Sum, in which a ranking abstraction is laid over bin densities for each class, and a classification is determined based on the sum of ranks over features. A novel extension of this method is also described based on an observed polarising of points by class when rank sum is applied to training data to convert it into 2D rank sum space. SVM is applied to this transformed data to produce models the parameters of which can be set according to trade-off curves to obtain a particular performance trade-off.