942 resultados para software, translation, validation tool, VMNET, Wikipedia, XML


Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper will focus on some aspects of translation based on blending distinct linguistic domains such as English Language and Portuguese in using false friends in the English class in tertiary level students, reflecting namely on: 1. the choice of a word suitable to the context in L2 ; 2. the difficulties encountered by choice of that word that could be misleading, by relying in a false L1 reality that is going to adulterate reality in the L2 domain; 3. the difficulty in making such type of distinctions due to the lack of linguistic and lexical knowledge. 4. the need to study the cause of these difficulties by working, not only with their peers, but also with their language teacher to develop strategies to diminish and if possible to eradicate this type of linguistic and, above all, translation problem by making an inventory of those types of mistakes. In relation to the first point it is necessary to know that translation tasks involve much more than literal concepts ( Ladmiral, 1975) : furthermore it is necessary and suitable to realise that lexicon relies in significant contexts (Coseriu 1966), which connects both domains, that, at first sight do not seem to be compatible. In other words, although students have the impression they dominate lexicon due to the fact that they possess at least seven years of foreign language exposure that doesn’t mean they master the particularities engaged in such a delicate task as translation is concerned. There are some chromaticisms in the words (false friends), that need to be researched and analysed later on by both students and language teachers. The reason for such state of affairs lies in their academic formation, of a mainly general stream, which has enabled them only for knowledge of the foreign language, but not for the translation as a tool as it is required only when they reach the tertiary level. Besides, for their translations they rely, most of the times, on glossaries, whose dominant language is portuguese of Brazil, which is, obviously, much different from the portuguese mother tongue reality and even more of English. So it seems necessary to use with caution the working tools (glossaries) that work as surpluses, but could bring translation problems as we will see.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The design and development of simulation models and tools for Demand Response (DR) programs are becoming more and more important for adequately taking the maximum advantages of DR programs use. Moreover, a more active consumers’ participation in DR programs can help improving the system reliability and decrease or defer the required investments. DemSi, a DR simulator, designed and implemented by the authors of this paper, allows studying DR actions and schemes in distribution networks. It undertakes the technical validation of the solution using realistic network simulation based on PSCAD. DemSi considers the players involved in DR actions, and the results can be analyzed from each specific player point of view.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this paper we present VERITAS, a tool that focus time maintenance, that is one of the most important processes in the engineering of the time during the development of KBS. The verification and validation (V&V) process is part of a wider process denominated knowledge maintenance, in which an enterprise systematically gathers, organizes, shares, and analyzes knowledge to accomplish its goals and mission. The V&V process states if the software requirements specifications have been correctly and completely fulfilled. The methodologies proposed in software engineering have showed to be inadequate for Knowledge Based Systems (KBS) validation and verification, since KBS present some particular characteristics. VERITAS is an automatic tool developed for KBS verification which is able to detect a large number of knowledge anomalies. It addresses many relevant aspects considered in real applications, like the usage of rule triggering selection mechanisms and temporal reasoning.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

CoDeSys "Controller Development Systems" is a development environment for programming in the area of automation controllers. It is an open source solution completely in line with the international industrial standard IEC 61131-3. All five programming languages for application programming as defined in IEC 61131-3 are available in the development environment. These features give professionals greater flexibility with regard to programming and allow control engineers have the ability to program for many different applications in the languages in which they feel most comfortable. Over 200 manufacturers of devices from different industrial sectors offer intelligent automation devices with a CoDeSys programming interface. In 2006, version 3 was released with new updates and tools. One of the great innovations of the new version of CoDeSys is object oriented programming. Object oriented programming (OOP) offers great advantages to the user for example when wanting to reuse existing parts of the application or when working on one application with several developers. For this reuse can be prepared a source code with several well known parts and this is automatically generated where necessary in a project, users can improve then the time/cost/quality management. Until now in version 2 it was necessary to have hardware interface called “Eni-Server” to have access to the generated XML code. Another of the novelties of the new version is a tool called Export PLCopenXML. This tool makes it possible to export the open XML code without the need of specific hardware. This type of code has own requisites to be able to comply with the standard described above. With XML code and with the knowledge how it works it is possible to do component-oriented development of machines with modular programming in an easy way. Eplan Engineering Center (EEC) is a software tool developed by Mind8 GmbH & Co. KG that allows configuring and generating automation projects. Therefore it uses modules of PLC code. The EEC already has a library to generate code for CoDeSys version 2. For version 3 and the constant innovation of drivers by manufacturers, it is necessary to implement a new library in this software. Therefore it is important to study the XML export to be then able to design any type of machine. The purpose of this master thesis is to study the new version of the CoDeSys XML taking into account all aspects and impact on the existing CoDeSys V2 models and libraries in the company Harro Höfliger Verpackungsmaschinen GmbH. For achieve this goal a small sample named “Traffic light” in CoDeSys version 2 will be done and then, using the tools of the new version it there will be a project with version 3 and also the EEC implementation for the automatically generated code.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Neonatal anthropometry is an inexpensive, noninvasive and convenient tool for bedside evaluation, especially in sick and fragile neonates. Anthropometry can be used in neonates as a tool for several purposes: diagnosis of foetal malnutrition and prediction of early postnatal complications; postnatal assessment of growth, body composition and nutritional status; prediction of long-term complications including metabolic syndrome; assessment of dysmorphology; and estimation of body surface. However, in this age group anthropometry has been notorious for its inaccuracy and the main concern is to make validated indices available. Direct measurements, such as body weight, length and body circumferences are the most commonly used measurements for nutritional assessment in clinical practice and in field studies. Body weight is the most reliable anthropometric measurement and therefore is often used alone in the assessment of the nutritional status, despite not reflecting body composition. Derived indices from direct measurements have been proposed to improve the accuracy of anthropometry. Equations based on body weight and length, mid-arm circumference/head circumference ratio, and upper-arm cross-sectional areas are among the most used derived indices to assess nutritional status and body proportionality, even though these indices require further validation for the estimation of body composition in neonates.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper presents the SmartClean tool. The purpose of this tool is to detect and correct the data quality problems (DQPs). Compared with existing tools, SmartClean has the following main advantage: the user does not need to specify the execution sequence of the data cleaning operations. For that, an execution sequence was developed. The problems are manipulated (i.e., detected and corrected) following that sequence. The sequence also supports the incremental execution of the operations. In this paper, the underlying architecture of the tool is presented and its components are described in detail. The tool's validity and, consequently, of the architecture is demonstrated through the presentation of a case study. Although SmartClean has cleaning capabilities in all other levels, in this paper are only described those related with the attribute value level.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Background: A common task in analyzing microarray data is to determine which genes are differentially expressed across two (or more) kind of tissue samples or samples submitted under experimental conditions. Several statistical methods have been proposed to accomplish this goal, generally based on measures of distance between classes. It is well known that biological samples are heterogeneous because of factors such as molecular subtypes or genetic background that are often unknown to the experimenter. For instance, in experiments which involve molecular classification of tumors it is important to identify significant subtypes of cancer. Bimodal or multimodal distributions often reflect the presence of subsamples mixtures. Consequently, there can be genes differentially expressed on sample subgroups which are missed if usual statistical approaches are used. In this paper we propose a new graphical tool which not only identifies genes with up and down regulations, but also genes with differential expression in different subclasses, that are usually missed if current statistical methods are used. This tool is based on two measures of distance between samples, namely the overlapping coefficient (OVL) between two densities and the area under the receiver operating characteristic (ROC) curve. The methodology proposed here was implemented in the open-source R software. Results: This method was applied to a publicly available dataset, as well as to a simulated dataset. We compared our results with the ones obtained using some of the standard methods for detecting differentially expressed genes, namely Welch t-statistic, fold change (FC), rank products (RP), average difference (AD), weighted average difference (WAD), moderated t-statistic (modT), intensity-based moderated t-statistic (ibmT), significance analysis of microarrays (samT) and area under the ROC curve (AUC). On both datasets all differentially expressed genes with bimodal or multimodal distributions were not selected by all standard selection procedures. We also compared our results with (i) area between ROC curve and rising area (ABCR) and (ii) the test for not proper ROC curves (TNRC). We found our methodology more comprehensive, because it detects both bimodal and multimodal distributions and different variances can be considered on both samples. Another advantage of our method is that we can analyze graphically the behavior of different kinds of differentially expressed genes. Conclusion: Our results indicate that the arrow plot represents a new flexible and useful tool for the analysis of gene expression profiles from microarrays.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Mestrado em Engenharia Electrotécnica – Sistemas Eléctricos de Energia

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Introduction Myocardial Perfusion Imaging (MPI) is a very important tool in the assessment of Coronary Artery Disease ( CAD ) patient s and worldwide data demonstrate an increasingly wider use and clinical acceptance. Nevertheless, it is a complex process and it is quite vulnerable concerning the amount and type of possible artefacts, some of them affecting seriously the overall quality and the clinical utility of the obtained data. One of the most in convenient artefacts , but relatively frequent ( 20% of the cases ) , is relate d with patient motion during image acquisition . Mostly, in those situations, specific data is evaluated and a decisi on is made between A) accept the results as they are , consider ing that t he “noise” so introduced does not affect too seriously the final clinical information, or B) to repeat the acquisition process . Another possib ility could be to use the “ Motion Correcti on Software” provided within the software package included in any actual gamma camera. The aim of this study is to compare the quality of the final images , obtained after the application of motion correction software and after the repetition of image acqui sition. Material and Methods Thirty cases of MPI affected by Motion Artefacts and repeated , were used. A group of three, independent (blinded for the differences of origin) expert Nuclear Medicine Clinicians had been invited to evaluate the 30 sets of thre e images - one set for each patient - being ( A) original image , motion uncorrected , (B) original image, motion corrected, and (C) second acquisition image, without motion . The results so obtained were statistically analysed . Results and Conclusion Results obtained demonstrate that the use of the Motion Correction Software is useful essentiall y if the amplitude of movement is not too important (with this specific quantification found hard to define precisely , due to discrepancies between clinicians and other factors , namely between one to another brand); when that is not the case and the amplitude of movement is too important , the n the percentage of agreement between clinicians is much higher and the repetition of the examination is unanimously considered ind ispensable.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Tese de Doutoramento, Ciências do Mar (Biologia Marinha)

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Background: The eukaryotic release factor 3 (eRF3) has been shown to affect both tubulin and actin cytoskeleton, suggesting a role in cytoskeleton assembly, mitotic spindle formation and chromosome segregation. Also, direct interactions between eRF3 and subunits of the cytosolic chaperonin CCT have been described. Moreover, both eRF3a and CCT subunits have been described to be up-regulated in cancer tissues. Our aim was to evaluate the hypothesis that eRF3 expression levels are correlated with the expression of genes encoding proteins involved in the tubulin folding pathways. Methods: Relative expression levels of eRF1, eRF3a/GSPT1, PFDN4, CCT2, CCT4, and TBCA genes in tumour samples relative to their adjacent normal tissues were investigated using real time-polymerase chain reaction in 20 gastric cancer patients. Results: The expression levels of eRF3a/GSPT1 were not correlated with the expression levels of the other genes studied. However, significant correlations were detected between the other genes, both within intestinal and diffuse type tumours. Conclusions: eRF3a/GSPT1 expression at the mRNA level is independent from both cell translation rates and from the expression of the genes involved in tubulin-folding pathways. The differences in the patterns of expression of the genes studied support the hypothesis of genetically independent pathways in the origin of intestinal and diffuse type gastric tumours.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The histone deacetylase inhibitors sodium butyrate (NaBu) and trichostatin A (TSA) exhibit anti-proliferative activity by causing cell cycle arrest and apoptosis. The mechanisms by which NaBu and TSA cause apoptosis and cell cycle arrest are not yet completely clarified, although these agents are known to modulate the expression of several genes including cell-cycle- and apoptosis-related genes. The enzymes involved in the process of translation have important roles in controlling cell growth and apoptosis, and several of these translation factors have been described as having a causal role in the development of cancer. The expression patterns of the translation mechanism, namely of the elongation factors eEF1A1 and eEF1A2, and of the termination factors eRF1 and eRF3, were studied in the breast cancer cell line MCF-7 by real-time quantitative reverse transcription-polymerase chain reaction after a 24-h treatment with NaBu and TSA. NaBu induced inhibition of translation factors' transcription, whereas TSA caused an increase in mRNA levels. Thus, these two agents may modulate the expression of translation factors through different pathways. We propose that the inhibition caused by NaBu may, in part, be responsible for the cell cycle arrest and apoptosis induced by this agent in MCF-7 cells.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Dissertação de natureza científica para obtenção do grau de Mestre em Engenharia Informática e de Computadores

Relevância:

30.00% 30.00%

Publicador:

Resumo:

O aumento da população Mundial, particularmente em Países emergentes como é o caso da China e da Índia, tem-se relevado um problema adicional no que confere às dificuldades associadas ao consumo mundial de energia, pois esta situação limita inequivocamente o acesso destes milhões de pessoas à energia eléctrica para os bens básicos de sobrevivência. Uma das muitas formas de se extinguir esta necessidade, começa a ser desenvolvida recorrendo ao uso de recursos renováveis como fontes de energia. Independentemente do local do mundo onde nos encontremos, essas fontes de energia são abundantes, inesgotáveis e gratuitas. O problema reside na forma como esses recursos renováveis são geridos em função das solicitações de carga que as instalações necessitam. Sistemas híbridos podem ser usados para produzir energia em qualquer parte do mundo. Historicamente este tipo de sistemas eram aplicados em locais isolados, mas nos dias que correm podem ser usados directamente conectados à rede, permitindo que se realize a venda de energia. Foi neste contexto que esta tese foi desenvolvida, com o objectivo de disponibilizar uma ferramenta informática capaz de calcular a rentabilidade de um sistema híbrido ligado à rede ou isolado. Contudo, a complexidade deste problema é muito elevada, pois existe uma extensa panóplia de características e distintos equipamentos que se pode adoptar. Assim, a aplicação informática desenvolvida teve de ser limitada e restringida aos dados disponíveis de forma a poder tornar-se genérica, mas ao mesmo tempo permitir ter uma aplicabilidade prática. O objectivo da ferramenta informática desenvolvida é apresentar de forma imediata os custos da implementação que um sistema híbrido pode acarretar, dependendo apenas de três variáveis distintas. A primeira variável terá de ter em consideração o local de instalação do sistema. Em segundo lugar é o tipo de ligação (isolado ou ligado à rede) e, por fim, o custo dos equipamentos (eólico, solar e restantes componentes) que serão introduzidos. Após a inserção destes dados a aplicação informática apresenta valores estimados de Payback e VAL.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Dissertação apresentada para a obtenção do Grau de Doutor em Informática pela Universidade Nova de Lisboa, Faculdade de Ciências e Tecnologia