906 resultados para software creation methodology


Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this article we describe the methodology developed for the semiautomatic annotation of EPEC-RolSem, a Basque corpus labeled at predicate level following the PropBank-VerbNet model. The methodology presented is the product of detailed theoretical study of the semantic nature of verbs in Basque and of their similarities and differences with verbs in other languages. As part of the proposed methodology, we are creating a Basque lexicon on the PropBank-VerbNet model that we have named the Basque Verb Index (BVI). Our work thus dovetails the general trend toward building lexicons from tagged corpora that is clear in work conducted for other languages. EPEC-RolSem and BVI are two important resources for the computational semantic processing of Basque; as far as the authors are aware, they are also the first resources of their kind developed for Basque. In addition, each entry in BVI is linked to the corresponding verb-entry in well-known resources like PropBank, VerbNet, WordNet, Levin’s Classification and FrameNet. We have also implemented several automatic processes to aid in creating and annotating the BVI, including processes designed to facilitate the task of manual annotation.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

ADMB2R is a collection of AD Model Builder routines for saving complex data structures into a file that can be read in the R statistics environment with a single command.1 ADMB2R provides both the means to transfer data structures significantly more complex than simple tables, and an archive mechanism to store data for future reference. We developed this software because we write and run computationally intensive numerical models in Fortran, C++, and AD Model Builder. We then analyse results with R. We desired to automate data transfer to speed diagnostics during working-group meetings. We thus developed the ADMB2R interface to write an R data object (of type list) to a plain-text file. The master list can contain any number of matrices, values, dataframes, vectors or lists, all of which can be read into R with a single call to the dget function. This allows easy transfer of structured data from compiled models to R. Having the capacity to transfer model data, metadata, and results has sharply reduced the time spent on diagnostics, and at the same time, our diagnostic capabilities have improved tremendously. The simplicity of this interface and the capabilities of R have enabled us to automate graph and table creation for formal reports. Finally, the persistent storage in files makes it easier to treat model results in analyses or meta-analyses devised months—or even years—later. We offer ADMB2R to others in the hope that they will find it useful. (PDF contains 30 pages)

Relevância:

20.00% 20.00%

Publicador:

Resumo:

C2R is a collection of C routines for saving complex data structures into a file that can be read in the R statistics environment with a single command.1 C2R provides both the means to transfer data structures significantly more complex than simple tables, and an archive mechanism to store data for future reference. We developed this software because we write and run computationally intensive numerical models in Fortran, C++, and AD Model Builder. We then analyse results with R. We desired to automate data transfer to speed diagnostics during working-group meetings. We thus developed the C2R interface to write an R data object (of type list) to a plain-text file. The master list can contain any number of matrices, values, dataframes, vectors or lists, all of which can be read into R with a single call to the dget function. This allows easy transfer of structured data from compiled models to R. Having the capacity to transfer model data, metadata, and results has sharply reduced the time spent on diagnostics, and at the same time, our diagnostic capabilities have improved tremendously. The simplicity of this interface and the capabilities of R have enabled us to automate graph and table creation for formal reports. Finally, the persistent storage in files makes it easier to treat model results in analyses or meta-analyses devised months—or even years—later. We offer C2R to others in the hope that they will find it useful. (PDF contains 27 pages)

Relevância:

20.00% 20.00%

Publicador:

Resumo:

For2R is a collection of Fortran routines for saving complex data structures into a file that can be read in the R statistics environment with a single command.1 For2R provides both the means to transfer data structures significantly more complex than simple tables, and an archive mechanism to store data for future reference. We developed this software because we write and run computationally intensive numerical models in Fortran, C++, and AD Model Builder. We then analyse results with R. We desired to automate data transfer to speed diagnostics during working-group meetings. We thus developed the For2R interface to write an R data object (of type list) to a plain-text file. The master list can contain any number of matrices, values, dataframes, vectors or lists, all of which can be read into R with a single call to the dget function. This allows easy transfer of structured data from compiled models to R. Having the capacity to transfer model data, metadata, and results has sharply reduced the time spent on diagnostics, and at the same time, our diagnostic capabilities have improved tremendously. The simplicity of this interface and the capabilities of R have enabled us to automate graph and table creation for formal reports. Finally, the persistent storage in files makes it easier to treat model results in analyses or meta-analyses devised months—or even years—later. We offer For2R to others in the hope that they will find it useful. (PDF contains 31 pages)

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Background Protein inference from peptide identifications in shotgun proteomics must deal with ambiguities that arise due to the presence of peptides shared between different proteins, which is common in higher eukaryotes. Recently data independent acquisition (DIA) approaches have emerged as an alternative to the traditional data dependent acquisition (DDA) in shotgun proteomics experiments. MSE is the term used to name one of the DIA approaches used in QTOF instruments. MSE data require specialized software to process acquired spectra and to perform peptide and protein identifications. However the software available at the moment does not group the identified proteins in a transparent way by taking into account peptide evidence categories. Furthermore the inspection, comparison and report of the obtained results require tedious manual intervention. Here we report a software tool to address these limitations for MSE data. Results In this paper we present PAnalyzer, a software tool focused on the protein inference process of shotgun proteomics. Our approach considers all the identified proteins and groups them when necessary indicating their confidence using different evidence categories. PAnalyzer can read protein identification files in the XML output format of the ProteinLynx Global Server (PLGS) software provided by Waters Corporation for their MSE data, and also in the mzIdentML format recently standardized by HUPO-PSI. Multiple files can also be read simultaneously and are considered as technical replicates. Results are saved to CSV, HTML and mzIdentML (in the case of a single mzIdentML input file) files. An MSE analysis of a real sample is presented to compare the results of PAnalyzer and ProteinLynx Global Server. Conclusions We present a software tool to deal with the ambiguities that arise in the protein inference process. Key contributions are support for MSE data analysis by ProteinLynx Global Server and technical replicates integration. PAnalyzer is an easy to use multiplatform and free software tool.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Contém uma síntese de ideias sobre Brasília e o Congresso Nacional, com foco na Câmara dos Deputados. Tem por intenção explorar as relações entre a cidade e o edifício, de modo a discutir sobre como ambos estão estreitamente associados, a ponto de ser difícil imaginar um apartado do outro.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The report introduces software sustainability, provides definitions, clearly demonstrates that software is not the same as data and illustrates aspects of sustainability in the software lifecycle. The recommendations state that improving software sustainability requires a number of changes: some technical and others societal, some small and others significant. We must start by raising awareness of researchers’ reliance on software. This goal will become easier if we recognise the valuable contribution that software makes to research – and reward those people who invest their time into developing reliable and reproducible software. The adoption of software has led to significant advances in research. But if we do not change our research practices, the continued rise in software use will be accompanied by a rise in retractions. Ultimately, anyone who is concerned about the reliability and reproducibility of research should be concerned about software sustainability. Beside highlighting the benefits of software sustainability and addressing the societal and technical barriers to software sustainability, the report provides access to expertise in software sustainability and outlines the role of funders. The report concludes with a short landscape of national activities in Europe and outside Europe. As a result of the workshop steps will be explored to establish European coordination and cooperation of national initiatives.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

El Cine Digital es aquel que utiliza la tecnología digital para grabar, distribuir y proyectar películas. En los años 90, el cine comenzó un proceso de transición, del soporte fílmico a la tecnología digital. Pero el salto definitivo se dio con las grandes superproducciones de principios del 2000. Con este cambio de era y debido a la rápida difusión del digital y la proliferación de formatos se creo el DCI (Digital Cinema Initiative), para cambiar el modo en que las personas consumen cine. Trabajando junto con los miembros del comité SMPTE (Organización americana encargada de crear los estándares de la industria audiovisual formada por ingenieros, técnicos y fabricantes) publicó un sistema de especificaciones que han adoptado las mayores productoras estadounidenses. Mediante este acuerdo, aseguraban la calidad técnica de las producciones, la compatibilidad entre sistemas y como no, su hegemonía particular. Entre las especificaciones técnicas que suscribieron que son la base actual del DCI figuran la resolución de fotograma, el espacio de color, la compresión de imagen, la encriptación y el método de empaquetado de archivos. Y hoy en día son un estándar en la masterización, distribución y en la proyección final en las salas de cine. Una de las grandes esperanzas que hay puestas en la tecnología digital es la democratización en el mundo del cine y la supuesta abolición de las barreras económicas a la hora de realizar películas, dado lo barato que puede resultar la grabación digital y la posibilidad de pasar el material a video y editarlo en un ordenador domestico. Independientemente de los sistemas de edición que se utilicen, el formato de archivo o incluso el códec usado, los servidores de cine digital solo aceptan un tipo de archivo llamado DCP (Digital Cinema Package). Hay que aclarar que el DCP es abierto, documentado y que esta basado en los estándares SMPTE. Por lo tanto, existen en el mercado actual herramientas de software libre que permiten crear un DCP válido según las normas del DCI y compatible con los actuales servidores de cine digital. El propósito de este proyecto principalmente es documentar desde un punto de vista técnico la creación de un archivo DCP y analizar las diferentes herramientas existentes en el mercado para poder realizarlo: tanto las de uso comercial, como las de software libre. Como base se partirá de las aplicaciones creadas por la empresa alemana Fraunhofer (EasyDCP Creator, Player) - es el software utilizado en el estudio REC - y de otras herramientas más rudimentarias y escritas en C++ como (asdcplib) de Cinecert.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Within the next few pages, I will try to give a wide description of the project that I have been doing for IK4-Ikerlan. For the last six months, I have been working in developing a socket-based application for Apple devices. These devices work under the iOS operative system, which is programmed in Objective-C, a language similar to C. Although I did not have the chance to develop this application for Apple TV, I was able to create an application for iPhone and another one for iPad. The only difference between both applications was the screen resolution, but we decided to make them separately, as it would be really hard to combine both resolutions, and wallpapers, everything in the same workspace. Finally, it is necessary to add that the main goal was not to create a new application for iOS, but to translate an Android application into iOS. To achieve this, it is required to translate Java code into Objective- C, which is the language used to develop applications for all kinds of Apple devices. Fortunately, there is a tool created by Google, which helped us with this exercise. This tool is called j2ObjC, and it is still being developed.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

La teleoperación o telerobótica es un campo de la robótica que se basa en el control remoto de robots esclavo por parte de un usuario encargado de gobernar, mediante un dispositivo maestro, la fuerza y movimiento del robot. Sobre dicho usuario recaen también las tareas de percepción del entorno, planificación y manipulación compleja. Concretamente se pretende desarrollar el control software necesario para teleoperar un manipulador esclavo, Kuka Lightweigh mediante un dispositivo háptico Phamton Omni, que se comporta como maestro, sin que afecten las diferencias dinámicas y estructurales existentes entre ambos dispositivos, aportando información adicional al operador para facilitar la operación. La principal motivación de la evolución de esta tecnología se debe a la necesidad de realizar trabajos en entornos hostiles, de difícil acceso, o perjudiciales para la salud del usuario.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Duración (en horas): Más de 50 horas. Destinatario: Estudiante y Docente