903 resultados para Repositórios de software


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Dissertação para obtenção do Grau de Mestre em Engenharia Informática

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Software Repository Mining (MSR) is a research area that analyses software repositories in order to derive relevant information for the research and practice of software engineering. The main goal of repository mining is to extract static information from repositories (e.g. code repository or change requisition system) into valuable information providing a way to support the decision making of software projects. On the other hand, another research area called Process Mining (PM) aims to find the characteristics of the underlying process of business organizations, supporting the process improvement and documentation. Recent works have been doing several analyses through MSR and PM techniques: (i) to investigate the evolution of software projects; (ii) to understand the real underlying process of a project; and (iii) create defect prediction models. However, few research works have been focusing on analyzing the contributions of software developers by means of MSR and PM techniques. In this context, this dissertation proposes the development of two empirical studies of assessment of the contribution of software developers to an open-source and a commercial project using those techniques. The contributions of developers are assessed through three different perspectives: (i) buggy commits; (ii) the size of commits; and (iii) the most important bugs. For the opensource project 12.827 commits and 8.410 bugs have been analyzed while 4.663 commits and 1.898 bugs have been analyzed for the commercial project. Our results indicate that, for the open source project, the developers classified as core developers have contributed with more buggy commits (although they have contributed with the majority of commits), more code to the project (commit size) and more important bugs solved while the results could not indicate differences with statistical significance between developer groups for the commercial project

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The main goal of Regression Test (RT) is to reuse the test suite of the latest version of a software in its current version, in order to maximize the value of the tests already developed and ensure that old features continue working after the new changes. Even with reuse, it is common that not all tests need to be executed again. Because of that, it is encouraged to use Regression Tests Selection (RTS) techniques, which aims to select from all tests, only those that reveal faults, this reduces costs and makes this an interesting practice for the testing teams. Several recent research works evaluate the quality of the selections performed by RTS techniques, identifying which one presents the best results, measured by metrics such as inclusion and precision. The RTS techniques should seek in the System Under Test (SUT) for tests that reveal faults. However, because this is a problem without a viable solution, they alternatively seek for tests that reveal changes, where faults may occur. Nevertheless, these changes may modify the execution flow of the algorithm itself, leading some tests no longer exercise the same stretch. In this context, this dissertation investigates whether changes performed in a SUT would affect the quality of the selection of tests performed by an RTS, if so, which features the changes present which cause errors, leading the RTS to include or exclude tests wrongly. For this purpose, a tool was developed using the Java language to automate the measurement of inclusion and precision averages achieved by a regression test selection technique for a particular feature of change. In order to validate this tool, an empirical study was conducted to evaluate the RTS technique Pythia, based on textual differencing, on a large web information system, analyzing the feature of types of tasks performed to evolve the SUT

Relevância:

70.00% 70.00%

Publicador:

Resumo:

Software product line engineering promotes large software reuse by developing a system family that shares a set of developed core features, and enables the selection and customization of a set of variabilities that distinguish each software product family from the others. In order to address the time-to-market, the software industry has been using the clone-and-own technique to create and manage new software products or product lines. Despite its advantages, the clone-and-own approach brings several difficulties for the evolution and reconciliation of the software product lines, especially because of the code conflicts generated by the simultaneous evolution of the original software product line, called Source, and its cloned products, called Target. This thesis proposes an approach to evolve and reconcile cloned products based on mining software repositories and code conflict analysis techniques. The approach provides support to the identification of different kinds of code conflicts – lexical, structural and semantics – that can occur during development task integration – bug correction, enhancements and new use cases – from the original evolved software product line to the cloned product line. We have also conducted an empirical study of characterization of the code conflicts produced during the evolution and merging of two large-scale web information system product lines. The results of our study demonstrate the approach potential to automatically or semi-automatically solve several existing code conflicts thus contributing to reduce the complexity and costs of the reconciliation of cloned software product lines.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

Soft skills and teamwork practices were identi ed as the main de ciencies of recent graduates in computer courses. This issue led to a realization of a qualitative research aimed at investigating the challenges faced by professors of those courses in conducting, monitoring and assessing collaborative software development projects. Di erent challenges were reported by teachers, including di culties in the assessment of students both in the collective and individual levels. In this context, a quantitative research was conducted with the aim to map soft skill of students to a set of indicators that can be extracted from software repositories using data mining techniques. These indicators are aimed at measuring soft skills, such as teamwork, leadership, problem solving and the pace of communication. Then, a peer assessment approach was applied in a collaborative software development course of the software engineering major at the Federal University of Rio Grande do Norte (UFRN). This research presents a correlation study between the students' soft skills scores and indicators based on mining software repositories. This study contributes: (i) in the presentation of professors' perception of the di culties and opportunities for improving management and monitoring practices in collaborative software development projects; (ii) in investigating relationships between soft skills and activities performed by students using software repositories; (iii) in encouraging the development of soft skills and the use of software repositories among software engineering students; (iv) in contributing to the state of the art of three important areas of software engineering, namely software engineering education, educational data mining and human aspects of software engineering.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

Soft skills and teamwork practices were identi ed as the main de ciencies of recent graduates in computer courses. This issue led to a realization of a qualitative research aimed at investigating the challenges faced by professors of those courses in conducting, monitoring and assessing collaborative software development projects. Di erent challenges were reported by teachers, including di culties in the assessment of students both in the collective and individual levels. In this context, a quantitative research was conducted with the aim to map soft skill of students to a set of indicators that can be extracted from software repositories using data mining techniques. These indicators are aimed at measuring soft skills, such as teamwork, leadership, problem solving and the pace of communication. Then, a peer assessment approach was applied in a collaborative software development course of the software engineering major at the Federal University of Rio Grande do Norte (UFRN). This research presents a correlation study between the students' soft skills scores and indicators based on mining software repositories. This study contributes: (i) in the presentation of professors' perception of the di culties and opportunities for improving management and monitoring practices in collaborative software development projects; (ii) in investigating relationships between soft skills and activities performed by students using software repositories; (iii) in encouraging the development of soft skills and the use of software repositories among software engineering students; (iv) in contributing to the state of the art of three important areas of software engineering, namely software engineering education, educational data mining and human aspects of software engineering.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Trabalho apresentado no âmbito do Mestrado em Engenharia Informática, como requisito parcial para obtenção do grau de Mestre em Engenharia Informática

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A capacidade de inovação é essencial à sobrevivência de empresas criativas de software. Assim, a criação e disseminação do conhecimento encontram cada vez mais espaço no planejamento e na estratégia de gestão dessas organizações. Este estudo, de natureza exploratória, efetuou entrevistas semi-estruturadas com profissionais gestores e técnicos da indústria criativa de software, tendo por objetivo analisar as técnicas e métodos presentes no processo de geração e disseminação do conhecimento, bem como as ferramentas de tecnologia da informação voltadas a essa finalidade. Os principais resultados mostram que há uma preocupação constante nessas organizações com a geração do conhecimento. As técnicas e métodos mais largamente empregados na criação e disseminação de conhecimento foram: brainstorming, reuniões presenciais, reuniões virtuais, treinamentos oficiais presenciais, treinamentos informais, cursos em universidades, compartilhamento de experiências, discussões e aplicativo de gestão de projetos de desenvolvimento de software. Já em relação às ferramentas utilizadas, somam-se: internet, intranet, aplicativos específicos, redes internas, repositórios de dados, diretórios específicos no banco de dados e sistemas de telefonia baseados na internet.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A great challenge of the Component Based Development is the creation of mechanisms to facilitate the finding of reusable assets that fulfill the requirements of a particular system under development. In this sense, some component repositories have been proposed in order to answer such a need. However, repositories need to represent the asset characteristics that can be taken into account by the consumers when choosing the more adequate assets for their needs. In such a context, the literature presents some models proposed to describe the asset characteristics, such as identification, classification, non-functional requirements, usage and deployment information and component interfaces. Nevertheless, the set of characteristics represented by those models is insufficient to describe information used before, during and after the asset acquisition. This information refers to negotiation, certification, change history, adopted development process, events, exceptions and so on. In order to overcome this gap, this work proposes an XML-based model to represent several characteristics, of different asset types, that may be employed in the component-based development. Besides representing metadata used by consumers, useful for asset discovering, acquisition and usage, this model, called X-ARM, also focus on helping asset developers activities. Since the proposed model represents an expressive amount of information, this work also presents a tool called X-Packager, developed with the goal of helping asset description with X-ARM

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Pós-graduação em Ciência da Computação - IBILCE

Relevância:

30.00% 30.00%

Publicador:

Resumo:

O principal objetivo deste artigo é contribuir com a adoção do modelo de referência MPS para Software (MPS-SW). Para isso, é proposta uma ontologia dos níveis G e F do modelo, visando apoiar a compreensão e implementação desses níveis. O modelo MPS-SW é adequado a micro, pequenas e médias empresas (mPME), assim como a empresas de grande porte. A proposta pretende mitigar os investimentos técnicos e financeiros na implantação e treinamento do modelo, principalmente nas mPME. Neste contexto, este artigo apresenta a metodologia utilizada para o desenvolvimento da ontologia, com a inclusão de conhecimentos de especialistas no modelo, conceitos do PMBOK (Project Management Body of Knowledge) e indicadores do BSC (Balanced Scorecard). A partir de testes centrados no usuário, foi gerada uma versão beta da ontologia, que está disponível em repositórios livres e gratuitos.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

O cenário atual de repositórios digitais mundialmente distribuídos estimula estudos diversificados com os quais esse trabalho visa contribuir, objetivando um levantamento dos repositórios de instituições de ensino superior no Brasil, verificando a eficácia de uma ferramenta experimental no tratamento e análise dos dados e usando como fontes os diretórios Registry of Open Access Repositories (ROAR), Directory of Open Access Repositories (OpenDOAR), Diretório Luso-Brasileiro de Periódicos e Repositórios de Acesso Livre e a lista L_repositories. A ferramenta experimental Google Fusion Tables foi aplicada nos dados dos repositórios institucionais pesquisados, categorizando suas principais características: Instituição mantenedora, Natureza da instituição, Local, Região geográfica, Software adotado e sua versão, adoção do padrão Dublin Core e quantidade de trabalhos disponibilizados na data do estudo. Foram identificados 49 repositórios que em agosto de 2013 disponibilizavam 396.881 itens, sendo as instituições federais as com maior povoamento e o repositório LUME o primeiro em volume de itens; a região Sudeste com o maior número de repositórios e volume de itens disponibilizados; o DSpace o software predominante, com maior utilização da versão 1.6.2 e o padrão de metadados Dublin Core em todas as aplicações desse software. Este estudo comprovou a eficácia e utilidade do FusionTables, permitindo caracterizar o panorama atual de repositórios de instituições de ensino superior no Brasil. Os resultados foram disponibilizados em um Catálogo de Repositórios de Instituições de Ensino Superior no Brasil e um Mapa interativo dos Repositórios de Instituições de Ensino Superior no Brasil.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This research deals with the use of a participatory design methodology to develop a repository of open educational resources, the Arcaz. Discusses key aspects of neutrality and determinism of technology within the context of Social Studies of Science and Technology and presents some concepts of critical theory of technology related to the democratic construction of technological artifacts. Discusses the philosophical heritage of the movements that led to the emergence of free software, open education and open educational resources and argues that participatory design share similar ideals. It presents concepts of human-computer interaction, interaction design and user centered design, important to enhance the user experience in information systems. It addresses the participatory design as a methodology that allows the democratic participation of users in the technological construction, promoting mutual learning and active voice for the participants. Develops a participatory design methodology adapted to the Arcaz context of use and provides the procedures for the meetings conducted to apply participatory design techniques to the repository and the results obtained. It concludes with a study of some of the interventions suggested in the system and orientations for future applications of participatory practices in the development of the repository and a list of best practices, focusing on ethical principles that should guide the participatory design.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This article aimed at comparing the accuracy of linear measurement tools of different commercial software packages. Eight fully edentulous dry mandibles were selected for this study. Incisor, canine, premolar, first molar and second molar regions were selected. Cone beam computed tomography (CBCT) images were obtained with i-CAT Next Generation. Linear bone measurements were performed by one observer on the cross-sectional images using three different software packages: XoranCat®, OnDemand3D® and KDIS3D®, all able to assess DICOM images. In addition, 25% of the sample was reevaluated for the purpose of reproducibility. The mandibles were sectioned to obtain the gold standard for each region. Intraclass coefficients (ICC) were calculated to examine the agreement between the two periods of evaluation; the one-way analysis of variance performed with the post-hoc Dunnett test was used to compare each of the software-derived measurements with the gold standard. The ICC values were excellent for all software packages. The least difference between the software-derived measurements and the gold standard was obtained with the OnDemand3D and KDIS3D (-0.11 and -0.14 mm, respectively), and the greatest, with the XoranCAT (+0.25 mm). However, there was no statistical significant difference between the measurements obtained with the different software packages and the gold standard (p> 0.05). In conclusion, linear bone measurements were not influenced by the software package used to reconstruct the image from CBCT DICOM data.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper presents SMarty, a variability management approach for UML-based software product lines (PL). SMarty is supported by a UML profile, the SMartyProfile, and a process for managing variabilities, the SMartyProcess. SMartyProfile aims at representing variabilities, variation points, and variants in UML models by applying a set of stereotypes. SMartyProcess consists of a set of activities that is systematically executed to trace, identify, and control variabilities in a PL based on SMarty. It also identifies variability implementation mechanisms and analyzes specific product configurations. In addition, a more comprehensive application of SMarty is presented using SEI's Arcade Game Maker PL. An evaluation of SMarty and related work are discussed.