931 resultados para Software Tools


Relevância:

70.00% 70.00%

Publicador:

Resumo:

This paper makes a contribution in bridging the theory and practice of the polyhedral model for designing parallel algorithms. Although the theory of polyhedral model is well developed, designers of massively parallel algorithms are unable to benefit from the theory due to the lack of software tools that incorporate the wide range of transformations that are possible in the model. The Uniformization tool that we developed was the first to integrate a number of techniques and to completely automate the transformation step allowing designers to explore a wide range of feasible designs from high-level specifications.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

A model based on graph isomorphisms is used to formalize software evolution. Step by step we narrow the search space by an informed selection of the attributes based on the current state-of-the-art in software engineering and generate a seed solution. We then traverse the resulting space using graph isomorphisms and other set operations over the vertex sets. The new solutions will preserve the desired attributes. The goal of defining an isomorphism based search mechanism is to construct predictors of evolution that can facilitate the automation of ’software factory’ paradigm. The model allows for automation via software tools implementing the concepts.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

A model based on graph isomorphisms is used to formalize software evolution. Step by step we narrow the search space by an informed selection of the attributes based on the current state-of-the-art in software engineering and generate a seed solution. We then traverse the resulting space using graph isomorphisms and other set operations over the vertex sets. The new solutions will preserve the desired attributes. The goal of defining an isomorphism based search mechanism is to construct predictors of evolution that can facilitate the automation of ’software factory’ paradigm. The model allows for automation via software tools implementing the concepts.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)

Relevância:

70.00% 70.00%

Publicador:

Resumo:

ALICE, that is an experiment held at CERN using the LHC, is specialized in analyzing lead-ion collisions. ALICE will study the properties of quarkgluon plasma, a state of matter where quarks and gluons, under conditions of very high temperatures and densities, are no longer confined inside hadrons. Such a state of matter probably existed just after the Big Bang, before particles such as protons and neutrons were formed. The SDD detector, one of the ALICE subdetectors, is part of the ITS that is composed by 6 cylindrical layers with the innermost one attached to the beam pipe. The ITS tracks and identifies particles near the interaction point, it also aligns the tracks of the articles detected by more external detectors. The two ITS middle layers contain the whole 260 SDD detectors. A multichannel readout board, called CARLOSrx, receives at the same time the data coming from 12 SDD detectors. In total there are 24 CARLOSrx boards needed to read data coming from all the SDD modules (detector plus front end electronics). CARLOSrx packs data coming from the front end electronics through optical link connections, it stores them in a large data FIFO and then it sends them to the DAQ system. Each CARLOSrx is composed by two boards. One is called CARLOSrx data, that reads data coming from the SDD detectors and configures the FEE; the other one is called CARLOSrx clock, that sends the clock signal to all the FEE. This thesis contains a description of the hardware design and firmware features of both CARLOSrx data and CARLOSrx clock boards, which deal with all the SDD readout chain. A description of the software tools necessary to test and configure the front end electronics will be presented at the end of the thesis.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

The need to effectively manage the documentation covering the entire production process, from the concept phase right through to market realise, constitutes a key issue in the creation of a successful and highly competitive product. For almost forty years the most commonly used strategies to achieve this have followed Product Lifecycle Management (PLM) guidelines. Translated into information management systems at the end of the '90s, this methodology is now widely used by companies operating all over the world in many different sectors. PLM systems and editor programs are the two principal types of software applications used by companies for their process aotomation. Editor programs allow to store in documents the information related to the production chain, while the PLM system stores and shares this information so that it can be used within the company and made it available to partners. Different software tools, which capture and store documents and information automatically in the PLM system, have been developed in recent years. One of them is the ''DirectPLM'' application, which has been developed by the Italian company ''Focus PLM''. It is designed to ensure interoperability between many editors and the Aras Innovator PLM system. In this dissertation we present ''DirectPLM2'', a new version of the previous software application DirectPLM. It has been designed and developed as prototype during the internship by Focus PLM. Its new implementation separates the abstract logic of business from the real commands implementation, previously strongly dependent on Aras Innovator. Thanks to its new design, Focus PLM can easily develop different versions of DirectPLM2, each one devised for a specific PLM system. In fact, the company can focus the development effort only on a specific set of software components which provides specialized functions interacting with that particular PLM system. This allows shorter Time-To-Market and gives the company a significant competitive advantage.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

As lipofilling of the female breast is becoming more popular in plastic surgery, the use of MRI to assess breast volume has been employed to control postoperative results. Therefore, we sought to evaluate the accuracy of magnetic resonance imaging (MRI)-based breast volumetry software tools by comparing the measurements of silicone implant augmented breasts with the actual implant volume specified by the manufacturer. MRI-based volume analysis was performed in eight bilaterally augmented patients (46 ± 9 years) with three different software programs (Brainlab© I plan 2.6 neuronavigation software; mass analysis, version 5.3, Medis©; and OsiriX© v.3.0.2. 32-bit). The implant volumes analysed by the BrainLab© software had a mean deviation of 2.2 ± 1.7% (r?=?0.99) relative to the implanted prosthesis. OsiriX© software analysis resulted in a mean deviation of 2.8 ± 3.0% (r?=?0.99) and the Medis© software had a mean deviation of 3.1 ± 3.0% (r?=?0.99). Overall, the volumes of all analysed breast implants correlated very well with the real implant volumes. Processing time was 10 min per breast with each system and 30 s (OsiriX©) to 5 min (BrainLab© and Medis©) per silicone implant. MRI-based volumetry is a powerful tool to calculate both native breast and silicone implant volume in situ. All software solutions performed well and the measurements were close to the actual implant sizes. The use of MRI breast volumetry may be helpful in: (1) planning reconstructive and aesthetic surgery of asymmetric breasts, (2) calculating implant size in patients with missing documentation of a previously implanted device and (3) assessing post-operative results objectively.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

eLearning supports the education in certain disciplines. Here, we report about novel eLearning concepts, techniques, and tools to support education in Software Engineering, a subdiscipline of computer science. We call this "Software Engineering eLearning". On the other side, software support is a substantial prerequisite for eLearning in any discipline. Thus, Software Engineering techniques have to be applied to develop and maintain those software systems. We call this "eLearning Software Engineering". Both aspects have been investigated in a large joint, BMBF-funded research project, termed MuSofT (Multimedia in Software Engineering). The main results are summarized in this paper.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

In this paper, the authors present results from a study on the usage rate of 2.0 tools and technologies among Spanish enterprises. The main objective of the study is to analyze, from the perceptions of executives, the influence of social software tools on a set of business processes. This analysis has been made using two graphic tools: the “2.0 Success Matrix” and the “Tool’s Footprint”. Both the review of literature and the empirical work have lead to important findings and conclusions.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

Software testing is a key aspect of software reliability and quality assurance in a context where software development constantly has to overcome mammoth challenges in a continuously changing environment. One of the characteristics of software testing is that it has a large intellectual capital component and can thus benefit from the use of the experience gained from past projects. Software testing can, then, potentially benefit from solutions provided by the knowledge management discipline. There are in fact a number of proposals concerning effective knowledge management related to several software engineering processes. Objective: We defend the use of a lesson learned system for software testing. The reason is that such a system is an effective knowledge management resource enabling testers and managers to take advantage of the experience locked away in the brains of the testers. To do this, the experience has to be gathered, disseminated and reused. Method: After analyzing the proposals for managing software testing experience, significant weaknesses have been detected in the current systems of this type. The architectural model proposed here for lesson learned systems is designed to try to avoid these weaknesses. This model (i) defines the structure of the software testing lessons learned; (ii) sets up procedures for lesson learned management; and (iii) supports the design of software tools to manage the lessons learned. Results: A different approach, based on the management of the lessons learned that software testing engineers gather from everyday experience, with two basic goals: usefulness and applicability. Conclusion: The architectural model proposed here lays the groundwork to overcome the obstacles to sharing and reusing experience gained in the software testing and test management. As such, it provides guidance for developing software testing lesson learned systems.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

Automatic grading of programming assignments is an important topic in academic research. It aims at improving the level of feedback given to students and optimizing the professor time. Several researches have reported the development of software tools to support this process. Then, it is helpfulto get a quickly and good sight about their key features. This paper reviews an ample set of tools forautomatic grading of programming assignments. They are divided in those most important mature tools, which have remarkable features; and those built recently, with new features. The review includes the definition and description of key features e.g. supported languages, used technology, infrastructure, etc. The two kinds of tools allow making a temporal comparative analysis. This analysis infrastructure, etc. The two kinds of tools allow making a temporal comparative analysis. This analysis shows good improvements in this research field, these include security, more language support, plagiarism detection, etc. On the other hand, the lack of a grading model for assignments is identified as an important gap in the reviewed tools. Thus, a characterization of evaluation metrics to grade programming assignments is provided as first step to get a model. Finally new paths in this research field are proposed.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

MEGARA (Multi-Espectrografo en GTC de Alta Resolucion para Astronomia) is an optical Integral-Field Unit (IFU) and Multi-Object Spectrograph (MOS) designed for the GTC 10.4 m telescope in La Palma. Its relatively complex layout makes necessary a set of tools to facilitate the observation preparation to the user. The MEGARA Observing Preparation Software Suite (MOPSS) consists on three software components designed to assist observers to optimally plan their observations with GTC/MEGARA: the Exposure Time Calculator, the Image Simulator, and the Fiber MOS Positioning tool. We describe these software tools and the status of their prototypes up to the date.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

Costs and environmental impacts are key elements in forest logistics and they must be integrated in forest decision-making. The evaluation of transportation fuel costs and carbon emissions depend on spatial and non-spatial data but in many cases the former type of data are dicult to obtain. On the other hand, the availability of software tools to evaluate transportation fuel consumption as well as costs and emissions of carbon dioxide is limited. We developed a software tool that combines two empirically validated models of truck transportation using Digital Elevation Model (DEM) data and an open spatial data tool, specically OpenStreetMap©. The tool generates tabular data and spatial outputs (maps) with information regarding fuel consumption, cost and CO2 emissions for four types of trucks. It also generates maps of the distribution of transport performance indicators (relation between beeline and real road distances). These outputs can be easily included in forest decision-making support systems. Finally, in this work we applied the tool in a particular case of forest logistics in north-eastern Portugal

Relevância:

70.00% 70.00%

Publicador:

Resumo:

The advantages of antennas that can resemble the shape of the body to which they are attached are obvious. However, electromagnetic modeling of such unusually shaped antennas can be difficult. In this paper, the commercially available software SolidWorks(TM) is used for accurately drawing complex shapes in conjunction with the electromagnetic software FEKO(TM) to model the EM behavior of conformal antennas. The application of SolidWorks and custom-written software allows all the required information that forms the analyzed structure to be automatically inserted into FEKO, and gives the user complete control over the antenna being modeled. This approach is illustrated by a number of simulation examples of single, wideband, multi-band planar and curved patch antennas.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

Species extinctions and the deterioration of other biodiversity features worldwide have led to the adoption of systematic conservation planning in many regions of the world. As a consequence, various software tools for conservation planning have been developed over the past twenty years. These, tools implement algorithms designed to identify conservation area networks for the representation and persistence of biodiversity features. Budgetary, ethical, and other sociopolitical constraints dictate that the prioritized sites represent biodiversity with minimum impact on human interests. Planning tools are typically also used to satisfy these criteria. This chapter reviews both the concepts and technical choices that underlie the development of these tools. Conservation planning problems can be formulated as optimization problems, and we evaluate the suitability of different algorithms for their solution. Finally, we also review some key issues associated with the use of these tools, such as computational efficiency, the effectiveness of taxa and abiotic parameters at choosing surrogates for biodiversity, the process of setting explicit targets of representation for biodiversity surrogates, and