910 resultados para software testing methods


Relevância:

90.00% 90.00%

Publicador:

Resumo:

Formal methods and software testing are tools to obtain and control software quality. When used together, they provide mechanisms for software specification, verification and error detection. Even though formal methods allow software to be mathematically verified, they are not enough to assure that a system is free of faults, thus, software testing techniques are necessary to complement the process of verification and validation of a system. Model Based Testing techniques allow tests to be generated from other software artifacts such as specifications and abstract models. Using formal specifications as basis for test creation, we can generate better quality tests, because these specifications are usually precise and free of ambiguity. Fernanda Souza (2009) proposed a method to define test cases from B Method specifications. This method used information from the machine s invariant and the operation s precondition to define positive and negative test cases for an operation, using equivalent class partitioning and boundary value analysis based techniques. However, the method proposed in 2009 was not automated and had conceptual deficiencies like, for instance, it did not fit in a well defined coverage criteria classification. We started our work with a case study that applied the method in an example of B specification from the industry. Based in this case study we ve obtained subsidies to improve it. In our work we evolved the proposed method, rewriting it and adding characteristics to make it compatible with a test classification used by the community. We also improved the method to support specifications structured in different components, to use information from the operation s behavior on the test case generation process and to use new coverage criterias. Besides, we have implemented a tool to automate the method and we have submitted it to more complex case studies

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The main goal of Regression Test (RT) is to reuse the test suite of the latest version of a software in its current version, in order to maximize the value of the tests already developed and ensure that old features continue working after the new changes. Even with reuse, it is common that not all tests need to be executed again. Because of that, it is encouraged to use Regression Tests Selection (RTS) techniques, which aims to select from all tests, only those that reveal faults, this reduces costs and makes this an interesting practice for the testing teams. Several recent research works evaluate the quality of the selections performed by RTS techniques, identifying which one presents the best results, measured by metrics such as inclusion and precision. The RTS techniques should seek in the System Under Test (SUT) for tests that reveal faults. However, because this is a problem without a viable solution, they alternatively seek for tests that reveal changes, where faults may occur. Nevertheless, these changes may modify the execution flow of the algorithm itself, leading some tests no longer exercise the same stretch. In this context, this dissertation investigates whether changes performed in a SUT would affect the quality of the selection of tests performed by an RTS, if so, which features the changes present which cause errors, leading the RTS to include or exclude tests wrongly. For this purpose, a tool was developed using the Java language to automate the measurement of inclusion and precision averages achieved by a regression test selection technique for a particular feature of change. In order to validate this tool, an empirical study was conducted to evaluate the RTS technique Pythia, based on textual differencing, on a large web information system, analyzing the feature of types of tasks performed to evolve the SUT

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Introduction: Computer software can be used to predict orthognathic surgery outcomes. The aim of this study was to subjectively compare the soft-tissue surgical simulations of 2 software programs. Methods: Standard profile pictures were taken of 10 patients with a Class III malocclusion and a concave facial profile who were scheduled for double-jaw orthognathic surgery. The patients had horizontal maxillary deficiency or horizontal mandibular excess. Two software programs (Dentofacial Planner Plus [Dentofacial Software, Toronto, Ontario, Canada] and Dolphin Imaging [version 9.0, Dolphin Imaging Software, Canoga Park, Calif]) were used to predict the postsurgical profiles. The predictive images were compared with the actual final photographs. One hundred one orthodontists, oral-maxillofacial surgeons, and general dentists evaluated the images and were asked whether they would use either software program to plan treatment for, or to educate, their patients. Results: Statistical analyses showed differences between the groups when each point was judged. Dolphin Imaging software had better prediction of nasal tip, chin, and submandibular area. Dentofacial Planner Plus software was better in predicting nasolabial angle, and upper and lower lips. The total profile comparison showed no statistical difference between the softwares. Conclusions: The 2 types of software are similar for obtaining 2-dimensional predictive profile images of patients with Class III malocclusion treated with orthognathic surgery. (Am J Orthod Dentofacial Orthop 2010; 137: 452.e1-452.e5)

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Objective. To determine the influence of cement thickness and ceramic/cement bonding on stresses and failure of CAD/CAM crowns, using both multi-physics finite element analysis and monotonic testing.Methods. Axially symmetric FEA models were created for stress analysis of a stylized monolithic crown having resin cement thicknesses from 50 to 500 mu m under occlusal loading. Ceramic-cement interface was modeled as bonded or not-bonded (cement-dentin as bonded). Cement polymerization shrinkage was simulated as a thermal contraction. Loads necessary to reach stresses for radial cracking from the intaglio surface were calculated by FEA. Experimentally, feldspathic CAD/CAM crowns based on the FEA model were machined having different occlusal cementation spaces, etched and cemented to dentin analogs. Non-bonding of etched ceramic was achieved using a thin layer of poly(dimethylsiloxane). Crowns were loaded to failure at 5 N/s, with radial cracks detected acoustically.Results. Failure loads depended on the bonding condition and the cement thickness for both FEA and physical testing. Average fracture loads for bonded crowns were: 673.5 N at 50 mu m cement and 300.6 N at 500 mu m. FEA stresses due to polymerization shrinkage increased with the cement thickness overwhelming the protective effect of bonding, as was also seen experimentally. At 50 mu m cement thickness, bonded crowns withstood at least twice the load before failure than non-bonded crowns.Significance. Occlusal "fit" can have structural implications for CAD/CAM crowns; pre-cementation spaces around 50-100 mu m being recommended from this study. Bonding benefits were lost at thickness approaching 450-500 mu m due to polymerization shrinkage stresses. (C) 2012 Academy of Dental Materials. Published by Elsevier Ltd. All rights reserved.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

A suit able decision-making on managing a contaminated site characterization program is strongly dependent of the diagnosis process. A detailed diagnosis can be done based on a Conceptual Site Model (CSM) elaboration using high resolution site characterization tools. The piezocone (CPTu) test is a high resolution tool which allows attaching several specific sensors, like the resistivity probe. This hybrid device is called the resistivity piezocone (RCPTu). A simulated geo-environmental site characterization program was performed on an erosion site using different tools (direct push tools soil samplers, hollow stem auger (HSA) drilling and RCPTu tests) to develop the CSM for a site similar to the Brazilian conditions. It was observed a good agreement between the site profiles interpreted by the different methods. The resistivity sensor attached to the piezocone improved the interpretation and the decision-making process on site was significantly better for the CSM elaboration. The RCPTu test data also allowed identifying the hydrogeological heterogeneities. The present study shows that the RCPTu test is also a useful and powerful tool to development an accurate CSM in a Brazilian condition, especially in an approach that prioritizes high resolution geo-environmental investigation. © 2013 Taylor & Francis Group.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Population growth, together with the gradual social ascent in Brazil, reflects at the growing need for better use of urban spaces. In this context, the amount of new buildings to meet the demand in property market, the needs for creating new roads and highways, among others, make the use of geotechnical works and, more specifically, retaining walls, more and more common. One of the simplest solutions for underground works is the use of retaining structures using tie back walls for soil support, therefore, the present work deals with this kind of structures. This paper proposes the use of FTOOL software testing in predicting deformations in tie back walls, by comparing simulations of the presented model to a real and measured deformation case in Guabirotuba Formation (PR). The results showed the importance of defining the parameters such as stiffness and curtain geometry, as well as the definition of representative loads acting on it. Also, it was pointed out that the passive response of the steel rods depends on the horizontal displacement of the wall. The study concluded that the program generates very representative results when compared to field data and seems to be a promising tool for tie back structures displacement predictions

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Software product line (SPL) engineering offers several advantages in the development of families of software products such as reduced costs, high quality and a short time to market. A software product line is a set of software intensive systems, each of which shares a common core set of functionalities, but also differs from the other products through customization tailored to fit the needs of individual groups of customers. The differences between products within the family are well-understood and organized into a feature model that represents the variability of the SPL. Products can then be built by generating and composing features described in the feature model. Testing of software product lines has become a bottleneck in the SPL development lifecycle, since many of the techniques used in their testing have been borrowed from traditional software testing and do not directly take advantage of the similarities between products. This limits the overall gains that can be achieved in SPL engineering. Recent work proposed by both industry and the research community for improving SPL testing has begun to consider this problem, but there is still a need for better testing techniques that are tailored to SPL development. In this thesis, I make two primary contributions to software product line testing. First I propose a new definition for testability of SPLs that is based on the ability to re-use test cases between products without a loss of fault detection effectiveness. I build on this idea to identify elements of the feature model that contribute positively and/or negatively towards SPL testability. Second, I provide a graph based testing approach called the FIG Basis Path method that selects products and features for testing based on a feature dependency graph. This method should increase our ability to re-use results of test cases across successive products in the family and reduce testing effort. I report the results of a case study involving several non-trivial SPLs and show that for these objects, the FIG Basis Path method is as effective as testing all products, but requires us to test no more than 24% of the products in the SPL.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Abstract Background Considering the increasing use of polymyxins to treat infections due to multidrug resistant Gram-negative in many countries, it is important to evaluate different susceptibility testing methods to this class of antibiotic. Methods Susceptibility of 109 carbapenem-resistant P. aeruginosa to polymyxins was tested comparing broth microdilution (reference method), disc diffusion, and Etest using the new interpretative breakpoints of Clinical and Laboratory Standards Institute. Results Twenty-nine percent of isolates belonged to endemic clone and thus, these strains were excluded of analysis. Among 78 strains evaluated, only one isolate was resistant to polymyxin B by the reference method (MIC: 8.0 μg/mL). Very major and major error rates of 1.2% and 11.5% were detected comparing polymyxin B disc diffusion with the broth microdilution (reference method). Agreement within 1 twofold dilution between Etest and the broth microdilution were 33% for polymyxin B and 79.5% for colistin. One major error and 48.7% minor errors were found comparing polymyxin B Etest with broth microdilution and only 6.4% minor errors with colistin. The concordance between Etest and the broth microdilution (reference method) was respectively 100% for colistin and 90% for polymyxin B. Conclusion Resistance to polymyxins seems to be rare among hospital carbapenem-resistant P. aeruginosa isolates over a six-year period. Our results showed, using the new CLSI criteria, that the disc diffusion susceptibility does not report major errors (false-resistant results) for colistin. On the other hand, showed a high frequency of minor errors and 1 very major error for polymyxin B. Etest presented better results for colistin than polymyxin B. Until these results are reproduced with a large number of polymyxins-resistant P. aeruginosa isolates, susceptibility to polymyxins should be confirmed by a reference method.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Nel primo capitolo si analizzeranno alcune basi del software testing e delle applicazioni web che ci permetteranno di introdurre l'end-to-end testing nel secondo capitolo dove se ne analizzeranno il funzionamento tecnico e le sue prospettive. Successivamente, nel terzo capitolo, verrà presentato l'uso dell'end-to-end testing applicandolo ad un caso concreto, il gestionale Buudis, per poi dimostrarne la sua efficacia ed efficienza nell'ultimo capitolo.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Petroleum supply and environmental pollution issues constantly increase interest in renewable low polluting alternative fuels. Published test results show decreased pollution with similar power output and fuel consumption from Internal Combustion Engines (ICE) burning alternative fuels. More specifically, diesel engines burning biodiesel derived from plant oils and animal fats not only reduce harmful exhaust emissions but are renewable and environmentally friendly. To validate these claims and assess the feasibility of alternative fuels, independent engine dynamometer and emissions testing was performed. A testing apparatus capable of making relevant measurements was designed, built, and used to test and determine the feasibility of biodiesel. The apparatus marks the addition of a valuable testing tool to the University and provides a foundation for future experiments. This thesis will discuss the background of biodiesel, testing methods, design and function of the testing apparatus, experimental results, relevant calculations, and conclusions.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

This paper describes the open source framework MARVIN for rapid application development in the field of biomedical and clinical research. MARVIN applications consist of modules that can be plugged together in order to provide the functionality required for a specific experimental scenario. Application modules work on a common patient database that is used to store and organize medical data as well as derived data. MARVIN provides a flexible input/output system with support for many file formats including DICOM, various 2D image formats and surface mesh data. Furthermore, it implements an advanced visualization system and interfaces to a wide range of 3D tracking hardware. Since it uses only highly portable libraries, MARVIN applications run on Unix/Linux, Mac OS X and Microsoft Windows.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

OBJECTIVE: The presence of minority nonnucleoside reverse transcriptase inhibitor (NNRTI)-resistant HIV-1 variants prior to antiretroviral therapy (ART) has been linked to virologic failure in treatment-naive patients. DESIGN: We performed a large retrospective study to determine the number of treatment failures that could have been prevented by implementing minority drug-resistant HIV-1 variant analyses in ART-naïve patients in whom no NNRTI resistance mutations were detected by routine resistance testing. METHODS: Of 1608 patients in the Swiss HIV Cohort Study, who have initiated first-line ART with two nucleoside reverse transcriptase inhibitors (NRTIs) and one NNRTI before July 2008, 519 patients were eligible by means of HIV-1 subtype, viral load and sample availability. Key NNRTI drug resistance mutations K103N and Y181C were measured by allele-specific PCR in 208 of 519 randomly chosen patients. RESULTS: Minority K103N and Y181C drug resistance mutations were detected in five out of 190 (2.6%) and 10 out of 201 (5%) patients, respectively. Focusing on 183 patients for whom virologic success or failure could be examined, virologic failure occurred in seven out of 183 (3.8%) patients; minority K103N and/or Y181C variants were present prior to ART initiation in only two of those patients. The NNRTI-containing, first-line ART was effective in 10 patients with preexisting minority NNRTI-resistant HIV-1 variant. CONCLUSION: As revealed in settings of case-control studies, minority NNRTI-resistant HIV-1 variants can have an impact on ART. However, the sole implementation of minority NNRTI-resistant HIV-1 variant analysis in addition to genotypic resistance testing (GRT) cannot be recommended in routine clinical settings. Additional associated risk factors need to be discovered.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

BACKGROUND Lung clearance index (LCI), a marker of ventilation inhomogeneity, is elevated early in children with cystic fibrosis (CF). However, in infants with CF, LCI values are found to be normal, although structural lung abnormalities are often detectable. We hypothesized that this discrepancy is due to inadequate algorithms of the available software package. AIM Our aim was to challenge the validity of these software algorithms. METHODS We compared multiple breath washout (MBW) results of current software algorithms (automatic modus) to refined algorithms (manual modus) in 17 asymptomatic infants with CF, and 24 matched healthy term-born infants. The main difference between these two analysis methods lies in the calculation of the molar mass differences that the system uses to define the completion of the measurement. RESULTS In infants with CF the refined manual modus revealed clearly elevated LCI above 9 in 8 out of 35 measurements (23%), all showing LCI values below 8.3 using the automatic modus (paired t-test comparing the means, P < 0.001). Healthy infants showed normal LCI values using both analysis methods (n = 47, paired t-test, P = 0.79). The most relevant reason for false normal LCI values in infants with CF using the automatic modus was the incorrect recognition of the end-of-test too early during the washout. CONCLUSION We recommend the use of the manual modus for the analysis of MBW outcomes in infants in order to obtain more accurate results. This will allow appropriate use of infant lung function results for clinical and scientific purposes.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

BACKGROUND AND OBJECTIVES Multiple-breath washout (MBW) is an attractive test to assess ventilation inhomogeneity, a marker of peripheral lung disease. Standardization of MBW is hampered as little data exists on possible measurement bias. We aimed to identify potential sources of measurement bias based on MBW software settings. METHODS We used unprocessed data from nitrogen (N2) MBW (Exhalyzer D, Eco Medics AG) applied in 30 children aged 5-18 years: 10 with CF, 10 formerly preterm, and 10 healthy controls. This setup calculates the tracer gas N2 mainly from measured O2 and CO2concentrations. The following software settings for MBW signal processing were changed by at least 5 units or >10% in both directions or completely switched off: (i) environmental conditions, (ii) apparatus dead space, (iii) O2 and CO2 signal correction, and (iv) signal alignment (delay time). Primary outcome was the change in lung clearance index (LCI) compared to LCI calculated with the settings as recommended. A change in LCI exceeding 10% was considered relevant. RESULTS Changes in both environmental and dead space settings resulted in uniform but modest LCI changes and exceeded >10% in only two measurements. Changes in signal alignment and O2 signal correction had the most relevant impact on LCI. Decrease of O2 delay time by 40 ms (7%) lead to a mean LCI increase of 12%, with >10% LCI change in 60% of the children. Increase of O2 delay time by 40 ms resulted in mean LCI decrease of 9% with LCI changing >10% in 43% of the children. CONCLUSIONS Accurate LCI results depend crucially on signal processing settings in MBW software. Especially correct signal delay times are possible sources of incorrect LCI measurements. Algorithms of signal processing and signal alignment should thus be optimized to avoid susceptibility of MBW measurements to this significant measurement bias.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Abstract. The ASSERT project de?ned new software engineering methods and tools for the development of critical embedded real-time systems in the space domain. The ASSERT model-driven engineering process was one of the achievements of the project and is based on the concept of property- preserving model transformations. The key element of this process is that non-functional properties of the software system must be preserved during model transformations. Properties preservation is carried out through model transformations compliant with the Ravenscar Pro?le and provides a formal basis to the process. In this way, the so-called Ravenscar Computational Model is central to the whole ASSERT process. This paper describes the work done in the HWSWCO study, whose main objective has been to address the integration of the Hardware/Software co-design phase in the ASSERT process. In order to do that, non-functional properties of the software system must also be preserved during hardware synthesis. Keywords : Ada 2005, Ravenscar pro?le, Hardware/Software co-design, real- time systems, high-integrity systems, ORK