954 resultados para Computer software - Quality control
Resumo:
The current study used statistical methods of quality control to evaluate the performance of a sewage treatment station. The concerned station is located in Cascavel city, Paraná State. The evaluated parameters were hydrogenionic potential, settleable solids, total suspended solids, chemical oxygen demand and biochemical oxygen demand in five days. Statistical analysis was performed through Shewhart control charts and process capability ratio. According to Shewhart charts, only the BOD(5.20) variable was under statistical control. Through capability ratios, we observed that except for pH the sewage treatment station is not capable to produce effluents under characteristics that fulfill specifications or standard launching required by environmental legislation.
Resumo:
Katselmoinnit ja tarkastusmenettelyt ovat osa ohjelmistotuotantoprosessin laadunvarmistusta. Staattisella tarkastamisella tarkoitetaan ohjelmistotuotteen visuaalista tarkastamista ohjelmistovirheiden havaitsemiseksi ja korjaamiseksi. Ohjelmiston lähdekoodin tarkastaminen voidaan suorittaa automaattisesti tarkoitukseen sopivalla ohjelmistolla l. analyysityökalulla. Tässä työssä toteutettiin analyysityökalu C#-kielisten lähdekoodien tarkastamiseen. Työkalulla suoritetussa kenttätestauksessa havaittiin tarkastettavissa ohjelmistoissa ohjelmiston ylläpitoon vaikuttavia puutteita. Lisäksi työssä tarkasteltiin katselmointeja osana ohjelmistotuotantoprosessin laadunvarmistusta sekä erilaisia ohjelmistovirheitä ja niiden lähteitä.
Resumo:
Objective The objective of this study is to assess the performance of cytopathology laboratories providing services to the Brazilian Unified Health System (Sistema Único de Saúde - SUS) in the State of Minas Gerais, Brazil. Methods This descriptive study uses data obtained from the Cervical Cancer Information System from January to December 2012. Three quality indicators were analyzed to assess the quality of cervical cytopathology tests: positivity index, percentage of atypical squamous cells (ASCs) in abnormal tests, and percentage of tests compatiblewith high-grade squamous intraepithelial lesions (HSILs). Laboratories were classified according to their production scale in tests per year≤5,000; from 5,001 to 10,000; from 10,001 to 15,000; and 15,001. Based on the collection of variables and the classification of laboratories according to production scale, we created and analyzed a database using Microsoft Office Excel 97-2003. Results In the Brazilian state of Minas Gerais, 146 laboratories provided services to the SUS in 2012 by performing a total of 1,277,018 cervical cytopathology tests. Half of these laboratories had production scales≤5,000 tests/year and accounted for 13.1% of all tests performed in the entire state; in turn, 13.7% of these laboratories presented production scales of > 15,001 tests/year and accounted for 49.2% of the total of tests performed in the entire state. The positivity indexes of most laboratories providing services to the SUS in 2012, regardless of production scale, were below or well below recommended limits. Of the 20 laboratories that performed more than 15,001 tests per year, only three presented percentages of tests compatible with HSILs above the lower limit recommended by the Brazilian Ministry of Health. Conclusion The majority of laboratories providing services to the SUS in Minas Gerais presented quality indicators outside the range recommended by the Brazilian Ministry of Health.
Resumo:
An empirical study was conducted in the area of software engineering to study relationships between development, testing and intended software quality. International standards served as a starting point of the study. For analysis a round of interviews was kept and transcribed. It was found that interaction between humans is critical, especially in transferring knowledge and standards’ processes. The standards are communicated through interaction and learning processes are involved before compliance. One of the results was that testing is the key to sufficient quality. The outcome was that successful interaction, sufficient testing and compliance with the standards combined with good motivation may provide most repeatable intended quality.
Resumo:
The augmented reality (AR) technology has applications in many fields as diverse as aeronautics, tourism, medicine, and education. In this review are summarized the current status of AR and it is proposed a new application of it in weed science. The basic algorithmic elements for AR implementation are already available to develop applications in the area of weed economic thresholds. These include algorithms for image recognition to identify and quantify weeds by species and software for herbicide selection based on weed density. Likewise, all hardware necessary for AR implementation in weed science are available at an affordable price for the user. Thus, the authors propose weed science can take a leading role integrating AR systems into weed economic thresholds software, thus, providing better opportunities for science and computer-based weed control decisions.
Resumo:
The goal of this thesis is to make a case study of test automation’s profitability in the development of embedded software in a real industrial setting. The cost-benefit analysis is done by considering the costs and benefits test automation causes to software development, before the software is released to customers. The potential benefits of test automation regarding software quality after customer release were not estimated. Test automation is a significant investment which often requires dedicated resources. When done accordingly, the investment in test automation can produce major cost savings by reducing the need for manual testing effort, especially if the software is developed with an agile development framework. It can reduce the cost of avoidable rework of software development, as test automation enables the detection of construction time defects in the earliest possible moment. Test automation also has many pitfalls such as test maintainability and testability of the software, and if those areas are neglected, the investment in test automation may become worthless or it may even produce negative results. The results of this thesis suggest that test automation is very profitable at the company under study.
Resumo:
Biokuvainformatiikan kehittäminen – mikroskopiasta ohjelmistoratkaisuihin – sovellusesimerkkinä α2β1-integriini Kun ihmisen genomi saatiin sekvensoitua vuonna 2003, biotieteiden päätehtäväksi tuli selvittää eri geenien tehtävät, ja erilaisista biokuvantamistekniikoista tuli keskeisiä tutkimusmenetelmiä. Teknologiset kehitysaskeleet johtivat erityisesti fluoresenssipohjaisten valomikroskopiatekniikoiden suosion räjähdysmäiseen kasvuun, mutta mikroskopian tuli muuntua kvalitatiivisesta tieteestä kvantitatiiviseksi. Tämä muutos synnytti uuden tieteenalan, biokuvainformatiikan, jonka on sanottu mahdollisesti mullistavan biotieteet. Tämä väitöskirja esittelee laajan, poikkitieteellisen työkokonaisuuden biokuvainformatiikan alalta. Väitöskirjan ensimmäinen tavoite oli kehittää protokollia elävien solujen neliulotteiseen konfokaalimikroskopiaan, joka oli yksi nopeimmin kasvavista biokuvantamismenetelmistä. Ihmisen kollageenireseptori α2β1-integriini, joka on tärkeä molekyyli monissa fysiologisissa ja patologisissa prosesseissa, oli sovellusesimerkkinä. Työssä saavutettiin selkeitä visualisointeja integriinien liikkeistä, yhteenkeräytymisestä ja solun sisään siirtymisestä, mutta työkaluja kuvainformaation kvantitatiiviseen analysointiin ei ollut. Väitöskirjan toiseksi tavoitteeksi tulikin tällaiseen analysointiin soveltuvan tietokoneohjelmiston kehittäminen. Samaan aikaan syntyi biokuvainformatiikka, ja kipeimmin uudella alalla kaivattiin erikoistuneita tietokoneohjelmistoja. Tämän väitöskirjatyön tärkeimmäksi tulokseksi muodostui näin ollen BioImageXD, uudenlainen avoimen lähdekoodin ohjelmisto moniulotteisten biokuvien visualisointiin, prosessointiin ja analysointiin. BioImageXD kasvoi yhdeksi alansa suurimmista ja monipuolisimmista. Se julkaistiin Nature Methods -lehden biokuvainformatiikkaa käsittelevässä erikoisnumerossa, ja siitä tuli tunnettu ja laajalti käytetty. Väitöskirjan kolmas tavoite oli soveltaa kehitettyjä menetelmiä johonkin käytännönläheisempään. Tehtiin keinotekoisia piidioksidinanopartikkeleita, joissa oli "osoitelappuina" α2β1-integriinin tunnistavia vasta-aineita. BioImageXD:n avulla osoitettiin, että nanopartikkeleilla on potentiaalia lääkkeiden täsmäohjaussovelluksissa. Tämän väitöskirjatyön yksi perimmäinen tavoite oli edistää uutta ja tuntematonta biokuvainformatiikan tieteenalaa, ja tämä tavoite saavutettiin erityisesti BioImageXD:n ja sen lukuisten julkaistujen sovellusten kautta. Väitöskirjatyöllä on merkittävää potentiaalia tulevaisuudessa, mutta biokuvainformatiikalla on vakavia haasteita. Ala on liian monimutkainen keskimääräisen biolääketieteen tutkijan hallittavaksi, ja alan keskeisin elementti, avoimen lähdekoodin ohjelmistokehitystyö, on aliarvostettu. Näihin seikkoihin tarvitaan useita parannuksia,
Resumo:
The adaptive behavior of human beings is usually supported by rapid monitoring of outstanding events in the environment. Some investigators have suggested that a primary attention deficit might trigger symptoms of schizophrenia. In addition, researchers have long discussed the relationship between schizophrenia and the schizophrenia-like psychosis of epilepsy (SLPE). On the basis of these considerations, the objective of the present study was to investigate attention performance of patients with both disorders. Patient age was 18 to 60 years, and all patients had received formal schooling for at least four years. Patients were excluded if they had any systemic disease with neurologic or psychiatric comorbidity, or a history of brain surgery. The computer-assisted TAVIS-2R test was applied to all patients and to a control group to evaluate and discriminate between selective, alternating and sustained attention. The TAVIS-2R test is divided into three parts: one for selective attention (5 min), the second for alternating attention (5 min), and the third for the evaluation of vigilance or sustained attention (10 min). The same computer software was used for statistical analysis of reaction time, omission errors, and commission errors. The sample consisted of 36 patients with schizophrenia, 28 with interictal SLPE, and 47 healthy controls. The results of the selective attention tests for both patient groups were significantly lower than that for controls. The patients with schizophrenia and SLPE performed differently in the alternating and sustained attention tests: patients with SLPE had alternating attention deficits, whereas patients with schizophrenia showed deficits in sustained attention. These quantitative results confirmed the qualitative clinical observations for both patient groups, that is, that patients with schizophrenia had difficulties in focusing attention, whereas those with epilepsy showed perseveration in attention focus.
Resumo:
This thesis studies the use of machine vision in RDF quality assurance and manufacturing. Currently machine vision is used in recycling and material detection and some commer- cial products are available in the market. In this thesis an on-line machine vision system is proposed for characterizing particle size. The proposed machine vision system is based on the mapping between image segmenta- tion and the ground truth of the particle size. The results shows that the implementation of such machine vision system is feasible.
Resumo:
Human beings have always strived to preserve their memories and spread their ideas. In the beginning this was always done through human interpretations, such as telling stories and creating sculptures. Later, technological progress made it possible to create a recording of a phenomenon; first as an analogue recording onto a physical object, and later digitally, as a sequence of bits to be interpreted by a computer. By the end of the 20th century technological advances had made it feasible to distribute media content over a computer network instead of on physical objects, thus enabling the concept of digital media distribution. Many digital media distribution systems already exist, and their continued, and in many cases increasing, usage is an indicator for the high interest in their future enhancements and enriching. By looking at these digital media distribution systems, we have identified three main areas of possible improvement: network structure and coordination, transport of content over the network, and the encoding used for the content. In this thesis, our aim is to show that improvements in performance, efficiency and availability can be done in conjunction with improvements in software quality and reliability through the use of formal methods: mathematical approaches to reasoning about software so that we can prove its correctness, together with the desirable properties. We envision a complete media distribution system based on a distributed architecture, such as peer-to-peer networking, in which different parts of the system have been formally modelled and verified. Starting with the network itself, we show how it can be formally constructed and modularised in the Event-B formalism, such that we can separate the modelling of one node from the modelling of the network itself. We also show how the piece selection algorithm in the BitTorrent peer-to-peer transfer protocol can be adapted for on-demand media streaming, and how this can be modelled in Event-B. Furthermore, we show how modelling one peer in Event-B can give results similar to simulating an entire network of peers. Going further, we introduce a formal specification language for content transfer algorithms, and show that having such a language can make these algorithms easier to understand. We also show how generating Event-B code from this language can result in less complexity compared to creating the models from written specifications. We also consider the decoding part of a media distribution system by showing how video decoding can be done in parallel. This is based on formally defined dependencies between frames and blocks in a video sequence; we have shown that also this step can be performed in a way that is mathematically proven correct. Our modelling and proving in this thesis is, in its majority, tool-based. This provides a demonstration of the advance of formal methods as well as their increased reliability, and thus, advocates for their more wide-spread usage in the future.
Resumo:
The development of new procedures for quickly obtaining accurate information on the physiological potential of seed lots is essential for developing quality control programs for the seed industry. In this study, the effectiveness of an automated system of seedling image analysis (Seed Vigor Imaging System - SVIS) in determining the physiological potential of sun hemp seeds and its relationship with electrical conductivity tests, were evaluated. SVIS evaluations were performed three and four days after sowing and data on the vigor index and the length and uniformity of seedling growth were collected. The electrical conductivity test was made on 50 seed replicates placed in containers with 75 mL of deionised water at 25 ºC and readings were taken after 1, 2, 4, 8 and 16 hours of imbibition. Electrical conductivity measurements at 4 or 8 hours and the use of the SVIS on 3-day old seedlings can effectively detect differences in vigor between different sun hemp seed lots.
Resumo:
Project scope is to utilize Six Sigma DMAIC approach and lean principles to improve production quality of the case company. Six Sigma tools and techniques are explored through a literature review and later used in the quality control phase. The focus is set on the Pareto analysis to demonstrate the most evident development areas in the production. Materials that are not delivered to the customer or materials that damaged during transportation comprise the biggest share of all feedbacks. The goal is set to reduce these feedbacks by 50 %. Production observation pointed out that not only material shortages but also over-production is a daily situation. As a result, an initial picking list where the purchased and own production components can be seen, is created, reduction of over- and underproduction and material marking improvement are seen the most competitive options so that the goal can be reached. The picking list development should still continue to make sure that the list can be used not only in the case study but also in the industrial scale. The reduction of material missing category can be evaluated reliably not sooner than in few years because it takes time to gather the needed statistical information.
Resumo:
The objectives of this master’s thesis were to understand the importance of bubbling fluidized bed (BFB) conditions and to find out how digital image processing and acoustic emission technology can help in monitoring the bed quality. An acoustic emission (AE) measurement system and a bottom ash camera system were evaluated in acquiring information about the bed conditions. The theory part of the study describes the fundamentals of BFB boiler and evaluates the characteristics of bubbling bed. Causes and effects of bed material coarsening are explained. The ways and methods to monitor the behaviour of BFB are determined. The study introduces the operating principles of AE technology and digital image processing. The empirical part of the study describes an experimental arrangement and results of a case study at an industrial BFB boiler. Sand consumption of the boiler was reduced by optimization of bottom ash handling and sand feeding. Furthermore, data from the AE measurement system and the bottom ash camera system was collected. The feasibility of these two systems was evaluated. The particle size of bottom ash and the changes in particle size distribution were monitored during the test period. Neither of the systems evaluated was ready to serve in bed quality control accurately or fast enough. Particle size distributions according to the bottom ash camera did not correspond to the results of manual sieving. Comprehensive interpretation of the collected AE data requires much experience. Both technologies do have potential and with more research and development they may enable acquiring reliable and real-time information about the bed conditions. This information could help to maintain disturbance-free combustion process and to optimize bottom ash handling system.
Resumo:
Health regulatory colleges promote quality practice and continued competence through Quality Assurance (QA) programs. For many colleges, a QA program includes the use of portfolios that incorporate self-directed learning. The purpose of this study was to determine some of the issues surrounding the effectiveness of QA portfolio programs. The literature review revealed that portfolios are valuable tools, but gaps in knowledge include a comparative analysis of QA programs and the perspective of regulatory college administrators. Data were collected through interviews with 6 administrators and a review of 14 portfolio models described on college websites. The results from the two data sources were applied to Robert Stake's responsive evaluation framework to identify issues related to the portfolio's effectiveness (Stake, 1967). The learning components of portfolios were analyzed through the humanist and constructivist lenses. All 14 portfolio models were found to have 3 main components: self-diagnosis, learning plan and activities, and self-evaluation. However, differences were uncovered in learners' autonomy in selecting learning activities, methods of portfolio evaluation, and the relationship between the portfolio and other QA components. The results revealed a dual philosophy of learning in portfolio models and an apparent contradiction between the needs of the individual learner and the organization. Paths for future research include the tenuous relationship between competence and learning, and the impact of technical approaches on selfdirected learning initiatives. A key recommendation is to acknowledge the unique identity of each profession so that health regulatory colleges can address legislative demands and learner needs.
Resumo:
Un objectif principal du génie logiciel est de pouvoir produire des logiciels complexes, de grande taille et fiables en un temps raisonnable. La technologie orientée objet (OO) a fourni de bons concepts et des techniques de modélisation et de programmation qui ont permis de développer des applications complexes tant dans le monde académique que dans le monde industriel. Cette expérience a cependant permis de découvrir les faiblesses du paradigme objet (par exemples, la dispersion de code et le problème de traçabilité). La programmation orientée aspect (OA) apporte une solution simple aux limitations de la programmation OO, telle que le problème des préoccupations transversales. Ces préoccupations transversales se traduisent par la dispersion du même code dans plusieurs modules du système ou l’emmêlement de plusieurs morceaux de code dans un même module. Cette nouvelle méthode de programmer permet d’implémenter chaque problématique indépendamment des autres, puis de les assembler selon des règles bien définies. La programmation OA promet donc une meilleure productivité, une meilleure réutilisation du code et une meilleure adaptation du code aux changements. Très vite, cette nouvelle façon de faire s’est vue s’étendre sur tout le processus de développement de logiciel en ayant pour but de préserver la modularité et la traçabilité, qui sont deux propriétés importantes des logiciels de bonne qualité. Cependant, la technologie OA présente de nombreux défis. Le raisonnement, la spécification, et la vérification des programmes OA présentent des difficultés d’autant plus que ces programmes évoluent dans le temps. Par conséquent, le raisonnement modulaire de ces programmes est requis sinon ils nécessiteraient d’être réexaminés au complet chaque fois qu’un composant est changé ou ajouté. Il est cependant bien connu dans la littérature que le raisonnement modulaire sur les programmes OA est difficile vu que les aspects appliqués changent souvent le comportement de leurs composantes de base [47]. Ces mêmes difficultés sont présentes au niveau des phases de spécification et de vérification du processus de développement des logiciels. Au meilleur de nos connaissances, la spécification modulaire et la vérification modulaire sont faiblement couvertes et constituent un champ de recherche très intéressant. De même, les interactions entre aspects est un sérieux problème dans la communauté des aspects. Pour faire face à ces problèmes, nous avons choisi d’utiliser la théorie des catégories et les techniques des spécifications algébriques. Pour apporter une solution aux problèmes ci-dessus cités, nous avons utilisé les travaux de Wiels [110] et d’autres contributions telles que celles décrites dans le livre [25]. Nous supposons que le système en développement est déjà décomposé en aspects et classes. La première contribution de notre thèse est l’extension des techniques des spécifications algébriques à la notion d’aspect. Deuxièmement, nous avons défini une logique, LA , qui est utilisée dans le corps des spécifications pour décrire le comportement de ces composantes. La troisième contribution consiste en la définition de l’opérateur de tissage qui correspond à la relation d’interconnexion entre les modules d’aspect et les modules de classe. La quatrième contribution concerne le développement d’un mécanisme de prévention qui permet de prévenir les interactions indésirables dans les systèmes orientés aspect.