976 resultados para quality estimation


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Introduction: Recently developed portable dental X-ray units increase the mobility of the forensic odontologists and allow more efficient X-ray work in a disaster field, especially when used in combination with digital sensors. This type of machines might also have potential for application in remote areas, military and humanitarian missions, dental care of patients with mobility limitation, as well as imaging in operating rooms. Objective: To evaluate radiographic image quality acquired by three portable X-ray devices in combination with four image receptors and to evaluate their medical physics parameters. Materials and methods: Images of five samples consisting of four teeth and one formalin-fixed mandible were acquired by one conventional wall-mounted X-ray unit, MinRay (R) 60/70 kVp, used as a clinical standard, and three portable dental X-ray devices: AnyRay (R) 60 kVp, Nomad (R) 60 kVp and Rextar (R) 70 kVp, in combination with a phosphor image plate (PSP), a CCD, or a CMOS sensor. Three observers evaluated images for standard image quality besides forensic diagnostic quality on a 4-point rating scale. Furthermore, all machines underwent tests for occupational as well as patient dosimetry. Results: Statistical analysis showed good quality imaging for all system, with the combination of Nomad (R) and PSP yielding the best score. A significant difference in image quality between the combination of the four X-ray devices and four sensors was established (p < 0.05). For patient safety, the exposure rate was determined and exit dose rates for MinRay (R) at 60 kVp, MinRay (R) at 70 kVp, AnyRay (R), Nomad (R) and Rextar (R) were 3.4 mGy/s, 4.5 mGy/s, 13.5 mGy/s, 3.8 mGy/s and 2.6 mGy/s respectively. The kVp of the AnyRay (R) system was the most stable, with a ripple of 3.7%. Short-term variations in the tube output of all the devices were less than 10%. AnyRay (R) presented higher estimated effective dose than other machines. Occupational dosimetry showed doses at the operator`s hand being lowest with protective shielding (Nomad (R): 0.1 mu Gy). It was also low while using remote control (distance > 1 m: Rextar (R) < 0.2 mu Gy, MinRay (R) < 0.1 mu Gy). Conclusions: The present study demonstrated the feasibility of three portable X-ray systems to be used for specific indications, based on acceptable image quality and sufficient accuracy of the machines and following the standard guidelines for radiation hygiene. (C) 2010 Elsevier Ireland Ltd. All rights reserved.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The aim of this paper was to estimate the return on investment in QMS (quality management systems) certification undertaken in Portuguese firms, according to the ISO 9000 series. A total of 426 certified Portuguese firms were surveyed. The response rate was 61.03 percent. The different payback periods were validated through statistical analysis and the relationship between expected and perceived payback periods was discussed. This study suggests that a firm’s sector of activity, size and degree of internationalization are related to the length of the investment in QMS certification recovery period. Furthermore, our findings suggest, that the time taken to obtain the certification is not directly related to the economic component of the certification. The majority of Portuguese firms (58.9%) took up to three years to recoup their investment and 35.5% of companies said they had not yet recovered the initial investment made. The recoup of investment was measured by the increase in the number of customers and consequent volume of deliveries, improved profitability and productivity of the company, improvement of competitive position and performance (cost savings), reduction in the number of external complaints and internal defects/scrap, achievement of some important clientele, among others. We compared our work to similar studies undertaken in other countries. This paper provides a contribution to the research related to the return on investment for costs related to the certification QMS according to ISO 9000. This paper provides a valuable contribution to the field and is one of the first studies to undertake this type of analysis in Portugal.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Os métodos clínicos que são realizados com recurso a tecnologias de imagiologia têm registado um aumento de popularidade nas últimas duas décadas. Os procedimentos tradicionais usados em cirurgia têm sido substituídos por métodos minimamente invasivos de forma a conseguir diminuir os custos associados e aperfeiçoar factores relacionados com a produtividade. Procedimentos clínicos modernos como a broncoscopia e a cardiologia são caracterizados por se focarem na minimização de acções invasivas, com os arcos em ‘C’ a adoptarem um papel relevante nesta área. Apesar de o arco em ‘C’ ser uma tecnologia amplamente utilizada no auxílio da navegação em intervenções minimamente invasivas, este falha na qualidade da informação fornecida ao cirurgião. A informação obtida em duas dimensões não é suficiente para proporcionar uma compreensão total da localização tridimensional da região de interesse, revelando-se como uma tarefa essencial o estabelecimento de um método que permita a aquisição de informação tridimensional. O primeiro passo para alcançar este objectivo foi dado ao definir um método que permite a estimativa da posição e orientação de um objecto em relação ao arco em ‘C’. De forma a realizar os testes com o arco em ‘C’, a geometria deste teve que ser inicialmente definida e a calibração do sistema feita. O trabalho desenvolvido e apresentado nesta tese foca-se num método que provou ser suficientemente sustentável e eficiente para se estabelecer como um ponto de partida no caminho para alcançar o objectivo principal: o desenvolvimento de uma técnica que permita o aperfeiçoamento da qualidade da informação adquirida com o arco em ‘C’ durante uma intervenção clínica.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Proceedings of the International Conference on Computer Vision Theory and Applications, 361-365, 2013, Barcelona, Spain

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Ship tracking systems allow Maritime Organizations that are concerned with the Safety at Sea to obtain information on the current location and route of merchant vessels. Thanks to Space technology in recent years the geographical coverage of the ship tracking platforms has increased significantly, from radar based near-shore traffic monitoring towards a worldwide picture of the maritime traffic situation. The long-range tracking systems currently in operations allow the storage of ship position data over many years: a valuable source of knowledge about the shipping routes between different ocean regions. The outcome of this Master project is a software prototype for the estimation of the most operated shipping route between any two geographical locations. The analysis is based on the historical ship positions acquired with long-range tracking systems. The proposed approach makes use of a Genetic Algorithm applied on a training set of relevant ship positions extracted from the long-term storage tracking database of the European Maritime Safety Agency (EMSA). The analysis of some representative shipping routes is presented and the quality of the results and their operational applications are assessed by a Maritime Safety expert.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Lean meat percentage (LMP) is an important carcass quality parameter. The aim of this work is to obtain a calibration equation for the Computed Tomography (CT) scans with the Partial Least Square Regression (PLS) technique in order to predict the LMP of the carcass and the different cuts and to study and compare two different methodologies of the selection of the variables (Variable Importance for Projection — VIP- and Stepwise) to be included in the prediction equation. The error of prediction with cross-validation (RMSEPCV) of the LMP obtained with PLS and selection based on VIP value was 0.82% and for stepwise selection it was 0.83%. The prediction of the LMP scanning only the ham had a RMSEPCV of 0.97% and if the ham and the loin were scanned the RMSEPCV was 0.90%. Results indicate that for CT data both VIP and stepwise selection are good methods. Moreover the scanning of only the ham allowed us to obtain a good prediction of the LMP of the whole carcass.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

BACKGROUND: Estimated glomerular filtration rate (eGFR) is an important diagnostic instrument in clinical practice. The National Kidney Foundation-Kidney Disease Quality Initiative (NKF-KDOQI) guidelines do not recommend using formulas developed for adults to estimate GFR in children; however, studies confirming these recommendations are scarce. The aim of our study was to evaluate the accuracy of the new Chronic Kidney Disease Epidemiology Collaboration (CKD-EPI) formula, the Modification of Diet in Renal Disease (MDRD) formula, and the Cockcroft-Gault formula in children with various stages of chronic kidney disease (CKD). METHODS: A total of 550 inulin clearance (iGFR) measurements for 391 children were analyzed. The cohort was divided into three groups: group 1, with iGFR >90 ml/min/1.73 m(2); group 2, with iGFR between 60 and 90 ml/min/1.73 m(2); group 3, with iGFR of <60 ml/min/1.73 m(2). RESULTS: All formulas overestimate iGFR with a significant bias (p < 0.001), present poor accuracies, and have poor Spearman correlations. For an accuracy of 10 %, only 11, 6, and 27 % of the eGFRs are accurate when using the MDRD, CKD-EPI, and Cockcroft-Gault formulas, respectively. For an accuracy of 30 %, these formulas do not reach the NKF-KDOQI guidelines for validation, with only 25, 20, and 70 % of the eGFRs, respectively, being accurate. CONCLUSIONS: Based on our results, the performances of all of these formulas are unreliable for eGFR in children across all CKD stages and cannot therefore be applied in the pediatric population group.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

BACKGROUND The demographic structure has a significant influence on the use of healthcare services, as does the size of the population denominators. Very few studies have been published on methods for estimating the real population such as tourist resorts. The lack of information about these problems means there is a corresponding lack of information about the behaviour of populational denominators (the floating population or tourist load) and the effect of this on the use of healthcare services. The objectives of the study were: a) To determine the Municipal Solid Waste (MSW) ratio, per person per day, among populations of known size; b) to estimate, by means of this ratio, the real population in an area where tourist numbers are very significant; and c) to determine the impact on the utilisation of hospital emergency healthcare services of the registered population, in comparison to the non-resident population, in two areas where tourist numbers are very significant. METHODS An ecological study design was employed. We analysed the Healthcare Districts of the Costa del Sol and the island of Menorca. Both are Spanish territories in the Mediterranean region. RESULTS In the two areas analysed, the correlation coefficient between the MSW ratio and admissions to hospital emergency departments exceeded 0.9, with p < 0.001. On the basis of MSW generation ratios, obtained for a control zone and also measured in neighbouring countries, we estimated the real population. For the summer months, when tourist activity is greatest and demand for emergency healthcare at hospitals is highest, this value was found to be double that of the registered population. CONCLUSION The MSW indicator, which is both ecological and indirect, can be used to estimate the real population in areas where population levels vary significantly during the year. This parameter is of interest in planning and dimensioning the provision of healthcare services.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper empirically explores the link between quality and concentration in a cross-section of manufactured goods. Using concentration data and product quality indicators, an ordered probit estimation explores the impact of concentration on quality that is defined as an index of quality characteristics. The results demonstrate that market concentration and quality are positively correlated across different industries. When industry concentration increases, the likelihood of the product being higher quality increases and the likelihood of observing a lower quality decreases.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We study model selection strategies based on penalized empirical loss minimization. We point out a tight relationship between error estimation and data-based complexity penalization: any good error estimate may be converted into a data-based penalty function and the performance of the estimate is governed by the quality of the error estimate. We consider several penalty functions, involving error estimates on independent test data, empirical {\sc vc} dimension, empirical {\sc vc} entropy, andmargin-based quantities. We also consider the maximal difference between the error on the first half of the training data and the second half, and the expected maximal discrepancy, a closely related capacity estimate that can be calculated by Monte Carlo integration. Maximal discrepancy penalty functions are appealing for pattern classification problems, since their computation is equivalent to empirical risk minimization over the training data with some labels flipped.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The soil water available to crops is defined by specific values of water potential limits. Underlying the estimation of hydro-physical limits, identified as permanent wilting point (PWP) and field capacity (FC), is the selection of a suitable method based on a multi-criteria analysis that is not always clear and defined. In this kind of analysis, the time required for measurements must be taken into consideration as well as other external measurement factors, e.g., the reliability and suitability of the study area, measurement uncertainty, cost, effort and labour invested. In this paper, the efficiency of different methods for determining hydro-physical limits is evaluated by using indices that allow for the calculation of efficiency in terms of effort and cost. The analysis evaluates both direct determination methods (pressure plate - PP and water activity meter - WAM) and indirect estimation methods (pedotransfer functions - PTFs). The PTFs must be validated for the area of interest before use, but the time and cost associated with this validation are not included in the cost of analysis. Compared to the other methods, the combined use of PP and WAM to determine hydro-physical limits differs significantly in time and cost required and quality of information. For direct methods, increasing sample size significantly reduces cost and time. This paper assesses the effectiveness of combining a general analysis based on efficiency indices and more specific analyses based on the different influencing factors, which were considered separately so as not to mask potential benefits or drawbacks that are not evidenced in efficiency estimation.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The knowledge of the relationship that links radiation dose and image quality is a prerequisite to any optimization of medical diagnostic radiology. Image quality depends, on the one hand, on the physical parameters such as contrast, resolution, and noise, and on the other hand, on characteristics of the observer that assesses the image. While the role of contrast and resolution is precisely defined and recognized, the influence of image noise is not yet fully understood. Its measurement is often based on imaging uniform test objects, even though real images contain anatomical backgrounds whose statistical nature is much different from test objects used to assess system noise. The goal of this study was to demonstrate the importance of variations in background anatomy by quantifying its effect on a series of detection tasks. Several types of mammographic backgrounds and signals were examined by psychophysical experiments in a two-alternative forced-choice detection task. According to hypotheses concerning the strategy used by the human observers, their signal to noise ratio was determined. This variable was also computed for a mathematical model based on the statistical decision theory. By comparing theoretical model and experimental results, the way that anatomical structure is perceived has been analyzed. Experiments showed that the observer's behavior was highly dependent upon both system noise and the anatomical background. The anatomy partly acts as a signal recognizable as such and partly as a pure noise that disturbs the detection process. This dual nature of the anatomy is quantified. It is shown that its effect varies according to its amplitude and the profile of the object being detected. The importance of the noisy part of the anatomy is, in some situations, much greater than the system noise. Hence, reducing the system noise by increasing the dose will not improve task performance. This observation indicates that the tradeoff between dose and image quality might be optimized by accepting a higher system noise. This could lead to a better resolution, more contrast, or less dose.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The goal of this study was to investigate the impact of computing parameters and the location of volumes of interest (VOI) on the calculation of 3D noise power spectrum (NPS) in order to determine an optimal set of computing parameters and propose a robust method for evaluating the noise properties of imaging systems. Noise stationarity in noise volumes acquired with a water phantom on a 128-MDCT and a 320-MDCT scanner were analyzed in the spatial domain in order to define locally stationary VOIs. The influence of the computing parameters in the 3D NPS measurement: the sampling distances bx,y,z and the VOI lengths Lx,y,z, the number of VOIs NVOI and the structured noise were investigated to minimize measurement errors. The effect of the VOI locations on the NPS was also investigated. Results showed that the noise (standard deviation) varies more in the r-direction (phantom radius) than z-direction plane. A 25 × 25 × 40 mm(3) VOI associated with DFOV = 200 mm (Lx,y,z = 64, bx,y = 0.391 mm with 512 × 512 matrix) and a first-order detrending method to reduce structured noise led to an accurate NPS estimation. NPS estimated from off centered small VOIs had a directional dependency contrary to NPS obtained from large VOIs located in the center of the volume or from small VOIs located on a concentric circle. This showed that the VOI size and location play a major role in the determination of NPS when images are not stationary. This study emphasizes the need for consistent measurement methods to assess and compare image quality in CT.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Aims: The psychometric properties of the EORTC QLQ-BN20, a brain cancer-specific HRQOL questionnaire, have been previously determined in an English-speaking sample of patients. This study examined the validity and reliability of the questionnaire in a multi-national, multi-lingual study. Methods: QLQ-BN20 data were selected from two completed phase III EORTC/NCIC clinical trials in brain cancer (N=891), including 12 languages. Experimental treatments were surgery followed by radiotherapy (RT) and adjuvant PCV chemotherapy or surgery followed by concomitant RT plus temozolomide (TMZ) chemotherapy and adjuvant TMZ chemotherapy. Standard treatment consisted of surgery and postoperative RT alone. The psychometrics of the QLQ-BN20 were examined by means of multi-trait scaling analyses, reliability estimation, known groups validity testing, and responsiveness analysis. Results: All QLQ-BN20 items correlated more strongly with their own scale (r>0.70) than with other QLQ-BN20 scales. Internal consistency reliability coefficients were high (all alpha0.70). Known-groups comparisons yielded positive results, with the QLQ-BN20 distinguishing between patients with differing levels of performance status and mental functioning. Responsiveness of the questionnaire to changes over time was acceptable. Conclusion: The QLQ-BN20 demonstrates adequate psychometric properties and can be recommended for use in conjunction with the QLQ-C30 in assessing the HRQOL of brain cancer patients in international studies.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Software engineering is criticized as not being engineering or 'well-developed' science at all. Software engineers seem not to know exactly how long their projects will last, what they will cost, and will the software work properly after release. Measurements have to be taken in software projects to improve this situation. It is of limited use to only collect metrics afterwards. The values of the relevant metrics have to be predicted, too. The predictions (i.e. estimates) form the basis for proper project management. One of the most painful problems in software projects is effort estimation. It has a clear and central effect on other project attributes like cost and schedule, and to product attributes like size and quality. Effort estimation can be used for several purposes. In this thesis only the effort estimation in software projects for project management purposes is discussed. There is a short introduction to the measurement issues, and some metrics relevantin estimation context are presented. Effort estimation methods are covered quite broadly. The main new contribution in this thesis is the new estimation model that has been created. It takes use of the basic concepts of Function Point Analysis, but avoids the problems and pitfalls found in the method. It is relativelyeasy to use and learn. Effort estimation accuracy has significantly improved after taking this model into use. A major innovation related to the new estimationmodel is the identified need for hierarchical software size measurement. The author of this thesis has developed a three level solution for the estimation model. All currently used size metrics are static in nature, but this new proposed metric is dynamic. It takes use of the increased understanding of the nature of the work as specification and design work proceeds. It thus 'grows up' along with software projects. The effort estimation model development is not possible without gathering and analyzing history data. However, there are many problems with data in software engineering. A major roadblock is the amount and quality of data available. This thesis shows some useful techniques that have been successful in gathering and analyzing the data needed. An estimation process is needed to ensure that methods are used in a proper way, estimates are stored, reported and analyzed properly, and they are used for project management activities. A higher mechanism called measurement framework is also introduced shortly. The purpose of the framework is to define and maintain a measurement or estimationprocess. Without a proper framework, the estimation capability of an organization declines. It requires effort even to maintain an achieved level of estimationaccuracy. Estimation results in several successive releases are analyzed. It isclearly seen that the new estimation model works and the estimation improvementactions have been successful. The calibration of the hierarchical model is a critical activity. An example is shown to shed more light on the calibration and the model itself. There are also remarks about the sensitivity of the model. Finally, an example of usage is shown.