977 resultados para cost estimation


Relevância:

30.00% 30.00%

Publicador:

Resumo:

The objective of this paper is to introduce a fourth-order cost function of the displaced frame difference (DFD) capable of estimatingmotion even for small regions or blocks. Using higher than second-orderstatistics is appropriate in case the image sequence is severely corruptedby additive Gaussian noise. Some results are presented and compared to those obtained from the mean kurtosis and the mean square error of the DFD.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Tämä diplomityö käsittelee työkaluja, jotka on suunniteltu kustannusten ennakointiin ja hinnan asetantaan. Aluksi on käyty läpi perinteisen ja toimintoperusteisen kustannuslaskennan perusteita. Näiden menetelmien välisiä eroja on tarkasteltu ja toimintoperusteisen kustannuslaskennan paremmin sopivuus nykypäivän yrityksille on perusteltu. Toisena käsitellään hinnoittelu. Hinnan merkitys, hinnoittelumenetelmät ja päätös lopullisesta hinnasta on käyty läpi. Hinnoittelun jälkeen esitellään kustannusjärjestelmät ja kustannusten arviointi. Nämä asiat todistavat, että tarkat kustannusarviot ovat elintärkeitä yritykselle. Tuotteen kustannusarviointi, hinnan asetanta ja tarjoaminen ovat erittäin merkityksellisiä asioita ottaen huomioon koko projektin elinkaaren ja tulevat tuotot. Nykyään on yleistä käyttää työkaluja kustannusarvioinnissa ja joskus myös hinnoittelussa. Työkalujen luotettavuus on tiedettävä, ennenkuin työkalut otetaan käyttöön. Myös työkalujen käyttäjät täytyy perehdyttää hyvin. Muuten yritys todennäköisesti kohtaa odottamattomia ja epämiellyttäviä yllätyksiä.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Software engineering is criticized as not being engineering or 'well-developed' science at all. Software engineers seem not to know exactly how long their projects will last, what they will cost, and will the software work properly after release. Measurements have to be taken in software projects to improve this situation. It is of limited use to only collect metrics afterwards. The values of the relevant metrics have to be predicted, too. The predictions (i.e. estimates) form the basis for proper project management. One of the most painful problems in software projects is effort estimation. It has a clear and central effect on other project attributes like cost and schedule, and to product attributes like size and quality. Effort estimation can be used for several purposes. In this thesis only the effort estimation in software projects for project management purposes is discussed. There is a short introduction to the measurement issues, and some metrics relevantin estimation context are presented. Effort estimation methods are covered quite broadly. The main new contribution in this thesis is the new estimation model that has been created. It takes use of the basic concepts of Function Point Analysis, but avoids the problems and pitfalls found in the method. It is relativelyeasy to use and learn. Effort estimation accuracy has significantly improved after taking this model into use. A major innovation related to the new estimationmodel is the identified need for hierarchical software size measurement. The author of this thesis has developed a three level solution for the estimation model. All currently used size metrics are static in nature, but this new proposed metric is dynamic. It takes use of the increased understanding of the nature of the work as specification and design work proceeds. It thus 'grows up' along with software projects. The effort estimation model development is not possible without gathering and analyzing history data. However, there are many problems with data in software engineering. A major roadblock is the amount and quality of data available. This thesis shows some useful techniques that have been successful in gathering and analyzing the data needed. An estimation process is needed to ensure that methods are used in a proper way, estimates are stored, reported and analyzed properly, and they are used for project management activities. A higher mechanism called measurement framework is also introduced shortly. The purpose of the framework is to define and maintain a measurement or estimationprocess. Without a proper framework, the estimation capability of an organization declines. It requires effort even to maintain an achieved level of estimationaccuracy. Estimation results in several successive releases are analyzed. It isclearly seen that the new estimation model works and the estimation improvementactions have been successful. The calibration of the hierarchical model is a critical activity. An example is shown to shed more light on the calibration and the model itself. There are also remarks about the sensitivity of the model. Finally, an example of usage is shown.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Competition in airline markets may be tough. In this context, network carriers have two alternative strategies to compete with low-cost carriers. First, they may establish a low-cost subsidiary. Second, they may try to reduce costs using the main brand. This paper examines a successful strategy of the first type implemented by Iberia in the Spanish domestic market. Our analysis of data and the estimation of a pricing equation show that Iberia has been able to charge lower prices than rivals with its low-cost subsidiary. The pricing policy of the Spanish network carrier has been particularly aggressive in less dense routes and shorter routes.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Contexte : Parmi les infections nosocomiales, le Staphylocoque méticilline résistant (MRSA) est le germe pathogène le plus couramment identifié dans les hôpitaux du monde entier. La stratégie de contrôle des MRSA au CHUV implique le dépistage des patients à risque. Avec la méthode de dépistage par culture, le temps d'attente est de plusieurs jours. Ceci occasionne des problèmes dans la gestion des flux des patients, principalement à cause des mesures d'isolement. Pour réduire le temps d'attente, l'hôpital envisage d'utiliser une méthode de diagnostic rapide par "polymerase chain reaction" (PCR). Méthodologie : Les données concernant les dépistages réalisés, dans trois services durant l'année 2007, ont été utilisées. Le nombre de jours d'isolement a d'abord été déterminé par patient et par service. Ensuite une analyse des coûts a été effectuée afin d'évaluer la différence des coûts entre les deux méthodes pour chaque service. Résultats : Le principal impact économique de la méthode par PCR dépend principalement du nombre de jours d'isolements évités par rapport à la méthode de culture. Aux services de soins, l'analyse a été menée sur 192 dépistages. Quand la différence de jours d'isolement est de deux jours, le coût des dépistages diminue de plus de 12kCHF et le nombre de jours d'isolement diminue de 384 jours. Au centre interdisciplinaire des urgences, sur 96 dépistages, le gain potentiel avec la méthode PCR est de 6kCHF avec une diminution de 192 jours d'isolement. Aux soins intensifs adultes, la méthode de dépistage par PCR est la méthode la plus rentable avec une diminution des coûts entre 4KCHF et 20K CHF et une diminution des jours d'isolement entre 170 et 310. Pour les trois services analysés, les résultats montrent un rapport coût-efficacité favorable pour la méthode PCR lorsque la diminution des jours d'isolement est supérieure à 1.3 jour. Quand la différence de jours d'isolement est inférieure à 1.3, il faut tenir compte d'autres paramètres, comme le coût de matériel qui doit être supérieur à 45.5 CHF, et du nombre d'analyses par dépistage, qui doit être inférieur à 3, pour que la PCR reste l'alternative la plus intéressante. Conclusions : La méthode par PCR montre des avantages potentiels importants, tant économiques qu'organisationnels qui limitent ou diminuent les contraintes liées à la stratégie de contrôle des MRSA au CHUV. [Auteure, p. 3]

Relevância:

30.00% 30.00%

Publicador:

Resumo:

[Table des matières] 1. Pourquoi s'intéresser à l'occupation inappropriée des lits de soins aigus au CHUV ?. - 1.1. Etat des lieux. - 1.1.1. Les chiffres du CHUV. - 1.1.2. La cellule de gestion des flux de patients. - 1.1.3. L'unité de patients en attente de placement. - 1.1.4. La pénurie de lits dans les EMS vaudois. - 1.1.5. Le vieillissement de la population vaudoise. - 1.2. Evidences nationales et internationales. - - 2. Estimation des coûts. - 2.1. Coûts chiffrables. - 2.1.1. Perte financière directe. - 2.1.2. Coûts des transferts pour engorgement. - 2.1.3. Coût d'opportunité. - 2.2. Coûts non chiffrables. - 2.2.1. Patients. - 2.2.2. Personnel médical. - 2.2.3. CHUV. - - 3. Propositions. - 3.1. Prises en charge alternatives. - 3.1.1. Les réseaux intégrés de services aux personnes âgées. - 3.1.2. Les courts séjours gériatriques. - 3.1.3. Autres solutions. - 3.2. Prévention. - 3.2.1. Prévention des chutes. - 3.2.2. La prévention par l'information aux personnes âgées. - 3.2.3. La prévention par l'information à l'ensemble de la population

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Sensor-based robot control allows manipulation in dynamic environments with uncertainties. Vision is a versatile low-cost sensory modality, but low sample rate, high sensor delay and uncertain measurements limit its usability, especially in strongly dynamic environments. Force is a complementary sensory modality allowing accurate measurements of local object shape when a tooltip is in contact with the object. In multimodal sensor fusion, several sensors measuring different modalities are combined to give a more accurate estimate of the environment. As force and vision are fundamentally different sensory modalities not sharing a common representation, combining the information from these sensors is not straightforward. In this thesis, methods for fusing proprioception, force and vision together are proposed. Making assumptions of object shape and modeling the uncertainties of the sensors, the measurements can be fused together in an extended Kalman filter. The fusion of force and visual measurements makes it possible to estimate the pose of a moving target with an end-effector mounted moving camera at high rate and accuracy. The proposed approach takes the latency of the vision system into account explicitly, to provide high sample rate estimates. The estimates also allow a smooth transition from vision-based motion control to force control. The velocity of the end-effector can be controlled by estimating the distance to the target by vision and determining the velocity profile giving rapid approach and minimal force overshoot. Experiments with a 5-degree-of-freedom parallel hydraulic manipulator and a 6-degree-of-freedom serial manipulator show that integration of several sensor modalities can increase the accuracy of the measurements significantly.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Background: Assessing of the costs of treating disease is necessary to demonstrate cost-effectiveness and to estimate the budget impact of new interventions and therapeutic innovations. However, there are few comprehensive studies on resource use and costs associated with lung cancer patients in clinical practice in Spain or internationally. The aim of this paper was to assess the hospital cost associated with lung cancer diagnosis and treatment by histology, type of cost and stage at diagnosis in the Spanish National Health Service. Methods: A retrospective, descriptive analysis on resource use and a direct medical cost analysis were performed. Resource utilisation data were collected by means of patient files from nine teaching hospitals. From a hospital budget impact perspective, the aggregate and mean costs per patient were calculated over the first three years following diagnosis or up to death. Both aggregate and mean costs per patient were analysed by histology, stage at diagnosis and cost type. Results: A total of 232 cases of lung cancer were analysed, of which 74.1% corresponded to non-small cell lung cancer (NSCLC) and 11.2% to small cell lung cancer (SCLC); 14.7% had no cytohistologic confirmation. The mean cost per patient in NSCLC ranged from 13,218 Euros in Stage III to 16,120 Euros in Stage II. The main cost components were chemotherapy (29.5%) and surgery (22.8%). Advanced disease stages were associated with a decrease in the relative weight of surgical and inpatient care costs but an increase in chemotherapy costs. In SCLC patients, the mean cost per patient was 15,418 Euros for limited disease and 12,482 Euros for extensive disease. The main cost components were chemotherapy (36.1%) and other inpatient costs (28.7%). In both groups, the Kruskall-Wallis test did not show statistically significant differences in mean cost per patient between stages. Conclusions: This study provides the costs of lung cancer treatment based on patient file reviews, with chemotherapy and surgery accounting for the major components of costs. This cost analysis is a baseline study that will provide a useful source of information for future studies on cost-effectiveness and on the budget impact of different therapeutic innovations in Spain.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In the current economy situation companies try to reduce their expenses. One of the solutions is to improve the energy efficiency of the processes. It is known that the energy consumption of pumping applications range from 20 up to 50% of the energy usage in the certain industrial plants operations. Some studies have shown that 30% to 50% of energy consumed by pump systems could be saved by changing the pump or the flow control method. The aim of this thesis is to create a mobile measurement system that can calculate a working point position of a pump drive. This information can be used to determine the efficiency of the pump drive operation and to develop a solution to bring pump’s efficiency to a maximum possible value. This can allow a great reduction in the pump drive’s life cycle cost. In the first part of the thesis, a brief introduction in the details of pump drive operation is given. Methods that can be used in the project are presented. Later, the review of available platforms for the project implementation is given. In the second part of the thesis, components of the project are presented. Detailed description for each created component is given. Finally, results of laboratory tests are presented. Acquired results are compared and analyzed. In addition, the operation of created system is analyzed and suggestions for the future development are given.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The determination of volumetric water content of soils is an important factor in irrigation management. Among the indirect methods for estimating, the time-domain reflectometry (TDR) technique has received a significant attention. Like any other technique, it has advantages and disadvantages, but its greatest disadvantage is the need of calibration and high cost of acquisition. The main goal of this study was to establish a calibration model for the TDR equipment, Trase System Model 6050X1, to estimate the volumetric water content in a Distroferric Red Latosol. The calibration was carried out in a laboratory with disturbed soil samples under study, packed in PVC columns of a volume of 0.0078m³. The TDR probes were handcrafted with three rods and 0.20m long. They were vertically installed in soil columns, with a total of five probes per column and sixteen columns. The weightings were carried out in a digital scale, while daily readings of dielectric constant were obtained in TDR equipment. The linear model θν = 0.0103 Ka + 0.1900 to estimate the studied volumetric water content showed an excellent coefficient of determination (0.93), enabling the use of probes in indirect estimation of soil moisture.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Design of flight control laws, verification of performance predictions, and the implementation of flight simulations are tasks that require a mathematical model of the aircraft dynamics. The dynamical models are characterized by coefficients (aerodynamic derivatives) whose values must be determined from flight tests. This work outlines the use of the Extended Kalman Filter (EKF) in obtaining the aerodynamic derivatives of an aircraft. The EKF shows several advantages over the more traditional least-square method (LS). Among these the most important are: there are no restrictions on linearity or in the form which the parameters appears in the mathematical model describing the system, and it is not required that these parameters be time invariant. The EKF uses the statistical properties of the process and the observation noise, to produce estimates based on the mean square error of the estimates themselves. Differently, the LS minimizes a cost function based on the plant output behavior. Results for the estimation of some longitudinal aerodynamic derivatives from simulated data are presented.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The recent emergence of low-cost RGB-D sensors has brought new opportunities for robotics by providing affordable devices that can provide synchronized images with both color and depth information. In this thesis, recent work on pose estimation utilizing RGBD sensors is reviewed. Also, a pose recognition system for rigid objects using RGB-D data is implemented. The implementation uses half-edge primitives extracted from the RGB-D images for pose estimation. The system is based on the probabilistic object representation framework by Detry et al., which utilizes Nonparametric Belief Propagation for pose inference. Experiments are performed on household objects to evaluate the performance and robustness of the system.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper presents a methodology for calculating the industrial equilibrium exchange rate, which is defined as the one enabling exporters of state-of-the-art manufactured goods to be competitive abroad. The first section highlights the causes and problems of overvalued exchange rates, particularly the Dutch disease issue, which is neutralized when the exchange rate strikes the industrial equilibrium level. This level is defined by the ratio between the unit labor cost in the country under consideration and in competing countries. Finally, the evolution of this exchange rate in the Brazilian economy is estimated.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

On s’intéresse ici aux erreurs de modélisation liées à l’usage de modèles de flammelette sous-maille en combustion turbulente non prémélangée. Le but de cette thèse est de développer une stratégie d’estimation d’erreur a posteriori pour déterminer le meilleur modèle parmi une hiérarchie, à un coût numérique similaire à l’utilisation de ces mêmes modèles. Dans un premier temps, une stratégie faisant appel à un estimateur basé sur les résidus pondérés est développée et testée sur un système d’équations d’advection-diffusion-réaction. Dans un deuxième temps, on teste la méthodologie d’estimation d’erreur sur un autre système d’équations, où des effets d’extinction et de réallumage sont ajoutés. Lorsqu’il n’y a pas d’advection, une analyse asymptotique rigoureuse montre l’existence de plusieurs régimes de combustion déjà observés dans les simulations numériques. Nous obtenons une approximation des paramètres de réallumage et d’extinction avec la courbe en «S», un graphe de la température maximale de la flamme en fonction du nombre de Damköhler, composée de trois branches et d’une double courbure. En ajoutant des effets advectifs, on obtient également une courbe en «S» correspondant aux régimes de combustion déjà identifiés. Nous comparons les erreurs de modélisation liées aux approximations asymptotiques dans les deux régimes stables et établissons une nouvelle hiérarchie des modèles en fonction du régime de combustion. Ces erreurs sont comparées aux estimations données par la stratégie d’estimation d’erreur. Si un seul régime stable de combustion existe, l’estimateur d’erreur l’identifie correctement ; si plus d’un régime est possible, on obtient une fac˛on systématique de choisir un régime. Pour les régimes où plus d’un modèle est approprié, la hiérarchie prédite par l’estimateur est correcte.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

So far, in the bivariate set up, the analysis of lifetime (failure time) data with multiple causes of failure is done by treating each cause of failure separately. with failures from other causes considered as independent censoring. This approach is unrealistic in many situations. For example, in the analysis of mortality data on married couples one would be interested to compare the hazards for the same cause of death as well as to check whether death due to one cause is more important for the partners’ risk of death from other causes. In reliability analysis. one often has systems with more than one component and many systems. subsystems and components have more than one cause of failure. Design of high-reliability systems generally requires that the individual system components have extremely high reliability even after long periods of time. Knowledge of the failure behaviour of a component can lead to savings in its cost of production and maintenance and. in some cases, to the preservation of human life. For the purpose of improving reliability. it is necessary to identify the cause of failure down to the component level. By treating each cause of failure separately with failures from other causes considered as independent censoring, the analysis of lifetime data would be incomplete. Motivated by this. we introduce a new approach for the analysis of bivariate competing risk data using the bivariate vector hazard rate of Johnson and Kotz (1975).