895 resultados para cost model
Resumo:
In July 1999, the Swinfen Charitable Trust in the UK established a telemedicine link in Bangladesh, between the Centre for the Rehabilitation of the Paralysed (CRP) in Dhaka and medical consultants abroad. This low-cost telemedicine system used a digital camera to capture still images, which were then transmitted by email. During the first 12 months, 27 telemedicine referrals were made. The following specialties were consulted: neurology (44%), orthopaedics (40%), rheumatology (8%), nephrology (4%) and paediatrics (4%). Initial email replies were received at the CRP within a day of referral in 70% of cases and within thee days in 100%, which shows that store-and-forward telemedicine can be both fast and reliable. Telemedicine consultation was complete within three days in 14 cases (52%) and within three weeks in 24 cases (89%). Referral was judged to be beneficial in 24 cases (89%), the benefits including establishment of the diagnosis, the provision of reassurance to the patient and referring doctor, and a change of management. Four patients (15% of the total) and their families were spared the considerable expense and unnecessary stress of travelling abroad for a second opinion, and the savings from this alone outweighed the set-up and running costs in Bangladesh. The latter are limited to an email account with an Internet service provider and the local-rate telephone call charges from the CRP. This successful telemedicine system is a model for further telemedicine projects in the developing world.
Resumo:
In many occupational safety interventions, the objective is to reduce the injury incidence as well as the mean claims cost once injury has occurred. The claims cost data within a period typically contain a large proportion of zero observations (no claim). The distribution thus comprises a point mass at 0 mixed with a non-degenerate parametric component. Essentially, the likelihood function can be factorized into two orthogonal components. These two components relate respectively to the effect of covariates on the incidence of claims and the magnitude of claims, given that claims are made. Furthermore, the longitudinal nature of the intervention inherently imposes some correlation among the observations. This paper introduces a zero-augmented gamma random effects model for analysing longitudinal data with many zeros. Adopting the generalized linear mixed model (GLMM) approach reduces the original problem to the fitting of two independent GLMMs. The method is applied to evaluate the effectiveness of a workplace risk assessment teams program, trialled within the cleaning services of a Western Australian public hospital.
Resumo:
This paper examines why practitioners and researchers get different estimates of equity value when they use a discounted cash flow (CF) model versus a residual income (RI) model. Both models are derived from the same underlying assumption -- that price is the present value of expected future net dividends discounted at the cost of equity capital -- but in practice and in research they frequently yield different estimates. We argue that the research literature devoted to comparing the accuracy of these two models is misguided; properly implemented, both models yield identical valuations for all firms in all years. We identify how prior research has applied inconsistent assumptions to the two models and show how these seemingly small errors cause surprisingly large differences in the value estimates. [ABSTRACT FROM AUTHOR]
Resumo:
Management are keen to maximize the life span of an information system because of the high cost, organizational disruption, and risk of failure associated with the re-development or replacement of an information system. This research investigates the effects that various factors have on an information system's life span by understanding how the factors affect an information system's stability. The research builds on a previously developed two-stage model of information system change whereby an information system is either in a stable state of evolution in which the information system's functionality is evolving, or in a state of revolution, in which the information system is being replaced because it is not providing the functionality expected by its users. A case study surveyed a number of systems within one organization. The aim was to test whether a relationship existed between the base value of the volatility index (a measure of the stability of an information system) and certain system characteristics. Data relating to some 3000 user change requests covering 40 systems over a 10-year period were obtained. The following factors were hypothesized to have significant associations with the base value of the volatility index: language level (generation of language of construction), system size, system age, and the timing of changes applied to a system. Significant associations were found in the hypothesized directions except that the timing of user changes was not associated with any change in the value of the volatility index. Copyright (C) 2002 John Wiley Sons, Ltd.
Resumo:
This trial compared the cost of an integrated home-based care model with traditional inpatient care for acute chronic obstructive pulmonary disease (COPD). 25 patients with acute COPD were randomised to either home or hospital management following request for hospital admission. The acute care at home group costs per separation ($745, CI95% $595-$895, n = 13) were significantly lower (p < 0.01) than the hospital group ($2543, CI95% $1766-$3321, n = 12). There was an improvement in lung function in the hospital-managed group at the Outpatient Department review, decreased anxiety in the Emergency Department in the home-managed group and equal patient satisfaction with care delivery. Acute care at home schemes can substitute for usual hospital care for some patients without adverse effects, and potentially release resources. A funding model that allows adequate resource delivery to the community will be needed if there is a move to devolve acute care to community providers.
Resumo:
Loss of magnetic medium solids from dense medium circuits is a substantial contributor to operating cost. Much of this loss is by way of wet drum magnetic separator effluent. A model of the separator would be useful for process design, optimisation and control. A review of the literature established that although various rules of thumb exist, largely based on empirical or anecdotal evidence, there is no model of magnetics recovery in a wet drum magnetic separator which includes as inputs all significant machine and operating variables. A series of trials, in both factorial experiments and in single variable experiments, was therefore carried out using a purpose built rig which featured a small industrial scale (700 mm lip length, 900 mm diameter) wet drum magnetic separator. A substantial data set of 191 trials was generated in the work. The results of the factorial experiments were used to identify the variables having a significant effect on magnetics recovery. Observations carried out as an adjunct to this work, as well as magnetic theory, suggests that the capture of magnetic particles in the wet drum magnetic separator is by a flocculation process. Such a process should be defined by a flocculation rate and a flocculation time; the latter being defined by the volumetric flowrate and the volume within the separation zone. A model based on this concept and containing adjustable parameters was developed. This model was then fitted to a randomly chosen 80% of the data, and validated by application to the remaining 20%. The model is shown to provide a satisfactory fit to the data over three orders of magnitude of magnetics loss. (C) 2003 Elsevier Science BY. All rights reserved.
Resumo:
Rapid prototyping (RP) is an approach for automatically building a physical object through solid freeform fabrication. Nowadays, RP has become a vital aspect of most product development processes, due to the significant competitive advantages it offers compared to traditional manual model making. Even in academic environments, it is important to be able to quickly create accurate physical representations of concept solutions. Some of these can be used for simple visual validation, while others can be employed for ergonomic assessment by potential users or even for physical testing. However, the cost of traditional RP methods prevents their use in most academic environments on a regular basis, and even for very preliminary prototypes in many small companies. That results in delaying the first physical prototypes to later stages, or creating very rough mock-ups which are not as useful as they could be. In this paper we propose an approach for rapid and inexpensive model-making, which was developed in an academic context, and which can be employed for a variety of objects.
Resumo:
This paper explores the main determinants of the use of the cost accounting system (CAS) in Portuguese local government (PLG). Regression analysis is used to study the fit of a model of accounting changes in PLG, focused on cost accounting systems oriented to activities and outputs. Based on survey data gathered from PLG, we have found that the use of information in decision-making and external reporting is still a mirage. We obtain evidence about the influence of the internal organizational context (especially the lack of support and difficulties in the CAS implementation) in the use for internal purposes, while the institutional environment (like external pressures to implement the CAS) appears to be more deterministic of the external use. Results strengthen the function of external reporting to legitimate the organization’s activities to external stakeholders. On the other hand, some control variables (like political competition, usefulness and experience) also evidence some explanatory power in the model. Some mixed results were found that appeal to further research in the future. Our empirical results contribute to understand the importance of interconnecting the contingency and institutional approaches to gain a clear picture of cost accounting changes in the public sector.
Resumo:
In this paper a realistic directional channel model that is an extension of the COST 273 channel model is presented. The model uses a cluster of scatterers and visibility region generation based strategy with increased realism, due to the introduction of terrain and clutter information. New approaches for path-loss prediction and line of sight modeling are considered, affecting the cluster path gain model implementation. The new model was implemented using terrain, clutter, street and user mobility information for the city of Lisbon, Portugal. Some of the model's outputs are presented, mainly path loss and small/large-scale fading statistics.
Resumo:
Toxic amides, such as acrylamide, are potentially harmful to Human health, so there is great interest in the fabrication of compact and economical devices to measure their concentration in food products and effluents. The CHEmically Modified Field Effect Transistor (CHEMFET) based onamorphous silicon technology is a candidate for this type of application due to its low fabrication cost. In this article we have used a semi-empirical modelof the device to predict its performance in a solution of interfering ions. The actual semiconductor unit of the sensor was fabricated by the PECVD technique in the top gate configuration. The CHEMFET simulation was performed based on the experimental current voltage curves of the semiconductor unit and on an empirical model of the polymeric membrane. Results presented here are useful for selection and design of CHEMFET membranes and provide an idea of the limitations of the amorphous CHEMFET device. In addition to the economical advantage, the small size of this prototype means it is appropriate for in situ operation and integration in a sensor array.
Resumo:
We study a model consisting of particles with dissimilar bonding sites ("patches"), which exhibits self-assembly into chains connected by Y-junctions, and investigate its phase behaviour by both simulations and theory. We show that, as the energy cost epsilon(j) of forming Y-junctions increases, the extent of the liquid-vapour coexistence region at lower temperatures and densities is reduced. The phase diagram thus acquires a characteristic "pinched" shape in which the liquid branch density decreases as the temperature is lowered. To our knowledge, this is the first model in which the predicted topological phase transition between a fluid composed of short chains and a fluid rich in Y-junctions is actually observed. Above a certain threshold for epsilon(j), condensation ceases to exist because the entropy gain of forming Y-junctions can no longer offset their energy cost. We also show that the properties of these phase diagrams can be understood in terms of a temperature-dependent effective valence of the patchy particles. (C) 2011 American Institute of Physics. [doi: 10.1063/1.3605703]
Resumo:
Electricity markets are complex environments, involving a large number of different entities, playing in a dynamic scene to obtain the best advantages and profits. MASCEM is a multi-agent electricity market simulator to model market players and simulate their operation in the market. Market players are entities with specific characteristics and objectives, making their decisions and interacting with other players. MASCEM provides several dynamic strategies for agents’ behavior. This paper presents a method that aims to provide market players with strategic bidding capabilities, allowing them to obtain the higher possible gains out of the market. This method uses a reinforcement learning algorithm to learn from experience how to choose the best from a set of possible bids. These bids are defined accordingly to the cost function that each producer presents.
Resumo:
This paper presents a distributed model predictive control (DMPC) for indoor thermal comfort that simultaneously optimizes the consumption of a limited shared energy resource. The control objective of each subsystem is to minimize the heating/cooling energy cost while maintaining the indoor temperature and used power inside bounds. In a distributed coordinated environment, the control uses multiple dynamically decoupled agents (one for each subsystem/house) aiming to achieve satisfaction of coupling constraints. According to the hourly power demand profile, each house assigns a priority level that indicates how much is willing to bid in auction for consume the limited clean resource. This procedure allows the bidding value vary hourly and consequently, the agents order to access to the clean energy also varies. Despite of power constraints, all houses have also thermal comfort constraints that must be fulfilled. The system is simulated with several houses in a distributed environment.
Resumo:
In general, modern networks are analysed by taking several Key Performance Indicators (KPIs) into account, their proper balance being required in order to guarantee a desired Quality of Service (QoS), particularly, cellular wireless heterogeneous networks. A model to integrate a set of KPIs into a single one is presented, by using a Cost Function that includes these KPIs, providing for each network node a single evaluation parameter as output, and reflecting network conditions and common radio resource management strategies performance. The proposed model enables the implementation of different network management policies, by manipulating KPIs according to users' or operators' perspectives, allowing for a better QoS. Results show that different policies can in fact be established, with a different impact on the network, e.g., with median values ranging by a factor higher than two.
Resumo:
The problem of selecting suppliers/partners is a crucial and important part in the process of decision making for companies that intend to perform competitively in their area of activity. The selection of supplier/partner is a time and resource-consuming task that involves data collection and a careful analysis of the factors that can positively or negatively influence the choice. Nevertheless it is a critical process that affects significantly the operational performance of each company. In this work, there were identified five broad selection criteria: Quality, Financial, Synergies, Cost, and Production System. Within these criteria, it was also included five sub-criteria. After the identification criteria, a survey was elaborated and companies were contacted in order to understand which factors have more weight in their decisions to choose the partners. Interpreted the results and processed the data, it was adopted a model of linear weighting to reflect the importance of each factor. The model has a hierarchical structure and can be applied with the Analytic Hierarchy Process (AHP) method or Value Analysis. The goal of the paper it's to supply a selection reference model that can represent an orientation/pattern for a decision making on the suppliers/partners selection process