932 resultados para Regular Linear System


Relevância:

30.00% 30.00%

Publicador:

Resumo:

The characteristics of the traditional linear economic model are high consumption, high emission and low efficiency. Economic development is still largely at the expense of the environment and requires a natural resource investment. This can realize rapid economic development but resource depletion and environmental pollution become increasingly serious. In the 1990's a new economic model, circular economics, began to enter our vision. The circular economy maximizes production and minimizes the impact of economic activities on the ecological environment through organizing the activities through the closed-loop feedback cycle of "resources - production - renewable resource". Circular economy is a better way to solve the contradictions between the economic development and resource shortages. Developing circular economy has become the major strategic initiatives to achieving sustainable development in countries all over the world. The evaluation of the development of circular economics is a necessary step for regional circular economy development. Having a quantitative evaluation of circular economy can better monitor and reveal the contradictions and problems in the process of the development of recycling economy. This thesis will: 1) Create an evaluation model framework and new types of industries and 2) Make an evaluation of the Shanghai circular economy currently to analyze the situation of Shanghai in the development of circular economy. I will then propose suggestions about the structure and development of Shanghai circular economy.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Wireless sensor network is an emerging research topic due to its vast and ever-growing applications. Wireless sensor networks are made up of small nodes whose main goal is to monitor, compute and transmit data. The nodes are basically made up of low powered microcontrollers, wireless transceiver chips, sensors to monitor their environment and a power source. The applications of wireless sensor networks range from basic household applications, such as health monitoring, appliance control and security to military application, such as intruder detection. The wide spread application of wireless sensor networks has brought to light many research issues such as battery efficiency, unreliable routing protocols due to node failures, localization issues and security vulnerabilities. This report will describe the hardware development of a fault tolerant routing protocol for railroad pedestrian warning system. The protocol implemented is a peer to peer multi-hop TDMA based protocol for nodes arranged in a linear zigzag chain arrangement. The basic working of the protocol was derived from Wireless Architecture for Hard Real-Time Embedded Networks (WAHREN).

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Despite widespread use of species-area relationships (SARs), dispute remains over the most representative SAR model. Using data of small-scale SARs of Estonian dry grassland communities, we address three questions: (1) Which model describes these SARs best when known artifacts are excluded? (2) How do deviating sampling procedures (marginal instead of central position of the smaller plots in relation to the largest plot; single values instead of average values; randomly located subplots instead of nested subplots) influence the properties of the SARs? (3) Are those effects likely to bias the selection of the best model? Our general dataset consisted of 16 series of nested-plots (1 cm(2)-100 m(2), any-part system), each of which comprised five series of subplots located in the four corners and the centre of the 100-m(2) plot. Data for the three pairs of compared sampling designs were generated from this dataset by subsampling. Five function types (power, quadratic power, logarithmic, Michaelis-Menten, Lomolino) were fitted with non-linear regression. In some of the communities, we found extremely high species densities (including bryophytes and lichens), namely up to eight species in 1 cm(2) and up to 140 species in 100 m(2), which appear to be the highest documented values on these scales. For SARs constructed from nested-plot average-value data, the regular power function generally was the best model, closely followed by the quadratic power function, while the logarithmic and Michaelis-Menten functions performed poorly throughout. However, the relative fit of the latter two models increased significantly relative to the respective best model when the single-value or random-sampling method was applied, however, the power function normally remained far superior. These results confirm the hypothesis that both single-value and random-sampling approaches cause artifacts by increasing stochasticity in the data, which can lead to the selection of inappropriate models.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

At the 111th German Medical Assembly in May 2008 in Ulm, Germany, a public debate on rationing of health care performances was started. Since the money in the German health care system is not enough to provide every diagnostic or therapy for every patient as a coverage of the compulsory medical insurances, a lot of specific health care performances have been rationed during the last years not to be covered by the regular medical insurance any more, such as, e. g., PSA measurements in urology or IOP measurements in ophthalmology. In contrast to the health care system in Scandinavia, where rationing of health care performances is publicly documented by the government, no similar public statements exist in Germany. Due to this, it is left to physicians to explain to their patients the "hidden" rationing of public health care performances, which also leads to an increase in individual health care performances (IGeL in Germany) to be paid for privately by the patient. It is undoubtedly true that not all medically possible performances need to be paid for by the health insurance; however, an official determination of these "out of pocket" health care performances is necessary. Therefore, it was the aim herein to work out possible "stop" criteria--according to the Scandinavian system--for common eye diseases and consecutive therapies, which need not be paid for or only be paid after a delay by the health insurances.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper treats the problem of setting the inventory level and optimizing the buffer allocation of closed-loop flow lines operating under the constant-work-in-process (CONWIP) protocol. We solve a very large but simple linear program that models an entire simulation run of a closed-loop flow line in discrete time to determine a production rate estimate of the system. This approach introduced in Helber, Schimmelpfeng, Stolletz, and Lagershausen (2011) for open flow lines with limited buffer capacities is extended to closed-loop CONWIP flow lines. Via this method, both the CONWIP level and the buffer allocation can be optimized simultaneously. The first part of a numerical study deals with the accuracy of the method. In the second part, we focus on the relationship between the CONWIP inventory level and the short-term profit. The accuracy of the method turns out to be best for such configurations that maximize production rate and/or short-term profit.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Master production schedule (MPS) plays an important role in an integrated production planning system. It converts the strategic planning defined in a production plan into the tactical operation execution. The MPS is also known as a tool for top management to control over manufacture resources and becomes input of the downstream planning levels such as material requirement planning (MRP) and capacity requirement planning (CRP). Hence, inappropriate decision on the MPS development may lead to infeasible execution, which ultimately causes poor delivery performance. One must ensure that the proposed MPS is valid and realistic for implementation before it is released to real manufacturing system. In practice, where production environment is stochastic in nature, the development of MPS is no longer simple task. The varying processing time, random event such as machine failure is just some of the underlying causes of uncertainty that may be hardly addressed at planning stage so that in the end the valid and realistic MPS is tough to be realized. The MPS creation problem becomes even more sophisticated as decision makers try to consider multi-objectives; minimizing inventory, maximizing customer satisfaction, and maximizing resource utilization. This study attempts to propose a methodology for MPS creation which is able to deal with those obstacles. This approach takes into account uncertainty and makes trade off among conflicting multi-objectives at the same time. It incorporates fuzzy multi-objective linear programming (FMOLP) and discrete event simulation (DES) for MPS development.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We have developed an assay for single strand DNA or RNA detection which is based on the homo-DNA templated Staudinger reduction of the profluorophore rhodamine-azide. The assay is based on a three component system, consisting of a homo-DNA/DNA hybrid probe, a set of homo-DNA reporter strands and the target DNA or RNA. We present two different formats of the assay (Omega probe and linear probe) in which the linear probe was found to perform best with catalytic turnover of the reporter strands (TON: 8) and a match/mismatch discrimination of up to 19. The advantage of this system is that the reporting (homo-DNA) and sensing (DNA) domain are decoupled from each other since the two pairing systems are bioorthogonal. This allows independent optimization of either domain which may lead to higher selectivity in in vivo imaging.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

OBJECTIVES We sought to analyze the time course of atrial fibrillation (AF) episodes before and after circular plus linear left atrial ablation and the percentage of patients with complete freedom from AF after ablation by using serial seven-day electrocardiograms (ECGs). BACKGROUND The curative treatment of AF targets the pathophysiological corner stones of AF (i.e., the initiating triggers and/or the perpetuation of AF). The pathophysiological complexity of both may not result in an "all-or-nothing" response but may modify number and duration of AF episodes. METHODS In patients with highly symptomatic AF, circular plus linear ablation lesions were placed around the left and right pulmonary veins, between the two circles, and from the left circle to the mitral annulus using the electroanatomic mapping system. Repetitive continuous 7-day ECGs administered before and after catheter ablation were used for rhythm follow-up. RESULTS In 100 patients with paroxysmal (n = 80) and persistent (n = 20) AF, relative duration of time spent in AF significantly decreased over time (35 +/- 37% before ablation, 26 +/- 41% directly after ablation, and 10 +/- 22% after 12 months). Freedom from AF stepwise increased in patients with paroxysmal AF and after 12 months measured at 88% or 74% depending on whether 24-h ECG or 7-day ECG was used. Complete pulmonary vein isolation was demonstrated in <20% of the circular lesions. CONCLUSIONS The results obtained in patients with AF treated with circular plus linear left atrial lesions strongly indicate that substrate modification is the main underlying pathophysiologic mechanism and that it results in a delayed cure instead of an immediate cure.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

BACKGROUND Mechanical autotransfusion systems for washed shed blood (WSB) were introduced to reduce the need for postoperative allogenic blood transfusions (ABTs). Although some authors have postulated decreased requirements for ABT by using autologous retransfusion devices, other trials, mostly evaluating retransfusion devices for unwashed shed blood (USB), verified a small or no benefit in reducing the need for postoperative ABT. Because of these contradictory findings it is still unclear whether autologous retransfusion systems for WSB can reduce transfusion requirements. QUESTIONS/PURPOSES We therefore asked whether one such autologous transfusion system for WSB can reduce the requirements for postoperative ABT. METHODS In a prospective, randomized, controlled study, we enrolled 151 patients undergoing TKA. In Group A (n=76 patients), the autotransfusion system was used for a total of 6 hours (intraoperatively and postoperatively) and the WSB was retransfused after processing. In Control Group B (n=75 patients), a regular drain without suction was used. We used signs of anemia and/or a hemoglobin value less than 8 g/dL as indications for transfusion. If necessary, we administered one or two units of allogenic blood. RESULTS Twenty-three patients (33%) in Group A, who received an average of 283 mL (range, 160-406 mL) of salvaged blood, needed a mean of 2.1 units of allogenic blood, compared with 23 patients (33%) in Control Group B who needed a mean of 2.1 units of allogenic blood. CONCLUSIONS We found the use of an autotransfusion system did not reduce the rate of postoperative ABTs. LEVEL OF EVIDENCE Level II, therapeutic study. See the Guidelines for Authors for a complete description of levels of evidence.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Both historical and idealized climate model experiments are performed with a variety of Earth system models of intermediate complexity (EMICs) as part of a community contribution to the Intergovernmental Panel on Climate Change Fifth Assessment Report. Historical simulations start at 850 CE and continue through to 2005. The standard simulations include changes in forcing from solar luminosity, Earth's orbital configuration, CO2, additional greenhouse gases, land use, and sulphate and volcanic aerosols. In spite of very different modelled pre-industrial global surface air temperatures, overall 20th century trends in surface air temperature and carbon uptake are reasonably well simulated when compared to observed trends. Land carbon fluxes show much more variation between models than ocean carbon fluxes, and recent land fluxes appear to be slightly underestimated. It is possible that recent modelled climate trends or climate–carbon feedbacks are overestimated resulting in too much land carbon loss or that carbon uptake due to CO2 and/or nitrogen fertilization is underestimated. Several one thousand year long, idealized, 2 × and 4 × CO2 experiments are used to quantify standard model characteristics, including transient and equilibrium climate sensitivities, and climate–carbon feedbacks. The values from EMICs generally fall within the range given by general circulation models. Seven additional historical simulations, each including a single specified forcing, are used to assess the contributions of different climate forcings to the overall climate and carbon cycle response. The response of surface air temperature is the linear sum of the individual forcings, while the carbon cycle response shows a non-linear interaction between land-use change and CO2 forcings for some models. Finally, the preindustrial portions of the last millennium simulations are used to assess historical model carbon-climate feedbacks. Given the specified forcing, there is a tendency for the EMICs to underestimate the drop in surface air temperature and CO2 between the Medieval Climate Anomaly and the Little Ice Age estimated from palaeoclimate reconstructions. This in turn could be a result of unforced variability within the climate system, uncertainty in the reconstructions of temperature and CO2, errors in the reconstructions of forcing used to drive the models, or the incomplete representation of certain processes within the models. Given the forcing datasets used in this study, the models calculate significant land-use emissions over the pre-industrial period. This implies that land-use emissions might need to be taken into account, when making estimates of climate–carbon feedbacks from palaeoclimate reconstructions.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

An Advanced Planning System (APS) offers support at all planning levels along the supply chain while observing limited resources. We consider an APS for process industries (e.g. chemical and pharmaceutical industries) consisting of the modules network design (for long–term decisions), supply network planning (for medium–term decisions), and detailed production scheduling (for short–term decisions). For each module, we outline the decision problem, discuss the specifi cs of process industries, and review state–of–the–art solution approaches. For the module detailed production scheduling, a new solution approach is proposed in the case of batch production, which can solve much larger practical problems than the methods known thus far. The new approach decomposes detailed production scheduling for batch production into batching and batch scheduling. The batching problem converts the primary requirements for products into individual batches, where the work load is to be minimized. We formulate the batching problem as a nonlinear mixed–integer program and transform it into a linear mixed–binary program of moderate size, which can be solved by standard software. The batch scheduling problem allocates the batches to scarce resources such as processing units, workers, and intermediate storage facilities, where some regular objective function like the makespan is to be minimized. The batch scheduling problem is modelled as a resource–constrained project scheduling problem, which can be solved by an efficient truncated branch–and–bound algorithm developed recently. The performance of the new solution procedures for batching and batch scheduling is demonstrated by solving several instances of a case study from process industries.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A patient classification system was developed integrating a patient acuity instrument with a computerized nursing distribution method based on a linear programming model. The system was designed for real-time measurement of patient acuity (workload) and allocation of nursing personnel to optimize the utilization of resources.^ The acuity instrument was a prototype tool with eight categories of patients defined by patient severity and nursing intensity parameters. From this tool, the demand for nursing care was defined in patient points with one point equal to one hour of RN time. Validity and reliability of the instrument was determined as follows: (1) Content validity by a panel of expert nurses; (2) predictive validity through a paired t-test analysis of preshift and postshift categorization of patients; (3) initial reliability by a one month pilot of the instrument in a practice setting; and (4) interrater reliability by the Kappa statistic.^ The nursing distribution system was a linear programming model using a branch and bound technique for obtaining integer solutions. The objective function was to minimize the total number of nursing personnel used by optimally assigning the staff to meet the acuity needs of the units. A penalty weight was used as a coefficient of the objective function variables to define priorities for allocation of staff.^ The demand constraints were requirements to meet the total acuity points needed for each unit and to have a minimum number of RNs on each unit. Supply constraints were: (1) total availability of each type of staff and the value of that staff member (value was determined relative to that type of staff's ability to perform the job function of an RN (i.e., value for eight hours RN = 8 points, LVN = 6 points); (2) number of personnel available for floating between units.^ The capability of the model to assign staff quantitatively and qualitatively equal to the manual method was established by a thirty day comparison. Sensitivity testing demonstrated appropriate adjustment of the optimal solution to changes in penalty coefficients in the objective function and to acuity totals in the demand constraints.^ Further investigation of the model documented: correct adjustment of assignments in response to staff value changes; and cost minimization by an addition of a dollar coefficient to the objective function. ^

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Objective. Essential hypertension affects 25% of the US adult population and is a leading contributor to morbidity and mortality. Because BP is a multifactorial phenotype that resists simple genetic analysis, intermediate phenotypes within the complex network of BP regulatory systems may be more accessible to genetic dissection. The Renin-Angiotensin System (RAS) is known to influence intermediate and long-term blood pressure regulation through alterations in vascular tone and renal sodium and fluid resorption. This dissertation examines associations between renin (REN), angiotensinogen (AGT), angiotensin-converting enzyme (ACE) and angiotensin II type 1 receptor (AT1) gene variation and interindividual differences in plasma hormone levels, renal hemodynamics, and BP homeostasis.^ Methods. A total of 150 unrelated men and 150 unrelated women, between 20.0 and 49.9 years of age and free of acute or chronic illness except for a history of hypertension (11 men and 7 women, all off medications), were studied after one week on a controlled sodium diet. RAS plasma hormone levels, renal hemodynamics and BP were determined prior to and during angiotensin II (Ang II) infusion. Individuals were genotyped by PCR for a variable number tandem repeat (VNTR) polymorphism in REN, and for the following restriction fragment length polymorphisms (RFLP): AGT M235T, ACE I/D, and AT1 A1166C. Associations between clinical measurements and allelic variation were examined using multiple linear regression statistical models.^ Results. Women homozygous for the AT1 1166C allele demonstrated higher intracellular levels of sodium (p = 0.044). Men homozygous for the AGT T235 allele demonstrated a blunted decrement in renal plasma flow in response to Ang II infusion (p = 0.0002). There were no significant associations between RAS gene variation and interindividual variation in RAS plasma hormone levels or BP.^ Conclusions. Rather than identifying new BP controlling genes or alleles, the study paradigm employed in this thesis (i.e., measured genes, controlled environments and interventions) may provide mechanistic insight into how candidate genes affect BP homeostasis. ^

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Computer vision-based food recognition could be used to estimate a meal's carbohydrate content for diabetic patients. This study proposes a methodology for automatic food recognition, based on the Bag of Features (BoF) model. An extensive technical investigation was conducted for the identification and optimization of the best performing components involved in the BoF architecture, as well as the estimation of the corresponding parameters. For the design and evaluation of the prototype system, a visual dataset with nearly 5,000 food images was created and organized into 11 classes. The optimized system computes dense local features, using the scale-invariant feature transform on the HSV color space, builds a visual dictionary of 10,000 visual words by using the hierarchical k-means clustering and finally classifies the food images with a linear support vector machine classifier. The system achieved classification accuracy of the order of 78%, thus proving the feasibility of the proposed approach in a very challenging image dataset.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Within the context of exoplanetary atmospheres, we present a comprehensive linear analysis of forced, damped, magnetized shallow water systems, exploring the effects of dimensionality, geometry (Cartesian, pseudo-spherical, and spherical), rotation, magnetic tension, and hydrodynamic and magnetic sources of friction. Across a broad range of conditions, we find that the key governing equation for atmospheres and quantum harmonic oscillators are identical, even when forcing (stellar irradiation), sources of friction (molecular viscosity, Rayleigh drag, and magnetic drag), and magnetic tension are included. The global atmospheric structure is largely controlled by a single key parameter that involves the Rossby and Prandtl numbers. This near-universality breaks down when either molecular viscosity or magnetic drag acts non-uniformly across latitude or a poloidal magnetic field is present, suggesting that these effects will introduce qualitative changes to the familiar chevron-shaped feature witnessed in simulations of atmospheric circulation. We also find that hydrodynamic and magnetic sources of friction have dissimilar phase signatures and affect the flow in fundamentally different ways, implying that using Rayleigh drag to mimic magnetic drag is inaccurate. We exhaustively lay down the theoretical formalism (dispersion relations, governing equations, and time-dependent wave solutions) for a broad suite of models. In all situations, we derive the steady state of an atmosphere, which is relevant to interpreting infrared phase and eclipse maps of exoplanetary atmospheres. We elucidate a pinching effect that confines the atmospheric structure to be near the equator. Our suite of analytical models may be used to develop decisively physical intuition and as a reference point for three-dimensional magnetohydrodynamic simulations of atmospheric circulation.