925 resultados para Multi factor affine processes
Resumo:
This paper describes the implementation of a multi-interface module (I2M) for automation of industrial processes, based on the IEEE1451 standard. Process automation with I2M can communicate through either wires or using wireless communication, without any hardware or software changes. We used FPGA resources to implement the I2M functions FPGA, with a NIOS II processor and ZigBee communication system (IEEE802.15), as well as RS232 serial standard. Part of the project was done in the SOPC Builder environment, which gave the designer flexibility and speed to implement the NIOS II-based microprocessor system. To test the I2M implementation, a didactic Industrial Hydraulic Module (MHI-01) was used to simulate two industrial processes to be controlled by the system proposed.
Resumo:
Over the last few years, Business Process Management (BPM) has achieved increasing popularity and dissemination. An analysis of the underlying assumptions of BPM shows that it pursues two apparently contradicting goals: on the one hand it aims at formalising work practices into business process models; on the other hand, it intends to confer flexibility to the organization - i.e. to maintain its ability to respond to new and unforeseen situations. This paper analyses the relationship between formalisation and flexibility in business process modelling by means of an empirical case study of a BPM project in an aircraft maintenance company. A qualitative approach is adopted based on the Actor-Network Theory. The paper offers two major contributions: (a) it illustrates the sociotechnical complexity involved in BPM initiatives; (b) it points towards a multidimensional understanding of the relation between formalization and flexibility in BPM projects.
Resumo:
Introduction. Patients with terminal heart failure have increased more than the available organs leading to a high mortality rate on the waiting list. Use of Marginal and expanded criteria donors has increased due to the heart shortage. Objective. We analyzed all heart transplantations (HTx) in Sao Paulo state over 8 years for donor profile and recipient risk factors. Method. This multi-institutional review collected HTx data from all institutions in the state of Sao Paulo, Brazil. From 2002 to 2008 (6 years), only 512 (28.8%) of 1777 available heart donors were accepted for transplantation. All medical records were analyzed retrospectively; none of the used donors was excluded, even those considered to be nonstandard. Results. The hospital mortality rate was 27.9% (n = 143) and the average follow-up time was 29.4 +/- 28.4 months. The survival rate was 55.5% (n = 285) at 6 years after HTx. Univariate analysis showed the following factors to impact survival: age (P = .0004), arterial hypertension (P = .4620), norepinephrine (P = .0450), cardiac arrest (P = .8500), diabetes mellitus (P = .5120), infection (P = .1470), CKMB (creatine kinase MB) (P = .8694), creatinine (P = .7225), and Na+ (P = .3273). On multivariate analysis, only age showed significance; logistic regression showed a significant cut-off at 40 years: organs from donors older than 40 years showed a lower late survival rates (P = .0032). Conclusions. Donor age older than 40 years represents an important risk factor for survival after HTx. Neither donor gender nor norepinephrine use negatively affected early survival.
Resumo:
Stylolites are rough paired surfaces, indicative of localized stress-induced dissolution under a non-hydrostatic state of stress, separated by a clay parting which is believed to be the residuum of the dissolved rock. These structures are the most frequent deformation pattern in monomineralic rocks and thus provide important information about low temperature deformation and mass transfer. The intriguing roughness of stylolites can be used to assess amount of volume loss and paleo-stress directions, and to infer the destabilizing processes during pressure solution. But there is little agreement on how stylolites form and why these localized pressure solution patterns develop their characteristic roughness.rnNatural bedding parallel and vertical stylolites were studied in this work to obtain a quantitative description of the stylolite roughness and understand the governing processes during their formation. Adapting scaling approaches based on fractal principles it is demonstrated that stylolites show two self affine scaling regimes with roughness exponents of 1.1 and 0.5 for small and large length scales separated by a crossover length at the millimeter scale. Analysis of stylolites from various depths proved that this crossover length is a function of the stress field during formation, as analytically predicted. For bedding parallel stylolites the crossover length is a function of the normal stress on the interface, but vertical stylolites show a clear in-plane anisotropy of the crossover length owing to the fact that the in-plane stresses (σ2 and σ3) are dissimilar. Therefore stylolite roughness contains a signature of the stress field during formation.rnTo address the origin of stylolite roughness a combined microstructural (SEM/EBSD) and numerical approach is employed. Microstructural investigations of natural stylolites in limestones reveal that heterogeneities initially present in the host rock (clay particles, quartz grains) are responsible for the formation of the distinctive stylolite roughness. A two-dimensional numerical model, i.e. a discrete linear elastic lattice spring model, is used to investigate the roughness evolving from an initially flat fluid filled interface induced by heterogeneities in the matrix. This model generates rough interfaces with the same scaling properties as natural stylolites. Furthermore two coinciding crossover phenomena in space and in time exist that separate length and timescales for which the roughening is either balanced by surface or elastic energies. The roughness and growth exponents are independent of the size, amount and the dissolution rate of the heterogeneities. This allows to conclude that the location of asperities is determined by a polimict multi-scale quenched noise, while the roughening process is governed by inherent processes i.e. the transition from a surface to an elastic energy dominated regime.rn
Resumo:
During the last decade, a multi-modal approach has been established in human experimental pain research for assessing pain thresholds and responses to various experimental pain modalities. Studies have concluded that differences in responses to pain stimuli are mainly related to variation between individuals rather than variation in response to different stimulus modalities. In a factor analysis of 272 consecutive volunteers (137 men and 135 women) who underwent tests with different experimental pain modalities, it was determined whether responses to different pain modalities represent distinct individual uncorrelated dimensions of pain perception. Volunteers underwent single painful electrical stimulation, repeated painful electrical stimulation (temporal summation), test for reflex receptive field, pressure pain stimulation, heat pain stimulation, cold pain stimulation, and a cold pressor test (ice water test). Five distinct factors were found representing responses to 5 distinct experimental pain modalities: pressure, heat, cold, electrical stimulation, and reflex-receptive fields. Each of the factors explained approximately 8% to 35% of the observed variance, and the 5 factors cumulatively explained 94% of the variance. The correlation between the 5 factors was near null (median ρ=0.00, range -0.03 to 0.05), with 95% confidence intervals for pairwise correlations between 2 factors excluding any relevant correlation. Results were almost similar for analyses stratified according to gender and age. Responses to different experimental pain modalities represent different specific dimensions and should be assessed in combination in future pharmacological and clinical studies to represent the complexity of nociception and pain experience.
Resumo:
Previous studies have shown that collective property rights offer higher flexibility than individual property and improve sustainable community-based forest management. Our case study, carried out in the Beni department of Bolivia, does not contradict this assertion, but shows that collective rights have been granted in areas where ecological contexts and market facilities were less favourable to intensive land use. Previous experiences suggest investigating political processes in order to understand the criteria according to which access rights were distributed. Based on remote sensing and on a multi-level land governance framework, our research confirms that land placed under collective rights, compared to individual property, is less affected by deforestation among Andean settlements. However, analysis of the historical process of land distribution in the area shows that the distribution of property rights is the result of a political process based on economic, spatial, and environmental strategies that are defined by multiple stakeholders. Collective titles were established in the more remote areas and distributed to communities with lower productive potentialities. Land rights are thus a secondary factor of forest cover change which results from diverse political compromises based on population distribution, accessibility, environmental perceptions, and expected production or extraction incomes.
Resumo:
One of the earliest accounts of duration perception by Karl von Vierordt implied a common process underlying the timing of intervals in the sub-second and the second range. To date, there are two major explanatory approaches for the timing of brief intervals: the Common Timing Hypothesis and the Distinct Timing Hypothesis. While the common timing hypothesis also proceeds from a unitary timing process, the distinct timing hypothesis suggests two dissociable, independent mechanisms for the timing of intervals in the sub-second and the second range, respectively. In the present paper, we introduce confirmatory factor analysis (CFA) to elucidate the internal structure of interval timing in the sub-second and the second range. Our results indicate that the assumption of two mechanisms underlying the processing of intervals in the second and the sub-second range might be more appropriate than the assumption of a unitary timing mechanism. In contrast to the basic assumption of the distinct timing hypothesis, however, these two timing mechanisms are closely associated with each other and share 77% of common variance. This finding suggests either a strong functional relationship between the two timing mechanisms or a hierarchically organized internal structure. Findings are discussed in the light of existing psychophysical and neurophysiological data.
Resumo:
A database containing the global and diffuse components of the surface solar hourly irradiation measured from 1 January 2004 to 31 December 2010 at eight stations of the Egyptian Meteorological Authority is presented. For three of these sites (Cairo, Aswan, and El-Farafra), the direct component is also available. In addition, a series of meteorological variables including surface pressure, relative humidity, temperature, wind speed and direction is provided at the same hourly resolution at all stations. The details of the experimental sites and instruments used for the acquisition are given. Special attention is paid to the quality of the data and the procedure applied to flag suspicious or erroneous measurements is described in details. Between 88 and 99% of the daytime measurements are validated by this quality control. Except at Barrani where the number is lower (13500), between 20000 and 29000 measurements of global and diffuse hourly irradiation are available at all sites for the 7-year period. Similarly, from 9000 to 13000 measurements of direct hourly irradiation values are provided for the three sites where this component is measured. With its high temporal resolution this consistent irradiation and meteorological database constitutes a reliable source to estimate the potential of solar energy in Egypt. It is also adapted to the study of high-frequency atmospheric processes such as the impact of aerosols on atmospheric radiative transfer. In the next future, it is planned to complete regularly the present 2004-2010 database.
Resumo:
First, this paper describes a future layered Air Traffic Management (ATM) system centred in the execution phase of flights. The layered ATM model is based on the work currently performed by SESAR [1] and takes into account the availability of accurate and updated flight information ?seen by all? across the European airspace. This shared information of each flight will be referred as Reference Business Trajectory (RBT). In the layered ATM system, exchanges of information will involve several actors (human or automatic), which will have varying time horizons, areas of responsibility and tasks. Second, the paper will identify the need to define the negotiation processes required to agree revisions to the RBT in the layered ATM system. Third, the final objective of the paper is to bring to the attention of researchers and engineers the communalities between multi-player games and Collaborative Decision Making processes (CDM) in a layered ATM system
Resumo:
The objective of this study was to propose a multi-criteria optimization and decision-making technique to solve food engineering problems. This technique was demostrated using experimental data obtained on osmotic dehydratation of carrot cubes in a sodium chloride solution. The Aggregating Functions Approach, the Adaptive Random Search Algorithm, and the Penalty Functions Approach were used in this study to compute the initial set of non-dominated or Pareto-optimal solutions. Multiple non-linear regression analysis was performed on a set of experimental data in order to obtain particular multi-objective functions (responses), namely water loss, solute gain, rehydration ratio, three different colour criteria of rehydrated product, and sensory evaluation (organoleptic quality). Two multi-criteria decision-making approaches, the Analytic Hierarchy Process (AHP) and the Tabular Method (TM), were used simultaneously to choose the best alternative among the set of non-dominated solutions. The multi-criteria optimization and decision-making technique proposed in this study can facilitate the assessment of criteria weights, giving rise to a fairer, more consistent, and adequate final compromised solution or food process. This technique can be useful to food scientists in research and education, as well as to engineers involved in the improvement of a variety of food engineering processes.
Resumo:
"June 1972."
Resumo:
"June 1960."
Resumo:
Mode of access: Internet.
Resumo:
This paper addresses advanced control of a biological nutrient removal (BNR) activated sludge process. Based on a previously validated distributed parameter model of the BNR activated sludge process, we present robust multivariable controller designs for the process, involving loop shaping of plant model, robust stability and performance analyses. Results from three design case studies showed that a multivariable controller with stability margins of 0.163, 0.492 and 1.062 measured by the normalised coprime factor, multiplicative and additive uncertainties respectively give the best results for meeting performance robustness specifications. The controller robustly stabilises effluent nutrients in the presence of uncertainties with the behaviour of phosphorus accumulating organisms as well as to effectively attenuate major disturbances introduced as step changes. This study also shows that, performance of the multivariable robust controller is superior to multi-loops SISO PI controllers for regulating the BNR activated sludge process in terms of robust stability and performance and controlling the process using inlet feed flowrate is infeasible. (C) 2003 Elsevier Ltd. All rights reserved.