834 resultados para INPUT-OUTPUT ANALYSIS
Resumo:
Qualquer tarefa motora ativa se dá pela ativação de uma população de unidades motoras. Porém, devido a diversas dificuldades, tanto técnicas quanto éticas, não é possível medir a entrada sináptica dos motoneurônios em humanos. Por essas razões, o uso de modelos computacionais realistas de um núcleo de motoneurônios e as suas respectivas fibras musculares tem um importante papel no estudo do controle humano dos músculos. Entretanto, tais modelos são complexos e uma análise matemática é difícil. Neste texto é apresentada uma abordagem baseada em identificação de sistemas de um modelo realista de um núcleo de unidades motoras, com o objetivo de obter um modelo mais simples capaz de representar a transdução das entradas do núcleo de unidades motoras na força do músculo associado ao núcleo. A identificação de sistemas foi baseada em um algoritmo de mínimos quadrados ortogonal para achar um modelo NARMAX, sendo que a entrada considerada foi a condutância sináptica excitatória dendrítica total dos motoneurônios e a saída foi a força dos músculos produzida pelo núcleo de unidades motoras. O modelo identificado reproduziu o comportamento médio da saída do modelo computacional realista, mesmo para pares de sinal de entrada-saída não usados durante o processo de identificação do modelo, como sinais de força muscular modulados senoidalmente. Funções de resposta em frequência generalizada do núcleo de motoneurônios foram obtidas do modelo NARMAX, e levaram a que se inferisse que oscilações corticais na banda-beta (20 Hz) podem influenciar no controle da geração de força pela medula espinhal, comportamento do núcleo de motoneurônios até então desconhecido.
Resumo:
La literatura económica ha centrado la atención en el offshoring de servicios y en su efecto sobre el nivel de empleo nacional, unido a importantes críticas en relación al impacto negativo que esta estrategia provoca en términos de destrucción de empleos nacionales. En este trabajo se analiza la relevancia que tiene el offshoring de servicios en la economía española y, en concreto, en las ramas de servicios y se estudia su efecto sobre el nivel de empleo de este sector. El análisis empírico se lleva a cabo estimando una función de demanda de trabajo con elasticidad de sustitución constante (CES), incluyendo en la misma el efecto de offshoring. Este estudio se realiza para el periodo previo a la crisis, 2000-2007, a partir de los datos contenidos en las Tablas Input-Output de la Contabilidad Nacional del INE.
Resumo:
"Supported in part by the Office of Naval Research. Contract no.N00014-67-A-0305-0007."
Resumo:
The Raf-MEK-ERK MAP kinase cascade transmits signals from activated receptors into the cell to regulate proliferation and differentiation. The cascade is controlled by the Ras GTPase, which recruits Raf from the cytosol to the plasma membrane for activation. In turn, MEK, ERK, and scaffold proteins translocate to the plasma membrane for activation. Here, we examine the input-output properties of the Raf-MEK-ERK MAP kinase module in mammalian cells activated in different cellular contexts. We show that the MAP kinase module operates as a molecular switch in vivo but that the input sensitivity of the module is determined by subcellular location. Signal output from the module is sensitive to low-level input only when it is activated at the plasma membrane. This is because the threshold for activation is low at the plasma membrane, whereas the threshold for activation is high in the cytosol. Thus, the circuit configuration of the module at the plasma membrane generates maximal outputs from low-level analog inputs, allowing cells to process and respond appropriately to physiological stimuli. These results reveal the engineering logic behind the recruitment of elements of the module from the cytosol to the membrane for activation.
Resumo:
Electronic communications devices intended for government or military applications must be rigorously evaluated to ensure that they maintain data confidentiality. High-grade information security evaluations require a detailed analysis of the device's design, to determine how it achieves necessary security functions. In practice, such evaluations are labour-intensive and costly, so there is a strong incentive to find ways to make the process more efficient. In this paper we show how well-known concepts from graph theory can be applied to a device's design to optimise information security evaluations. In particular, we use end-to-end graph traversals to eliminate components that do not need to be evaluated at all, and minimal cutsets to identify the smallest group of components that needs to be evaluated in depth.
Resumo:
Drawing on extensive academic research and theory on clusters and their analysis, the methodology employed in this pilot study (sponsored by the Welsh Assembly Government’s Economic Research Grants Assessment Board) seeks to create a framework for reviewing and monitoring clusters in Wales on an ongoing basis, and generate the information necessary for successful cluster development policy to occur. The multi-method framework developed and tested in the pilot study is designed to map existing Welsh sectors with cluster characteristics, uncover existing linkages, and better understand areas of strength and weakness. The approach adopted relies on synthesising both quantitative and qualitative evidence. Statistical measures, including the size of potential clusters, are united with other evidence on input-output derived inter-linkages within clusters and to other sectors in Wales and the UK, as well as the export and import intensity of the cluster. Multi Sector Qualitative Analysis is then designed for competencies/capacity, risk factors, markets, types and crucially, the perceived strengths of cluster structures and relationships. The approach outlined above can, with the refinements recommended through the review process, provide policy-makers with a valuable tool for reviewing and monitoring individual sectors and ameliorating problems in sectors likely to decline further.
Resumo:
As we enter the 21st Century, technologies originally developed for defense purposes such as computers and satellite communications appear to have become a driving force behind economic growth in the United States. Paradoxically, almost all previous econometric models suggest that the largely defense-oriented federal industrial R&D funding that helped create these technologies had no discernible effect on U.S. industrial productivity growth. This paper addresses this paradox by stressing that defense procurement as well as federal R&D expenditures were targeted to a few narrowly defined manufacturing sub-sectors that produced high tech weaponry. Analysis employing data from the NBER Manufacturing Productivity Database and the BEA' s Input Output tables then demonstrates that defense procurement policies did have significant effects on the productivity performance of disaggregated manufacturing industries because of a process of procurement-driven technological change.
Resumo:
This thesis is concerned with the study of a non-sequential identification technique, so that it may be applied to the identification of process plant mathematical models from process measurements with the greatest degree of accuracy and reliability. In order to study the accuracy of the technique under differing conditions, simple mathematical models were set up on a parallel hybrid. computer and these models identified from input/output measurements by a small on-line digital computer. Initially, the simulated models were identified on-line. However, this method of operation was found not suitable for a thorough study of the technique due to equipment limitations. Further analysis was carried out in a large off-line computer using data generated by the small on-line computer. Hence identification was not strictly on-line. Results of the work have shovm that the identification technique may be successfully applied in practice. An optimum sampling period is suggested, together with noise level limitations for maximum accuracy. A description of a double-effect evaporator is included in this thesis. It is proposed that the next stage in the work will be the identification of a mathematical model of this evaporator using the teclmique described.
Resumo:
This paper re-assesses three independently developed approaches that are aimed at solving the problem of zero-weights or non-zero slacks in Data Envelopment Analysis (DEA). The methods are weights restricted, non-radial and extended facet DEA models. Weights restricted DEA models are dual to envelopment DEA models with restrictions on the dual variables (DEA weights) aimed at avoiding zero values for those weights; non-radial DEA models are envelopment models which avoid non-zero slacks in the input-output constraints. Finally, extended facet DEA models recognize that only projections on facets of full dimension correspond to well defined rates of substitution/transformation between all inputs/outputs which in turn correspond to non-zero weights in the multiplier version of the DEA model. We demonstrate how these methods are equivalent, not only in their aim but also in the solutions they yield. In addition, we show that the aforementioned methods modify the production frontier by extending existing facets or creating unobserved facets. Further we propose a new approach that uses weight restrictions to extend existing facets. This approach has some advantages in computational terms, because extended facet models normally make use of mixed integer programming models, which are computationally demanding.
Resumo:
Liquid-liquid extraction has long been known as a unit operation that plays an important role in industry. This process is well known for its complexity and sensitivity to operation conditions. This thesis presents an attempt to explore the dynamics and control of this process using a systematic approach and state of the art control system design techniques. The process was studied first experimentally under carefully selected. operation conditions, which resembles the ranges employed practically under stable and efficient conditions. Data were collected at steady state conditions using adequate sampling techniques for the dispersed and continuous phases as well as during the transients of the column with the aid of a computer-based online data logging system and online concentration analysis. A stagewise single stage backflow model was improved to mimic the dynamic operation of the column. The developed model accounts for the variation in hydrodynamics, mass transfer, and physical properties throughout the length of the column. End effects were treated by addition of stages at the column entrances. Two parameters were incorporated in the model namely; mass transfer weight factor to correct for the assumption of no mass transfer in the. settling zones at each stage and the backmixing coefficients to handle the axial dispersion phenomena encountered in the course of column operation. The parameters were estimated by minimizing the differences between the experimental and the model predicted concentration profiles at steady state conditions using non-linear optimisation technique. The estimated values were then correlated as functions of operating parameters and were incorporated in·the model equations. The model equations comprise a stiff differential~algebraic system. This system was solved using the GEAR ODE solver. The calculated concentration profiles were compared to those experimentally measured. A very good agreement of the two profiles was achieved within a percent relative error of ±2.S%. The developed rigorous dynamic model of the extraction column was used to derive linear time-invariant reduced-order models that relate the input variables (agitator speed, solvent feed flowrate and concentration, feed concentration and flowrate) to the output variables (raffinate concentration and extract concentration) using the asymptotic method of system identification. The reduced-order models were shown to be accurate in capturing the dynamic behaviour of the process with a maximum modelling prediction error of I %. The simplicity and accuracy of the derived reduced-order models allow for control system design and analysis of such complicated processes. The extraction column is a typical multivariable process with agitator speed and solvent feed flowrate considered as manipulative variables; raffinate concentration and extract concentration as controlled variables and the feeds concentration and feed flowrate as disturbance variables. The control system design of the extraction process was tackled as multi-loop decentralised SISO (Single Input Single Output) as well as centralised MIMO (Multi-Input Multi-Output) system using both conventional and model-based control techniques such as IMC (Internal Model Control) and MPC (Model Predictive Control). Control performance of each control scheme was. studied in terms of stability, speed of response, sensitivity to modelling errors (robustness), setpoint tracking capabilities and load rejection. For decentralised control, multiple loops were assigned to pair.each manipulated variable with each controlled variable according to the interaction analysis and other pairing criteria such as relative gain array (RGA), singular value analysis (SVD). Loops namely Rotor speed-Raffinate concentration and Solvent flowrate Extract concentration showed weak interaction. Multivariable MPC has shown more effective performance compared to other conventional techniques since it accounts for loops interaction, time delays, and input-output variables constraints.
Resumo:
DEA literature continues apace but software has lagged behind. This session uses suitably selected data to present newly developed software which includes many of the most recent DEA models. The software enables the user to address a variety of issues not frequently found in existing DEA software such as: -Assessments under a variety of possible assumptions of returns to scale including NIRS and NDRS; -Scale elasticity computations; -Numerous Input/Output variables and truly unlimited number of assessment units (DMUs) -Panel data analysis -Analysis of categorical data (multiple categories) -Malmquist Index and its decompositions -Computations of Supper efficiency -Automated removal of super-efficient outliers under user-specified criteria; -Graphical presentation of results -Integrated statistical tests
Resumo:
Mathematical Subject Classification 2010:26A33, 33E99, 15A52, 62E15.
Resumo:
In recent years there has been a growing concern about the emission trade balance of countries. It is due to the fact that countries with an open economy are active players in the international trade, though trade is not only a major factor in forging a country’s economic structure anymore, but it does contribute to the movement of embodied emissions beyond the country borders. This issue is especially relevant from the carbon accounting policy’s point of view, as it is known that the production-based principle is in effect now in the Kyoto agreement. The study aims at revealing the interdependence of countries on international trade and its environmental impacts, and how the carbon accounting method plays a crucial role in evaluating a country’s environmental performance and its role in the climate mitigation processes. The input-output models are used in the methodology, as they provide an appropriate framework for this kind of environmental accounting; the analysis shows an international comparison of four European countries (Germany, the United Kingdom, the Netherlands, and Hungary) with extended trading activities and carbon emissions. Moving from the production-based approach in the climate policy, to the consumptionperspective principle and allocation [15], it would also help increasing the efficiency of emission reduction targets and the evaluation of the sustainability dimension and its impacts of international trade. The results of the study have shown that there is an importance of distinction between the two emission allocation approaches, both from global and local level point of view.
Resumo:
In recent years there has been growing concern about the emission trade balances of countries. This is due to the fact that countries with an open economy are active players in international trade. Trade is not only a major factor in forging a country’s economic structure, but contributes to the movement of embodied emissions beyond country borders. This issue is especially relevant from the carbon accounting policy and domestic production perspective, as it is known that the production-based principle is employed in the Kyoto agreement. The research described herein was designed to reveal the interdependence of countries on international trade and the corresponding embodied emissions both on national and on sectoral level and to illustrate the significance of the consumption-based emission accounting. It is presented here to what extent a consumption-based accounting would change the present system based on production-based accounting and allocation. The relationship of CO2 emission embodied in exports and embodied in imports is analysed here. International trade can blur the responsibility for the ecological effects of production and consumption and it can lengthen the link between consumption and its consequences. Input-output models are used in the methodology as they provide an appropriate framework for climate change accounting. The analysis comprises an international comparative study of four European countries (Germany, the United Kingdom, the Netherlands, and Hungary) with extended trading activities and carbon emissions. Moving from a production-based approach in climate policy to a consumption-based principle and allocation approach would help to increase the efficiency of emission reductions and would force countries to rethink their trading activities in order to decrease the environmental load of production activities. The results of this study show that it is important to distinguish between the two emission accounting approaches, both on the global and the local level.
Resumo:
If we classify variables in a program into various security levels, then a secure information flow analysis aims to verify statically that information in a program can flow only in ways consistent with the specified security levels. One well-studied approach is to formulate the rules of the secure information flow analysis as a type system. A major trend of recent research focuses on how to accommodate various sophisticated modern language features. However, this approach often leads to overly complicated and restrictive type systems, making them unfit for practical use. Also, problems essential to practical use, such as type inference and error reporting, have received little attention. This dissertation identified and solved major theoretical and practical hurdles to the application of secure information flow. ^ We adopted a minimalist approach to designing our language to ensure a simple lenient type system. We started out with a small simple imperative language and only added features that we deemed most important for practical use. One language feature we addressed is arrays. Due to the various leaking channels associated with array operations, arrays have received complicated and restrictive typing rules in other secure languages. We presented a novel approach for lenient array operations, which lead to simple and lenient typing of arrays. ^ Type inference is necessary because usually a user is only concerned with the security types for input/output variables of a program and would like to have all types for auxiliary variables inferred automatically. We presented a type inference algorithm B and proved its soundness and completeness. Moreover, algorithm B stays close to the program and the type system and therefore facilitates informative error reporting that is generated in a cascading fashion. Algorithm B and error reporting have been implemented and tested. ^ Lastly, we presented a novel framework for developing applications that ensure user information privacy. In this framework, core computations are defined as code modules that involve input/output data from multiple parties. Incrementally, secure flow policies are refined based on feedback from the type checking/inference. Core computations only interact with code modules from involved parties through well-defined interfaces. All code modules are digitally signed to ensure their authenticity and integrity. ^