929 resultados para capacity utilization rate
Resumo:
Body-size and temperature are the major factors explaining metabolic rate, and the additional factor of pH is a major driver at the biochemical level. These three factors have frequently been found to interact, complicating the formulation of broad models predicting metabolic rates and hence ecological functioning. In this first study of the effects of warming and ocean acidification, and their potential interaction, on metabolic rate across a broad body-size range (two-to-three orders of magnitude difference in body mass) we addressed the impact of climate change on the sea urchin Heliocidaris erythrogramma in context with climate projections for east Australia, an ocean warming hotspot. Urchins were gradually introduced to two temperatures (18 and 23 °C) and two pH (7.5 and 8.0), and maintained for two months. That a new physiological steady-state had been reached, otherwise know as acclimation, was validated through identical experimental trials separated by several weeks. The relationship between body-size, temperature and acidification on the metabolic rate of H. erythrogramma was strikingly stable. Both stressors caused increases in metabolic rate; 20% for temperature and 19% for pH. Combined effects were additive; a 44% increase in metabolism. Body-size had a highly stable relationship with metabolic rate regardless of temperature or pH. None of these diverse drivers of metabolism interacted or modulated the effects of the others, highlighting the partitioned nature of how each influences metabolic rate, and the importance of achieving a full acclimation state. Despite these increases in energetic demand there was very limited capacity for compensatory modulating of feeding rate; food consumption increased only in the very smallest specimens, and only in response to temperature, and not pH. Our data show that warming, acidification and body-size all substantially affect metabolism and are highly consistent and partitioned in their effects, and for H. erythrogramma near-future climate change will incur a substantial energetic cost.
Resumo:
This paper investigates how exchange rates affect the utilization of a free trade agreement (FTA) scheme in trading. Changes in exchange rates affect FTA utilization by two ways. The first way is by changing the excess profits gained by utilizing the FTA scheme, and the second way is by promoting the compliance of rules of origin. Our theoretical models predict that the depreciation of exporters' currency against that of importers enhances the likelihood of FTA utilization through those two channels. Furthermore, our empirical analysis, which is based on rich tariff-line-level data on the utilization of FTA schemes in Korea's imports from ASEAN countries, supports the theoretical prediction. We also show that the effects are smaller for more differentiated products.
Resumo:
This paper analyses how the internal resources of small- and medium-sized enterprises determine access (learning processes) to technology centres (TCs) or industrial research institutes (innovation infrastructure) in traditional low-tech clusters. These interactions basically represent traded (market-based) transactions, which constitute important sources of knowledge in clusters. The paper addresses the role of TCs in low-tech clusters, and uses semi-structured interviews with 80 firms in a manufacturing cluster. The results point out that producer–user interactions are the most frequent; thus, the higher the sector knowledge-intensive base, the more likely the utilization of the available research infrastructure becomes. Conversely, the sectors with less knowledge-intensive structures, i.e. less absorptive capacity (AC), present weak linkages to TCs, as they frequently prefer to interact with suppliers, who act as transceivers of knowledge. Therefore, not all the firms in a cluster can fully exploit the available research infrastructure, and their AC moderates this engagement. In addition, the existence of TCs is not sufficient since the active role of a firm's search strategies to undertake interactions and conduct openness to available sources of knowledge is also needed. The study has implications for policymakers and academia.
Resumo:
Esta tesis presenta un novedoso marco de referencia para el análisis y optimización del retardo de codificación y descodificación para vídeo multivista. El objetivo de este marco de referencia es proporcionar una metodología sistemática para el análisis del retardo en codificadores y descodificadores multivista y herramientas útiles en el diseño de codificadores/descodificadores para aplicaciones con requisitos de bajo retardo. El marco de referencia propuesto caracteriza primero los elementos que tienen influencia en el comportamiento del retardo: i) la estructura de predicción multivista, ii) el modelo hardware del codificador/descodificador y iii) los tiempos de proceso de cuadro. En segundo lugar, proporciona algoritmos para el cálculo del retardo de codificación/ descodificación de cualquier estructura arbitraria de predicción multivista. El núcleo de este marco de referencia consiste en una metodología para el análisis del retardo de codificación/descodificación multivista que es independiente de la arquitectura hardware del codificador/descodificador, completada con un conjunto de modelos que particularizan este análisis del retardo con las características de la arquitectura hardware del codificador/descodificador. Entre estos modelos, aquellos basados en teoría de grafos adquieren especial relevancia debido a su capacidad de desacoplar la influencia de los diferentes elementos en el comportamiento del retardo en el codificador/ descodificador, mediante una abstracción de su capacidad de proceso. Para revelar las posibles aplicaciones de este marco de referencia, esta tesis presenta algunos ejemplos de su utilización en problemas de diseño que afectan a codificadores y descodificadores multivista. Este escenario de aplicación cubre los siguientes casos: estrategias para el diseño de estructuras de predicción que tengan en consideración requisitos de retardo además del comportamiento tasa-distorsión; diseño del número de procesadores y análisis de los requisitos de velocidad de proceso en codificadores/ descodificadores multivista dado un retardo objetivo; y el análisis comparativo del comportamiento del retardo en codificadores multivista con diferentes capacidades de proceso e implementaciones hardware. ABSTRACT This thesis presents a novel framework for the analysis and optimization of the encoding and decoding delay for multiview video. The objective of this framework is to provide a systematic methodology for the analysis of the delay in multiview encoders and decoders and useful tools in the design of multiview encoders/decoders for applications with low delay requirements. The proposed framework characterizes firstly the elements that have an influence in the delay performance: i) the multiview prediction structure ii) the hardware model of the encoder/decoder and iii) frame processing times. Secondly, it provides algorithms for the computation of the encoding/decoding delay of any arbitrary multiview prediction structure. The core of this framework consists in a methodology for the analysis of the multiview encoding/decoding delay that is independent of the hardware architecture of the encoder/decoder, which is completed with a set of models that particularize this delay analysis with the characteristics of the hardware architecture of the encoder/decoder. Among these models, the ones based in graph theory acquire special relevance due to their capacity to detach the influence of the different elements in the delay performance of the encoder/decoder, by means of an abstraction of its processing capacity. To reveal possible applications of this framework, this thesis presents some examples of its utilization in design problems that affect multiview encoders and decoders. This application scenario covers the following cases: strategies for the design of prediction structures that take into consideration delay requirements in addition to the rate-distortion performance; design of number of processors and analysis of processor speed requirements in multiview encoders/decoders given a target delay; and comparative analysis of the encoding delay performance of multiview encoders with different processing capabilities and hardware implementations.
Resumo:
Governments are working in new policies to slow down total energy consumption and greenhouse gases (GHG) emissions, promoting the deployment of electric vehicles (EVs) in all countries. In order to facilitate this deployment and help to reduce the final costs of their batteries, additional utilization of EVs when those are parked has been proposed. EVs can be used to minimize the total electricity cost of buildings (named vehicle to building applications, V2B). In this paper an economic evaluation of EVs in the Building Energy Management System is shown. The optimal storage capacity and its equivalent number of EVs are determined. This value is then used for determining the optimal charging schedule to be applied to the batteries. From this schedule, the total expected profit is derived for the case of a real hotel in Spain.
Resumo:
Positive composite electrodes having LiNi0.5Mn1.5O4 spinel as active material, a blend of graphite and carbon black for increasing the electrode electrical conductivity and either polyvinyldenefluoride (PVDF) or a blend of PVDF with a small amount of Teflon® (1 wt%) for building up the electrode. They have been processed by tape casting on an aluminum foil as current collector using the doctor blade technique. Additionally, the component blends were either sonicated or not, and the processed electrodes were compacted or not under subsequent cold pressing. Composites electrodes with high weight, up to 17 mg/cm2, were prepared and studied as positive electrodes for lithium-ion batteries. The addition of Teflon® and the application of the sonication treatment lead to uniform electrodes that are well-adhered to the aluminum foil. Both parameters contribute to improve the capacity drained at high rates (5C). Additional compaction of the electrode/aluminum assemblies remarkably enhances the electrode rate capabilities. At 5C rate, remarkable capacity retentions between 80% and 90% are found for electrodes with weights in the range 3–17 mg/cm2, having Teflon® in their formulation, prepared after sonication of their component blends and compacted under 2 tonnes/cm2.
Resumo:
Optical communications receivers using wavelet signals processing is proposed in this paper for dense wavelength-division multiplexed (DWDM) systems and modal-division multiplexed (MDM) transmissions. The optical signal-to-noise ratio (OSNR) required to demodulate polarization-division multiplexed quadrature phase shift keying (PDM-QPSK) modulation format is alleviated with the wavelet denoising process. This procedure improves the bit error rate (BER) performance and increasing the transmission distance in DWDM systems. Additionally, the wavelet-based design relies on signal decomposition using time-limited basis functions allowing to reduce the computational cost in Digital-Signal-Processing (DSP) module. Attending to MDM systems, a new scheme of encoding data bits based on wavelets is presented to minimize the mode coupling in few-mode (FWF) and multimode fibers (MMF). The Shifted Prolate Wave Spheroidal (SPWS) functions are proposed to reduce the modal interference.
Resumo:
El objetivo principal de esta tesis ha sido el diseño y la optimización de receptores implementados con fibra óptica, para ser usados en redes ópticas de alta velocidad que empleen formatos de modulación de fase. En los últimos años, los formatos de modulación de fase (Phase Shift keying, PSK) han captado gran atención debido a la mejora de sus prestaciones respecto a los formatos de modulación convencionales. Principalmente, presentan una mejora de la eficiencia espectral y una mayor tolerancia a la degradación de la señal causada por la dispersión cromática, la dispersión por modo de polarización y los efectos no-lineales en la fibra óptica. En este trabajo, se analizan en detalle los formatos PSK, incluyendo sus variantes de modulación de fase diferencial (Differential Phase Shift Keying, DPSK), en cuadratura (Differential Quadrature Phase Shift Keying, DQPSK) y multiplexación en polarización (Polarization Multiplexing Differential Quadrature Phase Shift Keying, PM-DQPSK), con la finalidad de diseñar y optimizar los receptores que permita su demodulación. Para ello, se han analizado y desarrollado nuevas estructuras que ofrecen una mejora en las prestaciones del receptor y una reducción de coste comparadas con las actualmente disponibles. Para la demodulación de señales DPSK, en esta tesis, se proponen dos nuevos receptores basados en un interferómetro en línea Mach-Zehnder (MZI) implementado con tecnología todo-fibra. El principio de funcionamiento de los MZI todo-fibra propuestos se asienta en la interferencia modal que se produce en una fibra multimodo (MMF) cuando se situada entre dos monomodo (SMF). Este tipo de configuración (monomodo-multimodo-monomodo, SMS) presenta un buen ratio de extinción interferente si la potencia acoplada en la fibra multimodo se reparte, principal y equitativamente, entre dos modos dominantes. Con este objetivo, se han estudiado y demostrado tanto teórica como experimentalmente dos nuevas estructuras SMS que mejoran el ratio de extinción. Una de las propuestas se basa en emplear una fibra multimodo de índice gradual cuyo perfil del índice de refracción presenta un hundimiento en su zona central. La otra consiste en una estructura SMS con las fibras desalineadas y donde la fibra multimodo es una fibra de índice gradual convencional. Para las dos estructuras, mediante el análisis teórico desarrollado, se ha demostrado que el 80 – 90% de la potencia de entrada se acopla a los dos modos dominantes de la fibra multimodo y se consigue una diferencia inferior al 10% entre ellos. También se ha demostrado experimentalmente que se puede obtener un ratio de extinción de al menos 12 dB. Con el objeto de demostrar la capacidad de estas estructuras para ser empleadas como demoduladores de señales DPSK, se han realizado numerosas simulaciones de un sistema de transmisión óptico completo y se ha analizado la calidad del receptor bajo diferentes perspectivas, tales como la sensibilidad, la tolerancia a un filtrado óptico severo o la tolerancia a las dispersiones cromática y por modo de polarización. En todos los casos se ha concluido que los receptores propuestos presentan rendimientos comparables a los obtenidos con receptores convencionales. En esta tesis, también se presenta un diseño alternativo para la implementación de un receptor DQPSK, basado en el uso de una fibra mantenedora de la polarización (PMF). A través del análisi teórico y del desarrollo de simulaciones numéricas, se ha demostrado que el receptor DQPSK propuesto presenta prestaciones similares a los convencionales. Para complementar el trabajo realizado sobre el receptor DQPSK basado en PMF, se ha extendido el estudio de su principio de demodulación con el objeto de demodular señales PM-DQPSK, obteniendo como resultado la propuesta de una nueva estructura de demodulación. El receptor PM-DQPSK propuesto se basa en la estructura conjunta de una única línea de retardo junto con un rotador de polarización. Se ha analizado la calidad de los receptores DQPSK y PM-DQPSK bajo diferentes perspectivas, tales como la sensibilidad, la tolerancia a un filtrado óptico severo, la tolerancia a las dispersiones cromática y por modo de polarización o su comportamiento bajo condiciones no-ideales. En comparación con los receptores convencionales, nuestra propuesta exhibe prestaciones similares y además permite un diseño más simple que redunda en un coste potencialmente menor. En las redes de comunicaciones ópticas actuales se utiliza la tecnología de multimplexación en longitud de onda (WDM) que obliga al uso de filtros ópticos con bandas de paso lo más estrechas posibles y a emplear una serie de dispositivos que incorporan filtros en su arquitectura, tales como los multiplexores, demultiplexores, ROADMs, conmutadores y OXCs. Todos estos dispositivos conectados entre sí son equivalentes a una cadena de filtros cuyo ancho de banda se va haciendo cada vez más estrecho, llegando a distorsionar la forma de onda de las señales. Por esto, además de analizar el impacto del filtrado óptico en las señales de 40 Gbps DQPSK y 100 Gbps PM-DQPSK, este trabajo de tesis se completa estudiando qué tipo de filtro óptico minimiza las degradaciones causadas en la señal y analizando el número máximo de filtros concatenados que permiten mantener la calidad requerida al sistema. Se han estudiado y simulado cuatro tipos de filtros ópticos;Butterworth, Bessel, FBG y F-P. ABSTRACT The objective of this thesis is the design and optimization of optical fiber-based phase shift keying (PSK) demodulators for high-bit-rate optical networks. PSK modulation formats have attracted significant attention in recent years, because of the better performance with respect to conventional modulation formats. Principally, PSK signals can improve spectrum efficiency and tolerate more signal degradation caused by chromatic dispersion, polarization mode dispersion and nonlinearities in the fiber. In this work, many PSK formats were analyzed in detail, including the variants of differential phase modulation (Differential Phase Shift Keying, DPSK), in quadrature (Differential Quadrature Phase Shift Keying, DQPSK) and polarization multiplexing (Polarization Multiplexing Differential Quadrature Phase Shift Keying, PM-DQPSK), in order to design and optimize receivers enabling demodulations. Therefore, novel structures, which offer good receiver performances and a reduction in cost compared to the current structures, have been analyzed and developed. Two novel receivers based on an all-fiber in-line Mach-Zehnder interferometer (MZI) were proposed for DPSK signal demodulation in this thesis. The operating principle of the all-fiber MZI is based on the modal interference that occurs in a multimode fiber (MMF) when it is located between two single-mode fibers (SMFs). This type of configuration (Single-mode-multimode-single-mode, SMS) can provide a good extinction ratio if the incoming power from the SMF could be coupled equally into two dominant modes excited in the MMF. In order to improve the interference extinction ratio, two novel SMS structures have been studied and demonstrated, theoretically and experimentally. One of the two proposed MZIs is based on a graded-index multimode fiber (MMF) with a central dip in the index profile, located between two single-mode fibers (SMFs). The other one is based on a conventional graded-index MMF mismatch spliced between two SMFs. Theoretical analysis has shown that, in these two schemes, 80 – 90% of the incoming power can be coupled into the two dominant modes exited in the MMF, and the power difference between them is only ~10%. Experimental results show that interference extinction ratio of 12 dB could be obtained. In order to demonstrate the capacity of these two structures for use as DPSK signal demodulators, numerical simulations in a completed optical transmission system have been carried out, and the receiver quality has been analyzed under different perspectives, such as sensitivity, tolerance to severe optical filtering or tolerance to chromatic and polarization mode dispersion. In all cases, from the simulation results we can conclude that the two proposed receivers can provide performances comparable to conventional ones. In this thesis, an alternative design for the implementation of a DQPSK receiver, which is based on a polarization maintaining fiber (PMF), was also presented. To complement the work made for the PMF-based DQPSK receiver, the study of the demodulation principle has been extended to demodulate PM-DQPSK signals, resulting in the proposal of a novel demodulation structure. The proposed PM-DQPSK receiver is based on only one delay line and a polarization rotator. The quality of the proposed DQPSK and PM-DQPSK receivers under different perspectives, such as sensitivity, tolerance to severe optical filtering, tolerance to chromatic dispersion and polarization mode dispersion, or behavior under non-ideal conditions. Compared with the conventional receivers, our proposals exhibit similar performances but allow a simpler design which can potentially reduce the cost. The wavelength division multiplexing (WDM) technology used in current optical communications networks requires the use of optical filters with a passband as narrow as possible, and the use of a series of devices that incorporate filters in their architecture, such as multiplexers, demultiplexers, switches, reconfigurable add-drop multiplexers (ROADMs) and optical cross-connects (OXCs). All these devices connected together are equivalent to a chain of filters whose bandwidth becomes increasingly narrow, resulting in distortion to the waveform of the signals. Therefore, in addition to analyzing the impact of optical filtering on signal of 40 Gbps DQPSK and 100 Gbps PM-DQPSK, we study which kind of optical filter minimizes the signal degradation and analyze the maximum number of concatenated filters for maintaining the required quality of the system. Four types of optical filters, including Butterworth, Bessel, FBG and FP, have studied and simulated.
Resumo:
La conciencia de la crisis de la modernidad -que comienza ya a finales del siglo XIX- ha cobrado más experiencia debido al conocimiento de los límites del desarrollo económico, ya que como parecía razonable pensar, también los recursos naturales son finitos. En 1972, el Club de Roma analizó las distintas opciones disponibles para conseguir armonizar el desarrollo sostenible y las limitaciones medioambientales. Fue en 1987 cuando la Comisión Mundial para el Medio Ambiente y el Desarrollo de la ONU definía por primera vez el concepto de desarrollo sostenible. Definición que posteriormente fue incorporada en todos los programas de la ONU y sirvió de eje, por ejemplo, a la Cumbre de la Tierra celebrada en Río de Janeiro en 1992. Parece evidente que satisfacer la demanda energética, fundamentalmente desde la Revolución Industrial en el s XIX, trajo consigo un creciente uso de los combustibles fósiles, con la consiguiente emisión de los gases de efecto invernadero (GEI) y el aumento de la temperatura global media terrestre. Esta temperatura se incrementó en los últimos cien años en una media de 0.74ºC. La mayor parte del incremento observado desde la mitad del siglo XX en esta temperatura media se debe, con una probabilidad de al menos el 90%, al aumento observado en los GEI antropogénicos, siendo uno de ellos el CO2 que proviene de la transformación del carbono de los combustibles fósiles durante su combustión. Ante el creciente uso de los combustibles fósiles, los proyectos CAC, proyectos de captura, transporte y almacenamiento, se presentan como una contribución al desarrollo sostenible ya que se trata de una tecnología que permite mitigar el cambio climático. Para valorar si la tecnología CAC es sostenible, habrá que comprobar si existe o no capacidad para almacenar el CO2 en una cantidad mayor a la de producción y durante el tiempo necesario que impone la evolución de la concentración de CO2 en la atmósfera para mantenerla por debajo de las 450ppmv (concentración de CO2 que propone el Panel Intergubernamental para el Cambio Climático). El desarrollo de los proyectos CAC completos pasa por la necesaria selección de adecuados almacenes de CO2 que sean capaces de soportar los efectos de las presiones de inyección, así como asegurar la capacidad de dichos almacenes y la estanqueidad del CO2 en los mismos. La caracterización geológica de un acuífero susceptible de ser almacén de CO2 debe conducir a determinar las propiedades que dicho almacén posee para asegurar un volumen adecuado de almacenamiento, una inyectabilidad del CO2 en el mismo a un ritmo adecuado y la estanqueidad del CO2 en dicho acuífero a largo plazo. El presente trabajo pretende estudiar los parámetros que tienen influencia en el cálculo de la capacidad del almacén, para lo que en primer lugar se ha desarrollado la tecnología necesaria para llevar a cabo la investigación mediante ensayos de laboratorio. Así, se ha desarrollado una patente, "ATAP, equipo para ensayos petrofísicos (P201231913)", con la que se ha llevado a cabo la parte experimental de este trabajo para la caracterización de los parámetros que tienen influencia en el cálculo de la capacidad del almacén. Una vez desarrollada la tecnología, se aborda el estudio de los distintos parámetros que tienen influencia en la capacidad del almacén realizando ensayos con ATAP. Estos ensayos definen el volumen del almacenamiento, llegándose a la conclusión de que en la determinación de este volumen, juegan un papel importante el alcance de los mecanismos trampa, físicos o químicos, del CO2 en el almacén. Ensayos que definen la capacidad del almacén de "aceptar" o "rechazar" el CO2 inyectado, la inyectabilidad, y por último, ensayos encaminados a determinar posibles fugas que se pueden dar a través de los pozos de inyección, definidos estos como caminos preferenciales de fugas en un almacén subterráneo de CO2. Queda de este modo caracterizada la estanqueidad del CO2 en el acuífero a largo plazo y su influencia obvia en la determinación de la capacidad del almacén. Unido al propósito de la estimación de la capacidad del almacén, se encuentra el propósito de asegurar la estanqueidad de dichos almacenes en el tiempo, y adelantarse a la evolución de la pluma de CO2 en el interior de dichos almacenes. Para cumplir este propósito, se ha desarrollado un modelo dinámico a escala de laboratorio, mediante el programa ECLIPSE 300, con el fin de establecer una metodología para el cálculo de la capacidad estimada del almacén, así como el estudio de la evolución de la pluma de CO2 dentro del acuífero a lo largo del tiempo, partiendo de los resultados obtenidos en los ensayos realizados en ATAP y con la modelización de la probeta de roca almacén empleada en dichos ensayos. Presentamos por tanto un trabajo que establece las bases metodológicas para el estudio de la influencia de distintos parámetros petrofísicos en el cálculo de la capacidad del almacén unidos al desarrollo tecnológico de ATAP y su utilización para la determinación de dichos parámetros aplicables a cada acuífero concreto de estudio. ABSTRACT The crisis of modernity –which begins at the end of 19th Century- has been more important due to the knowledge of the limits of economic development, since it appeared to be thought reasonable, the natural resources are finite. In 1972, The Club of Rome analyzed the different options available in order to harmonize the sustainability and the environment development. It was in 1987 when The Global Commission on The Environment and the Development of UN, defined for the first time the concept of Sustainable Development. This definition that was fully incorporated in all the UN programs and it was useful as an axis, for example, in La Cumbre de la Tierra summit in Río de Janeiro in 1992. It seems obvious to satisfy energetic demand, basically after The Industrial Revolution in 19th Century, which represented an increasing use of fossil fuels, therefore greenhouse gases emission and the increasing of global average temperature. This temperature increased in the last 100 years up to 0.74ºC. The major part of the temperature increase is due to the increase observed in Greenhouse gases with human origin, at least with 90% of probability. The most important gas is the CO2 because of its quantity. In the face of the increasing use of fossil fuels, the CCS projects, Carbon Capture and Storage projects, appear as a contribution of sustainable development since it is a technology for avoiding the climate change. In order to evaluate if CCS technology is sustainable, it will be necessary to prove if the capacity for CO2 storage is available or not in a quantity greater than the production one and during the time necessary to keep the CO2 concentration in the atmosphere lower than 450ppmv (concentration imposed by IPCC). The development of full CCS projects goes through the selection of good CO2 storages that are able to support the effects of pressure injection, and assure the capacity of such storages and the watertightness of CO2. The geological characterization of the aquifer that could be potential CO2 storage should lead to determine the properties that such storage has in order to assure the adequate storage volume, the CO2 injectivity in a good rate, and the watertightness of the CO2 in the long term. The present work aims to study the parameters that have influence on the calculation of storage capacity, and for that purpose the appropriate technology has been developed for carrying out the research by mean of laboratory tests. Thus, a patent has been developed, "ATAP, equipo para ensayos petrofísicos (P201231913)", that has been used for developing the experimental part of this work. Once the technology has been developed, the study of different parameters, that have influence on the capacity of the storage, has been addressed developing different tests in ATAP. These tests define the storage volume which is related to the scope of different CO2 trap mechanisms, physical or chemical, in the storage. Tests that define the capacity of the storage to “accept” or “reject” the injected CO2, the injectivity, and tests led to determine possible leakages through injection wells. In this way we could talk about the watertightness in the aquifer in the long term and its influence on the storage capacity estimation. Together with the purpose of the storage capacity estimation, is the purpose of assuring the watertightness of such storages in the long term and anticipating the evolution of CO2 plume inside such aquifers. In order to fulfill this purpose, a dynamic model has been developed with ECLIPSE 300, for stablishing the methodology for the calculation of storage capacity estimation and the evolution of the CO2 plume, starting out with the tests carried out in ATAP. We present this work that establishes the methodology bases for the study of the influence of different petrophysics parameters in the calculation of the capacity of the storage together with the technological development of ATAP and its utilization for the determination of such parameters applicable to each aquifer.
Resumo:
Acknowledgements This article was based on the first author’s PhD which was financed by the Malawi Health Research Capacity Strengthening Initiative. We thank Mr Patrick Naphini formerly of the Ministry of Health and Mrs Mafase Sesani at CHAM Secretariat for helping with the data. We also thank Mr Jacob Mazalale for useful comments on the article.
Resumo:
Werner syndrome (WS) is an autosomal recessive disorder characterized by genomic instability and the premature onset of a number of age-related diseases. The gene responsible for WS encodes a member of the RecQ-like subfamily of DNA helicases. Here we show that its murine homologue maps to murine chromosome 8 in a region syntenic with the human WRN gene. We have deleted a segment of this gene and created Wrn-deficient embryonic stem (ES) cells and WS mice. While displaying reduced embryonic survival, live-born WS mice otherwise appear normal during their first year of life. Nonetheless, although several DNA repair systems are apparently intact in homozygous WS ES cells, such cells display a higher mutation rate and are significantly more sensitive to topoisomerase inhibitors (especially camptothecin) than are wild-type ES cells. Furthermore, mouse embryo fibroblasts derived from homozygous WS embryos show premature loss of proliferative capacity. At the molecular level, wild-type, but not mutant, WS protein copurifies through a series of centrifugation and chromatography steps with a multiprotein DNA replication complex.
Resumo:
During the aging process, mammals lose up to a third of their skeletal muscle mass and strength. Although the mechanisms underlying this loss are not entirely understood, we attempted to moderate the loss by increasing the regenerative capacity of muscle. This involved the injection of a recombinant adeno-associated virus directing overexpression of insulin-like growth factor I (IGF-I) in differentiated muscle fibers. We demonstrate that the IGF-I expression promotes an average increase of 15% in muscle mass and a 14% increase in strength in young adult mice, and remarkably, prevents aging-related muscle changes in old adult mice, resulting in a 27% increase in strength as compared with uninjected old muscles. Muscle mass and fiber type distributions were maintained at levels similar to those in young adults. We propose that these effects are primarily due to stimulation of muscle regeneration via the activation of satellite cells by IGF-I. This supports the hypothesis that the primary cause of aging-related impairment of muscle function is a cumulative failure to repair damage sustained during muscle utilization. Our results suggest that gene transfer of IGF-I into muscle could form the basis of a human gene therapy for preventing the loss of muscle function associated with aging and may be of benefit in diseases where the rate of damage to skeletal muscle is accelerated.
Simple neural networks for the amplification and utilization of small changes in neuron firing rates
Resumo:
I describe physiologically plausible “voter-coincidence” neural networks such that secondary “coincidence” neurons fire on the simultaneous receipt of sufficiently large sets of input pulses from primary sets of neurons. The networks operate such that the firing rate of the secondary, output neurons increases (or decreases) sharply when the mean firing rate of primary neurons increases (or decreases) to a much smaller degree. In certain sensory systems, signals that are generally smaller than the noise levels of individual primary detectors, are manifest in very small increases in the firing rates of sets of afferent neurons. For such systems, this kind of network can act to generate relatively large changes in the firing rate of secondary “coincidence” neurons. These differential amplification systems can be cascaded to generate sharp, “yes–no” spike signals that can direct behavioral responses.
Resumo:
We have developed and characterized a system to analyze light effects on auxin transport independent of photosynthetic effects. Polar transport of [3H]indole-3-acetic acid through hypocotyl segments from etiolated cucumber (Cucumis sativus L.) seedlings was increased in seedlings grown in dim-red light (DRL) (0.5 μmol m−2 s−1) relative to seedlings grown in darkness. Both transport velocity and transport intensity (export rate) were increased by at least a factor of 2. Tissue formed in DRL completely acquired the higher transport capacity within 50 h, but tissue already differentiated in darkness acquired only a partial increase in transport capacity within 50 h of DRL, indicating a developmental window for light induction of commitment to changes in auxin transport. This light-induced change probably manifests itself by alteration of function of the auxin efflux carrier, as revealed using specific transport inhibitors. Relative to dark controls, DRL-grown seedlings were differentially less sensitive to two inhibitors of polar auxin transport, N-(naphth-1-yl) phthalamic acid and 2,3,5-triiodobenzoic acid. On the basis of these data, we propose that the auxin efflux carrier is a key target of light regulation during photomorphogenesis.
Resumo:
Auditory responses in the caudomedial neostriatum (NCM) of the zebra finch (Taeniopygia guttata) forebrain habituate to repeated presentations of a novel conspecific song. This habituation is long lasting and specific to individual stimuli. We here test the acoustic and ethological basis of this stimulus-specific habituation by recording extracellular multiunit activity in the NCM of awake male and female zebra finches presented with a variety of conspecific and heterospecific vocalizations, white noise, and tones. Initial responses to conspecific song and calls and to human speech were higher than responses to the other stimuli. Immediate habituation rates were high for all novel stimuli except tones, which habituated at a lower rate. Habituation to conspecific calls and songs outlasted habituation to other stimuli. The extent of immediate habituation induced by a particular novel song was not diminished when other conspecific songs were presented in alternation. In addition, the persistence of habituation was not diminished by exposure to other songs before testing, nor was it influenced by gender or laterality. Our results suggest that the NCM is specialized for remembering the calls and songs of many individual conspecifics.