920 resultados para Methods Time Measurement (MTM)
Dynamic single cell measurements of kinase activity by synthetic kinase activity relocation sensors.
Resumo:
BACKGROUND: Mitogen activated protein kinases (MAPK) play an essential role in integrating extra-cellular signals and intra-cellular cues to allow cells to grow, adapt to stresses, or undergo apoptosis. Budding yeast serves as a powerful system to understand the fundamental regulatory mechanisms that allow these pathways to combine multiple signals and deliver an appropriate response. To fully comprehend the variability and dynamics of these signaling cascades, dynamic and quantitative single cell measurements are required. Microscopy is an ideal technique to obtain these data; however, novel assays have to be developed to measure the activity of these cascades. RESULTS: We have generated fluorescent biosensors that allow the real-time measurement of kinase activity at the single cell level. Here, synthetic MAPK substrates were engineered to undergo nuclear-to-cytoplasmic relocation upon phosphorylation of a nuclear localization sequence. Combination of fluorescence microscopy and automated image analysis allows the quantification of the dynamics of kinase activity in hundreds of single cells. A large heterogeneity in the dynamics of MAPK activity between individual cells was measured. The variability in the mating pathway can be accounted for by differences in cell cycle stage, while, in the cell wall integrity pathway, the response to cell wall stress is independent of cell cycle stage. CONCLUSIONS: These synthetic kinase activity relocation sensors allow the quantification of kinase activity in live single cells. The modularity of the architecture of these reporters will allow their application in many other signaling cascades. These measurements will allow to uncover new dynamic behaviour that previously could not be observed in population level measurements.
Resumo:
Sonar signal processing comprises of a large number of signal processing algorithms for implementing functions such as Target Detection, Localisation, Classification, Tracking and Parameter estimation. Current implementations of these functions rely on conventional techniques largely based on Fourier Techniques, primarily meant for stationary signals. Interestingly enough, the signals received by the sonar sensors are often non-stationary and hence processing methods capable of handling the non-stationarity will definitely fare better than Fourier transform based methods.Time-frequency methods(TFMs) are known as one of the best DSP tools for nonstationary signal processing, with which one can analyze signals in time and frequency domains simultaneously. But, other than STFT, TFMs have been largely limited to academic research because of the complexity of the algorithms and the limitations of computing power. With the availability of fast processors, many applications of TFMs have been reported in the fields of speech and image processing and biomedical applications, but not many in sonar processing. A structured effort, to fill these lacunae by exploring the potential of TFMs in sonar applications, is the net outcome of this thesis. To this end, four TFMs have been explored in detail viz. Wavelet Transform, Fractional Fourier Transfonn, Wigner Ville Distribution and Ambiguity Function and their potential in implementing five major sonar functions has been demonstrated with very promising results. What has been conclusively brought out in this thesis, is that there is no "one best TFM" for all applications, but there is "one best TFM" for each application. Accordingly, the TFM has to be adapted and tailored in many ways in order to develop specific algorithms for each of the applications.
Resumo:
The Bureau International des Poids et Mesures, the BIPM, was established by Article 1 of the Convention du Mètre, on 20 May 1875, and is charged with providing the basis for a single, coherent system of measurements to be used throughout the world. The decimal metric system, dating from the time of the French Revolution, was based on the metre and the kilogram. Under the terms of the 1875 Convention, new international prototypes of the metre and kilogram were made and formally adopted by the first Conférence Générale des Poids et Mesures (CGPM) in 1889. Over time this system developed, so that it now includes seven base units. In 1960 it was decided at the 11th CGPM that it should be called the Système International d’Unités, the SI (in English: the International System of Units). The SI is not static but evolves to match the world’s increasingly demanding requirements for measurements at all levels of precision and in all areas of science, technology, and human endeavour. This document is a summary of the SI Brochure, a publication of the BIPM which is a statement of the current status of the SI. The seven base units of the SI, listed in Table 1, provide the reference used to define all the measurement units of the International System. As science advances, and methods of measurement are refined, their definitions have to be revised. The more accurate the measurements, the greater the care required in the realization of the units of measurement.
Resumo:
The purpose of this study was to apply and compare two time-domain analysis procedures in the determination of oxygen uptake (VO2) kinetics in response to a pseudorandom binary sequence (PRBS) exercise test. PRBS exercise tests have typically been analysed in the frequency domain. However, the complex interpretation of frequency responses may have limited the application of this procedure in both sporting and clinical contexts, where a single time measurement would facilitate subject comparison. The relative potential of both a mean response time (MRT) and a peak cross-correlation time (PCCT) was investigated. This study was divided into two parts: a test-retest reliability study (part A), in which 10 healthy male subjects completed two identical PRBS exercise tests, and a comparison of the VO2 kinetics of 12 elite endurance runners (ER) and 12 elite sprinters (SR; part B). In part A, 95% limits of agreement were calculated for comparison between MRT and PCCT. The results of part A showed no significant difference between test and retest as assessed by MRT [mean (SD) 42.2 (4.2) s and 43.8 (6.9) s] or by PCCT [21.8 (3.7) s and 22.7 (4.5) s]. Measurement error (%) was lower for MRT in comparison with PCCT (16% and 25%, respectively). In part B of the study, the VO2 kinetics of ER were significantly faster than those of SR, as assessed by MRT [33.4 (3.4) s and 39.9 (7.1) s, respectively; P<0.01] and PCCT [20.9 (3.8) s and 24.8 (4.5) s; P < 0.05]. It is possible that either analysis procedure could provide a single test measurement Of VO2 kinetics; however, the greater reliability of the MRT data suggests that this method has more potential for development in the assessment Of VO2 kinetics by PRBS exercise testing.
Resumo:
In this paper, we propose a new method of measuring the very slow paramagnetic ion diffusion coefficient using a commercial high-resolution spectrometer. If there are distinct paramagnetic ions influencing the hydrogen nuclear magnetic relaxation time differently, their diffusion coefficients can be measured separately. A cylindrical phantom filled with Fricke xylenol gel solution and irradiated with gamma rays was used to validate the method. The Fricke xylenol gel solution was prepared with 270 Bloom porcine gelatin, the phantom was irradiated with gamma rays originated from a (60)Co source and a high-resolution 200 MHz nuclear magnetic resonance (NMR) spectrometer was used to obtain the phantom (1)H profile in the presence of a linear magnetic field gradient. By observing the temporal evolution of the phantom NMR profile, an apparent ferric ion diffusion coefficient of 0.50 mu m(2)/ms due to ferric ions diffusion was obtained. In any medical process where the ionizing radiation is used, the dose planning and the dose delivery are the key elements for the patient safety and success of treatment. These points become even more important in modern conformal radio therapy techniques, such as stereotactic radiosurgery, where the delivered dose in a single session of treatment can be an order of magnitude higher than the regular doses of radiotherapy. Several methods have been proposed to obtain the three-dimensional (3-D) dose distribution. Recently, we proposed an alternative method for the 3-D radiation dose mapping, where the ionizing radiation modifies the local relative concentration of Fe(2+)/Fe(3+) in a phantom containing Fricke gel and this variation is associated to the MR image intensity. The smearing of the intensity gradient is proportional to the diffusion coefficient of the Fe(3+) and Fe(2+) in the phantom. There are several methods for measurement of the ionic diffusion using NMR, however, they are applicable when the diffusion is not very slow.
Resumo:
We discuss the development and performance of a low-power sensor node (hardware, software and algorithms) that autonomously controls the sampling interval of a suite of sensors based on local state estimates and future predictions of water flow. The problem is motivated by the need to accurately reconstruct abrupt state changes in urban watersheds and stormwater systems. Presently, the detection of these events is limited by the temporal resolution of sensor data. It is often infeasible, however, to increase measurement frequency due to energy and sampling constraints. This is particularly true for real-time water quality measurements, where sampling frequency is limited by reagent availability, sensor power consumption, and, in the case of automated samplers, the number of available sample containers. These constraints pose a significant barrier to the ubiquitous and cost effective instrumentation of large hydraulic and hydrologic systems. Each of our sensor nodes is equipped with a low-power microcontroller and a wireless module to take advantage of urban cellular coverage. The node persistently updates a local, embedded model of flow conditions while IP-connectivity permits each node to continually query public weather servers for hourly precipitation forecasts. The sampling frequency is then adjusted to increase the likelihood of capturing abrupt changes in a sensor signal, such as the rise in the hydrograph – an event that is often difficult to capture through traditional sampling techniques. Our architecture forms an embedded processing chain, leveraging local computational resources to assess uncertainty by analyzing data as it is collected. A network is presently being deployed in an urban watershed in Michigan and initial results indicate that the system accurately reconstructs signals of interest while significantly reducing energy consumption and the use of sampling resources. We also expand our analysis by discussing the role of this approach for the efficient real-time measurement of stormwater systems.
Resumo:
In the tropical Atlantic Forest, 42 canopy gaps had their areas estimated using four different field methods of measurement: Runkle, Brokaw and Green [Runkle, J.R., 1981. Gap formation in some old-growth forests of the eastern United States. Ecology 62, 1041-1051; Brokaw, N.V.L., 1982. The definition of treefall gap and its effect on measures of forest dynamics. Biotropica 14, 158-160; Green, P.T., 1996. Canopy Gaps in rain forest on Christmas Island, Indian Ocean: size distribution and methods of measurement. J. Trop. Ecol. 12, 427-434] and a new method proposed in this work. It was found that within the same gap delimitation, average gap size varied from 56.0 up to 88.3 m(3) while total sum of gap area varied from 2351.3 to 3707.9 m(3) Differences among all methods and between pairs of method proved to be statistically significant. As a consequence, gap size-class distribution was also different between methods. When one method is held as a standard, deviation on average values of gap size ranged between 11.8 and 59.7% as deviations on single gap size can reach 172.8%. Implications on forest dynamics were expressed by the forest turnover rate that was 24% faster or 15% slower depending on the method adopted for gap measurement. Based on my results and on methods' evaluation, the use of a new method is proposed here for future research involving the measure of gap size in forest ecosystems. Finally, it is concluded that forest comparisons disregarding the influence of different methods of gap measurement should be reconsidered. (c) 2005 Elsevier B.V. All rights reserved.
Resumo:
This paper proposes to use a state-space technique to represent a frequency dependent line for simulating electromagnetic transients directly in time domain. The distributed nature of the line is represented by a multiple 1t section network made up of the lumped parameters and the frequency dependence of the per unit longitudinal parameters is matched by using a rational function. The rational function is represented by its equivalent circuit with passive elements. This passive circuit is then inserted in each 1t circuit of the cascade that represents the line. Because the system is very sparse, it is possible to use a sparsity technique to store only nonzero elements of this matrix for saving space and running time. The model was used to simulate the energization process of a 10 km length single-phase line. ©2008 IEEE.
Resumo:
Pós-graduação em Agronomia (Energia na Agricultura) - FCA
Resumo:
Ziel dieser Dissertation ist die experimentelle Charakterisierung und quantitative Beschreibung der Hybridisierung von komplementären Nukleinsäuresträngen mit oberflächengebundenen Fängermolekülen für die Entwicklung von integrierten Biosensoren. Im Gegensatz zu lösungsbasierten Verfahren ist mit Microarray Substraten die Untersuchung vieler Nukleinsäurekombinationen parallel möglich. Als biologisch relevantes Evaluierungssystem wurde das in Eukaryoten universell exprimierte Actin Gen aus unterschiedlichen Pflanzenspezies verwendet. Dieses Testsystem ermöglicht es, nahe verwandte Pflanzenarten auf Grund von geringen Unterschieden in der Gen-Sequenz (SNPs) zu charakterisieren. Aufbauend auf dieses gut studierte Modell eines House-Keeping Genes wurde ein umfassendes Microarray System, bestehend aus kurzen und langen Oligonukleotiden (mit eingebauten LNA-Molekülen), cDNAs sowie DNA und RNA Targets realisiert. Damit konnte ein für online Messung optimiertes Testsystem mit hohen Signalstärken entwickelt werden. Basierend auf den Ergebnissen wurde der gesamte Signalpfad von Nukleinsärekonzentration bis zum digitalen Wert modelliert. Die aus der Entwicklung und den Experimenten gewonnen Erkenntnisse über die Kinetik und Thermodynamik von Hybridisierung sind in drei Publikationen zusammengefasst die das Rückgrat dieser Dissertation bilden. Die erste Publikation beschreibt die Verbesserung der Reproduzierbarkeit und Spezifizität von Microarray Ergebnissen durch online Messung von Kinetik und Thermodynamik gegenüber endpunktbasierten Messungen mit Standard Microarrays. Für die Auswertung der riesigen Datenmengen wurden zwei Algorithmen entwickelt, eine reaktionskinetische Modellierung der Isothermen und ein auf der Fermi-Dirac Statistik beruhende Beschreibung des Schmelzüberganges. Diese Algorithmen werden in der zweiten Publikation beschrieben. Durch die Realisierung von gleichen Sequenzen in den chemisch unterschiedlichen Nukleinsäuren (DNA, RNA und LNA) ist es möglich, definierte Unterschiede in der Konformation des Riboserings und der C5-Methylgruppe der Pyrimidine zu untersuchen. Die kompetitive Wechselwirkung dieser unterschiedlichen Nukleinsäuren gleicher Sequenz und die Auswirkungen auf Kinetik und Thermodynamik ist das Thema der dritten Publikation. Neben der molekularbiologischen und technologischen Entwicklung im Bereich der Sensorik von Hybridisierungsreaktionen oberflächengebundener Nukleinsäuremolekülen, der automatisierten Auswertung und Modellierung der anfallenden Datenmengen und der damit verbundenen besseren quantitativen Beschreibung von Kinetik und Thermodynamik dieser Reaktionen tragen die Ergebnisse zum besseren Verständnis der physikalisch-chemischen Struktur des elementarsten biologischen Moleküls und seiner nach wie vor nicht vollständig verstandenen Spezifizität bei.
Resumo:
Identifying and comparing different steady states is an important task for clinical decision making. Data from unequal sources, comprising diverse patient status information, have to be interpreted. In order to compare results an expressive representation is the key. In this contribution we suggest a criterion to calculate a context-sensitive value based on variance analysis and discuss its advantages and limitations referring to a clinical data example obtained during anesthesia. Different drug plasma target levels of the anesthetic propofol were preset to reach and maintain clinically desirable steady state conditions with target controlled infusion (TCI). At the same time systolic blood pressure was monitored, depth of anesthesia was recorded using the bispectral index (BIS) and propofol plasma concentrations were determined in venous blood samples. The presented analysis of variance (ANOVA) is used to quantify how accurately steady states can be monitored and compared using the three methods of measurement.
Resumo:
Eph receptor tyrosine kinases and their ligands (ephrins) are key players during the development of the embryonic vasculature; however, their role and regulation in adult angiogenesis remain to be defined. Both receptors and ligands have been shown to be up-regulated in a variety of tumors. To address the hypothesis that hypoxia is an important regulator of Ephs/ephrins expression, we developed a mouse skin flap model of hypoxia. We demonstrate that our model truly represents segmental skin hypoxia by applying four independent methods: continuous measurement of partial cutaneous oxygen tension, monitoring of tissue lactate/pyruvate ratio, time course of hypoxia-inducible factor-1alpha (HIF-1alpha) induction, and localization of stabilized HIF-1alpha by immunofluorescence in the hypoxic skin flap. Our experiments indicate that hypoxia up-regulates not only HIF-1alpha and vascular endothelial growth factor (VEGF) expression, but also Ephs and ephrins of both A and B subclasses in the skin. In addition, we show that in Hep3B and PC-3 cells, the hypoxia-induced up-regulation of Ephs and ephrins is abrogated by small interfering RNA-mediated down-regulation of HIF-1alpha. These novel findings shed light on the role of this versatile receptor/ligand family in adult angiogenesis. Furthermore, our model offers considerable potential for analyzing distinct mechanisms of neovascularization in gene-targeted mice.
Resumo:
Existe un creciente interés internacional por el ahorro energético y la sostenibilidad en la edificación con importantes repercusiones en la Arquitectura. La inercia térmica es un parámetro fundamental para poder valorar energéticamente un edificio en condiciones reales. Para ello es necesario cambiar el enfoque tradicional de transmisión de calor en régimen estacionario por otro en régimen dinámico en el que se analizan las ondas térmicas y el flujo de calor oscilante que atraviesan los cerramientos. Los parámetros que definen la inercia térmica son: el espesor, la difusividad y el ciclo térmico. A su vez la difusividad está determinada por la conductividad térmica, la densidad y el calor específico del material. De estos parámetros la conductividad es el más complejo, variable y difícil de medir, especialmente en los cerramientos de tierra debido a su heterogeneidad y complejidad higrotérmica. En general, los métodos de medida de la conductividad o transmitancias en los paramentos presentan inconvenientes a la hora de medir un edificio construido con tierra: dificultades de implementación, el elevado coste o la fiabilidad de los resultados, principalmente. El Método de la Aguja Térmica (MAT) se basa en el principio de la evolución en el tiempo del calor emitido por una fuente lineal al insertarse en el seno de un material. Se ha escogido este método porque resulta práctico, de bajo coste y de fácil aplicación a gran escala pero tiene serios problemas de fiabilidad y exactitud. En esta tesis se desarrolla un método de medida de la conductividad térmica para Piezas de Albañilería de Tierra Cruda en laboratorio basado en el MAT, se mejora su fiabilidad, se analiza su incertidumbre, se compara con otros métodos de referencia y se aplica en adobes, Bloques de Tierra Comprimida y probetas de tierra estabilizada con distintas proporciones de paja. Este método servirá de base a una posterior aplicación in situ. Finalmente se proponen modelos matemáticos para mejorar la exactitud del dispositivo utilizado y para la estimación de la conductividad de cerramientos de tierra en función de su densidad. Con los resultados obtenidos se analizan las posibilidades de amortiguación y retardo de las ondas térmicas y capacidad de almacenaje de energía de los cerramientos en función de su densidad y humedad. There is growing international interest in energy saving and sustainability in buildings with significant impact on Architecture. Thermal inertia is a key parameter to assess energy in buildings in real conditions. This requires changing the traditional approach to heat transfer in steady state by another in dynamic regime which analyzes the thermal waves and oscillating heat flux passing through the external walls. The parameters defining the thermal inertia are: the thickness, the diffusivity and the thermal cycle. In turn, the diffusivity is determined by the thermal conductivity, density and specific heat of the material. Of these parameters, thermal conductivity is the most complex, variable and difficult to measure, especially in earth walls due to their heterogeneity and hygrothermal complexity. In general, the methods of measurement of conductivity and transmittance in walls have drawbacks when measuring a building with earth: implementation difficulties, high cost, or reliability of the results, mainly. The Thermal Needle Procedure (TNP) is based on the principle of evolution in time of heat from a line source when inserted within a material. This method was chosen because it is a practical, low cost and easy to implement on a large scale but has serious problems of reliability and accuracy. This thesis develops a laboratory method for measuring the thermal conductivity of Masonry Units Unfire Earth-based based on TNP, its uncertainty is analyzed, compared to other reference methods and applies in adobes, Compressed Earth Blocks and stabilized soil specimens with different proportions of straw. This method will form the basis of a subsequent application in situ. Finally, mathematical models are proposed to improve the accuracy of the device used, and to estimate the conductivity of earth enclosures depending on its density. With the results obtained earth enclosures are analyzed to estimate their possibilities of delay and buffer of termal waves and energy storage capacity according to their density and moisture.
Resumo:
In recent years, remote sensing imaging systems for the measurement of oceanic sea states have attracted renovated attention. Imaging technology is economical, non-invasive and enables a better understanding of the space-time dynamics of ocean waves over an area rather than at selected point locations of previous monitoring methods (buoys, wave gauges, etc.). We present recent progress in space-time measurement of ocean waves using stereo vision systems on offshore platforms, which focus on sea states with wavelengths in the range of 0.01 m to 10 m. Classical epipolar techniques and modern variational methods are reviewed to reconstruct the sea surface from the stereo pairs sequentially in time. The statistical and spectral properties of the resulting observed waves are analyzed. Current improvements of the variational methods are discussed as future lines of research.
Resumo:
In this work, a methodology is proposed to find the dynamic poles of a capacitive pressure transmitter in order to enhance and extend the online surveillance of this type of sensor based on the response time measurement by applying noise analysis techniques and the dynamic data system procedure. Several measurements taken from a pressurized water reactor have been analyzed. The methodology proposes an autoregressive fit whose order is determined by the sensor dynamic poles. Nevertheless, the signals that have been analyzed could not be filtered properly in order to remove the plant noise; thus, this was considered as an additional pair of complex conjugate poles. With this methodology we have come up with the numerical value of the sensor second real pole in spite of its low influence on the sensor dynamic response. This opens up a more accurate online sensor surveillance since the previous methods were achieved by considering one real pole only.