993 resultados para Sampling rate
Resumo:
Behavior is one of the most important indicators for assessing cattle health and well-being. The objective of this study was to develop and validate a novel algorithm to monitor locomotor behavior of loose-housed dairy cows based on the output of the RumiWatch pedometer (ITIN+HOCH GmbH, Fütterungstechnik, Liestal, Switzerland). Data of locomotion were acquired by simultaneous pedometer measurements at a sampling rate of 10 Hz and video recordings for manual observation later. The study consisted of 3 independent experiments. Experiment 1 was carried out to develop and validate the algorithm for lying behavior, experiment 2 for walking and standing behavior, and experiment 3 for stride duration and stride length. The final version was validated, using the raw data, collected from cows not included in the development of the algorithm. Spearman correlation coefficients were calculated between accelerometer variables and respective data derived from the video recordings (gold standard). Dichotomous data were expressed as the proportion of correctly detected events, and the overall difference for continuous data was expressed as the relative measurement error. The proportions for correctly detected events or bouts were 1 for stand ups, lie downs, standing bouts, and lying bouts and 0.99 for walking bouts. The relative measurement error and Spearman correlation coefficient for lying time were 0.09% and 1; for standing time, 4.7% and 0.96; for walking time, 17.12% and 0.96; for number of strides, 6.23% and 0.98; for stride duration, 6.65% and 0.75; and for stride length, 11.92% and 0.81, respectively. The strong to very high correlations of the variables between visual observation and converted pedometer data indicate that the novel RumiWatch algorithm may markedly improve automated livestock management systems for efficient health monitoring of dairy cows.
Resumo:
Many attempts have already been made to detect exomoons around transiting exoplanets, but the first confirmed discovery is still pending. The experiences that have been gathered so far allow us to better optimize future space telescopes for this challenge already during the development phase. In this paper we focus on the forthcoming CHaraterising ExOPlanet Satellite (CHEOPS), describing an optimized decision algorithm with step-by-step evaluation, and calculating the number of required transits for an exomoon detection for various planet moon configurations that can be observable by CHEOPS. We explore the most efficient way for such an observation to minimize the cost in observing time. Our study is based on PTV observations (photocentric transit timing variation) in simulated CHEOPS data, but the recipe does not depend on the actual detection method, and it can be substituted with, e.g., the photodynamical method for later applications. Using the current state-of-the-art level simulation of CHEOPS data we analyzed transit observation sets for different star planet moon configurations and performed a bootstrap analysis to determine their detection statistics. We have found that the detection limit is around an Earth-sized moon. In the case of favorable spatial configurations, systems with at least a large moon and a Neptune-sized planet, an 80% detection chance requires at least 5-6 transit observations on average. There is also a nonzero chance in the case of smaller moons, but the detection statistics deteriorate rapidly, while the necessary transit measurements increase quickly. After the CoRoT and Kepler spacecrafts, CHEOPS will be the next dedicated space telescope that will observe exoplanetary transits and characterize systems with known Doppler-planets. Although it has a smaller aperture than Kepler (the ratio of the mirror diameters is about 1/3) and is mounted with a CCD that is similar to Kepler's, it will observe brighter stars and operate with larger sampling rate; therefore, the detection limit for an exomoon can be the same as or better, which will make CHEOPS a competitive instruments in the quest for exomoons.
Resumo:
During the cruise the turbulence probe MSS050 was used for microstructure measurements. The profiler is produced by Sea and Sun Technology GmbH in co-operation with ISW Wassermesstechnik. The probe was equipped with 2 velocity microstructure shear sensors, a microstructure temperature sensor, standard CTD sensors for precision measurements, a vibration control sensor, a two component tilt sensor, a fluorescence sensor and surface detection sensor (SD) to indicate the water surface hit at rising measurements. The sampling rate for all sensors is 1024 samples per second, the resolution 16 bit. During the MSS measurements, the ship was moving with speed approx. 0.5-1.0 knots with respect to the water against the wind. In order to take into account the intermittent nature of marine turbulence, repeated MSS measurements were carried out in bursts of typically 10 profiles per station. The measurement interval was approximately 7 (10) min for a profile to 140 (200) dbar. During JC87 cruise 8 series of turbulence measurements were conducted. Fluorescence data was obtained on the 5 of them.
Resumo:
The seismic data were acquired north of the Knipovich Ridge on the western Svalbard margin during cruise MSM21/4. They were recorded using a Geometrics GeoEel streamer of either 120 channels (profiles p100-p208) or 88 channels (profiles p300-p805) with a group spacing of 1.56 m and a sampling rate of 2 kHz. A GI-Gun (2×1.7 l) with a main frequency of ~150 Hz was used as a source and operated at a shot interval of 6-8 s. Processing of profiles p100-p208 and p600-p805: Positions for each channel were calculated by backtracking along the profiles from the GI-Gun GPS positions. The shot gathers were analyzed for abnormal amplitudes below the seafloor reflection by comparing neighboring traces in different frequency bands within sliding time windows. To suppress surface-generated water noise, a tau-p filter was applied in the shot gather domain. Common mid-point (CMP) profiles were then generated through crooked-line binning with a CMP spacing of 1.5625 m. A zero-phase band-pass filter with corner frequencies of 60 Hz and 360 Hz was applied to the data. Based on regional velocity information from MCS data [Sarkar, 2012], an interpolated and extrapolated 3D interval velocity model was created below the digitized seafloor reflection of the high-resolution streamer data. This velocity model was used to apply a CMP stack and an amplitude-preserving Kirchhoff post-stack time migration. Processing of profiles p400-p500: Data were sampled at 0.5 ms and sorted into common midpoint (CMP) domain with a bin spacing of 5 m. Normal move out correction was carried out with a velocity of 1500 m s-1 and an Ormsby bandpass filter with corner frequencies at 40, 80, 600 and 1000 Hz was applied. The data were time migrated using the water velocity.
Resumo:
During summer seasons (from 2012 to 2015) high resolution observation were performed in the Civitavecchia coastal area (Northern Tyrrhenian sea, west coast). All sampling was carried out from a small boat (5m rigid inflatable) starting in the early morning typically around 06:00 a.m. , and lasting from 2 to 8 h, depending on the weather conditions. The purposes of the experiment was to observe the variations of both the coastal circulation and the water column in response to rotation of 180 ° in the wind direction. During surveys both current measurements and yo-yo time series were performed. Current data were acquired using an ADCP SonTeck (500 Khz, sampling interval from 20sec to 60 sec, average interval 50% sampling, cell thickness 1 meter) and the yo-yo time series employing a small instrument package (CTD). The CTD contained an Idronaut 316 Plus and a SeaPoint fluorometer. The sampling rate for the CTD was 10Hz, profiling with the CTD was done by allowing the instrument package to free-fall, at an average descent rate of 1 m/s. During the summer 2012, the sampling plan consisted in four stations spaced a quarter of a mile (St. 1 - 10 m; St. 2 - 20 m; St. 3 - 30 m; and St. 4 - 40 m), in which yo-yo time series were performed with an interval of 20 min. In order to study fluorescence of Chlorophyll a pathes distribution in coastal zone. Breeze induced circulation was the goal of the following summers surveys, in these current measurements and yo-yo time series were performed moored at a depth of 40 m. Offshore station (St. 1 -40m) has been chosen to perform measurement, basis of previously observations (2012 sampling surveys). It was decided as wind driven circulation and mixing phenomena are less influenced by seabed than the other stations. Acquired data have been processed by NEMO SeaDataNet software.
Resumo:
La temperatura es una preocupación que juega un papel protagonista en el diseño de circuitos integrados modernos. El importante aumento de las densidades de potencia que conllevan las últimas generaciones tecnológicas ha producido la aparición de gradientes térmicos y puntos calientes durante el funcionamiento normal de los chips. La temperatura tiene un impacto negativo en varios parámetros del circuito integrado como el retardo de las puertas, los gastos de disipación de calor, la fiabilidad, el consumo de energía, etc. Con el fin de luchar contra estos efectos nocivos, la técnicas de gestión dinámica de la temperatura (DTM) adaptan el comportamiento del chip en función en la información que proporciona un sistema de monitorización que mide en tiempo de ejecución la información térmica de la superficie del dado. El campo de la monitorización de la temperatura en el chip ha llamado la atención de la comunidad científica en los últimos años y es el objeto de estudio de esta tesis. Esta tesis aborda la temática de control de la temperatura en el chip desde diferentes perspectivas y niveles, ofreciendo soluciones a algunos de los temas más importantes. Los niveles físico y circuital se cubren con el diseño y la caracterización de dos nuevos sensores de temperatura especialmente diseñados para los propósitos de las técnicas DTM. El primer sensor está basado en un mecanismo que obtiene un pulso de anchura variable dependiente de la relación de las corrientes de fuga con la temperatura. De manera resumida, se carga un nodo del circuito y posteriormente se deja flotando de tal manera que se descarga a través de las corrientes de fugas de un transistor; el tiempo de descarga del nodo es la anchura del pulso. Dado que la anchura del pulso muestra una dependencia exponencial con la temperatura, la conversión a una palabra digital se realiza por medio de un contador logarítmico que realiza tanto la conversión tiempo a digital como la linealización de la salida. La estructura resultante de esta combinación de elementos se implementa en una tecnología de 0,35 _m. El sensor ocupa un área muy reducida, 10.250 nm2, y consume muy poca energía, 1.05-65.5nW a 5 muestras/s, estas cifras superaron todos los trabajos previos en el momento en que se publicó por primera vez y en el momento de la publicación de esta tesis, superan a todas las implementaciones anteriores fabricadas en el mismo nodo tecnológico. En cuanto a la precisión, el sensor ofrece una buena linealidad, incluso sin calibrar; se obtiene un error 3_ de 1,97oC, adecuado para tratar con las aplicaciones de DTM. Como se ha explicado, el sensor es completamente compatible con los procesos de fabricación CMOS, este hecho, junto con sus valores reducidos de área y consumo, lo hacen especialmente adecuado para la integración en un sistema de monitorización de DTM con un conjunto de monitores empotrados distribuidos a través del chip. Las crecientes incertidumbres de proceso asociadas a los últimos nodos tecnológicos comprometen las características de linealidad de nuestra primera propuesta de sensor. Con el objetivo de superar estos problemas, proponemos una nueva técnica para obtener la temperatura. La nueva técnica también está basada en las dependencias térmicas de las corrientes de fuga que se utilizan para descargar un nodo flotante. La novedad es que ahora la medida viene dada por el cociente de dos medidas diferentes, en una de las cuales se altera una característica del transistor de descarga |la tensión de puerta. Este cociente resulta ser muy robusto frente a variaciones de proceso y, además, la linealidad obtenida cumple ampliamente los requisitos impuestos por las políticas DTM |error 3_ de 1,17oC considerando variaciones del proceso y calibrando en dos puntos. La implementación de la parte sensora de esta nueva técnica implica varias consideraciones de diseño, tales como la generación de una referencia de tensión independiente de variaciones de proceso, que se analizan en profundidad en la tesis. Para la conversión tiempo-a-digital, se emplea la misma estructura de digitalización que en el primer sensor. Para la implementación física de la parte de digitalización, se ha construido una biblioteca de células estándar completamente nueva orientada a la reducción de área y consumo. El sensor resultante de la unión de todos los bloques se caracteriza por una energía por muestra ultra baja (48-640 pJ) y un área diminuta de 0,0016 mm2, esta cifra mejora todos los trabajos previos. Para probar esta afirmación, se realiza una comparación exhaustiva con más de 40 propuestas de sensores en la literatura científica. Subiendo el nivel de abstracción al sistema, la tercera contribución se centra en el modelado de un sistema de monitorización que consiste de un conjunto de sensores distribuidos por la superficie del chip. Todos los trabajos anteriores de la literatura tienen como objetivo maximizar la precisión del sistema con el mínimo número de monitores. Como novedad, en nuestra propuesta se introducen nuevos parámetros de calidad aparte del número de sensores, también se considera el consumo de energía, la frecuencia de muestreo, los costes de interconexión y la posibilidad de elegir diferentes tipos de monitores. El modelo se introduce en un algoritmo de recocido simulado que recibe la información térmica de un sistema, sus propiedades físicas, limitaciones de área, potencia e interconexión y una colección de tipos de monitor; el algoritmo proporciona el tipo seleccionado de monitor, el número de monitores, su posición y la velocidad de muestreo _optima. Para probar la validez del algoritmo, se presentan varios casos de estudio para el procesador Alpha 21364 considerando distintas restricciones. En comparación con otros trabajos previos en la literatura, el modelo que aquí se presenta es el más completo. Finalmente, la última contribución se dirige al nivel de red, partiendo de un conjunto de monitores de temperatura de posiciones conocidas, nos concentramos en resolver el problema de la conexión de los sensores de una forma eficiente en área y consumo. Nuestra primera propuesta en este campo es la introducción de un nuevo nivel en la jerarquía de interconexión, el nivel de trillado (o threshing en inglés), entre los monitores y los buses tradicionales de periféricos. En este nuevo nivel se aplica selectividad de datos para reducir la cantidad de información que se envía al controlador central. La idea detrás de este nuevo nivel es que en este tipo de redes la mayoría de los datos es inútil, porque desde el punto de vista del controlador sólo una pequeña cantidad de datos |normalmente sólo los valores extremos| es de interés. Para cubrir el nuevo nivel, proponemos una red de monitorización mono-conexión que se basa en un esquema de señalización en el dominio de tiempo. Este esquema reduce significativamente tanto la actividad de conmutación sobre la conexión como el consumo de energía de la red. Otra ventaja de este esquema es que los datos de los monitores llegan directamente ordenados al controlador. Si este tipo de señalización se aplica a sensores que realizan conversión tiempo-a-digital, se puede obtener compartición de recursos de digitalización tanto en tiempo como en espacio, lo que supone un importante ahorro de área y consumo. Finalmente, se presentan dos prototipos de sistemas de monitorización completos que de manera significativa superan la características de trabajos anteriores en términos de área y, especialmente, consumo de energía. Abstract Temperature is a first class design concern in modern integrated circuits. The important increase in power densities associated to recent technology evolutions has lead to the apparition of thermal gradients and hot spots during run time operation. Temperature impacts several circuit parameters such as speed, cooling budgets, reliability, power consumption, etc. In order to fight against these negative effects, dynamic thermal management (DTM) techniques adapt the behavior of the chip relying on the information of a monitoring system that provides run-time thermal information of the die surface. The field of on-chip temperature monitoring has drawn the attention of the scientific community in the recent years and is the object of study of this thesis. This thesis approaches the matter of on-chip temperature monitoring from different perspectives and levels, providing solutions to some of the most important issues. The physical and circuital levels are covered with the design and characterization of two novel temperature sensors specially tailored for DTM purposes. The first sensor is based upon a mechanism that obtains a pulse with a varying width based on the variations of the leakage currents on the temperature. In a nutshell, a circuit node is charged and subsequently left floating so that it discharges away through the subthreshold currents of a transistor; the time the node takes to discharge is the width of the pulse. Since the width of the pulse displays an exponential dependence on the temperature, the conversion into a digital word is realized by means of a logarithmic counter that performs both the timeto- digital conversion and the linearization of the output. The structure resulting from this combination of elements is implemented in a 0.35_m technology and is characterized by very reduced area, 10250 nm2, and power consumption, 1.05-65.5 nW at 5 samples/s, these figures outperformed all previous works by the time it was first published and still, by the time of the publication of this thesis, they outnumber all previous implementations in the same technology node. Concerning the accuracy, the sensor exhibits good linearity, even without calibration it displays a 3_ error of 1.97oC, appropriate to deal with DTM applications. As explained, the sensor is completely compatible with standard CMOS processes, this fact, along with its tiny area and power overhead, makes it specially suitable for the integration in a DTM monitoring system with a collection of on-chip monitors distributed across the chip. The exacerbated process fluctuations carried along with recent technology nodes jeop-ardize the linearity characteristics of the first sensor. In order to overcome these problems, a new temperature inferring technique is proposed. In this case, we also rely on the thermal dependencies of leakage currents that are used to discharge a floating node, but now, the result comes from the ratio of two different measures, in one of which we alter a characteristic of the discharging transistor |the gate voltage. This ratio proves to be very robust against process variations and displays a more than suficient linearity on the temperature |1.17oC 3_ error considering process variations and performing two-point calibration. The implementation of the sensing part based on this new technique implies several issues, such as the generation of process variations independent voltage reference, that are analyzed in depth in the thesis. In order to perform the time-to-digital conversion, we employ the same digitization structure the former sensor used. A completely new standard cell library targeting low area and power overhead is built from scratch to implement the digitization part. Putting all the pieces together, we achieve a complete sensor system that is characterized by ultra low energy per conversion of 48-640pJ and area of 0.0016mm2, this figure outperforms all previous works. To prove this statement, we perform a thorough comparison with over 40 works from the scientific literature. Moving up to the system level, the third contribution is centered on the modeling of a monitoring system consisting of set of thermal sensors distributed across the chip. All previous works from the literature target maximizing the accuracy of the system with the minimum number of monitors. In contrast, we introduce new metrics of quality apart form just the number of sensors; we consider the power consumption, the sampling frequency, the possibility to consider different types of monitors and the interconnection costs. The model is introduced in a simulated annealing algorithm that receives the thermal information of a system, its physical properties, area, power and interconnection constraints and a collection of monitor types; the algorithm yields the selected type of monitor, the number of monitors, their position and the optimum sampling rate. We test the algorithm with the Alpha 21364 processor under several constraint configurations to prove its validity. When compared to other previous works in the literature, the modeling presented here is the most complete. Finally, the last contribution targets the networking level, given an allocated set of temperature monitors, we focused on solving the problem of connecting them in an efficient way from the area and power perspectives. Our first proposal in this area is the introduction of a new interconnection hierarchy level, the threshing level, in between the monitors and the traditional peripheral buses that applies data selectivity to reduce the amount of information that is sent to the central controller. The idea behind this new level is that in this kind of networks most data are useless because from the controller viewpoint just a small amount of data |normally extreme values| is of interest. To cover the new interconnection level, we propose a single-wire monitoring network based on a time-domain signaling scheme that significantly reduces both the switching activity over the wire and the power consumption of the network. This scheme codes the information in the time domain and allows a straightforward obtention of an ordered list of values from the maximum to the minimum. If the scheme is applied to monitors that employ TDC, digitization resource sharing is achieved, producing an important saving in area and power consumption. Two prototypes of complete monitoring systems are presented, they significantly overcome previous works in terms of area and, specially, power consumption.
Resumo:
A iodação do sal de mesa é considerada o caminho mais eficiente para controlar os Distúrbios por Deficiência de Iodo. Em países tropicais, o elemento pode ser adicionado na forma de KIO3. Para garantir que os níveis ideais do ânion estejam disponíveis ao consumidor, o controle de qualidade do sal consiste numa estratégia fundamental. Sistemas em fluxo com multicomutação representam uma alternativa versátil para o desenvolvimento de procedimentos simples, rápidos e limpos, minimizando o consumo de reagentes e a geração de resíduos. Nesse contexto, um procedimento analítico utilizando sistema com multicomutação e detecção espectrofotométrica foi desenvolvido para a determinação de iodato em sal de mesa. A reação empregada foi baseada na formação de um composto roxo (540 nm) entre iodato (IO3-) e p-aminofenol (PAP) em meio ácido. O tempo de residência da zona de amostra no percurso analítico foi explorado a fim de favorecer a reação lenta e a frequência de amostragem para a melhoria do desempenho analítico. Foram selecionados 2 pulsos para inserção de amostra, 3 pulsos para reagente (PAP 0,25% (m/v) em HCl 0,025 mol L-1), 7 ciclos de amostragem, 200 pulsos de carregador (água), bolha de ar de 1 s (40 µL), reator de 70 cm (3 mm d.i.) e parada de fluxo de 480 s. Resposta linear foi observada entre 2,28x10-5 e 3,65x10-4 mol L-1, descrita pela equação A = 0,2443 + 2030 C, r = 0,997. Limite de detecção (99,7% de confiança), coeficiente de variação (n = 20) e frequência de amostragem foram estimados em 8,2x10-6 mol L-1, 0,42% e 70 determinações por hora, respectivamente. Houve consumo de 1,05 mg de PAP e geração de 0,70 mL de resíduos por determinação. As principais espécies concomitantes presentes na amostra não interferiram na determinação de iodato em concentrações até 8 vezes maiores que as usualmente encontradas. Estudos de adição e recuperação de iodato foram realizados pelo procedimento proposto, obtendo porcentagens de recuperação entre 88 e 104%. O procedimento analítico desenvolvido apresenta sensibilidade adequada para a determinação de iodato em amostra de sal de mesa e elevada frequência de amostragem quando comparado com procedimentos descritos na literatura
Resumo:
A ciência na qual se estuda a deformação de um fluido no qual é aplicada uma tensão de cisalhamento é conhecida como reologia e o equipamento utilizado para a realização dos ensaios é chamado de reômetro. Devido a impraticabilidade de uso de reômetros comerciais, diversos pesquisadores desenvolveram reômetros capazes de analisar suspensões de macropartículas, baseados nos mesmos princípios de funcionamento dos equipamentos já existentes. Em alguns casos, a medição do torque do motor é realizada pela aquisição da tensão, uma vez que esta é proporcional ao torque. Entretanto, para melhor compreensão do resultado e para evitar a possibilidade de conclusões precipitadas, vê-se necessária correta interpretação do sinal elétrico, precisando avaliar qual frequência do sinal é relevante para o ensaio e, também, qual a melhor taxa de amostragem. Além da aquisição, para que o ensaio reológico seja realizado com precisão, é indispensável ótimo controle da taxa ou tensão do motor e uma alternativa é a utilização de um servomotor e um servoconversor. No caso desse ser comercial é essencial saber configurá-lo. Para facilitar o usuário leigo, alguns pesquisadores desenvolveram softwares para controle do equipamento e análise dos dados. Assim, o presente trabalho tem como objetivo propor uma metodologia para compreender o sinal aquisitado de um reômetro servo controlado e desenvolvimento do software de análise para o tratamento dos dados obtidos a partir de ensaios reológicos. Verificou-se a melhor configuração do servocontrolador, a melhor taxa de amostragem, de no mínimo 20 amostras/segundo, e, também, desenvolveu-se um filtro digital passa-baixa do tipo FIR para remover a frequência indesejada. Além disso, foi desenvolvido um software utilizando uma rotina em Matlab e uma interface gráfica do usuário (Graphical User Interface - GUI), para o pós-processamento dos dados para auxiliar o usuário leigo no tratamento e interpretação do resultado, que se mostrou eficaz.
Resumo:
The published requirements for accurate measurement of heat transfer at the interface between two bodies have been reviewed. A strategy for reliable measurement has been established, based on the depth of the temperature sensors in the medium, on the inverse method parameters and on the time response of the sensors. Sources of both deterministic and stochastic errors have been investigated and a method to evaluate them has been proposed, with the help of a normalisation technique. The key normalisation variables are the duration of the heat input and the maximum heat flux density. An example of application of this technique in the field of high pressure die casting is demonstrated. The normalisation study, coupled with previous determination of the heat input duration, makes it possible to determine the optimum location for the sensors, along with an acceptable sampling rate and the thermocouples critical response-time (as well as eventual filter characteristics). Results from the gauge are used to assess the suitability of the initial design choices. In particular the unavoidable response time of the thermocouples is estimated by comparison with the normalised simulation. (c) 2006 Elsevier Ltd. All rights reserved.
Resumo:
We numerically investigate the combination of full-field detection and feed-forward equalizer (FFE) for adaptive chromatic dispersion compensation up to 2160 km in a 10 Gbit/s on-off keyed optical transmission system. The technique, with respect to earlier reports, incorporates several important implementation modules, including the algorithm for adaptive equalization of the gain imbalance between the two receiver chains, compensation of phase misalignment of the asymmetric Mach-Zehnder interferometer, and simplified implementation of field calculation. We also show that in addition to enabling fast adaptation and simplification of field calculation, full-field FFE exhibits enhanced tolerance to the sampling phase misalignment and reduced sampling rate when compared to the full-field implementation using a dispersive transmission line.
Resumo:
We investigate electronic mitigation of linear and non-linear fibre impairments and compare various digital signal processing techniques, including electronic dispersion compensation (EDC), single-channel back-propagation (SC-BP) and back-propagation with multiple channel processing (MC-BP) in a nine-channel 112 Gb/s PM-mQAM (m=4,16) WDM system, for reaches up to 6,320 km. We show that, for a sufficiently high local dispersion, SC-BP is sufficient to provide a significant performance enhancement when compared to EDC, and is adequate to achieve BER below FEC threshold. For these conditions we report that a sampling rate of two samples per symbol is sufficient for practical SC-BP, without significant penalties.
Resumo:
We numerically investigate the combination of full-field detection and feed-forward equalizer (FFE) for adaptive chromatic dispersion compensation up to 2160 km in a 10 Gbit/s on-off keyed optical transmission system. The technique, with respect to earlier reports, incorporates several important implementation modules, including the algorithm for adaptive equalization of the gain imbalance between the two receiver chains, compensation of phase misalignment of the asymmetric Mach-Zehnder interferometer, and simplified implementation of field calculation. We also show that in addition to enabling fast adaptation and simplification of field calculation, full-field FFE exhibits enhanced tolerance to the sampling phase misalignment and reduced sampling rate when compared to the full-field implementation using a dispersive transmission line.
Resumo:
2000 Mathematics Subject Classification: 62P10, 92C20
Resumo:
The purpose of this research is design considerations for environmental monitoring platforms for the detection of hazardous materials using System-on-a-Chip (SoC) design. Design considerations focus on improving key areas such as: (1) sampling methodology; (2) context awareness; and (3) sensor placement. These design considerations for environmental monitoring platforms using wireless sensor networks (WSN) is applied to the detection of methylmercury (MeHg) and environmental parameters affecting its formation (methylation) and deformation (demethylation). ^ The sampling methodology investigates a proof-of-concept for the monitoring of MeHg using three primary components: (1) chemical derivatization; (2) preconcentration using the purge-and-trap (P&T) method; and (3) sensing using Quartz Crystal Microbalance (QCM) sensors. This study focuses on the measurement of inorganic mercury (Hg) (e.g., Hg2+) and applies lessons learned to organic Hg (e.g., MeHg) detection. ^ Context awareness of a WSN and sampling strategies is enhanced by using spatial analysis techniques, namely geostatistical analysis (i.e., classical variography and ordinary point kriging), to help predict the phenomena of interest in unmonitored locations (i.e., locations without sensors). This aids in making more informed decisions on control of the WSN (e.g., communications strategy, power management, resource allocation, sampling rate and strategy, etc.). This methodology improves the precision of controllability by adding potentially significant information of unmonitored locations.^ There are two types of sensors that are investigated in this study for near-optimal placement in a WSN: (1) environmental (e.g., humidity, moisture, temperature, etc.) and (2) visual (e.g., camera) sensors. The near-optimal placement of environmental sensors is found utilizing a strategy which minimizes the variance of spatial analysis based on randomly chosen points representing the sensor locations. Spatial analysis is employed using geostatistical analysis and optimization occurs with Monte Carlo analysis. Visual sensor placement is accomplished for omnidirectional cameras operating in a WSN using an optimal placement metric (OPM) which is calculated for each grid point based on line-of-site (LOS) in a defined number of directions where known obstacles are taken into consideration. Optimal areas of camera placement are determined based on areas generating the largest OPMs. Statistical analysis is examined by using Monte Carlo analysis with varying number of obstacles and cameras in a defined space. ^
Resumo:
The Amyotrophic Lateral Sclerosis (ALS) is a neurodegenerative disease characterized by progressive muscle weakness that leads the patient to death, usually due to respiratory complications. Thus, as the disease progresses the patient will require noninvasive ventilation (NIV) and constant monitoring. This paper presents a distributed architecture for homecare monitoring of nocturnal NIV in patients with ALS. The implementation of this architecture used single board computers and mobile devices placed in patient’s homes, to display alert messages for caregivers and a web server for remote monitoring by the healthcare staff. The architecture used a software based on fuzzy logic and computer vision to capture data from a mechanical ventilator screen and generate alert messages with instructions for caregivers. The monitoring was performed on 29 patients for 7 con-tinuous hours daily during 5 days generating a total of 126000 samples for each variable monitored at a sampling rate of one sample per second. The system was evaluated regarding the rate of hits for character recognition and its correction through an algorithm for the detection and correction of errors. Furthermore, a healthcare team evaluated regarding the time intervals at which the alert messages were generated and the correctness of such messages. Thus, the system showed an average hit rate of 98.72%, and in the worst case 98.39%. As for the message to be generated, the system also agreed 100% to the overall assessment, and there was disagreement in only 2 cases with one of the physician evaluators.