935 resultados para Histogram of oriented gradients (HOG)
Resumo:
La temperatura es una preocupación que juega un papel protagonista en el diseño de circuitos integrados modernos. El importante aumento de las densidades de potencia que conllevan las últimas generaciones tecnológicas ha producido la aparición de gradientes térmicos y puntos calientes durante el funcionamiento normal de los chips. La temperatura tiene un impacto negativo en varios parámetros del circuito integrado como el retardo de las puertas, los gastos de disipación de calor, la fiabilidad, el consumo de energía, etc. Con el fin de luchar contra estos efectos nocivos, la técnicas de gestión dinámica de la temperatura (DTM) adaptan el comportamiento del chip en función en la información que proporciona un sistema de monitorización que mide en tiempo de ejecución la información térmica de la superficie del dado. El campo de la monitorización de la temperatura en el chip ha llamado la atención de la comunidad científica en los últimos años y es el objeto de estudio de esta tesis. Esta tesis aborda la temática de control de la temperatura en el chip desde diferentes perspectivas y niveles, ofreciendo soluciones a algunos de los temas más importantes. Los niveles físico y circuital se cubren con el diseño y la caracterización de dos nuevos sensores de temperatura especialmente diseñados para los propósitos de las técnicas DTM. El primer sensor está basado en un mecanismo que obtiene un pulso de anchura variable dependiente de la relación de las corrientes de fuga con la temperatura. De manera resumida, se carga un nodo del circuito y posteriormente se deja flotando de tal manera que se descarga a través de las corrientes de fugas de un transistor; el tiempo de descarga del nodo es la anchura del pulso. Dado que la anchura del pulso muestra una dependencia exponencial con la temperatura, la conversión a una palabra digital se realiza por medio de un contador logarítmico que realiza tanto la conversión tiempo a digital como la linealización de la salida. La estructura resultante de esta combinación de elementos se implementa en una tecnología de 0,35 _m. El sensor ocupa un área muy reducida, 10.250 nm2, y consume muy poca energía, 1.05-65.5nW a 5 muestras/s, estas cifras superaron todos los trabajos previos en el momento en que se publicó por primera vez y en el momento de la publicación de esta tesis, superan a todas las implementaciones anteriores fabricadas en el mismo nodo tecnológico. En cuanto a la precisión, el sensor ofrece una buena linealidad, incluso sin calibrar; se obtiene un error 3_ de 1,97oC, adecuado para tratar con las aplicaciones de DTM. Como se ha explicado, el sensor es completamente compatible con los procesos de fabricación CMOS, este hecho, junto con sus valores reducidos de área y consumo, lo hacen especialmente adecuado para la integración en un sistema de monitorización de DTM con un conjunto de monitores empotrados distribuidos a través del chip. Las crecientes incertidumbres de proceso asociadas a los últimos nodos tecnológicos comprometen las características de linealidad de nuestra primera propuesta de sensor. Con el objetivo de superar estos problemas, proponemos una nueva técnica para obtener la temperatura. La nueva técnica también está basada en las dependencias térmicas de las corrientes de fuga que se utilizan para descargar un nodo flotante. La novedad es que ahora la medida viene dada por el cociente de dos medidas diferentes, en una de las cuales se altera una característica del transistor de descarga |la tensión de puerta. Este cociente resulta ser muy robusto frente a variaciones de proceso y, además, la linealidad obtenida cumple ampliamente los requisitos impuestos por las políticas DTM |error 3_ de 1,17oC considerando variaciones del proceso y calibrando en dos puntos. La implementación de la parte sensora de esta nueva técnica implica varias consideraciones de diseño, tales como la generación de una referencia de tensión independiente de variaciones de proceso, que se analizan en profundidad en la tesis. Para la conversión tiempo-a-digital, se emplea la misma estructura de digitalización que en el primer sensor. Para la implementación física de la parte de digitalización, se ha construido una biblioteca de células estándar completamente nueva orientada a la reducción de área y consumo. El sensor resultante de la unión de todos los bloques se caracteriza por una energía por muestra ultra baja (48-640 pJ) y un área diminuta de 0,0016 mm2, esta cifra mejora todos los trabajos previos. Para probar esta afirmación, se realiza una comparación exhaustiva con más de 40 propuestas de sensores en la literatura científica. Subiendo el nivel de abstracción al sistema, la tercera contribución se centra en el modelado de un sistema de monitorización que consiste de un conjunto de sensores distribuidos por la superficie del chip. Todos los trabajos anteriores de la literatura tienen como objetivo maximizar la precisión del sistema con el mínimo número de monitores. Como novedad, en nuestra propuesta se introducen nuevos parámetros de calidad aparte del número de sensores, también se considera el consumo de energía, la frecuencia de muestreo, los costes de interconexión y la posibilidad de elegir diferentes tipos de monitores. El modelo se introduce en un algoritmo de recocido simulado que recibe la información térmica de un sistema, sus propiedades físicas, limitaciones de área, potencia e interconexión y una colección de tipos de monitor; el algoritmo proporciona el tipo seleccionado de monitor, el número de monitores, su posición y la velocidad de muestreo _optima. Para probar la validez del algoritmo, se presentan varios casos de estudio para el procesador Alpha 21364 considerando distintas restricciones. En comparación con otros trabajos previos en la literatura, el modelo que aquí se presenta es el más completo. Finalmente, la última contribución se dirige al nivel de red, partiendo de un conjunto de monitores de temperatura de posiciones conocidas, nos concentramos en resolver el problema de la conexión de los sensores de una forma eficiente en área y consumo. Nuestra primera propuesta en este campo es la introducción de un nuevo nivel en la jerarquía de interconexión, el nivel de trillado (o threshing en inglés), entre los monitores y los buses tradicionales de periféricos. En este nuevo nivel se aplica selectividad de datos para reducir la cantidad de información que se envía al controlador central. La idea detrás de este nuevo nivel es que en este tipo de redes la mayoría de los datos es inútil, porque desde el punto de vista del controlador sólo una pequeña cantidad de datos |normalmente sólo los valores extremos| es de interés. Para cubrir el nuevo nivel, proponemos una red de monitorización mono-conexión que se basa en un esquema de señalización en el dominio de tiempo. Este esquema reduce significativamente tanto la actividad de conmutación sobre la conexión como el consumo de energía de la red. Otra ventaja de este esquema es que los datos de los monitores llegan directamente ordenados al controlador. Si este tipo de señalización se aplica a sensores que realizan conversión tiempo-a-digital, se puede obtener compartición de recursos de digitalización tanto en tiempo como en espacio, lo que supone un importante ahorro de área y consumo. Finalmente, se presentan dos prototipos de sistemas de monitorización completos que de manera significativa superan la características de trabajos anteriores en términos de área y, especialmente, consumo de energía. Abstract Temperature is a first class design concern in modern integrated circuits. The important increase in power densities associated to recent technology evolutions has lead to the apparition of thermal gradients and hot spots during run time operation. Temperature impacts several circuit parameters such as speed, cooling budgets, reliability, power consumption, etc. In order to fight against these negative effects, dynamic thermal management (DTM) techniques adapt the behavior of the chip relying on the information of a monitoring system that provides run-time thermal information of the die surface. The field of on-chip temperature monitoring has drawn the attention of the scientific community in the recent years and is the object of study of this thesis. This thesis approaches the matter of on-chip temperature monitoring from different perspectives and levels, providing solutions to some of the most important issues. The physical and circuital levels are covered with the design and characterization of two novel temperature sensors specially tailored for DTM purposes. The first sensor is based upon a mechanism that obtains a pulse with a varying width based on the variations of the leakage currents on the temperature. In a nutshell, a circuit node is charged and subsequently left floating so that it discharges away through the subthreshold currents of a transistor; the time the node takes to discharge is the width of the pulse. Since the width of the pulse displays an exponential dependence on the temperature, the conversion into a digital word is realized by means of a logarithmic counter that performs both the timeto- digital conversion and the linearization of the output. The structure resulting from this combination of elements is implemented in a 0.35_m technology and is characterized by very reduced area, 10250 nm2, and power consumption, 1.05-65.5 nW at 5 samples/s, these figures outperformed all previous works by the time it was first published and still, by the time of the publication of this thesis, they outnumber all previous implementations in the same technology node. Concerning the accuracy, the sensor exhibits good linearity, even without calibration it displays a 3_ error of 1.97oC, appropriate to deal with DTM applications. As explained, the sensor is completely compatible with standard CMOS processes, this fact, along with its tiny area and power overhead, makes it specially suitable for the integration in a DTM monitoring system with a collection of on-chip monitors distributed across the chip. The exacerbated process fluctuations carried along with recent technology nodes jeop-ardize the linearity characteristics of the first sensor. In order to overcome these problems, a new temperature inferring technique is proposed. In this case, we also rely on the thermal dependencies of leakage currents that are used to discharge a floating node, but now, the result comes from the ratio of two different measures, in one of which we alter a characteristic of the discharging transistor |the gate voltage. This ratio proves to be very robust against process variations and displays a more than suficient linearity on the temperature |1.17oC 3_ error considering process variations and performing two-point calibration. The implementation of the sensing part based on this new technique implies several issues, such as the generation of process variations independent voltage reference, that are analyzed in depth in the thesis. In order to perform the time-to-digital conversion, we employ the same digitization structure the former sensor used. A completely new standard cell library targeting low area and power overhead is built from scratch to implement the digitization part. Putting all the pieces together, we achieve a complete sensor system that is characterized by ultra low energy per conversion of 48-640pJ and area of 0.0016mm2, this figure outperforms all previous works. To prove this statement, we perform a thorough comparison with over 40 works from the scientific literature. Moving up to the system level, the third contribution is centered on the modeling of a monitoring system consisting of set of thermal sensors distributed across the chip. All previous works from the literature target maximizing the accuracy of the system with the minimum number of monitors. In contrast, we introduce new metrics of quality apart form just the number of sensors; we consider the power consumption, the sampling frequency, the possibility to consider different types of monitors and the interconnection costs. The model is introduced in a simulated annealing algorithm that receives the thermal information of a system, its physical properties, area, power and interconnection constraints and a collection of monitor types; the algorithm yields the selected type of monitor, the number of monitors, their position and the optimum sampling rate. We test the algorithm with the Alpha 21364 processor under several constraint configurations to prove its validity. When compared to other previous works in the literature, the modeling presented here is the most complete. Finally, the last contribution targets the networking level, given an allocated set of temperature monitors, we focused on solving the problem of connecting them in an efficient way from the area and power perspectives. Our first proposal in this area is the introduction of a new interconnection hierarchy level, the threshing level, in between the monitors and the traditional peripheral buses that applies data selectivity to reduce the amount of information that is sent to the central controller. The idea behind this new level is that in this kind of networks most data are useless because from the controller viewpoint just a small amount of data |normally extreme values| is of interest. To cover the new interconnection level, we propose a single-wire monitoring network based on a time-domain signaling scheme that significantly reduces both the switching activity over the wire and the power consumption of the network. This scheme codes the information in the time domain and allows a straightforward obtention of an ordered list of values from the maximum to the minimum. If the scheme is applied to monitors that employ TDC, digitization resource sharing is achieved, producing an important saving in area and power consumption. Two prototypes of complete monitoring systems are presented, they significantly overcome previous works in terms of area and, specially, power consumption.
Crack mechanical failure in ceramic materials under ion irradiation: case of lithium niobate crystal
Resumo:
Swift heavy ion irradiation (ions with mass heavier than 15 and energy exceeding MeV/amu) transfer their energy mainly to the electronic system with small momentum transfer per collision. Therefore, they produce linear regions (columnar nano-tracks) around the straight ion trajectory, with marked modifications with respect to the virgin material, e.g., phase transition, amorphization, compaction, changes in physical or chemical properties. In the case of crystalline materials the most distinctive feature of swift heavy ion irradiation is the production of amorphous tracks embedded in the crystal. Lithium niobate is a relevant optical material that presents birefringence due to its anysotropic trigonal structure. The amorphous phase is certainly isotropic. In addition, its refractive index exhibits high contrast with those of the crystalline phase. This allows one to fabricate waveguides by swift ion irradiation with important technological relevance. From the mechanical point of view, the inclusion of an amorphous nano-track (with a density 15% lower than that of the crystal) leads to the generation of important stress/strain fields around the track. Eventually these fields are the origin of crack formation with fatal consequences for the integrity of the samples and the viability of the method for nano-track formation. For certain crystal cuts (X and Y), these fields are clearly anisotropic due to the crystal anisotropy. We have used finite element methods to calculate the stress/strain fields that appear around the ion- generated amorphous nano-tracks for a variety of ion energies and doses. A very remarkable feature for X cut-samples is that the maximum shear stress appears on preferential planes that form +/-45º with respect to the crystallographic planes. This leads to the generation of oriented surface cracks when the dose increases. The growth of the cracks along the anisotropic crystal has been studied by means of novel extended finite element methods, which include cracks as discontinuities. In this way we can study how the length and depth of a crack evolves as function of the ion dose. In this work we will show how the simulations compare with experiments and their application in materials modification by ion irradiation.
Resumo:
The study of temperature gradients in cold stores and containers is a critical issue in the food industry for the quality assurance of products during transport, as well as forminimizing losses. The objective of this work is to develop a new methodology of data analysis based on phase space graphs of temperature and enthalpy, collected by means of multidistributed, low cost and autonomous wireless sensors and loggers. A transoceanic refrigerated transport of lemons in a reefer container ship from Montevideo (Uruguay) to Cartagena (Spain) was monitored with a network of 39 semi-passive TurboTag RFID loggers and 13 i-button loggers. Transport included intermodal transit from transoceanic to short shipping vessels and a truck trip. Data analysis is carried out using qualitative phase diagrams computed on the basis of Takens?Ruelle reconstruction of attractors. Fruit stress is quantified in terms of the phase diagram area which characterizes the cyclic behaviour of temperature. Areas within the enthalpy phase diagram computed for the short sea shipping transport were 5 times higher than those computed for the long sea shipping, with coefficients of variation above 100% for both periods. This new methodology for data analysis highlights the significant heterogeneity of thermohygrometric conditions at different locations in the container.
Resumo:
Applying a brief repolarizing pre-pulse to a depolarized frog skeletal muscle fiber restores a small fraction of the transverse tubule membrane voltage sensors from the inactivated state. During a subsequent depolarizing test pulse we detected brief, highly localized elevations of myoplasmic Ca2+ concentration (Ca2+ “sparks”) initiated by restored voltage sensors in individual triads at all test pulse voltages. The latency histogram of these events gives the gating pattern of the sarcoplasmic reticulum (SR) calcium release channels controlled by the restored voltage sensors. Both event frequency and clustering of events near the start of the test pulse increase with test pulse depolarization. The macroscopic SR calcium release waveform, obtained from the spark latency histogram and the estimated open time of the channel or channels underlying a spark, exhibits an early peak and rapid marked decline during large depolarizations. For smaller depolarizations, the release waveform exhibits a smaller peak and a slower decline. However, the mean use time and mean amplitude of the individual sparks are quite similar at all test depolarizations and at all times during a given depolarization, indicating that the channel open times and conductances underlying sparks are essentially independent of voltage. Thus, the voltage dependence of SR Ca2+ release is due to changes in the frequency and pattern of occurrence of individual, voltage-independent, discrete release events.
Resumo:
We use residual-delay maps of observational field data for barometric pressure to demonstrate the structure of latitudinal gradients in nonlinearity in the atmosphere. Nonlinearity is weak and largely lacking in tropical and subtropical sites and increases rapidly into the temperate regions where the time series also appear to be much noisier. The degree of nonlinearity closely follows the meridional variation of midlatitude storm track frequency. We extract the specific functional form of this nonlinearity, a V shape in the lagged residuals that appears to be a basic feature of midlatitude synoptic weather systems associated with frontal passages. We present evidence that this form arises from the relative time scales of high-pressure versus low-pressure events. Finally, we show that this nonlinear feature is weaker in a well regarded numerical forecast model (European Centre for Medium-Range Forecasts) because small-scale temporal and spatial variation is smoothed out in the grided inputs. This is significant, in that it allows us to demonstrate how application of statistical corrections based on the residual-delay map may provide marked increases in local forecast accuracy, especially for severe weather systems.
Resumo:
Although it has been known for decades that the tight junctions of fluid-transporting epithelia are leaky to ions, it has not been possible to determine directly whether significant transjunctional water movement also occurs. An optical microscopic technique was developed for the direct visualization of the flow velocity profiles within the lateral intercellular spaces of a fluid-absorptive, cultured renal epithelium (MDCK) and used to determine the velocity of the fluid flow across the tight junction. The flow velocity within the lateral intercellular spaces fell to near zero adjacent to the tight junction, showing that significant transjunctional flow did not occur, even when transepithelial fluid movement was augmented by imposition of osmotic gradients.
Resumo:
Recent measurements of sedimentation equilibrium and sedimentation velocity have shown that the bacterial cell division protein FtsZ self-associates to form indefinitely long rod-like linear aggregates in the presence of GDP and Mg2+. In the present study, the newly developed technique of non-ideal tracer sedimentation equilibrium was used to measure the effect of high concentrations—up to 150 g/liter—of each of two inert “crowder” proteins, cyanmethemoglobin or BSA, on the thermodynamic activity and state of association of dilute FtsZ under conditions inhibiting (−Mg2+) and promoting (+Mg2+) FtsZ self-association. Analysis of equilibrium gradients of both FtsZ and crowder proteins indicates that, under the conditions of the present experiment, FtsZ interacts with each of the two crowder proteins essentially entirely via steric repulsion, which may be accounted for quantitatively by a simple model in which hemoglobin, albumin, and monomeric FtsZ are modeled as effective spherical hard particles, and each oligomeric species of FtsZ is modeled as an effective hard spherocylinder. The functional dependence of the sedimentation of FtsZ on the concentrations of FtsZ and either crowder indicates that, in the presence of high concentrations of crowder, both the weight-average degree of FtsZ self-association and the range of FtsZ oligomer sizes present in significant abundance are increased substantially.
Resumo:
Studies of retrograde amnesia are reviewed. First, the issues of temporal gradients of retrograde amnesia are discussed. Second, the question of the anatomical substrates of this syndrome are considered. Finally, some evidence for fractionation of different classes of memoranda within the retrograde time period are presented.
Resumo:
The active migration of neurons from their sites of origin to their final destinations requires the unidirectional translocation of the nuclei and somatic cytoplasm within the growing leading processes. To explore the cellular machinery underlying this translocation, we determined the polarity of microtubules situated within the leading and trailing processes of migrating cerebellar granule cells in situ. Our analysis reveals that the newly assembled positive ends of the microtubules in the leading process uniformly face the growing tip, while their disintegrating negative ends face the nucleus. In the trailing process, by contrast, microtubule arrays are of mixed polarity. We suggest that the dynamics of slow polymerization in combination with fast disintegration of oriented microtubules create "push" and "pull" forces that contribute to the piston-like saltatory displacement of the nucleus and cytoplasm within the membrane cylinder of the leading process of the migrating neuron.
Resumo:
Cardiac muscle contraction is triggered by a small and brief Ca2+ entry across the t-tubular membranes, which is believed to be locally amplified by release of Ca2+ from the adjacent junctional sarcoplasmic reticulum (SR). As Ca2+ diffusion is thought to be markedly attenuated in cells, it has been predicted that significant intrasarcomeric [Ca2+] gradients should exist during activation. To directly test for this, we measured [Ca2+] distribution in single cardiac myocytes using fluorescent [Ca2+] indicators and high speed, three-dimensional digital imaging microscopy and image deconvolution techniques. Steep cytosolic [Ca2+] gradients from the t-tubule region to the center of the sarcomere developed during the first 15 ms of systole. The steepness of these [Ca2+] gradients varied with treatments that altered Ca2+ release from internal stores. Electron probe microanalysis revealed a loss of Ca2+ from the junctional SR and an accumulation, principally in the A-band during activation. We propose that the prolonged existence of [Ca2+] gradients within the sarcomere reflects the relatively long period of Ca2+ release from the SR, the localization of Ca2+ binding sites and Ca2+ sinks remote from sites of release, and diffusion limitations within the sarcomere. The large [Ca2+] transient near the t-tubular/ junctional SR membranes is postulated to explain numerous features of excitation-contraction coupling in cardiac muscle.
Resumo:
Very-long-baseline radio interferometry (VLBI) imaging surveys have been undertaken since the late 1970s. The sample sizes were initially limited to a few tens of objects but the snapshot technique has now allowed samples containing almost 200 sources to be studied. The overwhelming majority of powerful compact sources are asymmetric corejects of one form or another, most of which exhibit apparent superluminal motion. However 5-10% of powerful flat-spectrum sources are 100-parsec (pc)-scale compact symmetric objects; these appear to form a continuum with the 1-kpc-scale double-lobed compact steep-spectrum sources, which make up 15-20% of lower frequency samples. It is likely that these sub-galactic-size symmetric sources are the precursors to the large-scale classical double sources. There is a surprising peak around 90 degrees in the histogram of misalignments between the dominant source axes on parsec and kiloparsec scales; this seems to be associated with sources exhibiting a high degree of relativistic beaming. VLBI snapshot surveys have great cosmological potential via measurements of both proper motion and angular size vs. redshift as well as searches for gravitational "millilensing."