953 resultados para Simulated annealing algorithms
Resumo:
La temperatura es una preocupación que juega un papel protagonista en el diseño de circuitos integrados modernos. El importante aumento de las densidades de potencia que conllevan las últimas generaciones tecnológicas ha producido la aparición de gradientes térmicos y puntos calientes durante el funcionamiento normal de los chips. La temperatura tiene un impacto negativo en varios parámetros del circuito integrado como el retardo de las puertas, los gastos de disipación de calor, la fiabilidad, el consumo de energía, etc. Con el fin de luchar contra estos efectos nocivos, la técnicas de gestión dinámica de la temperatura (DTM) adaptan el comportamiento del chip en función en la información que proporciona un sistema de monitorización que mide en tiempo de ejecución la información térmica de la superficie del dado. El campo de la monitorización de la temperatura en el chip ha llamado la atención de la comunidad científica en los últimos años y es el objeto de estudio de esta tesis. Esta tesis aborda la temática de control de la temperatura en el chip desde diferentes perspectivas y niveles, ofreciendo soluciones a algunos de los temas más importantes. Los niveles físico y circuital se cubren con el diseño y la caracterización de dos nuevos sensores de temperatura especialmente diseñados para los propósitos de las técnicas DTM. El primer sensor está basado en un mecanismo que obtiene un pulso de anchura variable dependiente de la relación de las corrientes de fuga con la temperatura. De manera resumida, se carga un nodo del circuito y posteriormente se deja flotando de tal manera que se descarga a través de las corrientes de fugas de un transistor; el tiempo de descarga del nodo es la anchura del pulso. Dado que la anchura del pulso muestra una dependencia exponencial con la temperatura, la conversión a una palabra digital se realiza por medio de un contador logarítmico que realiza tanto la conversión tiempo a digital como la linealización de la salida. La estructura resultante de esta combinación de elementos se implementa en una tecnología de 0,35 _m. El sensor ocupa un área muy reducida, 10.250 nm2, y consume muy poca energía, 1.05-65.5nW a 5 muestras/s, estas cifras superaron todos los trabajos previos en el momento en que se publicó por primera vez y en el momento de la publicación de esta tesis, superan a todas las implementaciones anteriores fabricadas en el mismo nodo tecnológico. En cuanto a la precisión, el sensor ofrece una buena linealidad, incluso sin calibrar; se obtiene un error 3_ de 1,97oC, adecuado para tratar con las aplicaciones de DTM. Como se ha explicado, el sensor es completamente compatible con los procesos de fabricación CMOS, este hecho, junto con sus valores reducidos de área y consumo, lo hacen especialmente adecuado para la integración en un sistema de monitorización de DTM con un conjunto de monitores empotrados distribuidos a través del chip. Las crecientes incertidumbres de proceso asociadas a los últimos nodos tecnológicos comprometen las características de linealidad de nuestra primera propuesta de sensor. Con el objetivo de superar estos problemas, proponemos una nueva técnica para obtener la temperatura. La nueva técnica también está basada en las dependencias térmicas de las corrientes de fuga que se utilizan para descargar un nodo flotante. La novedad es que ahora la medida viene dada por el cociente de dos medidas diferentes, en una de las cuales se altera una característica del transistor de descarga |la tensión de puerta. Este cociente resulta ser muy robusto frente a variaciones de proceso y, además, la linealidad obtenida cumple ampliamente los requisitos impuestos por las políticas DTM |error 3_ de 1,17oC considerando variaciones del proceso y calibrando en dos puntos. La implementación de la parte sensora de esta nueva técnica implica varias consideraciones de diseño, tales como la generación de una referencia de tensión independiente de variaciones de proceso, que se analizan en profundidad en la tesis. Para la conversión tiempo-a-digital, se emplea la misma estructura de digitalización que en el primer sensor. Para la implementación física de la parte de digitalización, se ha construido una biblioteca de células estándar completamente nueva orientada a la reducción de área y consumo. El sensor resultante de la unión de todos los bloques se caracteriza por una energía por muestra ultra baja (48-640 pJ) y un área diminuta de 0,0016 mm2, esta cifra mejora todos los trabajos previos. Para probar esta afirmación, se realiza una comparación exhaustiva con más de 40 propuestas de sensores en la literatura científica. Subiendo el nivel de abstracción al sistema, la tercera contribución se centra en el modelado de un sistema de monitorización que consiste de un conjunto de sensores distribuidos por la superficie del chip. Todos los trabajos anteriores de la literatura tienen como objetivo maximizar la precisión del sistema con el mínimo número de monitores. Como novedad, en nuestra propuesta se introducen nuevos parámetros de calidad aparte del número de sensores, también se considera el consumo de energía, la frecuencia de muestreo, los costes de interconexión y la posibilidad de elegir diferentes tipos de monitores. El modelo se introduce en un algoritmo de recocido simulado que recibe la información térmica de un sistema, sus propiedades físicas, limitaciones de área, potencia e interconexión y una colección de tipos de monitor; el algoritmo proporciona el tipo seleccionado de monitor, el número de monitores, su posición y la velocidad de muestreo _optima. Para probar la validez del algoritmo, se presentan varios casos de estudio para el procesador Alpha 21364 considerando distintas restricciones. En comparación con otros trabajos previos en la literatura, el modelo que aquí se presenta es el más completo. Finalmente, la última contribución se dirige al nivel de red, partiendo de un conjunto de monitores de temperatura de posiciones conocidas, nos concentramos en resolver el problema de la conexión de los sensores de una forma eficiente en área y consumo. Nuestra primera propuesta en este campo es la introducción de un nuevo nivel en la jerarquía de interconexión, el nivel de trillado (o threshing en inglés), entre los monitores y los buses tradicionales de periféricos. En este nuevo nivel se aplica selectividad de datos para reducir la cantidad de información que se envía al controlador central. La idea detrás de este nuevo nivel es que en este tipo de redes la mayoría de los datos es inútil, porque desde el punto de vista del controlador sólo una pequeña cantidad de datos |normalmente sólo los valores extremos| es de interés. Para cubrir el nuevo nivel, proponemos una red de monitorización mono-conexión que se basa en un esquema de señalización en el dominio de tiempo. Este esquema reduce significativamente tanto la actividad de conmutación sobre la conexión como el consumo de energía de la red. Otra ventaja de este esquema es que los datos de los monitores llegan directamente ordenados al controlador. Si este tipo de señalización se aplica a sensores que realizan conversión tiempo-a-digital, se puede obtener compartición de recursos de digitalización tanto en tiempo como en espacio, lo que supone un importante ahorro de área y consumo. Finalmente, se presentan dos prototipos de sistemas de monitorización completos que de manera significativa superan la características de trabajos anteriores en términos de área y, especialmente, consumo de energía. Abstract Temperature is a first class design concern in modern integrated circuits. The important increase in power densities associated to recent technology evolutions has lead to the apparition of thermal gradients and hot spots during run time operation. Temperature impacts several circuit parameters such as speed, cooling budgets, reliability, power consumption, etc. In order to fight against these negative effects, dynamic thermal management (DTM) techniques adapt the behavior of the chip relying on the information of a monitoring system that provides run-time thermal information of the die surface. The field of on-chip temperature monitoring has drawn the attention of the scientific community in the recent years and is the object of study of this thesis. This thesis approaches the matter of on-chip temperature monitoring from different perspectives and levels, providing solutions to some of the most important issues. The physical and circuital levels are covered with the design and characterization of two novel temperature sensors specially tailored for DTM purposes. The first sensor is based upon a mechanism that obtains a pulse with a varying width based on the variations of the leakage currents on the temperature. In a nutshell, a circuit node is charged and subsequently left floating so that it discharges away through the subthreshold currents of a transistor; the time the node takes to discharge is the width of the pulse. Since the width of the pulse displays an exponential dependence on the temperature, the conversion into a digital word is realized by means of a logarithmic counter that performs both the timeto- digital conversion and the linearization of the output. The structure resulting from this combination of elements is implemented in a 0.35_m technology and is characterized by very reduced area, 10250 nm2, and power consumption, 1.05-65.5 nW at 5 samples/s, these figures outperformed all previous works by the time it was first published and still, by the time of the publication of this thesis, they outnumber all previous implementations in the same technology node. Concerning the accuracy, the sensor exhibits good linearity, even without calibration it displays a 3_ error of 1.97oC, appropriate to deal with DTM applications. As explained, the sensor is completely compatible with standard CMOS processes, this fact, along with its tiny area and power overhead, makes it specially suitable for the integration in a DTM monitoring system with a collection of on-chip monitors distributed across the chip. The exacerbated process fluctuations carried along with recent technology nodes jeop-ardize the linearity characteristics of the first sensor. In order to overcome these problems, a new temperature inferring technique is proposed. In this case, we also rely on the thermal dependencies of leakage currents that are used to discharge a floating node, but now, the result comes from the ratio of two different measures, in one of which we alter a characteristic of the discharging transistor |the gate voltage. This ratio proves to be very robust against process variations and displays a more than suficient linearity on the temperature |1.17oC 3_ error considering process variations and performing two-point calibration. The implementation of the sensing part based on this new technique implies several issues, such as the generation of process variations independent voltage reference, that are analyzed in depth in the thesis. In order to perform the time-to-digital conversion, we employ the same digitization structure the former sensor used. A completely new standard cell library targeting low area and power overhead is built from scratch to implement the digitization part. Putting all the pieces together, we achieve a complete sensor system that is characterized by ultra low energy per conversion of 48-640pJ and area of 0.0016mm2, this figure outperforms all previous works. To prove this statement, we perform a thorough comparison with over 40 works from the scientific literature. Moving up to the system level, the third contribution is centered on the modeling of a monitoring system consisting of set of thermal sensors distributed across the chip. All previous works from the literature target maximizing the accuracy of the system with the minimum number of monitors. In contrast, we introduce new metrics of quality apart form just the number of sensors; we consider the power consumption, the sampling frequency, the possibility to consider different types of monitors and the interconnection costs. The model is introduced in a simulated annealing algorithm that receives the thermal information of a system, its physical properties, area, power and interconnection constraints and a collection of monitor types; the algorithm yields the selected type of monitor, the number of monitors, their position and the optimum sampling rate. We test the algorithm with the Alpha 21364 processor under several constraint configurations to prove its validity. When compared to other previous works in the literature, the modeling presented here is the most complete. Finally, the last contribution targets the networking level, given an allocated set of temperature monitors, we focused on solving the problem of connecting them in an efficient way from the area and power perspectives. Our first proposal in this area is the introduction of a new interconnection hierarchy level, the threshing level, in between the monitors and the traditional peripheral buses that applies data selectivity to reduce the amount of information that is sent to the central controller. The idea behind this new level is that in this kind of networks most data are useless because from the controller viewpoint just a small amount of data |normally extreme values| is of interest. To cover the new interconnection level, we propose a single-wire monitoring network based on a time-domain signaling scheme that significantly reduces both the switching activity over the wire and the power consumption of the network. This scheme codes the information in the time domain and allows a straightforward obtention of an ordered list of values from the maximum to the minimum. If the scheme is applied to monitors that employ TDC, digitization resource sharing is achieved, producing an important saving in area and power consumption. Two prototypes of complete monitoring systems are presented, they significantly overcome previous works in terms of area and, specially, power consumption.
Resumo:
Dynamic thermal management techniques require a collection of on-chip thermal sensors that imply a significant area and power overhead. Finding the optimum number of temperature monitors and their location on the chip surface to optimize accuracy is an NP-hard problem. In this work we improve the modeling of the problem by including area, power and networking constraints along with the consideration of three inaccuracy terms: spatial errors, sampling rate errors and monitor-inherent errors. The problem is solved by the simulated annealing algorithm. We apply the algorithm to a test case employing three different types of monitors to highlight the importance of the different metrics. Finally we present a case study of the Alpha 21364 processor under two different constraint scenarios.
Resumo:
Los modelos de simulación de cultivos permiten analizar varias combinaciones de laboreo-rotación y explorar escenarios de manejo. El modelo DSSAT fue evaluado bajo condiciones de secano en un experimento de campo de 16 años en la semiárida España central. Se evaluó el efecto del sistema de laboreo y las rotaciones basadas en cereales de invierno, en el rendimiento del cultivo y la calidad del suelo. Los modelos CERES y CROPGRO se utilizaron para simular el crecimiento y rendimiento del cultivo, mientras que el modelo DSSAT CENTURY se utilizó en las simulaciones de SOC y SN. Tanto las observaciones de campo como las simulaciones con CERES-Barley, mostraron que el rendimiento en grano de la cebada era mas bajo para el cereal continuo (BB) que para las rotaciones de veza (VB) y barbecho (FB) en ambos sistemas de laboreo. El modelo predijo más nitrógeno disponible en el laboreo convencional (CT) que en el no laboreo (NT) conduciendo a un mayor rendimiento en el CT. El SOC y el SN en la capa superficial del suelo, fueron mayores en NT que en CT, y disminuyeron con la profundidad en los valores tanto observados como simulados. Las mejores combinaciones para las condiciones de secano estudiadas fueron CT-VB y CT-FB, pero CT presentó menor contenido en SN y SOC que NT. El efecto beneficioso del NT en SOC y SN bajo condiciones Mediterráneas semiáridas puede ser identificado por observaciones de campo y por simulaciones de modelos de cultivos. La simulación del balance de agua en sistemas de cultivo es una herramienta útil para estudiar como el agua puede ser utilizado eficientemente. La comparación del balance de agua de DSSAT , con una simple aproximación “tipping bucket”, con el modelo WAVE más mecanicista, el cual integra la ecuación de Richard , es un potente método para valorar el funcionamiento del modelo. Los parámetros de suelo fueron calibrados usando el método de optimización global Simulated Annealing (SA). Un lisímetro continuo de pesada en suelo desnudo suministró los valores observados de drenaje y evapotranspiración (ET) mientras que el contenido de agua en el suelo (SW) fue suministrado por sensores de capacitancia. Ambos modelos funcionaron bien después de la optimización de los parámetros de suelo con SA, simulando el balance de agua en el suelo para el período de calibración. Para el período de validación, los modelos optimizados predijeron bien el contenido de agua en el suelo y la evaporación del suelo a lo largo del tiempo. Sin embargo, el drenaje fue predicho mejor con WAVE que con DSSAT, el cual presentó mayores errores en los valores acumulados. Esto podría ser debido a la naturaleza mecanicista de WAVE frente a la naturaleza más funcional de DSSAT. Los buenos resultados de WAVE indican que, después de la calibración, este puede ser utilizado como "benchmark" para otros modelos para periodos en los que no haya medidas de campo del drenaje. El funcionamiento de DSSAT-CENTURY en la simulación de SOC y N depende fuertemente del proceso de inicialización. Se propuso como método alternativo (Met.2) la inicialización de las fracciones de SOC a partir de medidas de mineralización aparente del suelo (Napmin). El Met.2 se comparó con el método de inicialización de Basso et al. (2011) (Met.1), aplicando ambos métodos a un experimento de campo de 4 años en un área en regadío de España central. Nmin y Napmin fueron sobreestimados con el Met.1, ya que la fracción estable obtenida (SOC3) en las capas superficiales del suelo fue más baja que con Met.2. El N lixiviado simulado fue similar en los dos métodos, con buenos resultados en los tratamientos de barbecho y cebada. El Met.1 subestimó el SOC en la capa superficial del suelo cuando se comparó con una serie observada de 12 años. El crecimiento y rendimiento del cultivo fueron adecuadamente simulados con ambos métodos, pero el N en la parte aérea de la planta y en el grano fueron sobreestimados con el Met.1. Los resultados variaron significativamente con las fracciones iniciales de SOC, resaltando la importancia del método de inicialización. El Met.2 ofrece una alternativa para la inicialización del modelo CENTURY, mejorando la simulación de procesos de N en el suelo. La continua emergencia de nuevas variedades de híbridos modernos de maíz limita la aplicación de modelos de simulación de cultivos, ya que estos nuevos híbridos necesitan ser calibrados en el campo para ser adecuados para su uso en los modelos. El desarrollo de relaciones basadas en la duración del ciclo, simplificaría los requerimientos de calibración facilitando la rápida incorporación de nuevos cultivares en DSSAT. Seis híbridos de maiz (FAO 300 hasta FAO 700) fueron cultivados en un experimento de campo de dos años en un área semiárida de regadío en España central. Los coeficientes genéticos fueron obtenidos secuencialmente, comenzando con los parámetros de desarrollo fenológico (P1, P2, P5 and PHINT), seguido de los parámetros de crecimiento del cultivo (G2 and G3). Se continuó el procedimiento hasta que la salida de las simulaciones estuvo en concordancia con las observaciones fenológicas de campo. Después de la calibración, los parámetros simulados se ajustaron bien a los parámetros observados, con bajos RMSE en todos los casos. Los P1 y P5 calibrados, incrementaron con la duración del ciclo. P1 fue una función lineal del tiempo térmico (TT) desde emergencia hasta floración y P5 estuvo linealmente relacionada con el TT desde floración a madurez. No hubo diferencias significativas en PHINT entre híbridos de FAO-500 a 700 , ya que tuvieron un número de hojas similar. Como los coeficientes fenológicos estuvieron directamente relacionados con la duración del ciclo, sería posible desarrollar rangos y correlaciones que permitan estimar dichos coeficientes a partir de la clasificación del ciclo. ABSTRACT Crop simulation models allow analyzing various tillage-rotation combinations and exploring management scenarios. DSSAT model was tested under rainfed conditions in a 16-year field experiment in semiarid central Spain. The effect of tillage system and winter cereal-based rotations on the crop yield and soil quality was evaluated. The CERES and CROPGRO models were used to simulate crop growth and yield, while the DSSAT CENTURY was used in the SOC and SN simulations. Both field observations and CERES-Barley simulations, showed that barley grain yield was lower for continuous cereal (BB) than for vetch (VB) and fallow (FB) rotations for both tillage systems. The model predicted higher nitrogen availability in the conventional tillage (CT) than in the no tillage (NT) leading to a higher yield in the CT. The SOC and SN in the top layer, were higher in NT than in CT, and decreased with depth in both simulated and observed values. The best combinations for the dry land conditions studied were CT-VB and CT-FB, but CT presented lower SN and SOC content than NT. The beneficial effect of NT on SOC and SN under semiarid Mediterranean conditions can be identified by field observations and by crop model simulations. The simulation of the water balance in cropping systems is a useful tool to study how water can be used efficiently. The comparison of DSSAT soil water balance, with a simpler “tipping bucket” approach, with the more mechanistic WAVE model, which integrates Richard’s equation, is a powerful method to assess model performance. The soil parameters were calibrated by using the Simulated Annealing (SA) global optimizing method. A continuous weighing lysimeter in a bare fallow provided the observed values of drainage and evapotranspiration (ET) while soil water content (SW) was supplied by capacitance sensors. Both models performed well after optimizing soil parameters with SA, simulating the soil water balance components for the calibrated period. For the validation period, the optimized models predicted well soil water content and soil evaporation over time. However, drainage was predicted better by WAVE than by DSSAT, which presented larger errors in the cumulative values. That could be due to the mechanistic nature of WAVE against the more functional nature of DSSAT. The good results from WAVE indicate that, after calibration, it could be used as benchmark for other models for periods when no drainage field measurements are available. The performance of DSSAT-CENTURY when simulating SOC and N strongly depends on the initialization process. Initialization of the SOC pools from apparent soil N mineralization (Napmin) measurements was proposed as alternative method (Met.2). Method 2 was compared to the Basso et al. (2011) initialization method (Met.1), by applying both methods to a 4-year field experiment in a irrigated area of central Spain. Nmin and Napmin were overestimated by Met.1, since the obtained stable pool (SOC3) in the upper layers was lower than from Met.2. Simulated N leaching was similar for both methods, with good results in fallow and barley treatments. Method 1 underestimated topsoil SOC when compared with a 12-year observed serial. Crop growth and yield were properly simulated by both methods, but N in shoots and grain were overestimated by Met.1. Results varied significantly with the initial SOC pools, highlighting the importance of the initialization procedure. Method 2 offers an alternative to initialize the CENTURY model, enhancing the simulation of soil N processes. The continuous emergence of new varieties of modern maize hybrids limits the application of crop simulation models, since these new hybrids should be calibrated in the field to be suitable for model use. The development of relationships based on the cycle duration, would simplify the calibration requirements facilitating the rapid incorporation of new cultivars into DSSAT. Six maize hybrids (FAO 300 through FAO 700) were grown in a 2-year field experiment in a semiarid irrigated area of central Spain. Genetic coefficients were obtained sequentially, starting with the phenological development parameters (P1, P2, P5 and PHINT), followed by the crop growth parameters (G2 and G3). The procedure was continued until the simulated outputs were in good agreement with the field phenological observations. After calibration, simulated parameters matched observed parameters well, with low RMSE in most cases. The calibrated P1 and P5 increased with the duration of the cycle. P1 was a linear function of the thermal time (TT) from emergence to silking and P5 was linearly related with the TT from silking to maturity . There were no significant differences in PHINT between hybrids from FAO-500 to 700 , as they had similar leaf number. Since phenological coefficients were directly related with the cycle duration, it would be possible to develop ranges and correlations which allow to estimate such coefficients from the cycle classification.
Resumo:
The CENTURY soil organic matter model was adapted for the DSSAT (Decision Support System for Agrotechnology Transfer), modular format in order to better simulate the dynamics of soil organic nutrient processes (Gijsman et al., 2002). The CENTURY model divides the soil organic carbon (SOC) into three hypothetical pools: microbial or active material (SOC1), intermediate (SOC2) and the largely inert and stable material (SOC3) (Jones et al., 2003). At the beginning of the simulation, CENTURY model needs a value of SOC3 per soil layer which can be estimated by the model (based on soil texture and management history) or given as an input. Then, the model assigns about 5% and 95% of the remaining SOC to SOC1 and SOC2, respectively. The model performance when simulating SOC and nitrogen (N) dynamics strongly depends on the initialization process. The common methods (e.g. Basso et al., 2011) to initialize SOC pools deal mostly with carbon (C) mineralization processes and less with N. Dynamics of SOM, SOC, and soil organic N are linked in the CENTURY-DSSAT model through the C/N ratio of decomposing material that determines either mineralization or immobilization of N (Gijsman et al., 2002). The aim of this study was to evaluate an alternative method to initialize the SOC pools in the DSSAT-CENTURY model from apparent soil N mineralization (Napmin) field measurements by using automatic inverse calibration (simulated annealing). The results were compared with the ones obtained by the iterative initialization procedure developed by Basso et al., 2011.
Resumo:
In this paper we focus on the selection of safeguards in a fuzzy risk analysis and management methodology for information systems (IS). Assets are connected by dependency relationships, and a failure of one asset may affect other assets. After computing impact and risk indicators associated with previously identified threats, we identify and apply safeguards to reduce risks in the IS by minimizing the transmission probabilities of failures throughout the asset network. However, as safeguards have associated costs, the aim is to select the safeguards that minimize costs while keeping the risk within acceptable levels. To do this, we propose a dynamic programming-based method that incorporates simulated annealing to tackle optimizations problems.
Resumo:
A 12 bp long GCN4-binding, self-complementary duplex DNA d(CATGACGTCATG)2 has been investigated by NMR spectroscopy to study the structure and dynamics of the molecule in aqueous solution. The NMR structure of the DNA obtained using simulated annealing and iterative relaxation matrix calculations compares quite closely with the X-ray structure of ATF/CREB DNA in complex with GCN4 protein (DNA-binding domain). The DNA is also seen to be curved in the free state and this has a significant bearing on recognition by the protein. The dynamic characteristics of the molecule have been studied by 13C relaxation measurements at natural abundance. A correlation has been observed between sequence-dependent dynamics and recognition by GCN4 protein.
Resumo:
A dumbbell double-stranded DNA decamer tethered with a hexaethylene glycol linker moiety (DDSDPEG), with a nick in the centre of one strand, has been synthesised. The standard NMR methods, E.COSY, TOCSY, NOESY and HMQC, were used to measure 1H, 31P and T1 spectral parameters. Molecular modelling using rMD-simulated annealing was used to compute the structure. Scalar couplings and dipolar contacts show that the molecule adopts a right-handed B-DNA helix in 38 mM phosphate buffer at pH 7. Its high melting temperature confirms the good base stacking and stability of the duplex. This is partly attributed to the presence of the PEG6 linker at both ends of the duplex that restricts the dynamics of the stem pentamers and thus stabilises the oligonucleotide. The inspection of the global parameters shows that the linker does not distort the B-DNA geometry. The computed structure suggests that the presence of the nick is not disturbing the overall tertiary structure, base pair geometry or duplex base pairing to a substantial extent. The nick has, however, a noticeable impact on the local geometry at the nick site, indicated clearly by NMR analysis and reflected in the conformational parameters of the computed structure. The 1H spectra also show much sharper resonances in the presence of K+ indicating that conformational heterogeneity of DDSDPEG is reduced in the presence of potassium as compared to sodium or caesium ions. At the same time the 1H resonances have longer T1 times. This parameter is suggested as a sensitive gauge of stabilisation.
Resumo:
Esta tese apresenta uma abordagem para a criação rápida de modelos em diferentes geometrias (complexas ou de alta simetria) com objetivo de calcular a correspondente intensidade espalhada, podendo esta ser utilizada na descrição de experimentos de es- palhamento à baixos ângulos. A modelagem pode ser realizada com mais de 100 geome- trias catalogadas em um Banco de Dados, além da possibilidade de construir estruturas a partir de posições aleatórias distribuídas na superfície de uma esfera. Em todos os casos os modelos são gerados por meio do método de elementos finitos compondo uma única geometria, ou ainda, compondo diferentes geometrias, combinadas entre si a partir de um número baixo de parâmetros. Para realizar essa tarefa foi desenvolvido um programa em Fortran, chamado de Polygen, que permite modelar geometrias convexas em diferentes formas, como sólidos, cascas, ou ainda com esferas ou estruturas do tipo DNA nas arestas, além de usar esses modelos para simular a curva de intensidade espalhada para sistemas orientados e aleatoriamente orientados. A curva de intensidade de espalhamento é calculada por meio da equação de Debye e os parâmetros que compõe cada um dos modelos, podem ser otimizados pelo ajuste contra dados experimentais, por meio de métodos de minimização baseados em simulated annealing, Levenberg-Marquardt e algorítmicos genéticos. A minimização permite ajustar os parâmetros do modelo (ou composição de modelos) como tamanho, densidade eletrônica, raio das subunidades, entre outros, contribuindo para fornecer uma nova ferramenta para modelagem e análise de dados de espalhamento. Em outra etapa desta tese, é apresentado o design de modelos atomísticos e a sua respectiva simulação por Dinâmica Molecular. A geometria de dois sistemas auto-organizado de DNA na forma de octaedro truncado, um com linkers de 7 Adeninas e outro com linkers de ATATATA, foram escolhidas para realizar a modelagem atomística e a simulação por Dinâmica Molecular. Para este sistema são apresentados os resultados de Root Mean Square Deviations (RMSD), Root Mean Square Fluctuations (RMSF), raio de giro, torção das hélices duplas de DNA além da avaliação das ligações de Hidrogênio, todos obtidos por meio da análise de uma trajetória de 50 ns.
Resumo:
A method is proposed for determining the optimal placement and controller design for multiple distributed actuators to reduce the vibrations of flexible structures. In particular, application of piezoceramic patches to a horizontally-slewing single-link flexible manipulator modeled using the assumed modes method is investigated. The optimization method uses simulated annealing and allows placement of any number of distributed actuators of unequal length, although piezoceramics of fixed equal lengths are used in the example. It also designs an linear-quadratic-regulator controller as part of the optimization procedure. The measures of performance used in the investigation to determine optimality are the total mass of the system and the time integral of the absolute value of the hub and tip position error. This study also varies the relative weightings for each of these performance measures to observe the effects on the controller designs and piezoceramic patch positions in the optimized solutions.
Resumo:
Alpha helices are key structural components of proteins and important recognition motifs in biology. New techniques for stabilizing short peptide helices could be valuable for studying protein folding, modeling proteins, creating artificial proteins, and may aid the design of inhibitors or mimics of protein function. We previously reported* that 5-15 residue peptides, corresponding to the Zn-binding domain of thermolysin, react with [Pd(en)(ONO,),]in DMF-d’ and 90% H,O 10% DzO to form a 22-membered [Pd(en)(H*ELTH*)]2+ macrocycle that is helical in solution and acts as a template in nucleating helicity in both Cand N- terminal directions within the longer sequences in DMF. ~f~~&g7$$& d&qx~m ~. y AC&q& In water, however, there was less a-helicity observed, testifying to #..q,& &$--Lb &l-- &.$;,J~p?:~~q&~+~~ ’ w w the difficulty of fixing intramolecular amide NH...OC H-bonds in 6,“;;” ( k.$ U”C.a , p d$. competition with the H-bond donor solvent water. To expand the utility of [Pd(en)(H*XXXH*)]*+ as a helix- @r4”8 & oJ#:& &G& @-qd ,‘d@-gyp promoting module in solution, we now report the result that Ac- ‘$4: %$yyy + H*ELTH*H*VTDH*-NH,(l), AC-H*ELTH*AVTDYH*ELTH*- NH, (2) and AC-H*AAAH*H*ELTH*H*VTDH*-NH* (3) react with multiple equivalents of [Pd(en)(ONO,),] to produce exclusively 4-6 respectively in both DMF-d7 and water (90% Hz0 10% D,O). Mass spectrometry, 15N- and 2D ‘H- NMR spectroscopy, and CD spectra were used to characterise the structures 4-6, and their three dimensional structures were calculated from NOE restraints using simulated annealing protocols. Results demonstrate (a) selective coordination of metal ions at (i, i+4) histidine positions in water and DMF, (b) incorporation of 2 and 3 a turn-mimicking modules [Pd(en)(HELTH)]2+ in lo-15 residue peptides, and (c) facile conversion of unstructured peptides into 3- and 4- turn helices of macrocycles, with well defined a-helicity throughout and more structure in DMF than in water.
Resumo:
A recent development of the Markov chain Monte Carlo (MCMC) technique is the emergence of MCMC samplers that allow transitions between different models. Such samplers make possible a range of computational tasks involving models, including model selection, model evaluation, model averaging and hypothesis testing. An example of this type of sampler is the reversible jump MCMC sampler, which is a generalization of the Metropolis-Hastings algorithm. Here, we present a new MCMC sampler of this type. The new sampler is a generalization of the Gibbs sampler, but somewhat surprisingly, it also turns out to encompass as particular cases all of the well-known MCMC samplers, including those of Metropolis, Barker, and Hastings. Moreover, the new sampler generalizes the reversible jump MCMC. It therefore appears to be a very general framework for MCMC sampling. This paper describes the new sampler and illustrates its use in three applications in Computational Biology, specifically determination of consensus sequences, phylogenetic inference and delineation of isochores via multiple change-point analysis.
Resumo:
sThe structure of a two-chain peptide formed by the treatment of the potent antimicrobial peptide microcin J25 (MccJ25) with thermolysin has been characterized by NMR spectroscopy and mass spectrometry. The native peptide is 21 amino acids in size and has the remarkable structural feature of a ring formed by linkage of the side chain of Glu8 to the N-terminus that is threaded by the C-terminal tail of the peptide. Thermolysin cleaves the peptide at the Phe10-Val11 amide bond, but the threading of the C-terminus through the N-terminal ring is so tight that the resultant two chains remain associated both in the solution and in the gas phases. The three-dimensional structure of the thermolysin-cleaved peptide derived using NMR spectroscopy and simulated annealing calculations has a well-defined core that comprises the N-terminal ring and the threading C-terminal tail. In contrast to the well-defined core, the newly formed termini at residues Phe10 and Val11 are disordered in solution. The C-terminal tail is associated to the ring both by hydrogen bonds stabilizing a short beta-sheet and by hydrophobic interactions. Moreover, unthreading of the tail through the ring is prevented by the bulky side chains of Phe19 and Tyr20, which flank the octapeptide ring. This noncovalent two-peptide complex that has a remarkable stability in solution and in highly denaturing conditions and that survives in the gas phase is the first example of such a two-chain peptide lacking disulfide or interchain covalent bonds.
Resumo:
A novel member of the human relaxin subclass of the insulin superfamily was recently discovered during a genomics database search and named relaxin-3. Like human relaxin-1 and relaxin-2, relaxin-3 is predicted to consist of a two-chain structure and three disulfide bonds in a disposition identical to that of insulin. To undertake detailed biophysical and biological characterization of the peptide, its chemical synthesis was undertaken. In contrast to human relaxin-1 and relaxin-2, however, relaxin-3 could not be successfully prepared by simple combination of the individual chains, thus necessitating recourse to the use of a regioselective disulfide bond formation strategy. Solid phase synthesis of the separate, selectively S-protected A and B chains followed by their purification and the subsequent stepwise formation of each of the three disulfides led to the successful acquisition of human relaxin-3. Comprehensive chemical characterization confirmed both the correct chain orientation and the integrity of the synthetic product. Relaxin-3 was found to bind to and activate native relaxin receptors in vitro and stimulate water drinking through central relaxin receptors in vivo. Recent studies have demonstrated that relaxin-3 will bind to and activate human LGR7, but not LGR8, in vitro. Secondary structural analysis showed it to adopt a less ordered confirmation than either relaxin-1 or relaxin-2, reflecting the presence in the former of a greater percentage of nonhelical forming amino acids. NMR spectroscopy and simulated annealing calculations were used to determine the three-dimensional structure of relaxin-3 and to identify key structural differences between the human relaxins.
Resumo:
In this work a superposition technique for designing gradient coils for the purpose of magnetic resonance imaging is outlined, which uses an optimized weight function superimposed upon an initial winding similar to that obtained from the target field method to generate the final wire winding. This work builds on the preliminary work performed in Part I on designing planar insertable gradient coils for high resolution imaging. The proposed superposition method for designing gradient coils results in coil patterns with relatively low inductances and the gradient coils can be used as inserts into existing magnetic resonance imaging hardware. The new scheme has the capacity to obtain images faster with more detail due to the deliver of greater magnetic held gradients. The proposed method for designing gradient coils is compared with a variant of the state-of-the-art target field method for planar gradient coils designs, and it is shown that the weighted superposition approach outperforms the well-known the classical method.
Resumo:
A primary purpose of this research is to design a gradient coil that is planar in construction and can be inserted within existing infrastructure. The proposed wave equation method for the design of gradient coils is novel within the field. it is comprehensively shown how this method can be used to design the planar x-, y-, and z-gradient wire windings to produce the required magnetic fields within a certain domain. The solution for the cylindrical gradient coil set is also elucidated. The wave equation technique is compared with the well-known target held method to gauge the quality of resultant design. In the case of the planar gradient coil design, it is shown that using the new method, a set of compact gradient coils with large field of view can be produced. The final design is considerably smaller in dimension when compared with the design obtained using the target field method, and therefore the manufacturing costs and materials required are somewhat reduced.