989 resultados para Calculation-based
Resumo:
It is still an open question how equilibrium warming in response to increasing radiative forcing - the specific equilibrium climate sensitivity S - depends on background climate. We here present palaeodata-based evidence on the state dependency of S, by using CO2 proxy data together with a 3-D ice-sheet-model-based reconstruction of land ice albedo over the last 5 million years (Myr). We find that the land ice albedo forcing depends non-linearly on the background climate, while any non-linearity of CO2 radiative forcing depends on the CO2 data set used. This non-linearity has not, so far, been accounted for in similar approaches due to previously more simplistic approximations, in which land ice albedo radiative forcing was a linear function of sea level change. The latitudinal dependency of ice-sheet area changes is important for the non-linearity between land ice albedo and sea level. In our set-up, in which the radiative forcing of CO2 and of the land ice albedo (LI) is combined, we find a state dependence in the calculated specific equilibrium climate sensitivity, S[CO2,LI], for most of the Pleistocene (last 2.1 Myr). During Pleistocene intermediate glaciated climates and interglacial periods, S[CO2,LI] is on average ~ 45 % larger than during Pleistocene full glacial conditions. In the Pliocene part of our analysis (2.6-5 Myr BP) the CO2 data uncertainties prevent a well-supported calculation for S[CO2,LI], but our analysis suggests that during times without a large land ice area in the Northern Hemisphere (e.g. before 2.82 Myr BP), the specific equilibrium climate sensitivity, S[CO2,LI], was smaller than during interglacials of the Pleistocene. We thus find support for a previously proposed state change in the climate system with the widespread appearance of northern hemispheric ice sheets. This study points for the first time to a so far overlooked non-linearity in the land ice albedo radiative forcing, which is important for similar palaeodata-based approaches to calculate climate sensitivity. However, the implications of this study for a suggested warming under CO2 doubling are not yet entirely clear since the details of necessary corrections for other slow feedbacks are not fully known and the uncertainties that exist in the ice-sheet simulations and global temperature reconstructions are large.
Resumo:
CIPWFULL is a user-friendly, stand-alone FORTRAN software program that is designed to calculate the comprehensive CIPW normative mineral composition of igneous rocks and strictly adheres to the original formulation of the CIPW protocol. This faithful adherence alleviates inaccuracies in normative mineral calculations by programs commonly used by petrologists. Additionally, several of the most important petrological and mineralogical parameters of igneous rocks are calculated by the program. Along with all the regular major oxide elements, all the significant minor elements whose contents can potentially effect the CIPW normative mineral composition are included. CIPWFULL also calculates oxidation ratios for igneous rock samples that have only one oxidation state of iron reported in the specimen analysis. It also provides an option for normalization of analyses to unity on a hydrous-free basis in order to facilitate comparison of norms among rock groups. Other capabilities of the program cater for rare situations, like the presence of cancrinite or exclusion from the norm calculation of rare rocks like carbonatite. Several mineralogical, petrological and discriminatory parameters and indexes are additionally calculated by the CIPWFULL program. The CIPWFULL program is very efficient and flexible and allows for a user-defined free-format input of all the chemical species, and it permits feeding of minor elements as parts per million or oxide percentages. Results of calculations are printed in a formatted ASCII text file and may be optionally casted into a space-delimited text files that are ready to be imported to general spreadsheet programs. CIPWFULL is DOS-based and is implemented on WINDOWS and mainframe platforms.
Resumo:
A plan to construct a canal through the Kra Isthmus in Southern Thailand has been proposed many times since the 17th century. The proposed canal would become an alternative route to the over-crowded Straits of Malacca. In this paper, we attempt to utilize a Geographical Information System (GIS) to calculate the realistic distances between ports that would be affected by the Kra Canal and to estimate the economic impact of the canal using a simulation model based on spatial economics. We find that China, India, Japan, and Europe gain the most from the construction of the canal, besides Thailand. On the other hand, the routes through the Straits of Malacca are largely beneficial to Malaysia, Brunei, and Indonesia, besides Singapore. Thus, it is beneficial for all ASEAN member countries that the Kra Canal and the Straits of Malacca coexist and complement one another.
Resumo:
This study presents a robust method for ground plane detection in vision-based systems with a non-stationary camera. The proposed method is based on the reliable estimation of the homography between ground planes in successive images. This homography is computed using a feature matching approach, which in contrast to classical approaches to on-board motion estimation does not require explicit ego-motion calculation. As opposed to it, a novel homography calculation method based on a linear estimation framework is presented. This framework provides predictions of the ground plane transformation matrix that are dynamically updated with new measurements. The method is specially suited for challenging environments, in particular traffic scenarios, in which the information is scarce and the homography computed from the images is usually inaccurate or erroneous. The proposed estimation framework is able to remove erroneous measurements and to correct those that are inaccurate, hence producing a reliable homography estimate at each instant. It is based on the evaluation of the difference between the predicted and the observed transformations, measured according to the spectral norm of the associated matrix of differences. Moreover, an example is provided on how to use the information extracted from ground plane estimation to achieve object detection and tracking. The method has been successfully demonstrated for the detection of moving vehicles in traffic environments.
Resumo:
A particle accelerator is any device that, using electromagnetic fields, is able to communicate energy to charged particles (typically electrons or ionized atoms), accelerating and/or energizing them up to the required level for its purpose. The applications of particle accelerators are countless, beginning in a common TV CRT, passing through medical X-ray devices, and ending in large ion colliders utilized to find the smallest details of the matter. Among the other engineering applications, the ion implantation devices to obtain better semiconductors and materials of amazing properties are included. Materials supporting irradiation for future nuclear fusion plants are also benefited from particle accelerators. There are many devices in a particle accelerator required for its correct operation. The most important are the particle sources, the guiding, focalizing and correcting magnets, the radiofrequency accelerating cavities, the fast deflection devices, the beam diagnostic mechanisms and the particle detectors. Most of the fast particle deflection devices have been built historically by using copper coils and ferrite cores which could effectuate a relatively fast magnetic deflection, but needed large voltages and currents to counteract the high coil inductance in a response in the microseconds range. Various beam stability considerations and the new range of energies and sizes of present time accelerators and their rings require new devices featuring an improved wakefield behaviour and faster response (in the nanoseconds range). This can only be achieved by an electromagnetic deflection device based on a transmission line. The electromagnetic deflection device (strip-line kicker) produces a transverse displacement on the particle beam travelling close to the speed of light, in order to extract the particles to another experiment or to inject them into a different accelerator. The deflection is carried out by the means of two short, opposite phase pulses. The diversion of the particles is exerted by the integrated Lorentz force of the electromagnetic field travelling along the kicker. This Thesis deals with a detailed calculation, manufacturing and test methodology for strip-line kicker devices. The methodology is then applied to two real cases which are fully designed, built, tested and finally installed in the CTF3 accelerator facility at CERN (Geneva). Analytical and numerical calculations, both in 2D and 3D, are detailed starting from the basic specifications in order to obtain a conceptual design. Time domain and frequency domain calculations are developed in the process using different FDM and FEM codes. The following concepts among others are analyzed: scattering parameters, resonating high order modes, the wakefields, etc. Several contributions are presented in the calculation process dealing specifically with strip-line kicker devices fed by electromagnetic pulses. Materials and components typically used for the fabrication of these devices are analyzed in the manufacturing section. Mechanical supports and connexions of electrodes are also detailed, presenting some interesting contributions on these concepts. The electromagnetic and vacuum tests are then analyzed. These tests are required to ensure that the manufactured devices fulfil the specifications. Finally, and only from the analytical point of view, the strip-line kickers are studied together with a pulsed power supply based on solid state power switches (MOSFETs). The solid state technology applied to pulsed power supplies is introduced and several circuit topologies are modelled and simulated to obtain fast and good flat-top pulses.
Resumo:
Se presenta un nuevo método de diseño conceptual en Ingeniería Aeronáutica basado el uso de modelos reducidos, también llamados modelos sustitutos (‘surrogates’). Los ingredientes de la función objetivo se calculan para cada indiviudo mediante la utilización de modelos sustitutos asociados a las distintas disciplinas técnicas que se construyen mediante definiciones de descomposición en valores singulares de alto orden (HOSVD) e interpolaciones unidimensionales. Estos modelos sustitutos se obtienen a partir de un número limitado de cálculos CFD. Los modelos sustitutos pueden combinarse, bien con un método de optimización global de tipo algoritmo genético, o con un método local de tipo gradiente. El método resultate es flexible a la par que mucho más eficiente, computacionalmente hablando, que los modelos convencionales basados en el cálculo directo de la función objetivo, especialmente si aparecen un gran número de parámetros de diseño y/o de modelado. El método se ilustra considerando una versión simplificada del diseño conceptual de un avión. Abstract An optimization method for conceptual design in Aeronautics is presented that is based on the use of surrogate models. The various ingredients in the target function are calculated for each individual using surrogates of the associated technical disciplines that are constructed via high order singular value decomposition and one dimensional interpolation. These surrogates result from a limited number of CFD calculated snapshots. The surrogates are combined with an optimization method, which can be either a global optimization method such as a genetic algorithm or a local optimization method, such as a gradient-like method. The resulting method is both flexible and much more computationally efficient than the conventional method based on direct calculation of the target function, especially if a large number of free design parameters and/or tunablemodeling parameters are present. The method is illustrated considering a simplified version of the conceptual design of an aircraft empennage.
Resumo:
This paper proposes an interleaved multiphase buck converter with minimum time control strategy for envelope amplifiers in high efficiency RF power amplifiers. The solution of the envelope amplifier is to combine the proposed converter with a linear regulator in series. High system efficiency can be obtained through modulating the supply voltage of the envelope amplifier with the fast output voltage variation of the converter working with several particular duty cycles that achieve total ripple cancellation. The transient model for minimum time control is explained, and the calculation of transient times that are pre-calculated and inserted into a look-up table is presented. The filter design trade-off that limits capability of envelope modulation is also discussed. The experimental results verify the fast voltage transient obtained with a 4-phase buck prototype.
Resumo:
In this paper, an interleaved multiphase buck converter with minimum time control strategy for envelope amplifiers in high efficiency RF power amplifiers is proposed. The solution for the envelope amplifier is to combine the proposed converter with a linear regulator in series. High efficiency of envelope amplifier can be obtained through modulating the supply voltage of the linear regulator. Instead of tracking the envelope, the buck converter has discrete output voltage that corresponding to particular duty cycles which achieve total ripple cancellation. The transient model for minimum time control is explained, and the calculation of transient times that are pre-calculated and inserted into a lookup table is presented. The filter design trade-off that limits capability of envelope modulation is also discussed. The experimental results verify the fast voltage transient obtained with a 4-phase buck prototype.
Resumo:
El objetivo de la presente tesis doctoral es el desarrollo de un nuevo concepto de biosensor óptico sin marcado, basado en una combinación de técnicas de caracterización óptica de interrogación vertical y estructuras sub-micrométricas fabricadas sobre chips de silicio. Las características más importantes de dicho dispositivo son su simplicidad, tanto desde el punto de vista de medida óptica como de introducción de las muestras a medir en el área sensible, aspectos que suelen ser críticos en la mayoría de sensores encontrados en la literatura. Cada uno de los aspectos relacionados con el diseño de un biosensor, que son fundamentalmente cuatro (diseño fotónico, caracterización óptica, fabricación y fluídica/inmovilización química) son desarrollados en detalle en los capítulos correspondientes. En la primera parte de la tesis se hace una introducción al concepto de biosensor, en qué consiste, qué tipos hay y cuáles son los parámetros más comunes usados para cuantificar su comportamiento. Posteriormente se realiza un análisis del estado del arte en la materia, enfocado en particular en el área de biosensores ópticos sin marcado. Se introducen también cuáles son las reacciones bioquímicas a estudiar (inmunoensayos). En la segunda parte se describe en primer lugar cuáles son las técnicas ópticas empleadas en la caracterización: Reflectometría, Elipsometría y Espectrometría; además de los motivos que han llevado a su empleo. Posteriormente se introducen diversos diseños de las denominadas "celdas optofluídicas", que son los dispositivos en los que se va a producir la interacción bioquímica. Se presentan cuatro dispositivos diferentes, y junto con ellos, se proponen diversos métodos de cálculo teórico de la respuesta óptica esperada. Posteriormente se procede al cálculo de la sensibilidad esperada para cada una de las celdas, así como al análisis de los procesos de fabricación de cada una de ellas y su comportamiento fluídico. Una vez analizados todos los aspectos críticos del comportamiento del biosensor, se puede realizar un proceso de optimización de su diseño. Esto se realiza usando un modelo de cálculo simplificado (modelo 1.5-D) que permite la obtención de parámetros como la sensibilidad y el límite de detección de un gran número de dispositivos en un tiempo relativamente reducido. Para este proceso se escogen dos de las celdas optofluídicas propuestas. En la parte final de la tesis se muestran los resultados experimentales obtenidos. En primer lugar, se caracteriza una celda basada en agujeros sub-micrométricos como sensor de índice de refracción, usando para ello diferentes líquidos orgánicos; dichos resultados experimentales presentan una buena correlación con los cálculos teóricos previos, lo que permite validar el modelo conceptual presentado. Finalmente, se realiza un inmunoensayo químico sobre otra de las celdas propuestas (pilares nanométricos de polímero SU-8). Para ello se utiliza el inmunoensayo de albumina de suero bovino (BSA) y su anticuerpo (antiBSA). Se detalla el proceso de obtención de la celda, la funcionalización de la superficie con los bioreceptores (en este caso, BSA) y el proceso de biorreconocimiento. Este proceso permite dar una primera estimación de cuál es el límite de detección esperable para este tipo de sensores en un inmunoensayo estándar. En este caso, se alcanza un valor de 2.3 ng/mL, que es competitivo comparado con otros ensayos similares encontrados en la literatura. La principal conclusión de la tesis es que esta tipología de dispositivos puede ser usada como inmunosensor, y presenta ciertas ventajas respecto a los actualmente existentes. Estas ventajas vienen asociadas, de nuevo, a su simplicidad, tanto a la hora de medir ópticamente, como dentro del proceso de introducción de los bioanalitos en el área sensora (depositando simplemente una gota sobre la micro-nano-estructura). Los cálculos teorícos realizados en los procesos de optimización sugieren a su vez que el comportamiento del sensor, medido en magnitudes como límite de detección biológico puede ser ampliamente mejorado con una mayor compactación de pilares, alcanzandose un valor mínimo de 0.59 ng/mL). The objective of this thesis is to develop a new concept of optical label-free biosensor, based on a combination of vertical interrogation optical techniques and submicron structures fabricated over silicon chips. The most important features of this device are its simplicity, both from the point of view of optical measurement and regarding to the introduction of samples to be measured in the sensing area, which are often critical aspects in the majority of sensors found in the literature. Each of the aspects related to the design of biosensors, which are basically four (photonic design, optical characterization, fabrication and fluid / chemical immobilization) are developed in detail in the relevant chapters. The first part of the thesis consists of an introduction to the concept of biosensor: which elements consists of, existing types and the most common parameters used to quantify its behavior. Subsequently, an analysis of the state of the art in this area is presented, focusing in particular in the area of label free optical biosensors. What are also introduced to study biochemical reactions (immunoassays). The second part describes firstly the optical techniques used in the characterization: reflectometry, ellipsometry and spectrometry; in addition to the reasons that have led to their use. Subsequently several examples of the so-called "optofluidic cells" are introduced, which are the devices where the biochemical interactions take place. Four different devices are presented, and their optical response is calculated by using various methods. Then is exposed the calculation of the expected sensitivity for each of the cells, and the analysis of their fabrication processes and fluidic behavior at the sub-micrometric range. After analyzing all the critical aspects of the biosensor, it can be performed a process of optimization of a particular design. This is done using a simplified calculation model (1.5-D model calculation) that allows obtaining parameters such as sensitivity and the detection limit of a large number of devices in a relatively reduced time. For this process are chosen two different optofluidic cells, from the four previously proposed. The final part of the thesis is the exposition of the obtained experimental results. Firstly, a cell based sub-micrometric holes is characterized as refractive index sensor using different organic fluids, and such experimental results show a good correlation with previous theoretical calculations, allowing to validate the conceptual model presented. Finally, an immunoassay is performed on another typology of cell (SU-8 polymer pillars). This immunoassay uses bovine serum albumin (BSA) and its antibody (antiBSA). The processes for obtaining the cell surface functionalization with the bioreceptors (in this case, BSA) and the biorecognition (antiBSA) are detailed. This immunoassay can give a first estimation of which are the expected limit of detection values for this typology of sensors in a standard immunoassay. In this case, it reaches a value of 2.3 ng/mL, which is competitive with other similar assays found in the literature. The main conclusion of the thesis is that this type of device can be used as immunosensor, and has certain advantages over the existing ones. These advantages are associated again with its simplicity, by the simpler coupling of light and in the process of introduction of bioanalytes into the sensing areas (by depositing a droplet over the micro-nano-structure). Theoretical calculations made in optimizing processes suggest that the sensor Limit of detection can be greatly improved with higher compacting of the lattice of pillars, reaching a minimum value of 0.59 ng/mL).
Resumo:
A novel formulation for the surface impedance characterization is introduced for the canonical problem of surface fields on a perfect electric conductor (PEC) circular cylinder with a dielectric coating due to a electric current source using the Uniform Theory of Diffraction (UTD) with an Impedance Boundary Condition (IBC). The approach is based on a TE/TM assumption of the surface fields from the original problem. Where this surface impedance fails, an optimization is performed to minimize the error in the SD Green?s function between the original problem and the equivalent one with the IBC. This asymptotic method, accurate for large separations between source and observer points, in combination with spectral domain (SD) Green?s functions for multidielectric coatings leads to a new hybrid SD-UTD with IBC to calculate mutual coupling among microstrip patches on a multilayer dielectric-coated PEC circular cylinder. Results are compared with the eigenfunction solution in SD, where a very good agreement is met.
Resumo:
Europe needs to restructure its energy system. The aim to decrease the reliance on fossil fuels to a higher dependence on renewable energy has now been imposed by The European Commission. In order to achieve this goal there is a great interest in Norway to become "The Green Battery of Europe". In the pursuit of this goal a GIS-tool was created to investigate the pump storage potential in Norway. The tool searches for possible connections between existing reservoirs and dams with the criteria selected by the user. The aim of this thesis was to test the tool and see if the results suggested were plausible, develop a cost calculation method for the PSH lines, and make suggestions for further development of the tool. During the process the tool presented many non-feasible pumped storage hydropower (PSH) connections. The area of Telemark was chosen for the more detailed study. The results were discussed and some improvements were suggested for further development of the tool. Also a sensitivity test was done to see which of the parameters set by the user are the most relevant for the PSH connection suggestion. From a range of the most promising PSH plants suggested by the tool, the one between Songavatn and Totak was chosen for a case study, where there already exists a power plant between both reservoirs. A new Pumped Storage Plant was designed with a power production of 1200 MW. There are still many topics open to discussion, such as how to deal with environmental restrictions, or how to deal with inflows and outflows of the reservoirs from the existing power plants. Consequently the GIS-tool can be a very useful tool to establish the best possible connections between existing reservoirs and dams, but it still needs a deep study and the creation of new parameters for the user.
Resumo:
Recently, a theoretical criterion to calculate the stability of an axial-flow compressor rotor has been presented in the scientific literature. This theoretical criterion was used for determining the locus of the stability line over the rotor map and for predicting the post-stall evolution of the constant-speed line of a rotor. The main objective of this paper is to improve the predictions of such a model. To do that, the paper proposes a different characterization of the characteristic azimuthal length and a calculation of the ratio of specific heats based on a polytropic exponent. Thanks to these new values, the model predicts two bifurcation points in the behaviour of the flow: the inception point of the instability and the surge point. Experimental data from a pure axial compressor are used to validate the model showing that the prediction of the flow coefficient at the surge point has an error inferior to 5%. For the rotor studied, the paper provides a quantitative and qualitative description of the inception of the instability and of the mechanism involved in the instable region of the compressor map. The paper also discusses the role of rotor efficiency in the position of the bifurcations and gives a sensitivity analysis of its position. Finally, it presents a discussion about how the model can explain the different behaviours exhibited by the same rotor when the flow coefficient is reduced