929 resultados para two input two output


Relevância:

100.00% 100.00%

Publicador:

Resumo:

BACKGROUND: In contrast to hypnosis, there is no surrogate parameter for analgesia in anesthetized patients. Opioids are titrated to suppress blood pressure response to noxious stimulation. The authors evaluated a novel model predictive controller for closed-loop administration of alfentanil using mean arterial blood pressure and predicted plasma alfentanil concentration (Cp Alf) as input parameters. METHODS: The authors studied 13 healthy patients scheduled to undergo minor lumbar and cervical spine surgery. After induction with propofol, alfentanil, and mivacurium and tracheal intubation, isoflurane was titrated to maintain the Bispectral Index at 55 (+/- 5), and the alfentanil administration was switched from manual to closed-loop control. The controller adjusted the alfentanil infusion rate to maintain the mean arterial blood pressure near the set-point (70 mmHg) while minimizing the Cp Alf toward the set-point plasma alfentanil concentration (Cp Alfref) (100 ng/ml). RESULTS: Two patients were excluded because of loss of arterial pressure signal and protocol violation. The alfentanil infusion was closed-loop controlled for a mean (SD) of 98.9 (1.5)% of presurgery time and 95.5 (4.3)% of surgery time. The mean (SD) end-tidal isoflurane concentrations were 0.78 (0.1) and 0.86 (0.1) vol%, the Cp Alf values were 122 (35) and 181 (58) ng/ml, and the Bispectral Index values were 51 (9) and 52 (4) before surgery and during surgery, respectively. The mean (SD) absolute deviations of mean arterial blood pressure were 7.6 (2.6) and 10.0 (4.2) mmHg (P = 0.262), and the median performance error, median absolute performance error, and wobble were 4.2 (6.2) and 8.8 (9.4)% (P = 0.002), 7.9 (3.8) and 11.8 (6.3)% (P = 0.129), and 14.5 (8.4) and 5.7 (1.2)% (P = 0.002) before surgery and during surgery, respectively. A post hoc simulation showed that the Cp Alfref decreased the predicted Cp Alf compared with mean arterial blood pressure alone. CONCLUSION: The authors' controller has a similar set-point precision as previous hypnotic controllers and provides adequate alfentanil dosing during surgery. It may help to standardize opioid dosing in research and may be a further step toward a multiple input-multiple output controller.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Target localization has a wide range of military and civilian applications in wireless mobile networks. Examples include battle-field surveillance, emergency 911 (E911), traffc alert, habitat monitoring, resource allocation, routing, and disaster mitigation. Basic localization techniques include time-of-arrival (TOA), direction-of-arrival (DOA) and received-signal strength (RSS) estimation. Techniques that are proposed based on TOA and DOA are very sensitive to the availability of Line-of-sight (LOS) which is the direct path between the transmitter and the receiver. If LOS is not available, TOA and DOA estimation errors create a large localization error. In order to reduce NLOS localization error, NLOS identifcation, mitigation, and localization techniques have been proposed. This research investigates NLOS identifcation for multiple antennas radio systems. The techniques proposed in the literature mainly use one antenna element to enable NLOS identifcation. When a single antenna is utilized, limited features of the wireless channel can be exploited to identify NLOS situations. However, in DOA-based wireless localization systems, multiple antenna elements are available. In addition, multiple antenna technology has been adopted in many widely used wireless systems such as wireless LAN 802.11n and WiMAX 802.16e which are good candidates for localization based services. In this work, the potential of spatial channel information for high performance NLOS identifcation is investigated. Considering narrowband multiple antenna wireless systems, two xvNLOS identifcation techniques are proposed. Here, the implementation of spatial correlation of channel coeffcients across antenna elements as a metric for NLOS identifcation is proposed. In order to obtain the spatial correlation, a new multi-input multi-output (MIMO) channel model based on rough surface theory is proposed. This model can be used to compute the spatial correlation between the antenna pair separated by any distance. In addition, a new NLOS identifcation technique that exploits the statistics of phase difference across two antenna elements is proposed. This technique assumes the phases received across two antenna elements are uncorrelated. This assumption is validated based on the well-known circular and elliptic scattering models. Next, it is proved that the channel Rician K-factor is a function of the phase difference variance. Exploiting Rician K-factor, techniques to identify NLOS scenarios are proposed. Considering wideband multiple antenna wireless systems which use MIMO-orthogonal frequency division multiplexing (OFDM) signaling, space-time-frequency channel correlation is exploited to attain NLOS identifcation in time-varying, frequency-selective and spaceselective radio channels. Novel NLOS identi?cation measures based on space, time and frequency channel correlation are proposed and their performances are evaluated. These measures represent a better NLOS identifcation performance compared to those that only use space, time or frequency.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Multi-input multi-output (MIMO) technology is an emerging solution for high data rate wireless communications. We develop soft-decision based equalization techniques for frequency selective MIMO channels in the quest for low-complexity equalizers with BER performance competitive to that of ML sequence detection. We first propose soft decision equalization (SDE), and demonstrate that decision feedback equalization (DFE) based on soft-decisions, expressed via the posterior probabilities associated with feedback symbols, is able to outperform hard-decision DFE, with a low computational cost that is polynomial in the number of symbols to be recovered, and linear in the signal constellation size. Building upon the probabilistic data association (PDA) multiuser detector, we present two new MIMO equalization solutions to handle the distinctive channel memory. With their low complexity, simple implementations, and impressive near-optimum performance offered by iterative soft-decision processing, the proposed SDE methods are attractive candidates to deliver efficient reception solutions to practical high-capacity MIMO systems. Motivated by the need for low-complexity receiver processing, we further present an alternative low-complexity soft-decision equalization approach for frequency selective MIMO communication systems. With the help of iterative processing, two detection and estimation schemes based on second-order statistics are harmoniously put together to yield a two-part receiver structure: local multiuser detection (MUD) using soft-decision Probabilistic Data Association (PDA) detection, and dynamic noise-interference tracking using Kalman filtering. The proposed Kalman-PDA detector performs local MUD within a sub-block of the received data instead of over the entire data set, to reduce the computational load. At the same time, all the inter-ference affecting the local sub-block, including both multiple access and inter-symbol interference, is properly modeled as the state vector of a linear system, and dynamically tracked by Kalman filtering. Two types of Kalman filters are designed, both of which are able to track an finite impulse response (FIR) MIMO channel of any memory length. The overall algorithms enjoy low complexity that is only polynomial in the number of information-bearing bits to be detected, regardless of the data block size. Furthermore, we introduce two optional performance-enhancing techniques: cross- layer automatic repeat request (ARQ) for uncoded systems and code-aided method for coded systems. We take Kalman-PDA as an example, and show via simulations that both techniques can render error performance that is better than Kalman-PDA alone and competitive to sphere decoding. At last, we consider the case that channel state information (CSI) is not perfectly known to the receiver, and present an iterative channel estimation algorithm. Simulations show that the performance of SDE with channel estimation approaches that of SDE with perfect CSI.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

We propose integrated optical structures that can be used as isolators and polarization splitters based on engineered photonic lattices. Starting from optical waveguide arrays that mimic Fock space (quantum state with a well-defined particle number) representation of a non-interacting two-site Bose Hubbard Hamiltonian, we show that introducing magneto-optic nonreciprocity to these structures leads to a superior optical isolation performance. In the forward propagation direction, an input TM polarized beam experiences a perfect state transfer between the input and output waveguide channels while surface Bloch oscillations block the backward transmission between the same ports. Our analysis indicates a large isolation ratio of 75 dB after a propagation distance of 8mm inside seven coupled waveguides. Moreover, we demonstrate that, a judicious choice of the nonreciprocity in this same geometry can lead to perfect polarization splitting.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The hippocampus receives input from upper levels of the association cortex and is implicated in many mnemonic processes, but the exact mechanisms by which it codes and stores information is an unresolved topic. This work examines the flow of information through the hippocampal formation while attempting to determine the computations that each of the hippocampal subfields performs in learning and memory. The formation, storage, and recall of hippocampal-dependent memories theoretically utilize an autoassociative attractor network that functions by implementing two competitive, yet complementary, processes. Pattern separation, hypothesized to occur in the dentate gyrus (DG), refers to the ability to decrease the similarity among incoming information by producing output patterns that overlap less than the inputs. In contrast, pattern completion, hypothesized to occur in the CA3 region, refers to the ability to reproduce a previously stored output pattern from a partial or degraded input pattern. Prior to addressing the functional role of the DG and CA3 subfields, the spatial firing properties of neurons in the dentate gyrus were examined. The principal cell of the dentate gyrus, the granule cell, has spatially selective place fields; however, the behavioral correlates of another excitatory cell, the mossy cell of the dentate polymorphic layer, are unknown. This report shows that putative mossy cells have spatially selective firing that consists of multiple fields similar to previously reported properties of granule cells. Other cells recorded from the DG had single place fields. Compared to cells with multiple fields, cells with single fields fired at a lower rate during sleep, were less likely to burst, and were more likely to be recorded simultaneously with a large population of neurons that were active during sleep and silent during behavior. These data suggest that single-field and multiple-field cells constitute at least two distinct cell classes in the DG. Based on these characteristics, we propose that putative mossy cells tend to fire in multiple, distinct locations in an environment, whereas putative granule cells tend to fire in single locations, similar to place fields of the CA1 and CA3 regions. Experimental evidence supporting the theories of pattern separation and pattern completion comes from both behavioral and electrophysiological tests. These studies specifically focused on the function of each subregion and made implicit assumptions about how environmental manipulations changed the representations encoded by the hippocampal inputs. However, the cell populations that provided these inputs were in most cases not directly examined. We conducted a series of studies to investigate the neural activity in the entorhinal cortex, dentate gyrus, and CA3 in the same experimental conditions, which allowed a direct comparison between the input and output representations. The results show that the dentate gyrus representation changes between the familiar and cue altered environments more than its input representations, whereas the CA3 representation changes less than its input representations. These findings are consistent with longstanding computational models proposing that (1) CA3 is an associative memory system performing pattern completion in order to recall previous memories from partial inputs, and (2) the dentate gyrus performs pattern separation to help store different memories in ways that reduce interference when the memories are subsequently recalled.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The purpose of this research is to examine the relative profitability of the firm within the nursing facility industry in Texas. An examination is made of the variables expected to affect profitability and of importance to the design and implementation of regulatory policy. To facilitate this inquiry, specific questions addressed are: (1) Do differences in ownership form affect profitability (defined as operating income before fixed costs)? (2) What impact does regional location have on profitability? (3) Do patient case-mix and access to care by Medicaid patients differ between proprietary and non-profit firms and facilities located in urban versus rural regions, and what association exists between these variables and profitability? (4) Are economies of scale present in the nursing home industry? (5) Do nursing facilities operate in a competitive output market characterized by the inability of a single firm to exhibit influence over market price?^ Prior studies have principally employed a cost function to assess efficiency differences between classifications of nursing facilities. The inherent weakness in this approach is that it only considers technical efficiency. Not both technical and price efficiency which are the two components of overall economic efficiency. One firm is more technically efficient compared to another if it is able to produce a given quantity of output at the least possible costs. Price efficiency means that scarce resources are being directed towards their most valued use. Assuming similar prices in both input and output markets, differences in overall economic efficiency between firm classes are assessed through profitability, hence a profit function.^ Using the framework of the profit function, data from 1990 Medicaid Costs Reports for Texas, and the analytic technique of Ordinary Least Squares Regression, the findings of the study indicated (1) similar profitability between nursing facilities organized as for-profit versus non-profit and located in urban versus rural regions, (2) an inverse association between both payor-mix and patient case-mix with profitability, (3) strong evidence for the presence of scale economies, and (4) existence of a competitive market structure. The paper concludes with implications regarding reimbursement methodology and construction moratorium policies in Texas. ^

Relevância:

100.00% 100.00%

Publicador:

Resumo:

A torque meter comprising hollow-keyed, input and output female shafts adapted to receive the male shafts of the power source and machine respectively. Each shaft has a circular flange whose face is perpendicular to the center line of the shafts. Each flange has a plurality of equally spaced cylindrical recesses machined into the inside face thereto adapted to receive conical inserts therein. Balls are contained by the conical inserts and transmit the rotational movement from the input to the output shaft. A stationary housing extends around the input and output shaft and has a transducer shell secured thereto. When force is applied to the input shaft to cause movement, the balls encounter torsional resistance which causes the balls to roll up the ramps of the conical seat inserts to separate the two torque flanges. The force transmitted through the balls causes rotation to the output shaft and produces tension to the stationary transducer shell. The stationary transducer shell is instrumented with semi-conductor strain gauges.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Objective: The present study offers a novel methodological contribution to the study of the configuration and dynamics of research groups, through a comparative perspective of the projects funded (inputs) and publication co-authorships (output). Method: A combination of bibliometric techniques and social network analysis was applied to a case study: the Departmento de Bibliotecología (DHUBI), Universidad Nacional de La Plata, Argentina, for the period 2000-2009. The results were interpreted statistically and staff members of the department, were interviewed. Results: The method makes it possible to distinguish groups, identify their members and reflect group make-up through an analytical strategy that involves the categorization of actors and the interdisciplinary and national or international projection of the networks that they configure. The integration of these two aspects (input and output) at different points in time over the analyzed period leads to inferences about group profiles and the roles of actors. Conclusions: The methodology presented is conducive to micro-level interpretations in a given area of study, regarding individual researchers or research groups. Because the comparative input-output analysis broadens the base of information and makes it possible to follow up, over time, individual and group trends, it may prove very useful for the management, promotion and evaluation of science

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Objective: The present study offers a novel methodological contribution to the study of the configuration and dynamics of research groups, through a comparative perspective of the projects funded (inputs) and publication co-authorships (output). Method: A combination of bibliometric techniques and social network analysis was applied to a case study: the Departmento de Bibliotecología (DHUBI), Universidad Nacional de La Plata, Argentina, for the period 2000-2009. The results were interpreted statistically and staff members of the department, were interviewed. Results: The method makes it possible to distinguish groups, identify their members and reflect group make-up through an analytical strategy that involves the categorization of actors and the interdisciplinary and national or international projection of the networks that they configure. The integration of these two aspects (input and output) at different points in time over the analyzed period leads to inferences about group profiles and the roles of actors. Conclusions: The methodology presented is conducive to micro-level interpretations in a given area of study, regarding individual researchers or research groups. Because the comparative input-output analysis broadens the base of information and makes it possible to follow up, over time, individual and group trends, it may prove very useful for the management, promotion and evaluation of science

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Objective: The present study offers a novel methodological contribution to the study of the configuration and dynamics of research groups, through a comparative perspective of the projects funded (inputs) and publication co-authorships (output). Method: A combination of bibliometric techniques and social network analysis was applied to a case study: the Departmento de Bibliotecología (DHUBI), Universidad Nacional de La Plata, Argentina, for the period 2000-2009. The results were interpreted statistically and staff members of the department, were interviewed. Results: The method makes it possible to distinguish groups, identify their members and reflect group make-up through an analytical strategy that involves the categorization of actors and the interdisciplinary and national or international projection of the networks that they configure. The integration of these two aspects (input and output) at different points in time over the analyzed period leads to inferences about group profiles and the roles of actors. Conclusions: The methodology presented is conducive to micro-level interpretations in a given area of study, regarding individual researchers or research groups. Because the comparative input-output analysis broadens the base of information and makes it possible to follow up, over time, individual and group trends, it may prove very useful for the management, promotion and evaluation of science

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Classical imaging optics has been developed over centuries in many areas, such as its paraxial imaging theory and practical design methods like multi-parametric optimization techniques. Although these imaging optical design methods can provide elegant solutions to many traditional optical problems, there are more and more new design problems, like solar concentrator, illumination system, ultra-compact camera, etc., that require maximum energy transfer efficiency, or ultra-compact optical structure. These problems do not have simple solutions from classical imaging design methods, because not only paraxial rays, but also non-paraxial rays should be well considered in the design process. Non-imaging optics is a newly developed optical discipline, which does not aim to form images, but to maximize energy transfer efficiency. One important concept developed from non-imaging optics is the “edge-ray principle”, which states that the energy flow contained in a bundle of rays will be transferred to the target, if all its edge rays are transferred to the target. Based on that concept, many CPC solar concentrators have been developed with efficiency close to the thermodynamic limit. When more than one bundle of edge-rays needs to be considered in the design, one way to obtain solutions is to use SMS method. SMS stands for Simultaneous Multiple Surface, which means several optical surfaces are constructed simultaneously. The SMS method was developed as a design method in Non-imaging optics during the 90s. The method can be considered as an extension to the Cartesian Oval calculation. In the traditional Cartesian Oval calculation, one optical surface is built to transform an input wave-front to an out-put wave-front. The SMS method however, is dedicated to solve more than 1 wave-fronts transformation problem. In the beginning, only 2 input wave-fronts and 2 output wave-fronts transformation problem was considered in the SMS design process for rotational optical systems or free-form optical systems. Usually “SMS 2D” method stands for the SMS procedure developed for rotational optical system, and “SMS 3D” method for the procedure for free-form optical system. Although the SMS method was originally employed in non-imaging optical system designs, it has been found during this thesis that with the improved capability to design more surfaces and control more input and output wave-fronts, the SMS method can also be applied to imaging system designs and possesses great advantage over traditional design methods. In this thesis, one of the main goals to achieve is to further develop the existing SMS-2D method to design with more surfaces and improve the stability of the SMS-2D and SMS-3D algorithms, so that further optimization process can be combined with SMS algorithms. The benefits of SMS plus optimization strategy over traditional optimization strategy will be explained in details for both rotational and free-form imaging optical system designs. Another main goal is to develop novel design concepts and methods suitable for challenging non-imaging applications, e.g. solar concentrator and solar tracker. This thesis comprises 9 chapters and can be grouped into two parts: the first part (chapter 2-5) contains research works in the imaging field, and the second part (chapter 6-8) contains works in the non-imaging field. In the first chapter, an introduction to basic imaging and non-imaging design concepts and theories is given. Chapter 2 presents a basic SMS-2D imaging design procedure using meridian rays. In this chapter, we will set the imaging design problem from the SMS point of view, and try to solve the problem numerically. The stability of this SMS-2D design procedure will also be discussed. The design concepts and procedures developed in this chapter lay the path for further improvement. Chapter 3 presents two improved SMS 3 surfaces’ design procedures using meridian rays (SMS-3M) and skew rays (SMS-1M2S) respectively. The major improvement has been made to the central segments selections, so that the whole SMS procedures become more stable compared to procedures described in Chapter 2. Since these two algorithms represent two types of phase space sampling, their image forming capabilities are compared in a simple objective design. Chapter 4 deals with an ultra-compact SWIR camera design with the SMS-3M method. The difficulties in this wide band camera design is how to maintain high image quality meanwhile reduce the overall system length. This interesting camera design provides a playground for the classical design method and SMS design methods. We will show designs and optical performance from both classical design method and the SMS design method. Tolerance study is also given as the end of the chapter. Chapter 5 develops a two-stage SMS-3D based optimization strategy for a 2 freeform mirrors imaging system. In the first optimization phase, the SMS-3D method is integrated into the optimization process to construct the two mirrors in an accurate way, drastically reducing the unknown parameters to only few system configuration parameters. In the second optimization phase, previous optimized mirrors are parameterized into Qbfs type polynomials and set up in code V. Code V optimization results demonstrates the effectiveness of this design strategy in this 2-mirror system design. Chapter 6 shows an etendue-squeezing condenser optics, which were prepared for the 2010 IODC illumination contest. This interesting design employs many non-imaging techniques such as the SMS method, etendue-squeezing tessellation, and groove surface design. This device has theoretical efficiency limit as high as 91.9%. Chapter 7 presents a freeform mirror-type solar concentrator with uniform irradiance on the solar cell. Traditional parabolic mirror concentrator has many drawbacks like hot-pot irradiance on the center of the cell, insufficient use of active cell area due to its rotational irradiance pattern and small acceptance angle. In order to conquer these limitations, a novel irradiance homogenization concept is developed, which lead to a free-form mirror design. Simulation results show that the free-form mirror reflector has rectangular irradiance pattern, uniform irradiance distribution and large acceptance angle, which confirm the viability of the design concept. Chapter 8 presents a novel beam-steering array optics design strategy. The goal of the design is to track large angle parallel rays by only moving optical arrays laterally, and convert it to small angle parallel output rays. The design concept is developed as an extended SMS method. Potential applications of this beam-steering device are: skylights to provide steerable natural illumination, building integrated CPV systems, and steerable LED illumination. Conclusion and future lines of work are given in Chapter 9. Resumen La óptica de formación de imagen clásica se ha ido desarrollando durante siglos, dando lugar tanto a la teoría de óptica paraxial y los métodos de diseño prácticos como a técnicas de optimización multiparamétricas. Aunque estos métodos de diseño óptico para formación de imagen puede aportar soluciones elegantes a muchos problemas convencionales, siguen apareciendo nuevos problemas de diseño óptico, concentradores solares, sistemas de iluminación, cámaras ultracompactas, etc. que requieren máxima transferencia de energía o dimensiones ultracompactas. Este tipo de problemas no se pueden resolver fácilmente con métodos clásicos de diseño porque durante el proceso de diseño no solamente se deben considerar los rayos paraxiales sino también los rayos no paraxiales. La óptica anidólica o no formadora de imagen es una disciplina que ha evolucionado en gran medida recientemente. Su objetivo no es formar imagen, es maximazar la eficiencia de transferencia de energía. Un concepto importante de la óptica anidólica son los “rayos marginales”, que se pueden utilizar para el diseño de sistemas ya que si todos los rayos marginales llegan a nuestra área del receptor, todos los rayos interiores también llegarán al receptor. Haciendo uso de este principio, se han diseñado muchos concentradores solares que funcionan cerca del límite teórico que marca la termodinámica. Cuando consideramos más de un haz de rayos marginales en nuestro diseño, una posible solución es usar el método SMS (Simultaneous Multiple Surface), el cuál diseña simultáneamente varias superficies ópticas. El SMS nació como un método de diseño para óptica anidólica durante los años 90. El método puede ser considerado como una extensión del cálculo del óvalo cartesiano. En el método del óvalo cartesiano convencional, se calcula una superficie para transformar un frente de onda entrante a otro frente de onda saliente. El método SMS permite transformar varios frentes de onda de entrada en frentes de onda de salida. Inicialmente, sólo era posible transformar dos frentes de onda con dos superficies con simetría de rotación y sin simetría de rotación, pero esta limitación ha sido superada recientemente. Nos referimos a “SMS 2D” como el método orientado a construir superficies con simetría de rotación y llamamos “SMS 3D” al método para construir superficies sin simetría de rotación o free-form. Aunque el método originalmente fue aplicado en el diseño de sistemas anidólicos, se ha observado que gracias a su capacidad para diseñar más superficies y controlar más frentes de onda de entrada y de salida, el SMS también es posible aplicarlo a sistemas de formación de imagen proporcionando una gran ventaja sobre los métodos de diseño tradicionales. Uno de los principales objetivos de la presente tesis es extender el método SMS-2D para permitir el diseño de sistemas con mayor número de superficies y mejorar la estabilidad de los algoritmos del SMS-2D y SMS-3D, haciendo posible combinar la optimización con los algoritmos. Los beneficios de combinar SMS y optimización comparado con el proceso de optimización tradicional se explican en detalle para sistemas con simetría de rotación y sin simetría de rotación. Otro objetivo importante de la tesis es el desarrollo de nuevos conceptos de diseño y nuevos métodos en el área de la concentración solar fotovoltaica. La tesis está estructurada en 9 capítulos que están agrupados en dos partes: la primera de ellas (capítulos 2-5) se centra en la óptica formadora de imagen mientras que en la segunda parte (capítulos 6-8) se presenta el trabajo del área de la óptica anidólica. El primer capítulo consta de una breve introducción de los conceptos básicos de la óptica anidólica y la óptica en formación de imagen. El capítulo 2 describe un proceso de diseño SMS-2D sencillo basado en los rayos meridianos. En este capítulo se presenta el problema de diseñar un sistema formador de imagen desde el punto de vista del SMS y se intenta obtener una solución de manera numérica. La estabilidad de este proceso se analiza con detalle. Los conceptos de diseño y los algoritmos desarrollados en este capítulo sientan la base sobre la cual se realizarán mejoras. El capítulo 3 presenta dos procedimientos para el diseño de un sistema con 3 superficies SMS, el primero basado en rayos meridianos (SMS-3M) y el segundo basado en rayos oblicuos (SMS-1M2S). La mejora más destacable recae en la selección de los segmentos centrales, que hacen más estable todo el proceso de diseño comparado con el presentado en el capítulo 2. Estos dos algoritmos representan dos tipos de muestreo del espacio de fases, su capacidad para formar imagen se compara diseñando un objetivo simple con cada uno de ellos. En el capítulo 4 se presenta un diseño ultra-compacto de una cámara SWIR diseñada usando el método SMS-3M. La dificultad del diseño de esta cámara de espectro ancho radica en mantener una alta calidad de imagen y al mismo tiempo reducir drásticamente sus dimensiones. Esta cámara es muy interesante para comparar el método de diseño clásico y el método de SMS. En este capítulo se presentan ambos diseños y se analizan sus características ópticas. En el capítulo 5 se describe la estrategia de optimización basada en el método SMS-3D. El método SMS-3D calcula las superficies ópticas de manera precisa, dejando sólo unos pocos parámetros libres para decidir la configuración del sistema. Modificando el valor de estos parámetros se genera cada vez mediante SMS-3D un sistema completo diferente. La optimización se lleva a cabo variando los mencionados parámetros y analizando el sistema generado. Los resultados muestran que esta estrategia de diseño es muy eficaz y eficiente para un sistema formado por dos espejos. En el capítulo 6 se describe un sistema de compresión de la Etendue, que fue presentado en el concurso de iluminación del IODC en 2010. Este interesante diseño hace uso de técnicas propias de la óptica anidólica, como el método SMS, el teselado de las lentes y el diseño mediante grooves. Este dispositivo tiene un límite teórica en la eficiencia del 91.9%. El capítulo 7 presenta un concentrador solar basado en un espejo free-form con irradiancia uniforme sobre la célula. Los concentradores parabólicos tienen numerosas desventajas como los puntos calientes en la zona central de la célula, uso no eficiente del área de la célula al ser ésta cuadrada y además tienen ángulos de aceptancia de reducido. Para poder superar estas limitaciones se propone un novedoso concepto de homogeneización de la irrandancia que se materializa en un diseño con espejo free-form. El análisis mediante simulación demuestra que la irradiancia es homogénea en una región rectangular y con mayor ángulo de aceptancia, lo que confirma la viabilidad del concepto de diseño. En el capítulo 8 se presenta un novedoso concepto para el diseño de sistemas afocales dinámicos. El objetivo del diseño es realizar un sistema cuyo haz de rayos de entrada pueda llegar con ángulos entre ±45º mientras que el haz de rayos a la salida sea siempre perpendicular al sistema, variando únicamente la posición de los elementos ópticos lateralmente. Las aplicaciones potenciales de este dispositivo son varias: tragaluces que proporcionan iluminación natural, sistemas de concentración fotovoltaica integrados en los edificios o iluminación direccionable con LEDs. Finalmente, el último capítulo contiene las conclusiones y las líneas de investigación futura.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The cutoff frequencies of an EMI filter are normally given by the noise attenuation requirements the filter has to fulfill. In order to select the component values of the filter elements, i.e. inductances and capacitances, an additional design criterium is needed. In this paper the effect of the EMI filter input and output impedances are considered. The input impedance influences the filters effect on the system displacement power factor and the output impedance plays a key role in the system stability. The effect of filter element values, the number of filter stages as well as additional damping networks are considered and a design procedure is provided. For this analysis a two-port description of the input filters employing ABCD-parameters is used.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

La difusión de TV3D actual utiliza formatos como el Side-by-Side o Top-and-Bottom, en los que cada par de imágenes, correspondiente a las vistas de los ojos derecho e izquierdo, se encapsula con la mitad de la resolución espacial en una sola imagen. Estas imágenes se muestran de manera casi simultánea de forma que el ojo humano compone una imagen con profundidad que se asemeja a la visión binocular natural. Desde hace un par de años las principales plataformas de televisión han empezado a crear canales con contenido 3D. La televisión 3D (TV3D) se ha introducido en los hogares gracias a los televisores estereoscópicos. Estos televisores, que son compatibles con los formatos antes mencionados, extraen de cada imagen sus dos vistas, recuperan la resolución original y presentan cada vista alternativamente en la pantalla, generando al mismo tiempo una señal de sincronismo para las gafas activas, creando de esta forma la sensación tridimensional de las imágenes. En este PFC se pretende realizar el diseño VHDL de un cambiador de formato que genere en tiempo real la secuencia de imágenes correspondiente a los ojos derecho e izquierdo, con resolución completa, a partir de una secuencia codificada en formato tipo Top-and-Bottom y el banco de test para su prueba. Este circuito se implementará como un periférico del procesador NIOS II de Altera. El diseño podría utilizarse como parte de un sistema que permita la visualización de las actuales emisiones de televisión 3D en un televisor convencional. La tecnología de referencia que se utilizará serán las FPGAs, más concretamente la tarjeta Cyclone III FPGA Starter Kit (EP3C25 FPGA) de Altera, junto a una tarjeta de ampliación de Microtronix con entrada y salida HDMI para video y audio. Además se pretende crear la documentación necesaria para el desarrollo de futuros trabajos relacionados con la televisión 3D. ABSTRACT Current TV3D broadcasting uses formats as Side-by-Side or Top-and-Bottom, where every single pair of images, corresponding to left and right eyes views, are encapsulated with half spatial resolution in one single image. These images are almost simultaneously displayed so that the human eye forms an image with depth resembling naturally binocular vision. From a couple of years the major TV platforms have begun to create 3D content channels. 3D Television (3DTV) has been introduced in homes through stereoscopic televisions. These televisions, which are compatible with the above formats, each image is extracted from the two views, and recover the original resolution and displays alternately each view in screen, while generating a synchronization signal for active glasses, thereby creating the three-dimensional sensation of the images. The main objective in this PFC is to make the design of an exchanger VHDL format in real time to generate the image sequence corresponding to the right and left eyes, with full resolution from an encoded sequence type format Top-and-Bottom and test bench for testing. This circuit is implemented as a Altera NIOS II processor peripheral.The design could be used as part of a system enabling the display of current television broadcasts 3D on a conventional television. The reference technology that will be use are FPGAs, more specifically Cyclone III FPGA Starter Card Kit (EP3C25 FPGA) Altera, along with an expansion card Microtronix with HDMI input and output video and audio. It also aims to create documentation for the development of future works related to 3D TV.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In this letter , we report a new method for óptical switching based on the electro-optical properties of liquid crystal materials and, in particular, of the nematic type. The basis of this new method is the use of twisted wedge structure that has not been reported before elsewhere. In the past several years , great efforts in integrated optics have been made to develop optical switching devices with fast speed by using electro-optic, acousto-optic or magneto -optic materials. A mechanically operated óptical switch made of grade-index rod 1enses and e1ectromagnets has been proposed too . Switches of this kind include one input and two output waveguides and, depending on the app1ied voltage, one incident light on the switch exits either in one or another of the two output waveguides.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

La informática teórica es una disciplina básica ya que la mayoría de los avances en informática se sustentan en un sólido resultado de esa materia. En los últimos a~nos debido tanto al incremento de la potencia de los ordenadores, como a la cercanía del límite físico en la miniaturización de los componentes electrónicos, resurge el interés por modelos formales de computación alternativos a la arquitectura clásica de von Neumann. Muchos de estos modelos se inspiran en la forma en la que la naturaleza resuelve eficientemente problemas muy complejos. La mayoría son computacionalmente completos e intrínsecamente paralelos. Por este motivo se les está llegando a considerar como nuevos paradigmas de computación (computación natural). Se dispone, por tanto, de un abanico de arquitecturas abstractas tan potentes como los computadores convencionales y, a veces, más eficientes: alguna de ellas mejora el rendimiento, al menos temporal, de problemas NPcompletos proporcionando costes no exponenciales. La representación formal de las redes de procesadores evolutivos requiere de construcciones, tanto independientes, como dependientes del contexto, dicho de otro modo, en general una representación formal completa de un NEP implica restricciones, tanto sintácticas, como semánticas, es decir, que muchas representaciones aparentemente (sintácticamente) correctas de casos particulares de estos dispositivos no tendrían sentido porque podrían no cumplir otras restricciones semánticas. La aplicación de evolución gramatical semántica a los NEPs pasa por la elección de un subconjunto de ellos entre los que buscar los que solucionen un problema concreto. En este trabajo se ha realizado un estudio sobre un modelo inspirado en la biología celular denominado redes de procesadores evolutivos [55, 53], esto es, redes cuyos nodos son procesadores muy simples capaces de realizar únicamente un tipo de mutación puntual (inserción, borrado o sustitución de un símbolo). Estos nodos están asociados con un filtro que está definido por alguna condición de contexto aleatorio o de pertenencia. Las redes están formadas a lo sumo de seis nodos y, teniendo los filtros definidos por una pertenencia a lenguajes regulares, son capaces de generar todos los lenguajes enumerables recursivos independientemente del grafo subyacente. Este resultado no es sorprendente ya que semejantes resultados han sido documentados en la literatura. Si se consideran redes con nodos y filtros definidos por contextos aleatorios {que parecen estar más cerca a las implementaciones biológicas{ entonces se pueden generar lenguajes más complejos como los lenguajes no independientes del contexto. Sin embargo, estos mecanismos tan simples son capaces de resolver problemas complejos en tiempo polinomial. Se ha presentado una solución lineal para un problema NP-completo, el problema de los 3-colores. Como primer aporte significativo se ha propuesto una nueva dinámica de las redes de procesadores evolutivos con un comportamiento no determinista y masivamente paralelo [55], y por tanto todo el trabajo de investigación en el área de la redes de procesadores se puede trasladar a las redes masivamente paralelas. Por ejemplo, las redes masivamente paralelas se pueden modificar de acuerdo a determinadas reglas para mover los filtros hacia las conexiones. Cada conexión se ve como un canal bidireccional de manera que los filtros de entrada y salida coinciden. A pesar de esto, estas redes son computacionalmente completas. Se pueden también implementar otro tipo de reglas para extender este modelo computacional. Se reemplazan las mutaciones puntuales asociadas a cada nodo por la operación de splicing. Este nuevo tipo de procesador se denomina procesador splicing. Este modelo computacional de Red de procesadores con splicing ANSP es semejante en cierto modo a los sistemas distribuidos en tubos de ensayo basados en splicing. Además, se ha definido un nuevo modelo [56] {Redes de procesadores evolutivos con filtros en las conexiones{ , en el cual los procesadores tan solo tienen reglas y los filtros se han trasladado a las conexiones. Dicho modelo es equivalente, bajo determinadas circunstancias, a las redes de procesadores evolutivos clásicas. Sin dichas restricciones el modelo propuesto es un superconjunto de los NEPs clásicos. La principal ventaja de mover los filtros a las conexiones radica en la simplicidad de la modelización. Otras aportaciones de este trabajo ha sido el dise~no de un simulador en Java [54, 52] para las redes de procesadores evolutivos propuestas en esta Tesis. Sobre el término "procesador evolutivo" empleado en esta Tesis, el proceso computacional descrito aquí no es exactamente un proceso evolutivo en el sentido Darwiniano. Pero las operaciones de reescritura que se han considerado pueden interpretarse como mutaciones y los procesos de filtrado se podrían ver como procesos de selección. Además, este trabajo no abarca la posible implementación biológica de estas redes, a pesar de ser de gran importancia. A lo largo de esta tesis se ha tomado como definición de la medida de complejidad para los ANSP, una que denotaremos como tama~no (considerando tama~no como el número de nodos del grafo subyacente). Se ha mostrado que cualquier lenguaje enumerable recursivo L puede ser aceptado por un ANSP en el cual el número de procesadores está linealmente acotado por la cardinalidad del alfabeto de la cinta de una máquina de Turing que reconoce dicho lenguaje L. Siguiendo el concepto de ANSP universales introducido por Manea [65], se ha demostrado que un ANSP con una estructura de grafo fija puede aceptar cualquier lenguaje enumerable recursivo. Un ANSP se puede considerar como un ente capaz de resolver problemas, además de tener otra propiedad relevante desde el punto de vista práctico: Se puede definir un ANSP universal como una subred, donde solo una cantidad limitada de parámetros es dependiente del lenguaje. La anterior característica se puede interpretar como un método para resolver cualquier problema NP en tiempo polinomial empleando un ANSP de tama~no constante, concretamente treinta y uno. Esto significa que la solución de cualquier problema NP es uniforme en el sentido de que la red, exceptuando la subred universal, se puede ver como un programa; adaptándolo a la instancia del problema a resolver, se escogerín los filtros y las reglas que no pertenecen a la subred universal. Un problema interesante desde nuestro punto de vista es el que hace referencia a como elegir el tama~no optimo de esta red.---ABSTRACT---This thesis deals with the recent research works in the area of Natural Computing {bio-inspired models{, more precisely Networks of Evolutionary Processors first developed by Victor Mitrana and they are based on P Systems whose father is Georghe Paun. In these models, they are a set of processors connected in an underlying undirected graph, such processors have an object multiset (strings) and a set of rules, named evolution rules, that transform objects inside processors[55, 53],. These objects can be sent/received using graph connections provided they accomplish constraints defined at input and output filters processors have. This symbolic model, non deterministic one (processors are not synchronized) and massive parallel one[55] (all rules can be applied in one computational step) has some important properties regarding solution of NP-problems in lineal time and of course, lineal resources. There are a great number of variants such as hybrid networks, splicing processors, etc. that provide the model a computational power equivalent to Turing machines. The origin of networks of evolutionary processors (NEP for short) is a basic architecture for parallel and distributed symbolic processing, related to the Connection Machine as well as the Logic Flow paradigm, which consists of several processors, each of them being placed in a node of a virtual complete graph, which are able to handle data associated with the respective node. All the nodes send simultaneously their data and the receiving nodes handle also simultaneously all the arriving messages, according to some strategies. In a series of papers one considers that each node may be viewed as a cell having genetic information encoded in DNA sequences which may evolve by local evolutionary events, that is point mutations. Each node is specialized just for one of these evolutionary operations. Furthermore, the data in each node is organized in the form of multisets of words (each word appears in an arbitrarily large number of copies), and all the copies are processed in parallel such that all the possible events that can take place do actually take place. Obviously, the computational process just described is not exactly an evolutionary process in the Darwinian sense. But the rewriting operations we have considered might be interpreted as mutations and the filtering process might be viewed as a selection process. Recombination is missing but it was asserted that evolutionary and functional relationships between genes can be captured by taking only local mutations into consideration. It is clear that filters associated with each node allow a strong control of the computation. Indeed, every node has an input and output filter; two nodes can exchange data if it passes the output filter of the sender and the input filter of the receiver. Moreover, if some data is sent out by some node and not able to enter any node, then it is lost. In this paper we simplify the ANSP model considered in by moving the filters from the nodes to the edges. Each edge is viewed as a two-way channel such that the input and output filters coincide. Clearly, the possibility of controlling the computation in such networks seems to be diminished. For instance, there is no possibility to loose data during the communication steps. In spite of this and of the fact that splicing is not a powerful operation (remember that splicing systems generates only regular languages) we prove here that these devices are computationally complete. As a consequence, we propose characterizations of two complexity classes, namely NP and PSPACE, in terms of accepting networks of restricted splicing processors with filtered connections. We proposed a uniform linear time solution to SAT based on ANSPFCs with linearly bounded resources. This solution should be understood correctly: we do not solve SAT in linear time and space. Since any word and auxiliary word appears in an arbitrarily large number of copies, one can generate in linear time, by parallelism and communication, an exponential number of words each of them having an exponential number of copies. However, this does not seem to be a major drawback since by PCR (Polymerase Chain Reaction) one can generate an exponential number of identical DNA molecules in a linear number of reactions. It is worth mentioning that the ANSPFC constructed above remains unchanged for any instance with the same number of variables. Therefore, the solution is uniform in the sense that the network, excepting the input and output nodes, may be viewed as a program according to the number of variables, we choose the filters, the splicing words and the rules, then we assign all possible values to the variables, and compute the formula.We proved that ANSP are computationally complete. Do the ANSPFC remain still computationally complete? If this is not the case, what other problems can be eficiently solved by these ANSPFCs? Moreover, the complexity class NP is exactly the class of all languages decided by ANSP in polynomial time. Can NP be characterized in a similar way with ANSPFCs?