864 resultados para Electronic alarm systems
Resumo:
The electronics industry is developing rapidly together with the increasingly complex problem of microelectronic equipment cooling. It has now become necessary for thermal design engineers to consider the problem of equipment cooling at some level. The use of Computational Fluid Dynamics (CFD) for such investigations is fast becoming a powerful and almost essential tool for the design, development and optimisation of engineering applications. However turbulence models remain a key issue when tackling such flow phenomena. The reliability of CFD analysis depends heavily on the turbulence model employed together with the wall functions implemented. In order to resolve the abrupt fluctuations experienced by the turbulent energy and other parameters located at near wall regions and shear layers a particularly fine computational mesh is necessary which inevitably increases the computer storage and run-time requirements. This paper will discuss results from an investigation into the accuract of currently used turbulence models. Also a newly formulated transitional hybrid turbulence model will be introduced with comparisonsaagainst experimental data.
Resumo:
A major percentage of the heat emitted from electronic packages can be extracted by air cooling whether by means of natural or forced convection. This flow of air throughout an electronic system and the heat extracted is highly dependable on the nature of turbulence present in the flow field. This paper will discuss results from an investigation into the accuracy of turbulence models to predict air cooling for electronic packages and systems.
Resumo:
This paper discusses an optimisation based decision support system and methodology for electronic packaging and product design and development which is capable of addressing in efficient manner specified environmental, reliability and cost requirements. A study which focuses on the design of a flip-chip package is presented. Different alternatives for the design of the flip-chip package are considered based on existing options for the applied underfill and volume of solder material used to form the interconnects. Variations in these design input parameters have simultaneous effect on package aspects such as cost, environmental impact and reliability. A decision system for the design of the flip-chip that uses numerical optimisation approach is used to identify the package optimal specification which satisfies the imposed requirements. The reliability aspect of interest is the fatigue of solder joints under thermal cycling. Transient nonlinear finite element analysis (FEA) is used to simulate the thermal fatigue damage in solder joints subject to thermal cycling. Simulation results are manipulated within design of experiments and response surface modelling framework to provide numerical model for reliability which can be used to quantify the package reliability. Assessment of the environmental impact of the package materials is performed by using so called Toxic Index (TI). In this paper we demonstrate the evaluation of the environmental impact only for underfill and lead-free solder materials. This evaluation is based on the amount of material per flip-chip package. Cost is the dominant factor in contemporary flip-chip packaging industry. In the optimisation based decision support system for the design of the flip-chip package, cost of materials which varies as a result of variations in the design parameters is considered.
Resumo:
Embedded electronic systems in vehicles are of rapidly increasing commercial importance for the automotive industry. While current vehicular embedded systems are extremely limited and static, a more dynamic configurable system would greatly simplify the integration work and increase quality of vehicular systems. This brings in features like separation of concerns, customised software configuration for individual vehicles, seamless connectivity, and plug-and-play capability. Furthermore, such a system can also contribute to increased dependability and resource optimization due to its inherent ability to adjust itself dynamically to changes in software, hardware resources, and environment condition. This paper describes the architectural approach to achieving the goals of dynamically self-configuring automotive embedded electronic systems by the EU research project DySCAS. The architecture solution outlined in this paper captures the application and operational contexts, expected features, middleware services, functions and behaviours, as well as the basic mechanisms and technologies. The paper also covers the architecture conceptualization by presenting the rationale, concerning the architecture structuring, control principles, and deployment concept. In this paper, we also present the adopted architecture V&V strategy and discuss some open issues in regards to the industrial acceptance.
Resumo:
An efficient method for calculating the electronic structure of systems that need a very fine sampling of the Brillouin zone is presented. The method is based on the variational optimization of a single (i.e., common to all points in the Brillouin zone) basis set for the expansion of the electronic orbitals. Considerations from k.p-approximation theory help to understand the efficiency of the method. The accuracy and the convergence properties of the method as a function of the optimal basis set size are analyzed for a test calculation on a 16-atom Na supercell.
Resumo:
The aim of this study is to compare the positioning accuracy at different gantry angles of two electronic portal imaging devices (EPIDs) support arm systems by using EPID difference images as a measure for displacement. This work presents a comparison of the mechanical performance of eight Varian aS500 (Varian Medical Systems, Palo Alto, CA) EPIDs, mounted using either the Varian Exact-arm or R-arm.
Resumo:
Real-space grids are a powerful alternative for the simulation of electronic systems. One of the main advantages of the approach is the flexibility and simplicity of working directly in real space where the different fields are discretized on a grid, combined with competitive numerical performance and great potential for parallelization. These properties constitute a great advantage at the time of implementing and testing new physical models. Based on our experience with the Octopus code, in this article we discuss how the real-space approach has allowed for the recent development of new ideas for the simulation of electronic systems. Among these applications are approaches to calculate response properties, modeling of photoemission, optimal control of quantum systems, simulation of plasmonic systems, and the exact solution of the Schrödinger equation for low-dimensionality systems.
Resumo:
Esta tese insere-se na área da simulação de circuitos de RF e microondas, e visa o estudo de ferramentas computacionais inovadoras que consigam simular, de forma eficiente, circuitos não lineares e muito heterogéneos, contendo uma estrutura combinada de blocos analógicos de RF e de banda base e blocos digitais, a operar em múltiplas escalas de tempo. Os métodos numéricos propostos nesta tese baseiam-se em estratégias multi-dimensionais, as quais usam múltiplas variáveis temporais definidas em domínios de tempo deformados e não deformados, para lidar, de forma eficaz, com as disparidades existentes entre as diversas escalas de tempo. De modo a poder tirar proveito dos diferentes ritmos de evolução temporal existentes entre correntes e tensões com variação muito rápida (variáveis de estado activas) e correntes e tensões com variação lenta (variáveis de estado latentes), são utilizadas algumas técnicas numéricas avançadas para operar dentro dos espaços multi-dimensionais, como, por exemplo, os algoritmos multi-ritmo de Runge-Kutta, ou o método das linhas. São também apresentadas algumas estratégias de partição dos circuitos, as quais permitem dividir um circuito em sub-circuitos de uma forma completamente automática, em função dos ritmos de evolução das suas variáveis de estado. Para problemas acentuadamente não lineares, são propostos vários métodos inovadores de simulação a operar estritamente no domínio do tempo. Para problemas com não linearidades moderadas é proposto um novo método híbrido frequência-tempo, baseado numa combinação entre a integração passo a passo unidimensional e o método seguidor de envolvente com balanço harmónico. O desempenho dos métodos é testado na simulação de alguns exemplos ilustrativos, com resultados bastante promissores. Uma análise comparativa entre os métodos agora propostos e os métodos actualmente existentes para simulação RF, revela ganhos consideráveis em termos de rapidez de computação.
Resumo:
Freshness and safety of muscle foods are generally considered as the most important parameters for the food industry. To address the rapid detection of meat spoilage microorganisms during aerobic or modified atmosphere storage, an electronic nose with the aid of fuzzy wavelet network has been considered in this research. The proposed model incorporates a clustering pre-processing stage for the definition of fuzzy rules. The dual purpose of the proposed modelling approach is not only to classify beef samples in the respective quality class (i.e. fresh, semi-fresh and spoiled), but also to predict their associated microbiological population directly from volatile compounds fingerprints. Comparison results against neural networks and neurofuzzy systems indicated that the proposed modelling scheme could be considered as a valuable detection methodology in food microbiology
Resumo:
The electronics industry is encountering thermal challenges and opportunities with lengthscales comparable to or much less than one micrometer. Examples include nanoscale phonon hotspots in transistors and the increasing temperature rise in onchip interconnects. Millimeter-scale hotspots on microprocessors, resulting from varying rates of power consumption, are being addressed using two-phase microchannel heat sinks. Nanoscale thermal data storage technology has received much attention recently. This paper provides an overview of these topics with a focus on related research at Stanford University.
Resumo:
Presentation guide for Tom and Paul - apologies for quality/lack-of-rehersal. It was only intended as a mic-check, but I got carried away.
Resumo:
Comparison of donor-acceptor electronic couplings calculated within two-state and three-state models suggests that the two-state treatment can provide unreliable estimates of Vda because of neglecting the multistate effects. We show that in most cases accurate values of the electronic coupling in a π stack, where donor and acceptor are separated by a bridging unit, can be obtained as Ṽ da = (E2 - E1) μ12 Rda + (2 E3 - E1 - E2) 2 μ13 μ23 Rda2, where E1, E2, and E3 are adiabatic energies of the ground, charge-transfer, and bridge states, respectively, μij is the transition dipole moments between the states i and j, and Rda is the distance between the planes of donor and acceptor. In this expression based on the generalized Mulliken-Hush approach, the first term corresponds to the coupling derived within a two-state model, whereas the second term is the superexchange correction accounting for the bridge effect. The formula is extended to bridges consisting of several subunits. The influence of the donor-acceptor energy mismatch on the excess charge distribution, adiabatic dipole and transition moments, and electronic couplings is examined. A diagnostic is developed to determine whether the two-state approach can be applied. Based on numerical results, we showed that the superexchange correction considerably improves estimates of the donor-acceptor coupling derived within a two-state approach. In most cases when the two-state scheme fails, the formula gives reliable results which are in good agreement (within 5%) with the data of the three-state generalized Mulliken-Hush model
Resumo:
A multi-layered architecture of self-organizing neural networks is being developed as part of an intelligent alarm processor to analyse a stream of power grid fault messages and provide a suggested diagnosis of the fault location. Feedback concerning the accuracy of the diagnosis is provided by an object-oriented grid simulator which acts as an external supervisor to the learning system. The utilization of artificial neural networks within this environment should result in a powerful generic alarm processor which will not require extensive training by a human expert to produce accurate results.
Resumo:
Introduction. Feature usage is a pre-requisite to realising the benefits of investments in feature rich systems. We propose that conceptualising the dependent variable 'system use' as 'level of use' and specifying it as a formative construct has greater value for measuring the post-adoption use of feature rich systems. We then validate the content of the construct as a first step in developing a research instrument to measure it. The context of our study is the post-adoption use of electronic medical records (EMR) by primary care physicians. Method. Initially, a literature review of the empirical context defines the scope based on prior studies. Having identified core features from the literature, they are further refined with the help of experts in a consensus seeking process that follows the Delphi technique. Results.The methodology was successfully applied to EMRs, which were selected as an example of feature rich systems. A review of EMR usage and regulatory standards provided the feature input for the first round of the Delphi process. A panel of experts then reached consensus after four rounds, identifying ten task-based features that would be indicators of level of use. Conclusions. To study why some users deploy more advanced features than others, theories of post-adoption require a rich formative dependent variable that measures level of use. We have demonstrated that a context sensitive literature review followed by refinement through a consensus seeking process is a suitable methodology to validate the content of this dependent variable. This is the first step of instrument development prior to statistical confirmation with a larger sample.