920 resultados para Computer input-output equipment.


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Model-based calibration of steady-state engine operation is commonly performed with highly parameterized empirical models that are accurate but not very robust, particularly when predicting highly nonlinear responses such as diesel smoke emissions. To address this problem, and to boost the accuracy of more robust non-parametric methods to the same level, GT-Power was used to transform the empirical model input space into multiple input spaces that simplified the input-output relationship and improved the accuracy and robustness of smoke predictions made by three commonly used empirical modeling methods: Multivariate Regression, Neural Networks and the k-Nearest Neighbor method. The availability of multiple input spaces allowed the development of two committee techniques: a 'Simple Committee' technique that used averaged predictions from a set of 10 pre-selected input spaces chosen by the training data and the "Minimum Variance Committee" technique where the input spaces for each prediction were chosen on the basis of disagreement between the three modeling methods. This latter technique equalized the performance of the three modeling methods. The successively increasing improvements resulting from the use of a single best transformed input space (Best Combination Technique), Simple Committee Technique and Minimum Variance Committee Technique were verified with hypothesis testing. The transformed input spaces were also shown to improve outlier detection and to improve k-Nearest Neighbor performance when predicting dynamic emissions with steady-state training data. An unexpected finding was that the benefits of input space transformation were unaffected by changes in the hardware or the calibration of the underlying GT-Power model.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Embedded siloxane polymer waveguides have shown promising results for use in optical backplanes. They exhibit high temperature stability, low optical absorption, and require common processing techniques. A challenging aspect of this technology is out-of-plane coupling of the waveguides. A multi-software approach to modeling an optical vertical interconnect (via) is proposed. This approach utilizes the beam propagation method to generate varied modal field distribution structures which are then propagated through a via model using the angular spectrum propagation technique. Simulation results show average losses between 2.5 and 4.5 dB for different initial input conditions. Certain configurations show losses of less than 3 dB and it is shown that in an input/output pair of vias, average losses per via may be lower than the targeted 3 dB.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper describes the open source framework MARVIN for rapid application development in the field of biomedical and clinical research. MARVIN applications consist of modules that can be plugged together in order to provide the functionality required for a specific experimental scenario. Application modules work on a common patient database that is used to store and organize medical data as well as derived data. MARVIN provides a flexible input/output system with support for many file formats including DICOM, various 2D image formats and surface mesh data. Furthermore, it implements an advanced visualization system and interfaces to a wide range of 3D tracking hardware. Since it uses only highly portable libraries, MARVIN applications run on Unix/Linux, Mac OS X and Microsoft Windows.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

It is a challenge to measure the impact of releasing data to the public since the effects may not be directly linked to particular open data activities or substantial impact may only occur several years after publishing the data. This paper proposes a framework to assess the impact of releasing open data by applying the Social Return on Investment (SROI) approach. SROI was developed for organizations intended to generate social and environmental benefits thus fitting the purpose of most open data initiatives. We link the four steps of SROI (input, output, outcome, impact) with the 14 high-value data categories of the G8 Open Data Charter to create a matrix of open data examples, activities, and impacts in each of the data categories. This Impact Monitoring Framework helps data providers to navigate the impact space of open data laying out the conceptual basis for further research.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This study focuses on the technological intensity of China's exports. It first introduces the method of decomposing gross exports by using the Asian international input–output tables. The empirical results indicate that the technological intensity of Chinese exports has been significantly overestimated due to its high dependency on import content, especially in high-technology exports, an area highly dominated by the electronic and electrical equipment sector. Furthermore, a significant portion of value added embodied in China's high-technology exports comes from services and high-technology manufacturers in neighboring economies, such as Japan, South Korea, and Taiwan.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The concept and logic of the "smile curve" in the context of global value chains has been widely used and discussed at the individual firm level, but rarely identified and investigated at the country and industry levels by using real data. This paper proposes an idea, based on an inter-country input-output model, to consistently measure both the strength and length of linkages between producers and consumers along global value chains. This idea allows for better identification and mapping of smile curves for countries and industries according to their positions and degrees of participation in a given conceptual value chain. Using the 1995-2011 World Input-Output Tables, several conceptual value chains are investigated, including exports of electrical and optical equipment from China and Mexico and exports of automobiles from Japan and Germany. The identified smile curves provide a very intuitive and visual image, which can significantly improve our understanding of the roles played by different countries and industries in global value chains. Further, the smile curves help identify the benefits gained by these countries and industries through their participation in global trade.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This article presents a new and computationally efficient method of analysis of a railway track modelled as a continuous beam of 2N spans supported by elastic vertical springs. The main feature of this method is its important reduction in computational effort with respect to standard matrix methods of structural analysis. In this article, the whole structure is considered to be a repetition of a single one. The analysis presented is applied to a simple railway track model, i.e. to a repetitive beam supported on vertical springs (sleepers). The proposed method of analysis is based on the general theory of spatially periodic structures. The main feature of this theory is the possibility to apply Discrete Fourier Transform (DFT) in order to reduce a large system of q(2N + 1) linear stiffness equilibrium equations to a set of 2N + 1 uncoupled systems of q equations each. In this way, a dramatic reduction of the computational effort of solving the large system of equations is achieved. This fact is particularly important in the analysis of railway track structures, in which N is a very large number (around several thousands), and q = 2, the vertical displacement and rotation, is very small. The proposed method allows us to easily obtain the exact solution given by Samartín [1], i.e. the continuous beam railway track response. The comparison between the proposed method and other methods of analysis of railway tracks, such as Lorente de Nó and Zimmermann-Timoshenko, clearly shows the accuracy of the obtained results for the proposed method, even for low values of N. In addition, identical results between the proposed and the Lorente methods have been found, although the proposed method seems to be of simpler application and computationally more efficient than the Lorente one. Small but significative differences occur between these two methods and the one developed by Zimmermann-Timoshenko. This article also presents a detailed sensitivity analysis of the vertical displacement of the sleepers. Although standard matrix methods of structural analysis can handle this railway model, one of the objectives of this article is to show the efficiency of DFT method with respect to standard matrix structural analysis. A comparative analysis between standard matrix structural analysis and the proposed method (DFT), in terms of computational time, input, output and also software programming, will be carried out. Finally, a URL link to a MatLab computer program list, based on the proposed method, is given

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Since the advent of the computer into the engineering field, the application of the numerical methods to the solution of engineering problems has grown very rapidly. Among the different computer methods of structural analysis the Finite Element (FEM) has been predominantly used. Shells and space structures are very attractive and have been constructed to solve a large variety of functional problems (roofs, industrial building, aqueducts, reservoirs, footings etc). In this type of structures aesthetics, structural efficiency and concept play a very important role. This class of structures can be divided into three main groups, namely continuous (concrete) shells, space frames and tension (fabric, pneumatic, cable etc )structures. In the following only the current applications of the FEM to the analysis of continuous shell structures will be discussed. However, some of the comments on this class of shells can be also applied to some extend to the others, but obviously specific computational problems will be restricted to the continuous shells. Different aspects, such as, the type of elements,input-output computational techniques etc, of the analysis of shells by the FEM will be described below. Clearly, the improvements and developments occurring in general for the FEM since its first appearance in the fifties have had a significative impact on the particular class of structures under discussion.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

La presente Tesis analiza y desarrolla metodología específica que permite la caracterización de sistemas de transmisión acústicos basados en el fenómeno del array paramétrico. Este tipo de estructuras es considerado como uno de los sistemas más representativos de la acústica no lineal con amplias posibilidades tecnológicas. Los arrays paramétricos aprovechan la no linealidad del medio aéreo para obtener en recepción señales en el margen sónico a partir de señales ultrasónicas en emisión. Por desgracia, este procedimiento implica que la señal transmitida y la recibida guardan una relación compleja, que incluye una fuerte ecualización así como una distorsión apreciable por el oyente. Este hecho reduce claramente la posibilidad de obtener sistemas acústicos de gran fidelidad. Hasta ahora, los esfuerzos tecnológicos dirigidos al diseño de sistemas comerciales han tratado de paliar esta falta de fidelidad mediante técnicas de preprocesado fuertemente dependientes de los modelos físicos teóricos. Estos están basados en la ecuación de propagación de onda no lineal. En esta Tesis se propone un nuevo enfoque: la obtención de una representación completa del sistema mediante series de Volterra que permita inferir un sistema de compensación computacionalmente ligero y fiable. La dificultad que entraña la correcta extracción de esta representación obliga a desarrollar una metodología completa de identificación adaptada a este tipo de estructuras. Así, a la hora de aplicar métodos de identificación se hace indispensable la determinación de ciertas características iniciales que favorezcan la parametrización del sistema. En esta Tesis se propone una metodología propia que extrae estas condiciones iniciales. Con estos datos, nos encontramos en disposición de plantear un sistema completo de identificación no lineal basado en señales pseudoaleatorias, que aumenta la fiabilidad de la descripción del sistema, posibilitando tanto la inferencia de la estructura basada en bloques subyacente, como el diseño de mecanismos de compensación adecuados. A su vez, en este escenario concreto en el que intervienen procesos de modulación, factores como el punto de trabajo o las características físicas del transductor, hacen inviables los algoritmos de caracterización habituales. Incluyendo el método de identificación propuesto. Con el fin de eliminar esta problemática se propone una serie de nuevos algoritmos de corrección que permiten la aplicación de la caracterización. Las capacidades de estos nuevos algoritmos se pondrán a prueba sobre un prototipo físico, diseñado a tal efecto. Para ello, se propondrán la metodología y los mecanismos de instrumentación necesarios para llevar a cabo el diseño, la identificación del sistema y su posible corrección, todo ello mediante técnicas de procesado digital previas al sistema de transducción. Los algoritmos se evaluarán en términos de error de modelado a partir de la señal de salida del sistema real frente a la salida sintetizada a partir del modelo estimado. Esta estrategia asegura la posibilidad de aplicar técnicas de compensación ya que éstas son sensibles a errores de estima en módulo y fase. La calidad del sistema final se evaluará en términos de fase, coloración y distorsión no lineal mediante un test propuesto a lo largo de este discurso, como paso previo a una futura evaluación subjetiva. ABSTRACT This Thesis presents a specific methodology for the characterization of acoustic transmission systems based on the parametric array phenomenon. These structures are well-known representatives of the nonlinear acoustics field and display large technological opportunities. Parametric arrays exploit the nonlinear behavior of air to obtain sonic signals at the receptors’side, which were generated within the ultrasonic range. The underlying physical process redunds in a complex relationship between the transmitted and received signals. This includes both a strong equalization and an appreciable distortion for a human listener. High fidelity, acoustic equipment based on this phenomenon is therefore difficult to design. Until recently, efforts devoted to this enterprise have focused in fidelity enhancement based on physically-informed, pre-processing schemes. These derive directly from the nonlinear form of the wave equation. However, online limited enhancement has been achieved. In this Thesis we propose a novel approach: the evaluation of a complete representation of the system through its projection onto the Volterra series, which allows the posterior inference of a computationally light and reliable compensation scheme. The main difficulty in the derivation of such representation strives from the need of a complete identification methodology, suitable for this particular type of structures. As an example, whenever identification techniques are involved, we require preliminary estimates on certain parameters that contribute to the correct parameterization of the system. In this Thesis we propose a methodology to derive such initial values from simple measures. Once these information is made available, a complete identification scheme is required for nonlinear systems based on pseudorandom signals. These contribute to the robustness and fidelity of the resulting model, and facilitate both the inference of the underlying structure, which we subdivide into a simple block-oriented construction, and the design of the corresponding compensation structure. In a scenario such as this where frequency modulations occur, one must control exogenous factors such as devices’ operation point and the physical properties of the transducer. These may conflict with the principia behind the standard identification procedures, as it is the case. With this idea in mind, the Thesis includes a series of novel correction algorithms that facilitate the application of the characterization results onto the system compensation. The proposed algorithms are tested on a prototype that was designed and built for this purpose. The methodology and instrumentation required for its design, the identification of the overall acoustic system and its correction are all based on signal processing techniques, focusing on the system front-end, i.e. prior to transduction. Results are evaluated in terms of input-output modelling error, considering a synthetic construction of the system. This criterion ensures that compensation techniques may actually be introduced, since these are highly sensible to estimation errors both on the envelope and the phase of the signals involved. Finally, the quality of the overall system will be evaluated in terms of phase, spectral color and nonlinear distortion; by means of a test protocol specifically devised for this Thesis, as a prior step for a future, subjective quality evaluation.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

As the telecommunications industry evolves over the next decade to provide the products and services that people will desire, several key technologies will become commonplace. Two of these, automatic speech recognition and text-to-speech synthesis, will provide users with more freedom on when, where, and how they access information. While these technologies are currently in their infancy, their capabilities are rapidly increasing and their deployment in today's telephone network is expanding. The economic impact of just one application, the automation of operator services, is well over $100 million per year. Yet there still are many technical challenges that must be resolved before these technologies can be deployed ubiquitously in products and services throughout the worldwide telephone network. These challenges include: (i) High level of accuracy. The technology must be perceived by the user as highly accurate, robust, and reliable. (ii) Easy to use. Speech is only one of several possible input/output modalities for conveying information between a human and a machine, much like a computer terminal or Touch-Tone pad on a telephone. It is not the final product. Therefore, speech technologies must be hidden from the user. That is, the burden of using the technology must be on the technology itself. (iii) Quick prototyping and development of new products and services. The technology must support the creation of new products and services based on speech in an efficient and timely fashion. In this paper I present a vision of the voice-processing industry with a focus on the areas with the broadest base of user penetration: speech recognition, text-to-speech synthesis, natural language processing, and speaker recognition technologies. The current and future applications of these technologies in the telecommunications industry will be examined in terms of their strengths, limitations, and the degree to which user needs have been or have yet to be met. Although noteworthy gains have been made in areas with potentially small user bases and in the more mature speech-coding technologies, these subjects are outside the scope of this paper.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Federal Railroad Administration, Office of Safety, Washington, D.C.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

"C00-2118-0005" (v. 1)

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Second language acquisition is a field that has fascinated linguists for numerous years and is a topic that is very much connected to how English teachers in Sweden try to teach the English language to the students in their classrooms. In 2009 Sundqvist examined what possible effects extramural English could have on learners' oral proficiency and their vocabulary. In her study she found out that extramural English “is an independent variable and a possible path to progress in English” (Sundqvist, 2009, p. i).  In 2014, three Swedish secondary- and upper secondary school teachers started a project for the Erasmus+. These three teachers tried to create better teaching conditions and to come up with new methods for teaching English. During their investigation they noticed that students who had only been in Sweden for four years or less, seemed to get less exposed to English in their spare time than native Swedish students, which created a disadvantage for them. Since the time when these two studies were carried out, the number of immigrants has increased drastically, which creates the need for further investigation within this area of second language acquisition. In this study, I therefore investigate how much and in what way students come in contact with the English language outside of school. I also examine if there are any differences between native Swedish students versus non-native Swedish students and if so, how this might affect the students and their grades in English. The study was conducted through the use of questionnaires and through observations of different teaching situations, including the participating teachers' methods and the participating students' reactions. The results show that there are differences between native- and non-native students when it comes to extramural English activities. The results also show that these differences seem to affect the students' grades in English, in favour of the native Swedish students. The native students tend to spend more time on extramural English activities, especially in connection to the Internet and computer games, than the non-native students. These results indicate that something needs to be done in order to compensate for the non-native students' disadvantage.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The Raf-MEK-ERK MAP kinase cascade transmits signals from activated receptors into the cell to regulate proliferation and differentiation. The cascade is controlled by the Ras GTPase, which recruits Raf from the cytosol to the plasma membrane for activation. In turn, MEK, ERK, and scaffold proteins translocate to the plasma membrane for activation. Here, we examine the input-output properties of the Raf-MEK-ERK MAP kinase module in mammalian cells activated in different cellular contexts. We show that the MAP kinase module operates as a molecular switch in vivo but that the input sensitivity of the module is determined by subcellular location. Signal output from the module is sensitive to low-level input only when it is activated at the plasma membrane. This is because the threshold for activation is low at the plasma membrane, whereas the threshold for activation is high in the cytosol. Thus, the circuit configuration of the module at the plasma membrane generates maximal outputs from low-level analog inputs, allowing cells to process and respond appropriately to physiological stimuli. These results reveal the engineering logic behind the recruitment of elements of the module from the cytosol to the membrane for activation.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Information security devices must preserve security properties even in the presence of faults. This in turn requires a rigorous evaluation of the system behaviours resulting from component failures, especially how such failures affect information flow. We introduce a compositional method of static analysis for fail-secure behaviour. Our method uses reachability matrices to identify potentially undesirable information flows based on the fault modes of the system's components.