915 resultados para Engineering design--Data processing
Resumo:
This paper presents a method for transforming the information of an engineering geological map into useful information for non-specialists involved in land-use planning. The method consists of classifying the engineering geological units in terms of land use capability and identifying the legal and the geologic restrictions that apply in the study area. Both informations are then superimposed over the land use and a conflict areas map is created. The analysis of these data leads to the identification of existing and forthcoming land use conflicts and enables the proposal of planning measures on a regional and local scale. The map for the regional planning was compiled at a 1:50,000 scale and encompasses the whole municipal land area where uses are mainly rural. The map for the local planning was compiled at a 1:10,000 scale and encompasses the urban area. Most of the classification and operations on maps used spatial analyst tools available in the Geographical Information System. The regional studies showed that the greater part of Analandia's territory presents appropriate land uses. The local-scale studies indicate that the majority of the densely occupied urban areas are in suitable land. Although the situation is in general positive, municipal policies should address the identified and expected land use conflicts, so that it can be further improved.
Resumo:
This paper proposes two new approaches for the sensitivity analysis of multiobjective design optimization problems whose performance functions are highly susceptible to small variations in the design variables and/or design environment parameters. In both methods, the less sensitive design alternatives are preferred over others during the multiobjective optimization process. While taking the first approach, the designer chooses the design variable and/or parameter that causes uncertainties. The designer then associates a robustness index with each design alternative and adds each index as an objective function in the optimization problem. For the second approach, the designer must know, a priori, the interval of variation in the design variables or in the design environment parameters, because the designer will be accepting the interval of variation in the objective functions. The second method does not require any law of probability distribution of uncontrollable variations. Finally, the authors give two illustrative examples to highlight the contributions of the paper.
Resumo:
The curriculum of the Bucknell University Chemical Engineering Department includes a required senior year capstone course titled Process Engineering, with an emphasis on process design. For the past ten years library research has been a significant component of the coursework, and students working in teams meet with the librarian throughout the semester to explore a wide variety of information resources required for their project. The assignment has been the same from 1989 to 1999. Teams of students are responsible for designing a safe, efficient, and profitable process for the dehydrogenation of ethylbenzene to styrene monomer. A series of written reports on their chosen process design is a significant course outcome. While the assignment and the specific chemical technology have not changed radically in the past decade, the process of research and discovery has evolved considerably. This paper describes the solutions offered in 1989 to meet the information needs of the chemical engineering students at Bucknell University, and the evolution in research brought about by online databases, electronic journals, and the Internet, making the process of discovery a completely different experience in 1999.
Resumo:
What motivates students to perform and pursue engineering design tasks? This study examines this question by way of three Learning Through Service (LTS) programs: 1) an on-going longitudinal study examining the impacts of service on engineering students, 2) an on-going analysis of an international senior design capstone program, and 3) an on-going evaluation of an international graduate-level research program. The evaluation of these programs incorporates both qualitative and quantitative methods, utilizing surveys, questionnaires, and interviews, which help to provide insight on what motivates students to do engineering design work. The quantitative methods were utilized in analyzing various instruments including: a Readiness assessment inventory, Intercultural Development Inventory, Sustainable Engineering through Service Learning survey, the Impacts of Service on Engineering Students’ survey, Motivational narratives, as well as some analysis for interview text. The results of these instruments help to provide some much needed insight on how prepared students are to participate in engineering programs. Additional qualitative methods include: Word clouds, Motivational narratives, as well as interview analysis. This thesis focused on how these instruments help to determine what motivates engineering students to pursue engineering design tasks. These instruments aim to collect some more in-depth information than the quantitative instruments will allow. Preliminary results suggest that of the 120 interviews analyzed Interest/Enjoyment, Application of knowledge and skills, as well as gaining knowledge are key motivating factors regardless of gender or academic level. Together these findings begin to shed light on what motivates students to perform engineering design tasks, which can be applied for better recruitment and retention in university programs.
Resumo:
Methodological evaluation of the proteomic analysis of cardiovascular-tissue material has been performed with a special emphasis on establishing examinations that allow reliable quantitative analysis of silver-stained readouts. Reliability, reproducibility, robustness and linearity were addressed and clarified. In addition, several types of normalization procedures were evaluated and new approaches are proposed. It has been found that the silver-stained readout offers a convenient approach for quantitation if a linear range for gel loading is defined. In addition, a broad range of a 10-fold input (loading 20-200 microg per gel) fulfills the linearity criteria, although at the lowest input (20 microg) a portion of protein species will remain undetected. The method is reliable and reproducible within a range of 65-200 microg input. The normalization procedure using the sum of all spot intensities from a silver-stained 2D pattern has been shown to be less reliable than other approaches, namely, normalization through median or through involvement of interquartile range. A special refinement of the normalization through virtual segmentation of pattern, and calculation of normalization factor for each stratum provides highly satisfactory results. The presented results not only provide evidence for the usefulness of silver-stained gels for quantitative evaluation, but they are directly applicable to the research endeavor of monitoring alterations in cardiovascular pathophysiology.
Resumo:
This paper shows the actual state of a compilation work on Thermal Control Design Data being done at Madrid (Lamf-ETSIA) under several ESTEC contracts, introducing a Handbook already issued, its additions and updatings.
Resumo:
PAMELA (Phased Array Monitoring for Enhanced Life Assessment) SHMTM System is an integrated embedded ultrasonic guided waves based system consisting of several electronic devices and one system manager controller. The data collected by all PAMELA devices in the system must be transmitted to the controller, who will be responsible for carrying out the advanced signal processing to obtain SHM maps. PAMELA devices consist of hardware based on a Virtex 5 FPGA with a PowerPC 440 running an embedded Linux distribution. Therefore, PAMELA devices, in addition to the capability of performing tests and transmitting the collected data to the controller, have the capability of perform local data processing or pre-processing (reduction, normalization, pattern recognition, feature extraction, etc.). Local data processing decreases the data traffic over the network and allows CPU load of the external computer to be reduced. Even it is possible that PAMELA devices are running autonomously performing scheduled tests, and only communicates with the controller in case of detection of structural damages or when programmed. Each PAMELA device integrates a software management application (SMA) that allows to the developer downloading his own algorithm code and adding the new data processing algorithm to the device. The development of the SMA is done in a virtual machine with an Ubuntu Linux distribution including all necessary software tools to perform the entire cycle of development. Eclipse IDE (Integrated Development Environment) is used to develop the SMA project and to write the code of each data processing algorithm. This paper presents the developed software architecture and describes the necessary steps to add new data processing algorithms to SMA in order to increase the processing capabilities of PAMELA devices.An example of basic damage index estimation using delay and sum algorithm is provided.
Resumo:
Nowadays, devices that monitor the health of structures consume a lot of power and need a lot of time to acquire, process, and send the information about the structure to the main processing unit. To decrease this time, fast electronic devices are starting to be used to accelerate this processing. In this paper some hardware algorithms implemented in an electronic logic programming device are described. The goal of this implementation is accelerate the process and diminish the information that has to be send. By reaching this goal, the time the processor needs for treating all the information is reduced and so the power consumption is reduced too.
Resumo:
We describe the use of singular value decomposition in transforming genome-wide expression data from genes × arrays space to reduced diagonalized “eigengenes” × “eigenarrays” space, where the eigengenes (or eigenarrays) are unique orthonormal superpositions of the genes (or arrays). Normalizing the data by filtering out the eigengenes (and eigenarrays) that are inferred to represent noise or experimental artifacts enables meaningful comparison of the expression of different genes across different arrays in different experiments. Sorting the data according to the eigengenes and eigenarrays gives a global picture of the dynamics of gene expression, in which individual genes and arrays appear to be classified into groups of similar regulation and function, or similar cellular state and biological phenotype, respectively. After normalization and sorting, the significant eigengenes and eigenarrays can be associated with observed genome-wide effects of regulators, or with measured samples, in which these regulators are overactive or underactive, respectively.
Resumo:
This report sheds light on the fundamental questions and underlying tensions between current policy objectives, compliance strategies and global trends in online personal data processing, assessing the existing and future framework in terms of effective regulation and public policy. Based on the discussions among the members of the CEPS Digital Forum and independent research carried out by the rapporteurs, policy conclusions are derived with the aim of making EU data protection policy more fit for purpose in today’s online technological context. This report constructively engages with the EU data protection framework, but does not provide a textual analysis of the EU data protection reform proposal as such.