983 resultados para Experimental results


Relevância:

60.00% 60.00%

Publicador:

Resumo:

Technology plays a double role in Education: it can act as a facilitator in the teaching/learning process and it can be the very subject of that process in Science & Engineering courses. This is especially true when students perform laboratory activities where they interact with equipment and objects under experimentation. In this context, technology can also play a facilitator role if it allows students to perform experiments in a remote fashion, through the Internet, in a so-called weblab or remote laboratory. No doubt, the Internet has been revolutionizing the educational process in many aspects, and it can be stated that remote laboratories are just an angle of that on-going revolution. As any other educational tool or resource, the i) pedagogical approach and the ii) technology used in the development of a remote laboratory can dictate its general success or its ephemeral existence. By pedagogical approach we consider the way remote experiments address the process by which students acquire experimental skills and link experimental results to theoretical concepts. In respect to technology, we discuss different specification and implementation alternatives, to show the case where the adoption of a family of standards would positively contribute to a larger acceptance and utilization of remote laboratories, and also to a wider collaboration in their development.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Solvatochromic UV-Vis shifts of four indicators (4-nitroaniline, 4-nitroanisole, 4-nitrophenol and N,N-dimethy-1-4-nitro aniline) have been measured at 298.15 K in the ternary mixture methano1/1-propanol/acetonitrile (MeOH/1-PrOH/MeCN) in a total of 22 mole fractions, along with 18 additional mole fractions for each of the corresponding binary mixtures, MeOH/1-PrOH, 1-PrOH/MeCN and MeOH/MeCN. These values, combined with our previous experimental results for 2,6-dipheny1-4-(2,4,6-triphenylpyridinium-1-yl)phenolate (Reichardt's betaine dye) in the same mixtures, permitted the computation of the Kamlet-Taft solvent parameters, alpha, beta, and pi*. The rationalization of the spectroscopic behavior of each probe within each mixture's whole mole fraction range was achieved through the use of the Bosch and Roses preferential solvation model. The applied model allowed the identification of synergistic behaviors in MeCN/alcohol mixtures and thus to infer the existence of solvent complexes in solution. Also, the addition of small amounts of MeCN to the binary mixtures was seen to cause a significant variation in pi*, whereas the addition of alcohol to MeCN mixtures always lead to a sudden change in a and The behavior of these parameters in the ternary mixture was shown to be mainly determined by the contributions of the underlying binary mixtures. (C) 2014 Elsevier B.V. All rights reserved.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

On-chip debug (OCD) features are frequently available in modern microprocessors. Their contribution to shorten the time-to-market justifies the industry investment in this area, where a number of competing or complementary proposals are available or under development, e.g. NEXUS, CJTAG, IJTAG. The controllability and observability features provided by OCD infrastructures provide a valuable toolbox that can be used well beyond the debugging arena, improving the return on investment rate by diluting its cost across a wider spectrum of application areas. This paper discusses the use of OCD features for validating fault tolerant architectures, and in particular the efficiency of various fault injection methods provided by enhanced OCD infrastructures. The reference data for our comparative study was captured on a workbench comprising the 32-bit Freescale MPC-565 microprocessor, an iSYSTEM IC3000 debugger (iTracePro version) and the Winidea 2005 debugging package. All enhanced OCD infrastructures were implemented in VHDL and the results were obtained by simulation within the same fault injection environment. The focus of this paper is on the comparative analysis of the experimental results obtained for various OCD configurations and debugging scenarios.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Thrust ball bearings lubricated with several different greases were tested on a modified Four-Ball Machine, where the Four-Ball arrangement was replaced by a bearing assembly. The friction torque and operating temperatures in a thrust ball bearing were measured during the tests. At the end of each test a grease sample was analyzed through ferrographic techniques in order to quantify and evaluate bearing wear. A rolling bearing friction torque model was used and the coefficient of friction in full film lubrication was determined for each grease, depending on the operating conditions. The experimental results obtained showed that grease formulation had a very significant influence on friction torque and operating temperature. The friction torque depends on the viscosity of the grease base oil, on its nature (mineral, ester, PAO, etc.), on the coefficient of friction in full film conditions, but also on the interaction between grease thickener and base oil, which affected contact replenishment and contact starvation, and thus influenced the friction torque.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Dissertação apresentada na Faculdade de Ciências e Tecnologia da Universidade Nova de Lisboa para a obtenção do Grau de Mestre em Engenharia Informática

Relevância:

60.00% 60.00%

Publicador:

Resumo:

We present the modeling efforts on antenna design and frequency selection to monitor brain temperature during prolonged surgery using noninvasive microwave radiometry. A tapered log-spiral antenna design is chosen for its wideband characteristics that allow higher power collection from deep brain. Parametric analysis with the software HFSS is used to optimize antenna performance for deep brain temperature sensing. Radiometric antenna efficiency (eta) is evaluated in terms of the ratio of power collected from brain to total power received by the antenna. Anatomical information extracted from several adult computed tomography scans is used to establish design parameters for constructing an accurate layered 3-D tissue phantom. This head phantom includes separate brain and scalp regions, with tissue equivalent liquids circulating at independent temperatures on either side of an intact skull. The optimized frequency band is 1.1-1.6 GHz producing an average antenna efficiency of 50.3% from a two turn log-spiral antenna. The entire sensor package is contained in a lightweight and low-profile 2.8 cm diameter by 1.5 cm high assembly that can be held in place over the skin with an electromagnetic interference shielding adhesive patch. The calculated radiometric equivalent brain temperature tracks within 0.4 degrees C of the measured brain phantom temperature when the brain phantom is lowered 10. C and then returned to the original temperature (37 degrees C) over a 4.6-h experiment. The numerical and experimental results demonstrate that the optimized 2.5-cm log-spiral antenna is well suited for the noninvasive radiometric sensing of deep brain temperature.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

A genetic algorithm used to design radio-frequency binary-weighted differential switched capacitor arrays (RFDSCAs) is presented in this article. The algorithm provides a set of circuits all having the same maximum performance. This article also describes the design, implementation, and measurements results of a 0.25 lm BiCMOS 3-bit RFDSCA. The experimental results show that the circuit presents the expected performance up to 40 GHz. The similarity between the evolutionary solutions, circuit simulations, and measured results indicates that the genetic synthesis method is a very useful tool for designing optimum performance RFDSCAs.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

In the field of appearance-based robot localization, the mainstream approach uses a quantized representation of local image features. An alternative strategy is the exploitation of raw feature descriptors, thus avoiding approximations due to quantization. In this work, the quantized and non-quantized representations are compared with respect to their discriminativity, in the context of the robot global localization problem. Having demonstrated the advantages of the non-quantized representation, the paper proposes mechanisms to reduce the computational burden this approach would carry, when applied in its simplest form. This reduction is achieved through a hierarchical strategy which gradually discards candidate locations and by exploring two simplifying assumptions about the training data. The potential of the non-quantized representation is exploited by resorting to the entropy-discriminativity relation. The idea behind this approach is that the non-quantized representation facilitates the assessment of the distinctiveness of features, through the entropy measure. Building on this finding, the robustness of the localization system is enhanced by modulating the importance of features according to the entropy measure. Experimental results support the effectiveness of this approach, as well as the validity of the proposed computation reduction methods.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

A design methodology for monolithic integration of inductor based DC-DC converters is proposed in this paper. A power loss model of the power stage, including the drive circuits, is defined in order to optimize efficiency. Based on this model and taking as reference a 0.35 mu m CMOS process, a buck converter was designed and fabricated. For a given set of operating conditions the defined power loss model allows to optimize the design parameters for the power stage, including the gate-driver tapering factor and the width of the power MOSFETs. Experimental results obtained from a buck converter at 100 MHz switching frequency are presented to validate the proposed methodology.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Many learning problems require handling high dimensional datasets with a relatively small number of instances. Learning algorithms are thus confronted with the curse of dimensionality, and need to address it in order to be effective. Examples of these types of data include the bag-of-words representation in text classification problems and gene expression data for tumor detection/classification. Usually, among the high number of features characterizing the instances, many may be irrelevant (or even detrimental) for the learning tasks. It is thus clear that there is a need for adequate techniques for feature representation, reduction, and selection, to improve both the classification accuracy and the memory requirements. In this paper, we propose combined unsupervised feature discretization and feature selection techniques, suitable for medium and high-dimensional datasets. The experimental results on several standard datasets, with both sparse and dense features, show the efficiency of the proposed techniques as well as improvements over previous related techniques.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Data analytic applications are characterized by large data sets that are subject to a series of processing phases. Some of these phases are executed sequentially but others can be executed concurrently or in parallel on clusters, grids or clouds. The MapReduce programming model has been applied to process large data sets in cluster and cloud environments. For developing an application using MapReduce there is a need to install/configure/access specific frameworks such as Apache Hadoop or Elastic MapReduce in Amazon Cloud. It would be desirable to provide more flexibility in adjusting such configurations according to the application characteristics. Furthermore the composition of the multiple phases of a data analytic application requires the specification of all the phases and their orchestration. The original MapReduce model and environment lacks flexible support for such configuration and composition. Recognizing that scientific workflows have been successfully applied to modeling complex applications, this paper describes our experiments on implementing MapReduce as subworkflows in the AWARD framework (Autonomic Workflow Activities Reconfigurable and Dynamic). A text mining data analytic application is modeled as a complex workflow with multiple phases, where individual workflow nodes support MapReduce computations. As in typical MapReduce environments, the end user only needs to define the application algorithms for input data processing and for the map and reduce functions. In the paper we present experimental results when using the AWARD framework to execute MapReduce workflows deployed over multiple Amazon EC2 (Elastic Compute Cloud) instances.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Feature selection is a central problem in machine learning and pattern recognition. On large datasets (in terms of dimension and/or number of instances), using search-based or wrapper techniques can be cornputationally prohibitive. Moreover, many filter methods based on relevance/redundancy assessment also take a prohibitively long time on high-dimensional. datasets. In this paper, we propose efficient unsupervised and supervised feature selection/ranking filters for high-dimensional datasets. These methods use low-complexity relevance and redundancy criteria, applicable to supervised, semi-supervised, and unsupervised learning, being able to act as pre-processors for computationally intensive methods to focus their attention on smaller subsets of promising features. The experimental results, with up to 10(5) features, show the time efficiency of our methods, with lower generalization error than state-of-the-art techniques, while being dramatically simpler and faster.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

The behavior of tandem pin heterojunctions based on a-SiC: H alloys is investigated under different optical and electrical bias conditions. The devices are optimized to act as optically selective wavelength filters. Depending on the device configuration (optical gaps, thickness, sequence of cells in the stack structure) and on the applied voltage (positive or negative) and optical bias (wavelength, intensity, frequency) it is possible to combine the wavelength discrimination function with the self amplification of the signal. This wavelength nonlinearity allows the amplification or the rejection of a weak signal-impulse. The device works as an active tunable optical filter for wavelength selection and can be used as an add/drop multiplexer (ADM) which enables data to enter and leave an optical network bit stream without having to demultiplex the stream. Results show that, even under weak transient input signals, the background wavelength controls the output signal. This nonlinearity, due to the transient asymmetrical light penetration of the input channels across the device together with the modification on the electrical field profile due to the optical bias, allows tuning an input channel without demultiplexing the stream. This high optical nonlinearity makes the optimized devices attractive for the amplification of all optical signals. Transfer characteristics effects due to changes in steady state light, control d.c. voltage and applied light pulses are presented. Based on the experimental results and device configuration an optoelectronic model is developed. The transfer characteristics effects due to changes in steady state light, dc control voltage or applied light pulses are simulated and compared with the experimental data. A good agreement was achieved.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Este artigo apresenta uma nova abordagem (MM-GAV-FBI), aplicável ao problema da programação de projectos com restrições de recursos e vários modos de execução por actividade, problema conhecido na literatura anglo-saxónica por MRCPSP. Cada projecto tem um conjunto de actividades com precedências tecnológicas definidas e um conjunto de recursos limitados, sendo que cada actividade pode ter mais do que um modo de realização. A programação dos projectos é realizada com recurso a um esquema de geração de planos (do inglês Schedule Generation Scheme - SGS) integrado com uma metaheurística. A metaheurística é baseada no paradigma dos algoritmos genéticos. As prioridades das actividades são obtidas a partir de um algoritmo genético. A representação cromossómica utilizada baseia-se em chaves aleatórias. O SGS gera planos não-atrasados. Após a obtenção de uma solução é aplicada uma melhoria local. O objectivo da abordagem é encontrar o melhor plano (planning), ou seja, o plano que tenha a menor duração temporal possível, satisfazendo as precedências das actividades e as restrições de recursos. A abordagem proposta é testada num conjunto de problemas retirados da literatura da especialidade e os resultados computacionais são comparados com outras abordagens. Os resultados computacionais validam o bom desempenho da abordagem, não apenas em termos de qualidade da solução, mas também em termos de tempo útil.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

In this study, the tensile strength of single-lap joints (SLJs) between similar and dissimilar adherends bonded with an acrylic adhesive was evaluated experimentally and numerically. The adherend materials included polyethylene (PE), polypropylene (PP), carbon-epoxy (CFRP), and glass-polyester (GFRP) composites. The following adherend combinations were tested: PE/PE, PE/PP, PE/CFRP, PE/GFRP, PP/PP, CFRP/CFRP, and GFRP/GFRP. One of the objectives of this work was to assess the influence of the adherends stiffness on the strength of the joints since it significantly affects the peel stresses magnitude in the adhesive layer. The experimental results were also used to validate a new mixed-mode cohesive damage model developed to simulate the adhesive layer. Thus, the experimental results were compared with numerical simulations performed in ABAQUS®, including a developed mixed-mode (I+II) cohesive damage model, based on the indirect use of fracture mechanics and implemented within interface finite elements. The cohesive laws present a trapezoidal shape with an increasing stress plateau, to reproduce the behaviour of the ductile adhesive used. A good agreement was found between the experimental and numerical results.