929 resultados para Visualization technique
Resumo:
Information technology in construction (ITC) has been gaining wide acceptance and is being implemented in the construction research domains as a tool to assist decision makers. Most of the research into visualization technologies (VT) has been on the wide range of 3D and simulation applications suitable for construction processes. Despite its development with interoperability and standardization of products, VT usage has remained very low when it comes to communicating and addressing the needs of building end-users (BEU). This paper argues that building end users are a source of experience and expertise that can be brought into the briefing stage for the evaluation of design proposals. It also suggests that the end user is a source of new ideas promoting innovation. In this research a positivistic methodology that includes the comparison of 3D models and the traditional 2D methods is proposed. It will help to identify "how much", if anything, a non-spatial specialist can gain in terms Of "understanding" of a particular design proposal presented, using both methods.
Resumo:
The incorporation of caseins and whey proteins into acid gels produced from unheated and heat treated skimmed milk was studied by confocal scanning laser microscopy (CSLM) using fluorescent labelled proteins. Bovine casein micelles were labelled using Alexa Fluor 594, while whey proteins were labelled using Alexa Fluor 488. Samples of the labelled protein solutions were introduced into aliquots of pasteurised skim milk, and skim milk heated to 90 degrees C for 2 min and 95 degrees C for 8 min. The milk was acidified at 40 degrees C to a final pH of 4.4 using 20 g gluconodelta-lactone/l (GDL). The formation of gels was observed with CSLM at two wavelengths (488 nm and 594 nm), and also by visual and rheological methods. In the control milk, as pH decreased distinct casein aggregates appeared, and as further pH reduction occurred, the whey proteins could be seen to coat the casein aggregates. With the heated milks, the gel structure was formed of continuous strands consisting of both casein and whey protein. The formation of the gel network was correlated with an increase in the elastic modulus for all three treatments, in relation to the severity of heat treatment. This model system allows the separate observation of the caseins and whey proteins, and the study of the interactions between the two protein fractions during the formation of the acid gel structure, on a real-time basis. The system could therefore be a valuable tool in the study of structure formation in yoghurt and other dairy protein systems.
Resumo:
Introduction A high saturated fatty acid intake is a well recognized risk factor for coronary heart disease development. More recently a high intake of n-6 polyunsaturated fatty acids (PUFA) in combination with a low intake of the long chain n-3 PUFA, eicosapentaenoic acid and docosahexaenoic acid has also been implicated as an important risk factor. Aim To compare total dietary fat and fatty acid intake measured by chemical analysis of duplicate diets with nutritional database analysis of estimated dietary records, collected over the same 3-day study period. Methods Total fat was analysed using soxhlet extraction and subsequently the individual fatty acid content of the diet was determined by gas chromatography. Estimated dietary records were analysed using a nutrient database which was supplemented with a selection of dishes commonly consumed by study participants. Results Bland & Altman statistical analysis demonstrated a lack of agreement between the two dietary assessment techniques for determining dietary fat and fatty acid intake. Conclusion The lack of agreement observed between dietary evaluation techniques may be attributed to inadequacies in either or both assessment techniques. This study highlights the difficulties that may be encountered when attempting to accurately evaluate dietary fat intake among the population.
Resumo:
Pullpipelining, a pipeline technique where data is pulled from successor stages from predecessor stages is proposed Control circuits using a synchronous, a semi-synchronous and an asynchronous approach are given. Simulation examples for a DLX generic RISC datapath show that common control pipeline circuit overhead is avoided using the proposal. Applications to linear systolic arrays in cases when computation is finished at early stages in the array are foreseen. This would allow run-time data-driven digital frequency modulation of synchronous pipelined designs. This has applications to implement algorithms exhibiting average-case processing time using a synchronous approach.
Resumo:
Very large scale scheduling and planning tasks cannot be effectively addressed by fully automated schedule optimisation systems, since many key factors which govern 'fitness' in such cases are unformalisable. This raises the question of an interactive (or collaborative) approach, where fitness is assigned by the expert user. Though well-researched in the domains of interactively evolved art and music, this method is as yet rarely used in logistics. This paper concerns a difficulty shared by all interactive evolutionary systems (IESs), but especially those used for logistics or design problems. The difficulty is that objective evaluation of IESs is severely hampered by the need for expert humans in the loop. This makes it effectively impossible to, for example, determine with statistical confidence any ranking among a decent number of configurations for the parameters and strategy choices. We make headway into this difficulty with an Automated Tester (AT) for such systems. The AT replaces the human in experiments, and has parameters controlling its decision-making accuracy (modelling human error) and a built-in notion of a target solution which may typically be at odds with the solution which is optimal in terms of formalisable fitness. Using the AT, plausible evaluations of alternative designs for the IES can be done, allowing for (and examining the effects of) different levels of user error. We describe such an AT for evaluating an IES for very large scale planning.
Resumo:
This paper investigates the application of the Hilbert spectrum (HS), which is a recent tool for the analysis of nonlinear and nonstationary time-series, to the study of electromyographic (EMG) signals. The HS allows for the visualization of the energy of signals through a joint time-frequency representation. In this work we illustrate the use of the HS in two distinct applications. The first is for feature extraction from EMG signals. Our results showed that the instantaneous mean frequency (IMNF) estimated from the HS is a relevant feature to clinical practice. We found that the median of the IMNF reduces when the force level of the muscle contraction increases. In the second application we investigated the use of the HS for detection of motor unit action potentials (MUAPs). The detection of MUAPs is a basic step in EMG decomposition tools, which provide relevant information about the neuromuscular system through the morphology and firing time of MUAPs. We compared, visually, how MUAP activity is perceived on the HS with visualizations provided by some traditional (e.g. scalogram, spectrogram, Wigner-Ville) time-frequency distributions. Furthermore, an alternative visualization to the HS, for detection of MUAPs, is proposed and compared to a similar approach based on the continuous wavelet transform (CWT). Our results showed that both the proposed technique and the CWT allowed for a clear visualization of MUAP activity on the time-frequency distributions, whereas results obtained with the HS were the most difficult to interpret as they were extremely affected by spurious energy activity. (c) 2008 Elsevier Inc. All rights reserved.
Resumo:
A simple and practical technique for assessing the risks, that is, the potential for error, and consequent loss, in software system development, acquired during a requirements engineering phase is described. The technique uses a goal-based requirements analysis as a framework to identify and rate a set of key issues in order to arrive at estimates of the feasibility and adequacy of the requirements. The technique is illustrated and how it has been applied to a real systems development project is shown. How problems in this project could have been identified earlier is shown, thereby avoiding costly additional work and unhappy users.
Resumo:
This paper presents a clocking pipeline technique referred to as a single-pulse pipeline (PP-Pipeline) and applies it to the problem of mapping pipelined circuits to a Field Programmable Gate Array (FPGA). A PP-pipeline replicates the operation of asynchronous micropipelined control mechanisms using synchronous-orientated logic resources commonly found in FPGA devices. Consequently, circuits with an asynchronous-like pipeline operation can be efficiently synthesized using a synchronous design methodology. The technique can be extended to include data-completion circuitry to take advantage of variable data-completion processing time in synchronous pipelined designs. It is also shown that the PP-pipeline reduces the clock tree power consumption of pipelined circuits. These potential applications are demonstrated by post-synthesis simulation of FPGA circuits. (C) 2004 Elsevier B.V. All rights reserved.