954 resultados para harvester simulator
Resumo:
This paper shows that radiometer channel radiances for cloudy atmospheric conditions can be simulated with an optimised frequency grid derived under clear-sky conditions. A new clear-sky optimised grid is derived for AVHRR channel 5 ð12 m m, 833 cm �1 Þ. For HIRS channel 11 ð7:33 m m, 1364 cm �1 Þ and AVHRR channel 5, radiative transfer simulations using an optimised frequency grid are compared with simulations using a reference grid, where the optimised grid has roughly 100–1000 times less frequencies than the full grid. The root mean square error between the optimised and the reference simulation is found to be less than 0.3 K for both comparisons, with the magnitude of the bias less than 0.03 K. The simulations have been carried out with the radiative transfer model Atmospheric Radiative Transfer Simulator (ARTS), version 2, using a backward Monte Carlo module for the treatment of clouds. With this module, the optimised simulations are more than 10 times faster than the reference simulations. Although the number of photons is the same, the smaller number of frequencies reduces the overhead for preparing the optical properties for each frequency. With deterministic scattering solvers, the relative decrease in runtime would be even more. The results allow for new radiative transfer applications, such as the development of new retrievals, because it becomes much quicker to carry out a large number of simulations. The conclusions are applicable to any downlooking infrared radiometer.
Resumo:
A study has been carried out to assess the importance of radiosonde corrections in improving the agreement between satellite and radiosonde measurements of upper-tropospheric humidity. Infrared [High Resolution Infrared Radiation Sounder (HIRS)-12] and microwave [Advanced Microwave Sounding Unit (AMSU)-18] measurements from the NOAA-17 satellite were used for this purpose. The agreement was assessed by comparing the satellite measurements against simulated measurements using collocated radiosonde profiles of the Atmospheric Radiation Measurement (ARM) Program undertaken at tropical and midlatitude sites. The Atmospheric Radiative Transfer Simulator (ARTS) was used to simulate the satellite radiances. The comparisons have been done under clear-sky conditions, separately for daytime and nighttime soundings. Only Vaisala RS92 radiosonde sensors were used and an empirical correction (EC) was applied to the radiosonde measurements. The EC includes correction for mean calibration bias and for solar radiation error, and it removes radiosonde bias relative to three instruments of known accuracy. For the nighttime dataset, the EC significantly reduces the bias from 0.63 to 20.10 K in AMSU-18 and from 1.26 to 0.35 K in HIRS-12. The EC has an even greater impact on the daytime dataset with a bias reduction from 2.38 to 0.28 K in AMSU-18 and from 2.51 to 0.59 K in HIRS-12. The present study promises a more accurate approach in future radiosonde-based studies in the upper troposphere.
Resumo:
Future extreme-scale high-performance computing systems will be required to work under frequent component failures. The MPI Forum's User Level Failure Mitigation proposal has introduced an operation, MPI_Comm_shrink, to synchronize the alive processes on the list of failed processes, so that applications can continue to execute even in the presence of failures by adopting algorithm-based fault tolerance techniques. This MPI_Comm_shrink operation requires a fault tolerant failure detection and consensus algorithm. This paper presents and compares two novel failure detection and consensus algorithms. The proposed algorithms are based on Gossip protocols and are inherently fault-tolerant and scalable. The proposed algorithms were implemented and tested using the Extreme-scale Simulator. The results show that in both algorithms the number of Gossip cycles to achieve global consensus scales logarithmically with system size. The second algorithm also shows better scalability in terms of memory and network bandwidth usage and a perfect synchronization in achieving global consensus.
Resumo:
Background: Daily consumption of Concord grape juice (CGJ) over three to four months has been shown to improve memory function in adults with mild cognitive impairment, and reduce blood pressure in hypertensive adults. These benefits are likely due to the high concentration of polyphenols in CGJ. Increased stress can impair cognitive function and elevate blood pressure. Thus we examined the potential beneficial effect of CGJ in individuals experiencing somewhat stressful demanding lifestyles. Objective: To examine the effects of twelve weeks’ daily consumption of CGJ on cognitive function, driving performance, and blood pressure in healthy, middle-aged working mothers. Design: Twenty five healthy mothers of pre-teen children, aged 40-50 years, who were employed for > 30 hours/week consumed 12oz (355ml) CGJ (containing 777mg total polyphenols) or an energy, taste and appearance matched placebo daily for twelve weeks according to a randomised, crossover design with a four week washout. Verbal and spatial memory, executive function, attention, blood pressure and mood were assessed at baseline, six weeks and twelve weeks. Immediately following the cognitive battery, a subsample of seventeen females completed a driving performance assessment in the University of Leeds Driving Simulator. The twenty five minute driving task required participants to match the speed and direction of a lead vehicle. Results: Significant improvements in immediate spatial memory and driving performance were observed following CGJ relative to placebo. There was evidence of an enduring effect of CGJ such that participants who received CGJ in arm 1 maintained better performance in the placebo arm. Conclusions: Cognitive benefits associated with chronic consumption of flavonoid-rich grape juice are not exclusive to adults with mild cognitive impairment. Moreover, these cognitive benefits are apparent in complex everyday tasks such as driving. Effects may persist beyond cessation of flavonoid consumption and future studies should carefully consider the length of washout within crossover designs.
Resumo:
The assessment of routing protocols for mobile wireless networks is a difficult task, because of the networks` dynamic behavior and the absence of benchmarks. However, some of these networks, such as intermittent wireless sensors networks, periodic or cyclic networks, and some delay tolerant networks (DTNs), have more predictable dynamics, as the temporal variations in the network topology can be considered as deterministic, which may make them easier to study. Recently, a graph theoretic model-the evolving graphs-was proposed to help capture the dynamic behavior of such networks, in view of the construction of least cost routing and other algorithms. The algorithms and insights obtained through this model are theoretically very efficient and intriguing. However, there is no study about the use of such theoretical results into practical situations. Therefore, the objective of our work is to analyze the applicability of the evolving graph theory in the construction of efficient routing protocols in realistic scenarios. In this paper, we use the NS2 network simulator to first implement an evolving graph based routing protocol, and then to use it as a benchmark when comparing the four major ad hoc routing protocols (AODV, DSR, OLSR and DSDV). Interestingly, our experiments show that evolving graphs have the potential to be an effective and powerful tool in the development and analysis of algorithms for dynamic networks, with predictable dynamics at least. In order to make this model widely applicable, however, some practical issues still have to be addressed and incorporated into the model, like adaptive algorithms. We also discuss such issues in this paper, as a result of our experience.
Resumo:
Internet protocol TV (IPTV) is predicted to be the key technology winner in the future. Efforts to accelerate the deployment of IPTV centralized model which is combined of VHO, encoders, controller, access network and Home network. Regardless of whether the network is delivering live TV, VOD, or Time-shift TV, all content and network traffic resulting from subscriber requests must traverse the entire network from the super-headend all the way to each subscriber's Set-Top Box (STB).IPTV services require very stringent QoS guarantees When IPTV traffic shares the network resources with other traffic like data and voice, how to ensure their QoS and efficiently utilize the network resources is a key and challenging issue. For QoS measured in the network-centric terms of delay jitter, packet losses and bounds on delay. The main focus of this thesis is on the optimized bandwidth allocation and smooth datatransmission. The proposed traffic model for smooth delivering video service IPTV network with its QoS performance evaluation. According to Maglaris et al [5] First, analyze the coding bit rate of a single video source. Various statistical quantities are derived from bit rate data collected with a conditional replenishment inter frame coding scheme. Two correlated Markov process models (one in discrete time and one incontinuous time) are shown to fit the experimental data and are used to model the input rates of several independent sources into a statistical multiplexer. Preventive control mechanism which is to be include CAC, traffic policing used for traffic control.QoS has been evaluated of common bandwidth scheduler( FIFO) by use fluid models with Markovian queuing method and analysis the result by using simulator andanalytically, Which is measured the performance of the packet loss, overflow and mean waiting time among the network users.
Resumo:
IPTV is now offered by several operators in Europe, US and Asia using broadcast video over private IP networks that are isolated from Internet. IPTV services rely ontransmission of live (real-time) video and/or stored video. Video on Demand (VoD)and Time-shifted TV are implemented by IP unicast and Broadcast TV (BTV) and Near video on demand are implemented by IP multicast. IPTV services require QoS guarantees and can tolerate no more than 10-6 packet loss probability, 200 ms delay, and 50 ms jitter. Low delay is essential for satisfactory trick mode performance(pause, resume,fast forward) for VoD, and fast channel change time for BTV. Internet Traffic Engineering (TE) is defined in RFC 3272 and involves both capacity management and traffic management. Capacity management includes capacityplanning, routing control, and resource management. Traffic management includes (1)nodal traffic control functions such as traffic conditioning, queue management, scheduling, and (2) other functions that regulate traffic flow through the network orthat arbitrate access to network resources. An IPTV network architecture includes multiple networks (core network, metronetwork, access network and home network) that connects devices (super head-end, video hub office, video serving office, home gateway, set-top box). Each IP router in the core and metro networks implements some queueing and packet scheduling mechanism at the output link controller. Popular schedulers in IP networks include Priority Queueing (PQ), Class-Based Weighted Fair Queueing (CBWFQ), and Low Latency Queueing (LLQ) which combines PQ and CBWFQ.The thesis analyzes several Packet Scheduling algorithms that can optimize the tradeoff between system capacity and end user performance for the traffic classes. Before in the simulator FIFO,PQ,GPS queueing methods were implemented inside. This thesis aims to implement the LLQ scheduler inside the simulator and to evaluate the performance of these packet schedulers. The simulator is provided by ErnstNordström and Simulator was built in Visual C++ 2008 environmentand tested and analyzed in MatLab 7.0 under windows VISTA.
Resumo:
This thesis work concerns about the Performance evolution of peer to peer networks, where we used different distribution technique’s of peer distribution like Weibull, Lognormal and Pareto distribution process. Then we used a network simulator to evaluate the performance of these three distribution techniques.During the last decade the Internet has expanded into a world-wide network connecting millions of hosts and users and providing services for everyone. Many emerging applications are bandwidth-intensive in their nature; the size of downloaded files including music and videos can be huge, from ten megabits to many gigabits. The efficient use of network resources is thus crucial for the survivability of the Internet. Traffic engineering (TE) covers a range of mechanisms for optimizing operational networks from the traffic perspective. The time scale in traffic engineering varies from the short-term network control to network planning over a longer time period.Here in this thesis work we considered the peer distribution technique in-order to minimise the peer arrival and service process with three different techniques, where we calculated the congestion parameters like blocking time for each peer before entering into the service process, waiting time for a peers while the other peer has been served in the service block and the delay time for each peer. Then calculated the average of each process and graphs have been plotted using Matlab to analyse the results
Resumo:
The aim of this work was to design a set of rules for levodopa infusion dose adjustment in Parkinson’s disease based on a simulation experiments. Using this simulator, optimal infusions dose in different conditions were calculated. There are seven conditions (-3 to +3)appearing in a rating scale for Parkinson’s disease patients. By finding mean of the differences between conditions and optimal dose, two sets of rules were designed. The set of rules was optimized by several testing. Usefulness for optimizing the titration procedure of new infusion patients based on rule-based reasoning was investigated. Results show that both of the number of the steps and the errors for finding optimal dose was shorten by new rules. At last, the dose predicted with new rules well on each single occasion of majority of patients in simulation experiments.
Resumo:
This Thesis project is a part of the all-round automation of production of concentrating solar PV/T systems Absolicon X10. ABSOLICON Solar Concentrator AB has been invented and started production of the prospective solar concentrated system Absolicon X10. The aims of this Thesis project are designing, assembling, calibrating and putting in operation the automatic measurement system intended to evaluate the shape of concentrating parabolic reflectors.On the basis of the requirements of the company administration and needs of real production process the operation conditions for the Laser testing rig were formulated. The basic concept to use laser radiation was defined.At the first step, the complex design of the whole system was made and division on the parts was defined. After the preliminary conducted simulations the function and operation conditions of the all parts were formulated.At the next steps, the detailed design of all the parts was conducted. Most components were ordered from respective companies. Some of the mechanical components were made in the workshop of the company. All parts of the Laser-testing rig were assembled and tested. Software part, which controls the Laser-testing rig work, was created on the LabVIEW basis. To tune and test software part the special simulator was designed and assembled.When all parts were assembled in the complete system, the Laser-testing rig was tested, calibrated and tuned.In the workshop of Absolicon AB, the trial measurements were conducted and Laser-testing rig was installed in the production line at the plant in Soleftea.
Resumo:
This Thesis project is a part of the research conducted in Solar industry. ABSOLICON Solar Concentrator AB has invented and started production of the prospective solar concentrated system Absolicon X10. The aims of this Thesis project are designing, assembling, calibrating and putting in operation the automatic measurement system intended to evaluate distribution of density of solar radiation in the focal line of the concentrated parabolic reflectors and to measure radiation from the artificial source of light being a calibration-testing tool.On the basis of the requirements of the company’s administration and needs of designing the concentrated reflectors the operation conditions for the Sun-Walker were formulated. As the first step, the complex design of the whole system was made and division on the parts was specified. After the preliminary conducted simulation of the functions and operation conditions of the all parts were formulated.As the next steps, the detailed design of all the parts was made. Most components were ordered from respective companies. Some of the mechanical components were made in the workshop of the company. All parts of the Sun-Walker were assembled and tested. The software part, which controls the Sun-Walker work and conducts measurements of solar irradiation, was created on the LabVIEW basis. To tune and test the software part, the special simulator was designed and assembled.When all parts were assembled in the complete system, the Sun-Walker was tested, calibrated and tuned.
Resumo:
The context of this report and the IRIDIA laboratory are described in the preface. Evolutionary Robotics and the box-pushing task are presented in the introduction.The building of a test system supporting Evolutionary Robotics experiments is then detailed. This system is made of a robot simulator and a Genetic Algorithm. It is used to explore the possibility of evolving box-pushing behaviours. The bootstrapping problem is explained, and a novel approach for dealing with it is proposed, with results presented.Finally, ideas for extending this approach are presented in the conclusion.
Resumo:
Internet protocol TV (IPTV) is predicted to be the key technology winner in the future. Efforts to accelerate the deployment of IPTV centralized model which is combined of VHO, encoders, controller, access network and Home network. Regardless of whether the network is delivering live TV, VOD, or Time-shift TV, all content and network traffic resulting from subscriber requests must traverse the entire network from the super-headend all the way to each subscriber's Set-Top Box (STB). IPTV services require very stringent QoS guarantees When IPTV traffic shares the network resources with other traffic like data and voice, how to ensure their QoS and efficiently utilize the network resources is a key and challenging issue. For QoS measured in the network-centric terms of delay jitter, packet losses and bounds on delay. The main focus of this thesis is on the optimized bandwidth allocation and smooth data transmission. The proposed traffic model for smooth delivering video service IPTV network with its QoS performance evaluation. According to Maglaris et al [5] first, analyze the coding bit rate of a single video source. Various statistical quantities are derived from bit rate data collected with a conditional replenishment inter frame coding scheme. Two correlated Markov process models (one in discrete time and one in continuous time) are shown to fit the experimental data and are used to model the input rates of several independent sources into a statistical multiplexer. Preventive control mechanism which is to be including CAC, traffic policing used for traffic control. QoS has been evaluated of common bandwidth scheduler( FIFO) by use fluid models with Markovian queuing method and analysis the result by using simulator and analytically, Which is measured the performance of the packet loss, overflow and mean waiting time among the network users.
Resumo:
The logistics of the aerospace aftermarket raises a number of very interesting challenges from the perspective of electronic contracting. This is a highly dynamic domain, where contracts are established between airlines and engine manufacturers, as well as between engine manufacturers all the way down the supply lines, providing a particularly illustrative showcase for the technologies developed in the CONTRACT project. In this paper, we describe such a domain, as well as our modelling of it as a multiagent simulator where the CONTRACT framework has been used to monitor for compliance with norms.
Resumo:
O presente trabalho é uma análise experimental que procura investigar os efeitos dos fatores carga por eixo, pressão de inflação e tipo de pneu sobre o desempenho de pavimentos. Os dados e informações existentes sobre o quadro de cargas por eixo praticada pela frota de veículos pesados é razoável. A ação conjunta dos referidos fatores é ainda uma questão pouco explorada. Assim, para compensar a escassez de dados sobre pressão de inflação e tipo de construção de pneus utilizados na frota de carga, programou-se uma pesquisa de campo na rodovia estadual RS/240. As informações oriundas da pesquisa demonstram que há um acréscimo generalizado no valor da pressão dos pneus e um aumento de uso dos pneus de fabricação radial em relação aos pneus convencionais. Os dados de campo subsidiaram a programação de um experimento fatorial cruzado executado na Área de Pesquisas e Testes de Pavimentos da UFRGS/DAER. A variação dos níveis dos fatores deu-se através de um simulador linear de tráfego atuando sobre uma pista experimental com 20 m de comprimento por 3,5 m de largura. Tendo como resposta estrutural do pavimento a máxima deflexão superficial recuperável medida com uma viga Benkelmam, determinou-se como significantes os efeitos dos fatores carga por eixo e pressão de enchimento dos pneus. Os cálculos estatísticos indicam também que não há diferenças significativas entre os pneus tipo 9.00R20 e 9.00x20 e que todas interações não exercem efeitos significativos sobre a variável de resposta. Em seqüência, determinaram-se as áreas de contato do rodado duplo do simulador de tráfego com a superfície do pavimento ensaiado para as combinações dos níveis dos fatores. Pode-se, então, comparar área de contato medida com área circular calculada, considerada em muitos modelos de dimensionamento de pavimentos. Relacionou-se a variação da pressão de contato com a deflexão recuperável e procedeu-se uma comparação da mesma com a pressão de inflação nominal dos pneus. Apresenta-se um modelo de análise do desempenho do pavimento em função da carga por eixo e da pressão de inflação, nos limites do experimento. Os dados decorrentes do experimento viabilizaram a determinação dos Fatores de Equivalência de Cargas para os níveis dos fatores, considerando-se o pavimento ensaiado. Avaliou-se, via evolução das deflexões e dos Fatores de Equivalência de Cargas, a redução da vida do pavimento, obtendo-se, para a combinação de níveis mais críticos de carregamento, resultados significativos de até 88 % de redução. Propõe-se, por último, uma formulação para o Fator de Equivalência de Cargas que considere também a ação da pressão de inflação.