913 resultados para Optimal trigger speed
Resumo:
J.A. Ferreira Neto, E.C. Santos Junior, U. Fra Paleo, D. Miranda Barros, and M.C.O. Moreira. 2011. Optimal subdivision of land in agrarian reform projects: an analysis using genetic algorithms. Cien. Inv. Agr. 38(2): 169-178. The objective of this manuscript is to develop a new procedure to achieve optimal land subdivision using genetic algorithms (GA). The genetic algorithm was tested in the rural settlement of Veredas, located in Minas Gerais, Brazil. This implementation was based on the land aptitude and its productivity index. The sequence of tests in the study was carried out in two areas with eight different agricultural aptitude classes, including one area of 391.88 ha subdivided into 12 lots and another of 404.1763 ha subdivided into 14 lots. The effectiveness of the method was measured using the shunting line standard value of a parceled area lot`s productivity index. To evaluate each parameter, a sequence of 15 calculations was performed to record the best individual fitness average (MMI) found for each parameter variation. The best parameter combination found in testing and used to generate the new parceling with the GA was the following: 320 as the generation number, a population of 40 individuals, 0.8 mutation tax, and a 0.3 renewal tax. The solution generated rather homogeneous lots in terms of productive capacity.
Resumo:
The topology of real-world complex networks, such as in transportation and communication, is always changing with time. Such changes can arise not only as a natural consequence of their growth, but also due to major modi. cations in their intrinsic organization. For instance, the network of transportation routes between cities and towns ( hence locations) of a given country undergo a major change with the progressive implementation of commercial air transportation. While the locations could be originally interconnected through highways ( paths, giving rise to geographical networks), transportation between those sites progressively shifted or was complemented by air transportation, with scale free characteristics. In the present work we introduce the path-star transformation ( in its uniform and preferential versions) as a means to model such network transformations where paths give rise to stars of connectivity. It is also shown, through optimal multivariate statistical methods (i.e. canonical projections and maximum likelihood classification) that while the US highways network adheres closely to a geographical network model, its path-star transformation yields a network whose topological properties closely resembles those of the respective airport transportation network.
Resumo:
The purpose of this study was to comparatively evaluate the response of human pulps after cavity preparation with different devices. Deep class I cavities were prepared in sound mandibular premolars using either a high-speed air-turbine handpiece (Group 1) or an Er: YAG laser (Group 2). Following total acid etching and the application of an adhesive system, all cavities were restored with composite resin. Fifteen days after the clinical procedure, the teeth were extracted and processed for analysis under optical microscopy. In Group 1 in which the average for the remaining dentin thickness (RDT) between the cavity floor and the coronal pulp was 909.5 mu m, a discrete inflammatory response occurred in only one specimen with an RDT of 214 mu m. However, tissue disorganization occurred in most specimens. In Group 2 (average RDT = 935.2 mu m), the discrete inflammatory pulp response was observed in only one specimen (average RDT = 413 mu m). It may be concluded that the high-speed air-turbine handpiece caused greater structural alterations in the pulp, although without inducing inflammatory processes.
Resumo:
The surface detector array of the Pierre Auger Observatory consists of 1600 water-Cherenkov detectors, for the study of extensive air showers (EAS) generated by ultra-high-energy cosmic rays. We describe the trigger hierarchy, from the identification of candidate showers at the level of a single detector, amongst a large background (mainly random single cosmic ray muons), up to the selection of real events and the rejection of random coincidences. Such trigger makes the surface detector array fully efficient for the detection of EAS with energy above 3 x 10(18) eV, for all zenith angles between 0 degrees and 60 degrees, independently of the position of the impact point and of the mass of the primary particle. In these range of energies and angles, the exposure of the surface array can be determined purely on the basis of the geometrical acceptance. (C) 2009 Elsevier B.V. All rights reserved.
Resumo:
To get an optimal print result with rotogravure on improved newsprint, MF-paper, knowledge aboutwhat adjustment that should be done through the whole production is needed. This degree project investigatesthe reproduction’s effect on the final print result.A test print is done at Quebecor World Nordic Interprint´s regular print unit. The evaluation of theprint shows how important the use of a unit adjusted ICC-profile is for the print quality. A comparisonbetween the two separation methods GCR and UCR shows that Interprint´s present ICC-profile, usingUCR as separation method, gives less missing dots. When the cylinder that prints the black ink isengraving whit a slower speed than normal the result shows that the colour becomes a little darker, thedetails becomes clear but the difference is very small.
Resumo:
To connect different electrical, network and data devices with the minimum cost and shortest path, is a complex job. In huge buildings, where the devices are placed at different locations on different floors and only some specific routes are available to pass the cables and buses, the shortest path search becomes more complex. The aim of this thesis project is, to develop an application which indentifies the best path to connect all objects or devices by following the specific routes.To address the above issue we adopted three algorithms Greedy Algorithm, Simulated Annealing and Exhaustive search and analyzed their results. The given problem is similar to Travelling Salesman Problem. Exhaustive search is a best algorithm to solve this problem as it checks each and every possibility and give the accurate result but it is an impractical solution because of huge time consumption. If no. of objects increased from 12 it takes hours to search the shortest path. Simulated annealing is emerged with some promising results with lower time cost. As of probabilistic nature, Simulated annealing could be non optimal but it gives a near optimal solution in a reasonable duration. Greedy algorithm is not a good choice for this problem. So, simulated annealing is proved best algorithm for this problem. The project has been implemented in C-language which takes input and store output in an Excel Workbook
Resumo:
The cost of a road construction over its service life is a function of design, quality of construction as well as maintenance strategies and operations. An optimal life-cycle cost for a road requires evaluations of the above mentioned components. Unfortunately, road designers often neglect a very important aspect, namely, the possibility to perform future maintenance activities. Focus is mainly directed towards other aspects such as investment costs, traffic safety, aesthetic appearance, regional development and environmental effects. This doctoral thesis presents the results of a research project aimed to increase consideration of road maintenance aspects in the planning and design process. The following subgoals were established: Identify the obstacles that prevent adequate consideration of future maintenance during the road planning and design process; and Examine optimisation of life-cycle costs as an approach towards increased efficiency during the road planning and design process. The research project started with a literature review aimed at evaluating the extent to which maintenance aspects are considered during road planning and design as an improvement potential for maintenance efficiency. Efforts made by road authorities to increase efficiency, especially maintenance efficiency, were evaluated. The results indicated that all the evaluated efforts had one thing in common, namely ignorance of the interrelationship between geometrical road design and maintenance as an effective tool to increase maintenance efficiency. Focus has mainly been on improving operating practises and maintenance procedures. This fact might also explain why some efforts to increase maintenance efficiency have been less successful. An investigation was conducted to identify the problems and difficulties, which obstruct due consideration of maintainability during the road planning and design process. A method called “Change Analysis” was used to analyse data collected during interviews with experts in road design and maintenance. The study indicated a complex combination of problems which result in inadequate consideration of maintenance aspects when planning and designing roads. The identified problems were classified into six categories: insufficient consulting, insufficient knowledge, regulations and specifications without consideration of maintenance aspects, insufficient planning and design activities, inadequate organisation and demands from other authorities. Several urgent needs for changes to eliminate these problems were identified. One of the problems identified in the above mentioned study as an obstacle for due consideration of maintenance aspects during road design was the absence of a model for calculating life-cycle costs for roads. Because of this lack of knowledge, the research project focused on implementing a new approach for calculating and analysing life-cycle costs for roads with emphasis on the relationship between road design and road maintainability. Road barriers were chosen as an example. The ambition is to develop this approach to cover other road components at a later stage. A study was conducted to quantify repair rates for barriers and associated repair costs as one of the major maintenance costs for road barriers. A method called “Case Study Research Method” was used to analyse the effect of several factors on barrier repairs costs, such as barrier type, road type, posted speed and seasonal effect. The analyses were based on documented data associated with 1625 repairs conducted in four different geographical regions in Sweden during 2006. A model for calculation of average repair costs per vehicle kilometres was created. Significant differences in the barrier repair costs were found between the studied barrier types. In another study, the injuries associated with road barrier collisions and the corresponding influencing factors were analysed. The analyses in this study were based on documented data from actual barrier collisions between 2005 and 2008 in Sweden. The result was used to calculate the cost for injuries associated with barrier collisions as a part of the socio-economic cost for road barriers. The results showed significant differences in the number of injuries associated with collisions with different barrier types. To calculate and analyse life-cycle costs for road barriers a new approach was developed based on a method called “Activity-based Life-cycle Costing”. By modelling uncertainties, the presented approach gives a possibility to identify and analyse factors crucial for optimising life-cycle costs. The study showed a great potential to increase road maintenance efficiency through road design. It also showed that road components with low investment costs might not be the best choice when including maintenance and socio-economic aspects. The difficulties and problems faced during the collection of data for calculating life-cycle costs for road barriers indicated a great need for improving current data collecting and archiving procedures. The research focused on Swedish road planning and design. However, the conclusions can be applied to other Nordic countries, where weather conditions and road design practices are similar. The general methodological approaches used in this research project may be applied also to other studies.
Resumo:
This report presents a new way of control engineering. Dc motor speed controlled by three controllers PID, pole placement and Fuzzy controller and discusses the advantages and disadvantages of each controller for different conditions under loaded and unloaded scenarios using software Matlab. The brushless series wound Dc motor is very popular in industrial application and control systems because of the high torque density, high efficiency and small size. First suitable equations are developed for DC motor. PID controller is developed and tuned in order to get faster step response. The simulation results of PID controller provide very good results and the controller is further tuned in order to decrease its overshoot error which is common in PID controllers. Further it is purposed that in industrial environment these controllers are better than others controllers as PID controllers are easy to tuned and cheap. Pole placement controller is the best example of control engineering. An addition of integrator reduced the noise disturbances in pole placement controller and this makes it a good choice for industrial applications. The fuzzy controller is introduce with a DC chopper to make the DC motor speed control smooth and almost no steady state error is observed. Another advantage is achieved in fuzzy controller that the simulations of three different controllers are compared and concluded from the results that Fuzzy controller outperforms to PID controller in terms of steady state error and smooth step response. While Pole placement controller have no comparison in terms of controls because designer can change the step response according to nature of control systems, so this controller provide wide range of control over a system. Poles location change the step response in a sense that if poles are near to origin then step response of motor is fast. Finally a GUI of these three controllers are developed which allow the user to select any controller and change its parameters according to the situation.
Resumo:
Denna avhandling tar sin utgångspunkt i ett ifrågasättande av effektiviteten i EU:s konditionalitetspolitik avseende minoritetsrättigheter. Baserat på den rationalistiska teoretiska modellen, External Incentives Model of Governance, syftar denna hypotesprövande avhandling till att förklara om tidsavståndet på det potentiella EU medlemskapet påverkar lagstiftningsnivån avseende minoritetsspråksrättigheter. Mätningen av nivån på lagstiftningen avseende minoritetsspråksrättigheter begränsas till att omfatta icke-diskriminering, användning av minoritetsspråk i officiella sammanhang samt minoriteters språkliga rättigheter i utbildningen. Metodologiskt används ett jämförande angreppssätt både avseende tidsramen för studien, som sträcker sig mellan 2003 och 2010, men även avseende urvalet av stater. På basis av det \"mest lika systemet\" kategoriseras staterna i tre grupper efter deras olika tidsavstånd från det potentiella EU medlemskapet. Hypotesen som prövas är följande: ju kortare tidsavstånd till det potentiella EU medlemskapet desto större sannolikhet att staternas lagstiftningsnivå inom de tre områden som studeras har utvecklats till en hög nivå. Studien visar att hypotesen endast bekräftas delvis. Resultaten avseende icke-diskriminering visar att sambandet mellan tidsavståndet och nivån på lagstiftningen har ökat markant under den undersökta tidsperioden. Detta samband har endast stärkts mellan kategorin av stater som ligger tidsmässigt längst bort ett potentiellt EU medlemskap och de två kategorier som ligger närmare respektive närmast ett potentiellt EU medlemskap. Resultaten avseende användning av minoritetsspråk i officiella sammanhang och minoriteters språkliga rättigheter i utbildningen visar inget respektive nästan inget samband mellan tidsavståndet och utvecklingen på lagstiftningen mellan 2003 och 2010.