926 resultados para Complex engineering problems


Relevância:

30.00% 30.00%

Publicador:

Resumo:

This dissertation proposes statistical methods to formulate, estimate and apply complex transportation models. Two main problems are part of the analyses conducted and presented in this dissertation. The first method solves an econometric problem and is concerned with the joint estimation of models that contain both discrete and continuous decision variables. The use of ordered models along with a regression is proposed and their effectiveness is evaluated with respect to unordered models. Procedure to calculate and optimize the log-likelihood functions of both discrete-continuous approaches are derived, and difficulties associated with the estimation of unordered models explained. Numerical approximation methods based on the Genz algortithm are implemented in order to solve the multidimensional integral associated with the unordered modeling structure. The problems deriving from the lack of smoothness of the probit model around the maximum of the log-likelihood function, which makes the optimization and the calculation of standard deviations very difficult, are carefully analyzed. A methodology to perform out-of-sample validation in the context of a joint model is proposed. Comprehensive numerical experiments have been conducted on both simulated and real data. In particular, the discrete-continuous models are estimated and applied to vehicle ownership and use models on data extracted from the 2009 National Household Travel Survey. The second part of this work offers a comprehensive statistical analysis of free-flow speed distribution; the method is applied to data collected on a sample of roads in Italy. A linear mixed model that includes speed quantiles in its predictors is estimated. Results show that there is no road effect in the analysis of free-flow speeds, which is particularly important for model transferability. A very general framework to predict random effects with few observations and incomplete access to model covariates is formulated and applied to predict the distribution of free-flow speed quantiles. The speed distribution of most road sections is successfully predicted; jack-knife estimates are calculated and used to explain why some sections are poorly predicted. Eventually, this work contributes to the literature in transportation modeling by proposing econometric model formulations for discrete-continuous variables, more efficient methods for the calculation of multivariate normal probabilities, and random effects models for free-flow speed estimation that takes into account the survey design. All methods are rigorously validated on both real and simulated data.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Human and robots have complementary strengths in performing assembly operations. Humans are very good at perception tasks in unstructured environments. They are able to recognize and locate a part from a box of miscellaneous parts. They are also very good at complex manipulation in tight spaces. The sensory characteristics of the humans, motor abilities, knowledge and skills give the humans the ability to react to unexpected situations and resolve problems quickly. In contrast, robots are very good at pick and place operations and highly repeatable in placement tasks. Robots can perform tasks at high speeds and still maintain precision in their operations. Robots can also operate for long periods of times. Robots are also very good at applying high forces and torques. Typically, robots are used in mass production. Small batch and custom production operations predominantly use manual labor. The high labor cost is making it difficult for small and medium manufacturers to remain cost competitive in high wage markets. These manufactures are mainly involved in small batch and custom production. They need to find a way to reduce the labor cost in assembly operations. Purely robotic cells will not be able to provide them the necessary flexibility. Creating hybrid cells where humans and robots can collaborate in close physical proximities is a potential solution. The underlying idea behind such cells is to decompose assembly operations into tasks such that humans and robots can collaborate by performing sub-tasks that are suitable for them. Realizing hybrid cells that enable effective human and robot collaboration is challenging. This dissertation addresses the following three computational issues involved in developing and utilizing hybrid assembly cells: - We should be able to automatically generate plans to operate hybrid assembly cells to ensure efficient cell operation. This requires generating feasible assembly sequences and instructions for robots and human operators, respectively. Automated planning poses the following two challenges. First, generating operation plans for complex assemblies is challenging. The complexity can come due to the combinatorial explosion caused by the size of the assembly or the complex paths needed to perform the assembly. Second, generating feasible plans requires accounting for robot and human motion constraints. The first objective of the dissertation is to develop the underlying computational foundations for automatically generating plans for the operation of hybrid cells. It addresses both assembly complexity and motion constraints issues. - The collaboration between humans and robots in the assembly cell will only be practical if human safety can be ensured during the assembly tasks that require collaboration between humans and robots. The second objective of the dissertation is to evaluate different options for real-time monitoring of the state of human operator with respect to the robot and develop strategies for taking appropriate measures to ensure human safety when the planned move by the robot may compromise the safety of the human operator. In order to be competitive in the market, the developed solution will have to include considerations about cost without significantly compromising quality. - In the envisioned hybrid cell, we will be relying on human operators to bring the part into the cell. If the human operator makes an error in selecting the part or fails to place it correctly, the robot will be unable to correctly perform the task assigned to it. If the error goes undetected, it can lead to a defective product and inefficiencies in the cell operation. The reason for human error can be either confusion due to poor quality instructions or human operator not paying adequate attention to the instructions. In order to ensure smooth and error-free operation of the cell, we will need to monitor the state of the assembly operations in the cell. The third objective of the dissertation is to identify and track parts in the cell and automatically generate instructions for taking corrective actions if a human operator deviates from the selected plan. Potential corrective actions may involve re-planning if it is possible to continue assembly from the current state. Corrective actions may also involve issuing warning and generating instructions to undo the current task.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Re-creating and understanding the origin of life represents one of the major challenges facing the scientific community. We will never know exactly how life started on planet Earth, however, we can reconstruct the most likely chemical pathways that could have contributed to the formation of the first living systems. Traditionally, prebiotic chemistry has investigated the formation of modern life’s precursors and their self-organisation under very specific conditions thought to be ‘plausible’. So far, this approach has failed to produce a living system from the bottom-up. In the work presented herein, two different approaches are employed to explore the transition from inanimate to living matter. The development of microfluidic technology during the last decades has changed the way traditional chemical and biological experiments are performed. Microfluidics allows the handling of low volumes of reagents with very precise control. The use of micro-droplets generated within microfluidic devices is of particular interest to the field of Origins of Life and Artificial Life. Whilst many efforts have been made aiming to construct cell-like compartments from modern biological constituents, these are usually very difficult to handle. However, microdroplets can be easily generated and manipulated at kHz rates, making it suitable for high-throughput experimentation and analysis of compartmentalised chemical reactions. Therefore, we decided to develop a microfluidic device capable of manipulating microdroplets in such a way that they could be efficiently mixed, split and sorted within iterative cycles. Since no microfluidic technology had been developed before in the Cronin Group, the first chapter of this thesis describes the soft lithographic methods and techniques developed to fabricate microfluidic devices. Also, special attention is placed on the generation of water-in-oil microdroplets, and the subsequent modules required for the manipulation of the droplets such as: droplet fusers, splitters, sorters and single/multi-layer micromechanical valves. Whilst the first part of this thesis describes the development of a microfluidic platform to assist chemical evolution, finding a compatible set of chemical building blocks capable of reacting to form complex molecules with endowed replicating or catalytic activity was challenging. Abstract 10 Hence, the second part of this thesis focuses on potential chemistry that will ultimately possess the properties mentioned above. A special focus is placed on the formation of peptide bonds from unactivated amino acids, despite being one of the greatest challenges in prebiotic chemistry. As opposed to classic prebiotic experiments, in which a specific set of conditions is studied to fit a particular hypothesis, we took a different approach: we explored the effects of several parameters at once on a model polymerisation reaction, without constraints on hypotheses on the nature of optimum conditions or plausibility. This was facilitated by development of a new high-throughput automated platform, allowing the exploration of a much larger number of parameters. This led us to discover that peptide bond formation is less challenging than previously imagined. Having established the right set of conditions under which peptide bond formation was enhanced, we then explored the co-oligomerisation between different amino acids, aiming for the formation of heteropeptides with different structure or function. Finally, we studied the effect of various environmental conditions (rate of evaporation, presence of salts or minerals) in the final product distribution of our oligomeric products.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Developments in theory and experiment have raised the prospect of an electronic technology based on the discrete nature of electron tunnelling through a potential barrier. This thesis deals with novel design and analysis tools developed to study such systems. Possible devices include those constructed from ultrasmall normal tunnelling junctions. These exhibit charging effects including the Coulomb blockade and correlated electron tunnelling. They allow transistor-like control of the transfer of single carriers, and present the prospect of digital systems operating at the information theoretic limit. As such, they are often referred to as single electronic devices. Single electronic devices exhibit self quantising logic and good structural tolerance. Their speed, immunity to thermal noise, and operating voltage all scale beneficially with junction capacitance. For ultrasmall junctions the possibility of room temperature operation at sub picosecond timescales seems feasible. However, they are sensitive to external charge; whether from trapping-detrapping events, externally gated potentials, or system cross-talk. Quantum effects such as charge macroscopic quantum tunnelling may degrade performance. Finally, any practical system will be complex and spatially extended (amplifying the above problems), and prone to fabrication imperfection. This summarises why new design and analysis tools are required. Simulation tools are developed, concentrating on the basic building blocks of single electronic systems; the tunnelling junction array and gated turnstile device. Three main points are considered: the best method of estimating capacitance values from physical system geometry; the mathematical model which should represent electron tunnelling based on this data; application of this model to the investigation of single electronic systems. (DXN004909)

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Complex functions, generally feature some interesting peculiarities, seen as extensions real functions, complementing the study of real analysis. However, the visualization of some complex functions properties requires the simultaneous visualization of two-dimensional spaces. The multiple Windows of GeoGebra, combined with its ability of algebraic computation with complex numbers, allow the study of the functions defined from ℂ to ℂ through traditional techniques and by the use of Domain Colouring. Here, we will show how we can use GeoGebra for the study of complex functions, using several representations and creating tools which complement the tools already provided by the software. Our proposals designed for students of the first year of engineering and science courses can and should be used as an educational tool in collaborative learning environments. The main advantage in its use in individual terms is the promotion of the deductive reasoning (conjecture / proof). In performed the literature review few references were found involving this educational topic and by the use of a single software.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The goal of Vehicle Routing Problems (VRP) and their variations is to transport a set of orders with the minimum number of vehicles at least cost. Most approaches are designed to solve specific problem variations independently, whereas in real world applications, different constraints are handled concurrently. This research extends solutions obtained for the traveling salesman problem with time windows to a much wider class of route planning problems in logistics. The work describes a novel approach that:  supports a heterogeneous fleet of vehicles  dynamically reduces the number of vehicles  respects individual capacity restrictions  satisfies pickup and delivery constraints  takes Hamiltonian paths (rather than cycles) The proposed approach uses Monte-Carlo Tree Search and in particular Nested Rollout Policy Adaptation. For the evaluation of the work, real data from the industry was obtained and tested and the results are reported.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This thesis work deals with a mathematical description of flow in polymeric pipe and in a specific peristaltic pump. This study involves fluid-structure interaction analysis in presence of complex-turbulent flows treated in an arbitrary Lagrangian-Eulerian (ALE) framework. The flow simulations are performed in COMSOL 4.4, as 2D axial symmetric model, and ABAQUS 6.14.1, as 3D model with symmetric boundary conditions. In COMSOL, the fluid and structure problems are coupled by monolithic algorithm, while ABAQUS code links ABAQUS CFD and ABAQUS Standard solvers with single block-iterative partitioned algorithm. For the turbulent features of the flow, the fluid model in both codes is described by RNG k-ϵ. The structural model is described, on the basis of the pipe material, by Elastic models or Hyperelastic Neo-Hookean models with Rayleigh damping properties. In order to describe the pulsatile fluid flow after the pumping process, the available data are often defective for the fluid problem. Engineering measurements are normally able to provide average pressure or velocity at a cross-section. This problem has been analyzed by McDonald's and Womersley's work for average pressure at fixed cross section by Fourier analysis since '50, while nowadays sophisticated techniques including Finite Elements and Finite Volumes exist to study the flow. Finally, we set up peristaltic pipe simulations in ABAQUS code, by using the same model previously tested for the fl uid and the structure.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Social network sites (SNS), such as Facebook, Google+ and Twitter, have attracted hundreds of millions of users daily since their appearance. Within SNS, users connect to each other, express their identity, disseminate information and form cooperation by interacting with their connected peers. The increasing popularity and ubiquity of SNS usage and the invaluable user behaviors and connections give birth to many applications and business models. We look into several important problems within the social network ecosystem. The first one is the SNS advertisement allocation problem. The other two are related to trust mechanisms design in social network setting, including local trust inference and global trust evaluation. In SNS advertising, we study the problem of advertisement allocation from the ad platform's angle, and discuss its differences with the advertising model in the search engine setting. By leveraging the connection between social networks and hyperbolic geometry, we propose to solve the problem via approximation using hyperbolic embedding and convex optimization. A hyperbolic embedding method, \hcm, is designed for the SNS ad allocation problem, and several components are introduced to realize the optimization formulation. We show the advantages of our new approach in solving the problem compared to the baseline integer programming (IP) formulation. In studying the problem of trust mechanisms in social networks, we consider the existence of distrust (i.e. negative trust) relationships, and differentiate between the concept of local trust and global trust in social network setting. In the problem of local trust inference, we propose a 2-D trust model. Based on the model, we develop a semiring-based trust inference framework. In global trust evaluation, we consider a general setting with conflicting opinions, and propose a consensus-based approach to solve the complex problem in signed trust networks.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Since the emergence of software engineering in the late 1960's as a response to the software crisis, researchers throughout the world are trying to give theoretical support to this discipline. Several points of view have to be reviewed in order to complete this task. In the middle 70's Frederick Brooks Jr. coined the term "silver bullet" suggesting the solution to several problems rela-ted to software engineering and, hence, we adopted such a metaphor as a symbol for this book. Methods, modeling, and teaching are the insights reviewed in this book. Some work related to these topies is presented by software engineering researchers, led by Ivar Jacobson, one of the most remarkable researchers in this area. We hope our work will contribute to advance in giving the theoretieal support that software engineering needs.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Motion planning, or trajectory planning, commonly refers to a process of converting high-level task specifications into low-level control commands that can be executed on the system of interest. For different applications, the system will be different. It can be an autonomous vehicle, an Unmanned Aerial Vehicle(UAV), a humanoid robot, or an industrial robotic arm. As human machine interaction is essential in many of these systems, safety is fundamental and crucial. Many of the applications also involve performing a task in an optimal manner within a given time constraint. Therefore, in this thesis, we focus on two aspects of the motion planning problem. One is the verification and synthesis of the safe controls for autonomous ground and air vehicles in collision avoidance scenarios. The other part focuses on the high-level planning for the autonomous vehicles with the timed temporal constraints. In the first aspect of our work, we first propose a verification method to prove the safety and robustness of a path planner and the path following controls based on reachable sets. We demonstrate the method on quadrotor and automobile applications. Secondly, we propose a reachable set based collision avoidance algorithm for UAVs. Instead of the traditional approaches of collision avoidance between trajectories, we propose a collision avoidance scheme based on reachable sets and tubes. We then formulate the problem as a convex optimization problem seeking control set design for the aircraft to avoid collision. We apply our approach to collision avoidance scenarios of quadrotors and fixed-wing aircraft. In the second aspect of our work, we address the high level planning problems with timed temporal logic constraints. Firstly, we present an optimization based method for path planning of a mobile robot subject to timed temporal constraints, in a dynamic environment. Temporal logic (TL) can address very complex task specifications such as safety, coverage, motion sequencing etc. We use metric temporal logic (MTL) to encode the task specifications with timing constraints. We then translate the MTL formulae into mixed integer linear constraints and solve the associated optimization problem using a mixed integer linear program solver. We have applied our approach on several case studies in complex dynamical environments subjected to timed temporal specifications. Secondly, we also present a timed automaton based method for planning under the given timed temporal logic specifications. We use metric interval temporal logic (MITL), a member of the MTL family, to represent the task specification, and provide a constructive way to generate a timed automaton and methods to look for accepting runs on the automaton to find an optimal motion (or path) sequence for the robot to complete the task.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Dissertação de Mestrado, Engenharia Informática, Faculdade de Ciências e Tecnologia, Universidade do Algarve, 2014

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The objective of reservoir engineering is to manage fields of oil production in order to maximize the production of hydrocarbons according to economic and physical restrictions. The deciding of a production strategy is a complex activity involving several variables in the process. Thus, a smart system, which assists in the optimization of the options for developing of the field, is very useful in day-to-day of reservoir engineers. This paper proposes the development of an intelligent system to aid decision making, regarding the optimization of strategies of production in oil fields. The intelligence of this system will be implemented through the use of the technique of reinforcement learning, which is presented as a powerful tool in problems of multi-stage decision. The proposed system will allow the specialist to obtain, in time, a great alternative (or near-optimal) for the development of an oil field known

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Background: Complex chronic diseases are a challenge for the current configuration of Health services. Case management is a service frequently provided for people with chronic conditions and despite its effectiveness in many outcomes, such as mortality or readmissions, uncertainty remains about the most effective form of team organization, structures, and the nature of the interventions. Many processes and outcomes of case management for people with complex chronic conditions cannot be addressed with the information provided by electronic clinical records. Registries are frequently used to deal with this weakness. The aim of this study was to generate a registry-based information system of patients receiving case management to identify their clinical characteristics, their context of care, events identified during their follow-up, interventions developed by case managers, and services used. Methods and design: The study was divided into three phases, covering the detection of information needs, the design and its implementation in the healthcare system, using literature review and expert consensus methods to select variables that would be included in the registry. Objective: To describe the essential characteristics of the provision of ca re lo people who receive case management (structure, process and outcomes), with special emphasis on those with complex chronic diseases. Study population: Patients from any District of Primary Care, who initiate the utilization of case management services, to avoid information bias that may occur when including subjects who have already been received the service, and whose outcomes and characteristics could not be properly collected. Results: A total of 102 variables representing structure, processes and outcomes of case management were selected for their inclusion in the registry after the consensus phase. Total sample was composed of 427 patients, of which 211 (49.4%) were women and 216 (50.6%) were men. The average functional level (Barthel lndex) was 36.18 (SD 29.02), cognitive function (Pfeiffer) showed an average of 4.37 {SD 6.57), Chat1son Comorbidity lndex, obtained a mean of 3.03 (SD 2.7) and Social Support (Duke lndex) was 34.2 % (SD 17.57). More than half of patients include in the Registry, correspond lo immobilized or transitional care for patients discharged from hospital (66.5 %). The patient's educational level was low or very low (50.4%). Caregivers overstrain (Caregiver stress index), obtained an average value of 6.09% (SD 3.53). Only 1.2 % of patients had declared their advanced directives, 58.6 had not defined the tutelage and the vast majority lived at home 98.8 %. Regarding the major events recorded at RANGE Registry, 25.8 % of the selected patients died in the first three months, 8.2 % suffered a hospital admission at least once time, 2.3%, two times, and 1.2% three times, 7.5% suffered a fall, 8.7% had pressure ulcer, 4.7% had problems with medication, and 3.3 % were institutionalized. Stroke is the more prevalent health problem recorded (25.1%), followed by hypertension (11.1%) and COPD (11.1%). Patients registered by NCMs had as main processes diabetes (16.8%) and dementia (11.3 %). The most frequent nursing diagnoses referred to the self-care deficit in various activities of daily living. Regarding to nursing interventions, described by the Nursing Intervention Classification (NIC), dementia management is the most used intervention, followed by mutual goal setting, caregiver and emotional support. Conclusions: The patient profile who receive case management services is a chronic complex patient with severe dependence, cognitive impairment, normal social support, low educational level, health problems such as stroke, hypertension or COPD, diabetes or dementia, and has an informal caregiver. At the first follow up, mortality was 19.2%, and a discrete rate of readmissions and falls.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Atmospheric scattering plays a crucial rule in degrading the performance of electro optical imaging systems operating in the visible and infra-red spectral bands, and hence limits the quality of the acquired images, either through reduction of contrast or increase of image blur. The exact nature of light scattering by atmospheric media is highly complex and depends on the types, orientations, sizes and distributions of particles constituting these media, as well as wavelengths, polarization states and directions of the propagating radiation. Here we follow the common approach for solving imaging and propagation problems by treating the propagating light through atmospheric media as composed of two main components: a direct (unscattered), and a scattered component. In this work we developed a detailed model of the effects of absorption and scattering by haze and fog atmospheric aerosols on the optical radiation propagating from the object plane to an imaging system, based on the classical theory of EM scattering. This detailed model is then used to compute the average point spread function (PSF) of an imaging system which properly accounts for the effects of the diffraction, scattering, and the appropriate optical power level of both the direct and the scattered radiation arriving at the pupil of the imaging system. Also, the calculated PSF, properly weighted for the energy contributions of the direct and scattered components is used, in combination with a radiometric model, to estimate the average number of the direct and scattered photons detected at the sensor plane, which are then used to calculate the image spectrum signal to- noise ratio (SNR) in the visible near infra-red (NIR) and mid infra-red (MIR) spectral wavelength bands. Reconstruction of images degraded by atmospheric scattering and measurement noise is then performed, up to the limit imposed by the noise effective cutoff spatial frequency of the image spectrum SNR. Key results of this research are as follows: A mathematical model based on Mie scattering theory for how scattering from aerosols affects the overall point spread function (PSF) of an imaging system was developed, coded in MATLAB, and demonstrated. This model along with radiometric theory was used to predict the limiting resolution of an imaging system as a function of the optics, scattering environment, and measurement noise. Finally, image reconstruction algorithms were developed and demonstrated which mitigate the effects of scattering-induced blurring to within the limits imposed by noise.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The focus of the current dissertation is to study qualitatively the underlying physics of vortex-shedding and wake dynamics in long aspect-ratio aerodynamics in incompressible viscous flow through the use of the KLE method. We carried out a long series of numerical experiments in the cases of flow around the cylinder at low Reynolds numbers. The study of flow at low Reynolds numbers provides an insight in the fluid physics and also plays a critical role when applying to stalled turbine rotors. Many of the conclusions about the qualitative nature of the physical mechanisms characterizing vortex formation, shedding and further interaction analyzed here at low Re could be extended to other Re regimes and help to understand the separation of the boundary layers in airfoils and other aerodynamic surfaces. In the long run, it aims to provide a better understanding of the complex multi-physics problems involving fluid-structure-control interaction through improved mathematical computational models of the multi-physics process. Besides the scientific conclusions produced, the research work on streamlined and bluff-body condition will also serve as a valuable guide for the future design of blade aerodynamics and the placement of wind turbines and hydrakinetic turbines, increasing the efficiency in the use of expensive workforce, supplies, and infrastructure. After the introductory section describing the main fields of application of wind power and hydrokinetic turbines, we describe the main features and theoretical background of the numerical method used here. Then, we present the analysis of the numerical experimentation results for the oscillatory regime right before the onset of vortex shedding for circular cylinders. We verified the wake length of the closed near-wake behind the cylinder and analysed the decay of the wake at the wake formation region, and then studied the St-Re relationship at the Reynolds numbers before the wake sheds compared to the experimental data. We found a theoretical model that describes the time evolution of the amplitude of fluctuations in the vorticity field on the twin vortex wake, which accurately matches the numerical results in terms of the frequency of the oscillation and rate of decay. We also proposed a model based on an analog circuit that is able to interpret the concerning flow by reducing the number of degrees of freedom. It follows the idea of the non-linear oscillator and resembles the dynamics mechanism of the closed near-wake with a common configured sine wave oscillator. This low-dimensional circuital model may also help to understand the underlying physical mechanisms, related to vorticity transport, that give origin to those oscillations.