890 resultados para Branch and Bound algorithm
Resumo:
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)
Resumo:
This thesis gathers the work carried out by the author in the last three years of research and it concerns the study and implementation of algorithms to coordinate and control a swarm of mobile robots moving in unknown environments. In particular, the author's attention is focused on two different approaches in order to solve two different problems. The first algorithm considered in this work deals with the possibility of decomposing a main complex task in many simple subtasks by exploiting the decentralized implementation of the so called \emph{Null Space Behavioral} paradigm. This approach to the problem of merging different subtasks with assigned priority is slightly modified in order to handle critical situations that can be detected when robots are moving through an unknown environment. In fact, issues can occur when one or more robots got stuck in local minima: a smart strategy to avoid deadlock situations is provided by the author and the algorithm is validated by simulative analysis. The second problem deals with the use of concepts borrowed from \emph{graph theory} to control a group differential wheel robots by exploiting the Laplacian solution of the consensus problem. Constraints on the swarm communication topology have been introduced by the use of a range and bearing platform developed at the Distributed Intelligent Systems and Algorithms Laboratory (DISAL), EPFL (Lausanne, CH) where part of author's work has been carried out. The control algorithm is validated by demonstration and simulation analysis and, later, is performed by a team of four robots engaged in a formation mission. To conclude, the capabilities of the algorithm based on the local solution of the consensus problem for differential wheel robots are demonstrated with an application scenario, where nine robots are engaged in a hunting task.
Resumo:
Im Rahmen dieser Arbeit wurde eine Methode entwickelt, Perylendiimidfarbstoffe mit Oligonucleotiden in der Lösung zu verknüpfen. Das Ziel der Arbeit war die nicht-kovalente Synthese von Perylendiimid-DNA- und Protein- supramolekularen Strukturen. Dabei werden die molekularen Erkennungseigenschaften von DNA und Proteinen zunutze gemacht. Insgesamt drei Themenbereiche wurden dabei betrachtet: 1. Synthese und Hybridisierung von symmetrischen und asymmetrischen Perylendiimid-bis(oligonucleotid)-konjugaten für die Bildung supramolekularer Strukturen, 2. Erzeugung von Oberflächenstrukturen auf der Basis von Streptavidin-Perylendiimid-Komplexen, 3. Synthese wasserlöslicher Rylenfarbstoffe für Anwendungen in biologischen Systemen. Zur Synthese und Hybridisierung von Perylendiimid-Oligonucleotid-Konjugaten wurde eine neue Idee verfolgt und erfolgreich realisiert. Dabei handelt es sich um die Synthese von Perylendiimid-DNA-Polymeren durch nicht-kovalente Bindungen. Die Basis des entwickelten Konzepts ist die Ausnutzung der Erkennungseigenschaften der DNA, um Perylendiimidmoleküle in eine lineare Makrostruktur zu organisieren, was sonst nur durch komplizierte chemische Polymersynthese zugänglich wäre. Die Selbstorganisation von zwei komplementären Perylendiimid-bis(oligonucleotid)-konjugaten (PODN1 und PODN2), die an der 5`-Position verknüpft sind, führte zu einem linearen Perylendiimid-DNA-Polymer in der Form von …ABABABAB…., das mit Hilfe von Gelelektrophorese charakterisiert wurde. Eindrucksvoll war auch die erfolgreiche Kopplung des hydrophoben Perylendiimids mit zwei unterschiedlichen Oligonucleotidsequenzen in der Lösung, um asymmetrische Perylendiimid-bis(oligonucleotid)-konjugate zu synthetisieren. Mit solchen asymmetrischen Konjugaten konnte die programmierbare Selbstorganisation der Perylendiimid-Oligonucleotide zu einer definierten Polymerstruktur realisiert werden. Die Synthese von PDI-(biotin)2 wurde vorgestellt. Durch die spezifische Erkennungseigenschaft zwischen Biotin und Streptavidin ist es möglich, eine Oberflächenstruktur zu bilden. Die Immobilisierungsexperimente zeigten, dass das PDI (biotin)2 Streptavidin erkennen und binden kann. Dabei konnte eine multischichtige Nanostruktur (5 Doppelschichten) auf einer Goldoberfläche.
Resumo:
Präsentiert wird ein vollständiger, exakter und effizienter Algorithmus zur Berechnung des Nachbarschaftsgraphen eines Arrangements von Quadriken (Algebraische Flächen vom Grad 2). Dies ist ein wichtiger Schritt auf dem Weg zur Berechnung des vollen 3D Arrangements. Dabei greifen wir auf eine bereits existierende Implementierung zur Berechnung der exakten Parametrisierung der Schnittkurve von zwei Quadriken zurück. Somit ist es möglich, die exakten Parameterwerte der Schnittpunkte zu bestimmen, diese entlang der Kurven zu sortieren und den Nachbarschaftsgraphen zu berechnen. Wir bezeichnen unsere Implementierung als vollständig, da sie auch die Behandlung aller Sonderfälle wie singulärer oder tangentialer Schnittpunkte einschließt. Sie ist exakt, da immer das mathematisch korrekte Ergebnis berechnet wird. Und schließlich bezeichnen wir unsere Implementierung als effizient, da sie im Vergleich mit dem einzigen bisher implementierten Ansatz gut abschneidet. Implementiert wurde unser Ansatz im Rahmen des Projektes EXACUS. Das zentrale Ziel von EXACUS ist es, einen Prototypen eines zuverlässigen und leistungsfähigen CAD Geometriekerns zu entwickeln. Obwohl wir das Design unserer Bibliothek als prototypisch bezeichnen, legen wir dennoch größten Wert auf Vollständigkeit, Exaktheit, Effizienz, Dokumentation und Wiederverwendbarkeit. Über den eigentlich Beitrag zu EXACUS hinaus, hatte der hier vorgestellte Ansatz durch seine besonderen Anforderungen auch wesentlichen Einfluss auf grundlegende Teile von EXACUS. Im Besonderen hat diese Arbeit zur generischen Unterstützung der Zahlentypen und der Verwendung modularer Methoden innerhalb von EXACUS beigetragen. Im Rahmen der derzeitigen Integration von EXACUS in CGAL wurden diese Teile bereits erfolgreich in ausgereifte CGAL Pakete weiterentwickelt.
Resumo:
Oggetto di indagine del lavoro è il movimento ambientalista e culturale delle Città in Transizione che rappresentano esperimenti di ri-localizzazione delle risorse volte a preparare le comunità (paesi, città, quartieri) ad affrontare la duplice sfida del cambiamento climatico e del picco del petrolio. A partire dal Regno Unito, la rete delle Transition Towns si è in pochi anni estesa significativamente e conta oggi più di mille iniziative. L’indagine di tale movimento ha richiesto in prima battuta di focalizzare l’attenzione sul campo disciplinare della sociologia dell’ambiente. L’attenzione si è concentrata sul percorso di riconoscimento che ha reso la sociologia dell’ambiente una branca autonoma e sul percorso teorico-concettuale che ha caratterizzato la profonda spaccatura paradigmatica proposta da Catton e Dunlap, che hanno introdotto nel dibattito sociologico il Nuovo Paradigma Ecologico, prendendo le distanze dalla tradizionale visione antropocentrica della sociologia classica. Vengono poi presentate due delle più influenti prospettive teoriche della disciplina, quella del Treadmill of Production e la più attuale teoria della modernizzazione ecologica. La visione che viene adottata nel lavoro di tesi è quella proposta da Spaargaren, fautore della teoria della modernizzazione ecologica, secondo il quale la sociologia dell’ambiente può essere collocata in uno spazio intermedio che sta tra le scienze ambientali e la sociologia generale, evidenziando una vocazione interdisciplinare richiamata anche dal dibattito odierno sulla sostenibilità. Ma le evidenze empiriche progressivamente scaturite dallo studio di questo movimento che si autodefinisce culturale ed ambientalista hanno richiesto una cornice teorica più ampia, quella della modernità riflessiva e della società del rischio, in grado di fornire categorie concettuali spendibili nella descrizione dei problemi ambientali e nell’indagine del mutamento socio-culturale e dei suoi attori. I riferimenti empirici dello studio sono tre specifiche realtà locali in Transizione: York in Transition per il Regno Unito, Monteveglio (Bo) e Scandiano (Re) in Transizione per l’Italia.
Resumo:
Der Austausch der NO2-Konzentration zwischen der Atmosphäre und verschiedenen Bäumen (Betula pendula, Fagus sylvatica, Quercus robur, Quercus ilex und Pinus sylvestris) wurde mit einer Dynamischen Küvette gemessen. Die NO2-Konzentrationen wurden mit einem CLD 780 TR Analysator in Verbindung mit einem PLC 762 gemessen. Die experimentellen Untersuchungen wurden im Dunkeln und unter zwei Lichtintensitäts-Niveaus (PAR, 450 und 900 µmol m-2 s-1) und sechs verschiedene NO2-Konzentrationen zwischen 0 - 5.0 ppb durchgeführt. Der stomatäre Einfluss wurde unter Einsatz des Hormons Abscisinsäure untersucht. Die Umgebungsparameter (Lufttemperatur und Luftfeuchtigkeit) wurden konstant gehalten. Die Daten zeigten klar und deutlich den dominanten Einfluss der jeweiligen Baumspezies auf die NO2-Konzentrationen innerhalb der Küvette. Die Ergebnisse dieser Arbeit belegen bei allen Spezies eine lineare Abhängigkeit der NO2-Austauschrate mit der NO2-Umgebungskozentration und mit der stomatären Leitfähigkeit. Das Vorhandensein eines Kompensationspunkt wird nicht bestätigt.
Resumo:
Future wireless communications systems are expected to be extremely dynamic, smart and capable to interact with the surrounding radio environment. To implement such advanced devices, cognitive radio (CR) is a promising paradigm, focusing on strategies for acquiring information and learning. The first task of a cognitive systems is spectrum sensing, that has been mainly studied in the context of opportunistic spectrum access, in which cognitive nodes must implement signal detection techniques to identify unused bands for transmission. In the present work, we study different spectrum sensing algorithms, focusing on their statistical description and evaluation of the detection performance. Moving from traditional sensing approaches we consider the presence of practical impairments, and analyze algorithm design. Far from the ambition of cover the broad spectrum of spectrum sensing, we aim at providing contributions to the main classes of sensing techniques. In particular, in the context of energy detection we studied the practical design of the test, considering the case in which the noise power is estimated at the receiver. This analysis allows to deepen the phenomenon of the SNR wall, providing the conditions for its existence and showing that presence of the SNR wall is determined by the accuracy of the noise power estimation process. In the context of the eigenvalue based detectors, that can be adopted by multiple sensors systems, we studied the practical situation in presence of unbalances in the noise power at the receivers. Then, we shift the focus from single band detectors to wideband sensing, proposing a new approach based on information theoretic criteria. This technique is blind and, requiring no threshold setting, can be adopted even if the statistical distribution of the observed data in not known exactly. In the last part of the thesis we analyze some simple cooperative localization techniques based on weighted centroid strategies.
Resumo:
The Hilbert transform is an important tool in both pure and applied mathematics. It is largely used in the field of signal processing. Lately has been used in mathematical finance as the fast Hilbert transform method is an efficient and accurate algorithm for pricing discretely monitored barrier and Bermudan style options. The purpose of this report is to show the basic properties of the Hilbert transform and to check the domain of definition of this operator.
Resumo:
In questa tesi viene analizzato un problema di ottimizzazione proposto da alcuni esercizi commerciali che hanno la necessita` di selezionare e disporre i propri ar- ticoli in negozio. Il problema nasce dall’esigenza di massimizzare il profitto com- plessivo atteso dei prodotti in esposizione, trovando per ognuno una locazione sugli scaffali. I prodotti sono suddivisi in dipartimenti, dai quali solo un ele- mento deve essere selezionato ed esposto. In oltre si prevede la possibilita` di esprimere vincoli sulla locazione e compatibilita` dei prodotti. Il problema risul- tante `e una generalizzazione dei gia` noti Multiple-Choice Knapsack Problem e Multiple Knapsack Problem. Dopo una ricerca esaustiva in letteratura si `e ev- into che questo problema non `e ancora stato studiato. Si `e quindi provveduto a formalizzare il problema mediante un modello di programmazione lineare intera. Si propone un algoritmo esatto per la risoluzione del problema basato su column generation e branch and price. Sono stati formulati quattro modelli differenti per la risoluzione del pricing problem su cui si basa il column generation, per individuare quale sia il piu` efficiente. Tre dei quattro modelli proposti hanno performance comparabili, mentre l’ultimo si `e rivelato piu` inefficiente. Dai risul- tati ottenuti si evince che il metodo risolutivo proposto `e adatto a istanze di dimensione medio-bassa.
Resumo:
Introduction: Handlebar injuries in children may lead to severe organ lesions despite minimal initial signs and without visible skin bruise. We present our experiences applying a diagnostic and therapeutic algorithm for blunt abdominal trauma, and present the history of two selected cases. Materials and methods: We retrospectively assessed the charts of children below 16 years of age, only who were observed for 24 h or more in our institution due to a handlebar injury between 2004 and 2011. All children were treated according to an institutional algorithm. Results: 40 patients with a median age of 9.5 years were included. Diagnosed lesions were: ruptures of the liver (n = 6), spleen (n = 5), kidney (n = 1), and pancreas (n = 2), small bowel perforation (n = 3), and hernias of the abdominal (n = 2) or thoracic wall (n = 1). Surgical interventions were performed in 8 patients. The outcome was favorable in all the cases. Overall median hospitalization duration was 4.5 days (range 1–19 days). The overall duration between the accident and arrival at our emergency unit was 2.75 h (median, range 1–19 h). 20 children presented directly at our emergency unit after a median of 1.7 h (range 1–19.5 h). 20 children were referred by a family physician or a primary hospital after a median of 4.0 h (range 1–46 h). Conclusion: Handlebar injuries in children resulted in serious trunk lesions in half of the present patient series. The spectrum of injuries in handlebar accidents varies widely, especially injuries to the abdomen can unmask often only in the course. We advocate close observation of patients with thoracic and abdominal handlebar injuries which may be regarded as blunt stab wounds. An institutional algorithm for blunt abdominal trauma management is supportive for emergency care in patients with handlebar injuries.
Resumo:
Intraneural Ganglion Cysts expand within in a nerve, causing neurological deficits in afflicted patients. Modeling the propagation of these cysts, originating in the articular branch and then expanding radially outward, will help prove articular theory, and ultimately allow for more purposeful treatment of this condition. In Finite Element Analysis, traditional Lagrangian meshing methods fail to model the excessive deformation that occurs in the propagation of these cysts. This report explores the method of manual adaptive remeshing as a method to allow for the use of Lagrangian meshing, while circumventing the severe mesh distortions typical of using a Lagrangian mesh with a large deformation. Manual adaptive remeshing is the process of remeshing a deformed meshed part and then reapplying loads in order to achieve a larger deformation than a single mesh can achieve without excessive distortion. The methods of manual adaptive remeshing described in this Master’s Report are sufficient in modeling large deformations.
Resumo:
BACKGROUND: This study investigated the role of a negative FAST in the diagnostic and therapeutic algorithm of multiply injured patients with liver or splenic lesions. METHODS: A retrospective analysis of 226 multiply injured patients with liver or splenic lesions treated at Bern University Hospital, Switzerland. RESULTS: FAST failed to detect free fluid or organ lesions in 45 of 226 patients with spleen or liver injuries (sensitivity 80.1%). Overall specificity was 99.5%. The positive and negative predictive values were 99.4% and 83.3%. The overall likelihood ratios for a positive and negative FAST were 160.2 and 0.2. Grade III-V organ lesions were detected more frequently than grade I and II lesions. Without the additional diagnostic accuracy of a CT scan, the mean ISS of the FAST-false-negative patients would be significantly underestimated and 7 previously unsuspected intra-abdominal injuries would have been missed. CONCLUSION: FAST is an expedient tool for the primary assessment of polytraumatized patients to rule out high grade intra-abdominal injuries. However, the low overall diagnostic sensitivity of FAST may lead to underestimated injury patterns and delayed complications may occur. Hence, in hemodynamically stable patients with abdominal trauma, an early CT scan should be considered and one must be aware of the potential shortcomings of a "negative FAST".
Resumo:
Fuzzy community detection is to identify fuzzy communities in a network, which are groups of vertices in the network such that the membership of a vertex in one community is in [0,1] and that the sum of memberships of vertices in all communities equals to 1. Fuzzy communities are pervasive in social networks, but only a few works have been done for fuzzy community detection. Recently, a one-step forward extension of Newman’s Modularity, the most popular quality function for disjoint community detection, results into the Generalized Modularity (GM) that demonstrates good performance in finding well-known fuzzy communities. Thus, GMis chosen as the quality function in our research. We first propose a generalized fuzzy t-norm modularity to investigate the effect of different fuzzy intersection operators on fuzzy community detection, since the introduction of a fuzzy intersection operation is made feasible by GM. The experimental results show that the Yager operator with a proper parameter value performs better than the product operator in revealing community structure. Then, we focus on how to find optimal fuzzy communities in a network by directly maximizing GM, which we call it Fuzzy Modularity Maximization (FMM) problem. The effort on FMM problem results into the major contribution of this thesis, an efficient and effective GM-based fuzzy community detection method that could automatically discover a fuzzy partition of a network when it is appropriate, which is much better than fuzzy partitions found by existing fuzzy community detection methods, and a crisp partition of a network when appropriate, which is competitive with partitions resulted from the best disjoint community detections up to now. We address FMM problem by iteratively solving a sub-problem called One-Step Modularity Maximization (OSMM). We present two approaches for solving this iterative procedure: a tree-based global optimizer called Find Best Leaf Node (FBLN) and a heuristic-based local optimizer. The OSMM problem is based on a simplified quadratic knapsack problem that can be solved in linear time; thus, a solution of OSMM can be found in linear time. Since the OSMM algorithm is called within FBLN recursively and the structure of the search tree is non-deterministic, we can see that the FMM/FBLN algorithm runs in a time complexity of at least O (n2). So, we also propose several highly efficient and very effective heuristic algorithms namely FMM/H algorithms. We compared our proposed FMM/H algorithms with two state-of-the-art community detection methods, modified MULTICUT Spectral Fuzzy c-Means (MSFCM) and Genetic Algorithm with a Local Search strategy (GALS), on 10 real-world data sets. The experimental results suggest that the H2 variant of FMM/H is the best performing version. The H2 algorithm is very competitive with GALS in producing maximum modularity partitions and performs much better than MSFCM. On all the 10 data sets, H2 is also 2-3 orders of magnitude faster than GALS. Furthermore, by adopting a simply modified version of the H2 algorithm as a mutation operator, we designed a genetic algorithm for fuzzy community detection, namely GAFCD, where elite selection and early termination are applied. The crossover operator is designed to make GAFCD converge fast and to enhance GAFCD’s ability of jumping out of local minimums. Experimental results on all the data sets show that GAFCD uncovers better community structure than GALS.
Resumo:
Im operativen Betrieb einer Stückgutspeditionsanlage entscheidet der Betriebslenker bzw. der Disponent in einem ersten Schritt darüber, an welche Tore die Fahrzeuge zur Be- und Entladung andocken sollen. Darüber hinaus muss er für jede Tour ein Zeitfenster ausweisen innerhalb dessen sie das jeweilige Tor belegt. Durch die örtliche und zeitliche Fahrzeug-Tor-Zuordnung wird der für den innerbetrieblichen Umschlagprozess erforderliche Ressourcenaufwand in Form von zu fahrenden Wegstrecken oder aber Gabelstaplerstunden bestimmt. Ein Ziel der Planungsaufgabe ist somit, die Zuordnung der Fahrzeuge an die Tore so vorzunehmen, dass dabei minimale innerbetriebliche Wegstrecken entstehen. Dies führt zu einer minimalen Anzahl an benötigten Umschlagmittelressourcen. Darüber hinaus kann es aber auch zweckmäßig sein, die Fahrzeuge möglichst früh an die Tore anzudocken. Jede Tour verfügt über einen individuellen Fahrplan, der Auskunft über den Ankunftszeitpunkt sowie den Abfahrtszeitpunkt der jeweiligen Tour von der Anlage gibt. Nur innerhalb dieses Zeitfensters darf der Disponent die Tour einem der Tore zuweisen. Geschieht die Zuweisung nicht sofort nach Ankunft in der Anlage, so muss das Fahrzeug auf einer Parkfläche warten. Eine Minimierung der Wartezeiten ist wünschenswert, damit das Gelände der Anlage möglichst nicht durch zuviele Fahrzeuge gleichzeitig belastet wird. Es kann vor allem aber auch im Hinblick auf das Reservieren der Tore für zeitkritische Touren sinnvoll sein, Fahrzeuge möglichst früh abzufertigen. Am Lehrstuhl Verkehrssysteme und -logistik (VSL) der Universität Dortmund wurde die Entscheidungssituation im Rahmen eines Forschungsprojekts bei der Stiftung Industrieforschung in Anlehnung an ein zeitdiskretes Mehrgüterflussproblem mit unsplittable flow Bedingungen modelliert. Die beiden Zielsetzungen wurden dabei in einer eindimensionalen Zielfunktion integriert. Das resultierende Mixed Integer Linear Programm (MILP) wurde programmiert und für mittlere Szenarien durch Eingabe in den Optimization Solver CPlex mit dem dort implementierten exakten Branch-and-Cut Verfahren gelöst. Parallel wurde im Rahmen einer Kooperation zwischen dem Lehrstuhl VSL und dem Unternehmen hafa Docking Systems, einem der weltweit führenden Tor und Rampenhersteller, für die gleiche Planungsaufgabe ein heuristisches Scheduling Verfahren sowie ein Dispositionsleitstand namens LoadDock Navigation entwickelt. Der Dispositionsleitstand dient der optimalen Steuerung der Torbelegungen in logistischen Anlagen. In dem Leitstand wird planerische Intelligenz in Form des heuristischen Schedulingverfahrens, technische Neuerungen in der Rampentechnik in Form von Sensoren und das Expertenwissen des Disponenten in einem Tool verbunden. Das mathematische Modell sowie der Prototyp mit der integrierten Heuristik werden im Rahmen dieses Artikels vorgestellt.
Resumo:
We present in this paper several contributions on the collision detection optimization centered on hardware performance. We focus on the broad phase which is the first step of the collision detection process and propose three new ways of parallelization of the well-known Sweep and Prune algorithm. We first developed a multi-core model takes into account the number of available cores. Multi-core architecture enables us to distribute geometric computations with use of multi-threading. Critical writing section and threads idling have been minimized by introducing new data structures for each thread. Programming with directives, like OpenMP, appears to be a good compromise for code portability. We then proposed a new GPU-based algorithm also based on the "Sweep and Prune" that has been adapted to multi-GPU architectures. Our technique is based on a spatial subdivision method used to distribute computations among GPUs. Results show that significant speed-up can be obtained by passing from 1 to 4 GPUs in a large-scale environment.