899 resultados para the SIMPLE algorithm


Relevância:

90.00% 90.00%

Publicador:

Resumo:

Cutting and packing problems arise in a variety of industries, including garment, wood and shipbuilding. Irregular shape packing is a special case which admits irregular items and is much more complex due to the geometry of items. In order to ensure that items do not overlap and no item from the layout protrudes from the container, the collision free region concept was adopted. It represents all possible translations for a new item to be inserted into a container with already placed items. To construct a feasible layout, collision free region for each item is determined through a sequence of Boolean operations over polygons. In order to improve the speed of the algorithm, a parallel version of the layout construction was proposed and it was applied to a simulated annealing algorithm used to solve bin packing problems. Tests were performed in order to determine the speed improvement of the parallel version over the serial algorithm

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Recently, a rising interest in political and economic integration/disintegration issues has been developed in the political economy field. This growing strand of literature partly draws on traditional issues of fiscal federalism and optimum public good provision and focuses on a trade-off between the benefits of centralization, arising from economies of scale or externalities, and the costs of harmonizing policies as a consequence of the increased heterogeneity of individual preferences in an international union or in a country composed of at least two regions. This thesis stems from this strand of literature and aims to shed some light on two highly relevant aspects of the political economy of European integration. The first concerns the role of public opinion in the integration process; more precisely, how economic benefits and costs of integration shape citizens' support for European Union (EU) membership. The second is the allocation of policy competences among different levels of government: European, national and regional. Chapter 1 introduces the topics developed in this thesis by reviewing the main recent theoretical developments in the political economy analysis of integration processes. It is structured as follows. First, it briefly surveys a few relevant articles on economic theories of integration and disintegration processes (Alesina and Spolaore 1997, Bolton and Roland 1997, Alesina et al. 2000, Casella and Feinstein 2002) and discusses their relevance for the study of the impact of economic benefits and costs on public opinion attitude towards the EU. Subsequently, it explores the links existing between such political economy literature and theories of fiscal federalism, especially with regard to normative considerations concerning the optimal allocation of competences in a union. Chapter 2 firstly proposes a model of citizens’ support for membership of international unions, with explicit reference to the EU; subsequently it tests the model on a panel of EU countries. What are the factors that influence public opinion support for the European Union (EU)? In international relations theory, the idea that citizens' support for the EU depends on material benefits deriving from integration, i.e. whether European integration makes individuals economically better off (utilitarian support), has been common since the 1970s, but has never been the subject of a formal treatment (Hix 2005). A small number of studies in the 1990s have investigated econometrically the link between national economic performance and mass support for European integration (Eichenberg and Dalton 1993; Anderson and Kalthenthaler 1996), but only making informal assumptions. The main aim of Chapter 2 is thus to propose and test our model with a view to providing a more complete and theoretically grounded picture of public support for the EU. Following theories of utilitarian support, we assume that citizens are in favour of membership if they receive economic benefits from it. To develop this idea, we propose a simple political economic model drawing on the recent economic literature on integration and disintegration processes. The basic element is the existence of a trade-off between the benefits of centralisation and the costs of harmonising policies in presence of heterogeneous preferences among countries. The approach we follow is that of the recent literature on the political economy of international unions and the unification or break-up of nations (Bolton and Roland 1997, Alesina and Wacziarg 1999, Alesina et al. 2001, 2005a, to mention only the relevant). The general perspective is that unification provides returns to scale in the provision of public goods, but reduces each member state’s ability to determine its most favoured bundle of public goods. In the simple model presented in Chapter 2, support for membership of the union is increasing in the union’s average income and in the loss of efficiency stemming from being outside the union, and decreasing in a country’s average income, while increasing heterogeneity of preferences among countries points to a reduced scope of the union. Afterwards we empirically test the model with data on the EU; more precisely, we perform an econometric analysis employing a panel of member countries over time. The second part of Chapter 2 thus tries to answer the following question: does public opinion support for the EU really depend on economic factors? The findings are broadly consistent with our theoretical expectations: the conditions of the national economy, differences in income among member states and heterogeneity of preferences shape citizens’ attitude towards their country’s membership of the EU. Consequently, this analysis offers some interesting policy implications for the present debate about ratification of the European Constitution and, more generally, about how the EU could act in order to gain more support from the European public. Citizens in many member states are called to express their opinion in national referenda, which may well end up in rejection of the Constitution, as recently happened in France and the Netherlands, triggering a European-wide political crisis. These events show that nowadays understanding public attitude towards the EU is not only of academic interest, but has a strong relevance for policy-making too. Chapter 3 empirically investigates the link between European integration and regional autonomy in Italy. Over the last few decades, the double tendency towards supranationalism and regional autonomy, which has characterised some European States, has taken a very interesting form in this country, because Italy, besides being one of the founding members of the EU, also implemented a process of decentralisation during the 1970s, further strengthened by a constitutional reform in 2001. Moreover, the issue of the allocation of competences among the EU, the Member States and the regions is now especially topical. The process leading to the drafting of European Constitution (even if then it has not come into force) has attracted much attention from a constitutional political economy perspective both on a normative and positive point of view (Breuss and Eller 2004, Mueller 2005). The Italian parliament has recently passed a new thorough constitutional reform, still to be approved by citizens in a referendum, which includes, among other things, the so called “devolution”, i.e. granting the regions exclusive competence in public health care, education and local police. Following and extending the methodology proposed in a recent influential article by Alesina et al. (2005b), which only concentrated on the EU activity (treaties, legislation, and European Court of Justice’s rulings), we develop a set of quantitative indicators measuring the intensity of the legislative activity of the Italian State, the EU and the Italian regions from 1973 to 2005 in a large number of policy categories. By doing so, we seek to answer the following broad questions. Are European and regional legislations substitutes for state laws? To what extent are the competences attributed by the European treaties or the Italian Constitution actually exerted in the various policy areas? Is their exertion consistent with the normative recommendations from the economic literature about their optimum allocation among different levels of government? The main results show that, first, there seems to be a certain substitutability between EU and national legislations (even if not a very strong one), but not between regional and national ones. Second, the EU concentrates its legislative activity mainly in international trade and agriculture, whilst social policy is where the regions and the State (which is also the main actor in foreign policy) are more active. Third, at least two levels of government (in some cases all of them) are significantly involved in the legislative activity in many sectors, even where the rationale for that is, at best, very questionable, indicating that they actually share a larger number of policy tasks than that suggested by the economic theory. It appears therefore that an excessive number of competences are actually shared among different levels of government. From an economic perspective, it may well be recommended that some competences be shared, but only when the balance between scale or spillover effects and heterogeneity of preferences suggests so. When, on the contrary, too many levels of government are involved in a certain policy area, the distinction between their different responsibilities easily becomes unnecessarily blurred. This may not only leads to a slower and inefficient policy-making process, but also risks to make it too complicate to understand for citizens, who, on the contrary, should be able to know who is really responsible for a certain policy when they vote in national,local or European elections or in referenda on national or European constitutional issues.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The main aim of this Ph.D. dissertation is the study of clustering dependent data by means of copula functions with particular emphasis on microarray data. Copula functions are a popular multivariate modeling tool in each field where the multivariate dependence is of great interest and their use in clustering has not been still investigated. The first part of this work contains the review of the literature of clustering methods, copula functions and microarray experiments. The attention focuses on the K–means (Hartigan, 1975; Hartigan and Wong, 1979), the hierarchical (Everitt, 1974) and the model–based (Fraley and Raftery, 1998, 1999, 2000, 2007) clustering techniques because their performance is compared. Then, the probabilistic interpretation of the Sklar’s theorem (Sklar’s, 1959), the estimation methods for copulas like the Inference for Margins (Joe and Xu, 1996) and the Archimedean and Elliptical copula families are presented. In the end, applications of clustering methods and copulas to the genetic and microarray experiments are highlighted. The second part contains the original contribution proposed. A simulation study is performed in order to evaluate the performance of the K–means and the hierarchical bottom–up clustering methods in identifying clusters according to the dependence structure of the data generating process. Different simulations are performed by varying different conditions (e.g., the kind of margins (distinct, overlapping and nested) and the value of the dependence parameter ) and the results are evaluated by means of different measures of performance. In light of the simulation results and of the limits of the two investigated clustering methods, a new clustering algorithm based on copula functions (‘CoClust’ in brief) is proposed. The basic idea, the iterative procedure of the CoClust and the description of the written R functions with their output are given. The CoClust algorithm is tested on simulated data (by varying the number of clusters, the copula models, the dependence parameter value and the degree of overlap of margins) and is compared with the performance of model–based clustering by using different measures of performance, like the percentage of well–identified number of clusters and the not rejection percentage of H0 on . It is shown that the CoClust algorithm allows to overcome all observed limits of the other investigated clustering techniques and is able to identify clusters according to the dependence structure of the data independently of the degree of overlap of margins and the strength of the dependence. The CoClust uses a criterion based on the maximized log–likelihood function of the copula and can virtually account for any possible dependence relationship between observations. Many peculiar characteristics are shown for the CoClust, e.g. its capability of identifying the true number of clusters and the fact that it does not require a starting classification. Finally, the CoClust algorithm is applied to the real microarray data of Hedenfalk et al. (2001) both to the gene expressions observed in three different cancer samples and to the columns (tumor samples) of the whole data matrix.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Large scale wireless adhoc networks of computers, sensors, PDAs etc. (i.e. nodes) are revolutionizing connectivity and leading to a paradigm shift from centralized systems to highly distributed and dynamic environments. An example of adhoc networks are sensor networks, which are usually composed by small units able to sense and transmit to a sink elementary data which are successively processed by an external machine. Recent improvements in the memory and computational power of sensors, together with the reduction of energy consumptions, are rapidly changing the potential of such systems, moving the attention towards datacentric sensor networks. A plethora of routing and data management algorithms have been proposed for the network path discovery ranging from broadcasting/floodingbased approaches to those using global positioning systems (GPS). We studied WGrid, a novel decentralized infrastructure that organizes wireless devices in an adhoc manner, where each node has one or more virtual coordinates through which both message routing and data management occur without reliance on either flooding/broadcasting operations or GPS. The resulting adhoc network does not suffer from the deadend problem, which happens in geographicbased routing when a node is unable to locate a neighbor closer to the destination than itself. WGrid allow multidimensional data management capability since nodes' virtual coordinates can act as a distributed database without needing neither special implementation or reorganization. Any kind of data (both single and multidimensional) can be distributed, stored and managed. We will show how a location service can be easily implemented so that any search is reduced to a simple query, like for any other data type. WGrid has then been extended by adopting a replication methodology. We called the resulting algorithm WRGrid. Just like WGrid, WRGrid acts as a distributed database without needing neither special implementation nor reorganization and any kind of data can be distributed, stored and managed. We have evaluated the benefits of replication on data management, finding out, from experimental results, that it can halve the average number of hops in the network. The direct consequence of this fact are a significant improvement on energy consumption and a workload balancing among sensors (number of messages routed by each node). Finally, thanks to the replications, whose number can be arbitrarily chosen, the resulting sensor network can face sensors disconnections/connections, due to failures of sensors, without data loss. Another extension to {WGrid} is {W*Grid} which extends it by strongly improving network recovery performance from link and/or device failures that may happen due to crashes or battery exhaustion of devices or to temporary obstacles. W*Grid guarantees, by construction, at least two disjoint paths between each couple of nodes. This implies that the recovery in W*Grid occurs without broadcasting transmissions and guaranteeing robustness while drastically reducing the energy consumption. An extensive number of simulations shows the efficiency, robustness and traffic road of resulting networks under several scenarios of device density and of number of coordinates. Performance analysis have been compared to existent algorithms in order to validate the results.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

[EN]A new parallel algorithm for simultaneous untangling and smoothing of tetrahedral meshes is proposed in this paper. We provide a detailed analysis of its performance on shared-memory many-core computer architectures. This performance analysis includes the evaluation of execution time, parallel scalability, load balancing, and parallelism bottlenecks. Additionally, we compare the impact of three previously published graph coloring procedures on the performance of our parallel algorithm. We use six benchmark meshes with a wide range of sizes. Using these experimental data sets, we describe the behavior of the parallel algorithm for different data sizes. We demonstrate that this algorithm is highly scalable when it runs on two different high-performance many-core computers with up to 128 processors...

Relevância:

90.00% 90.00%

Publicador:

Resumo:

[EN]We describe a method for studying the classicalexperiment of the simple pendulum, consisting of a body of magnetic material oscillating through a thin conducting coil (magnetic pendulum), which according to Faraday’s law of induction generates a fluctuating current in the coil that can be transferred into a periodic signal in an oscilloscope. The set up described here allows to study the motion of the pendulum beyond what is normally considered in more basic settings, including a detailed analysis of both small and large oscillations, and the determination of the value of the acceleration of gravity.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

This thesis gathers the work carried out by the author in the last three years of research and it concerns the study and implementation of algorithms to coordinate and control a swarm of mobile robots moving in unknown environments. In particular, the author's attention is focused on two different approaches in order to solve two different problems. The first algorithm considered in this work deals with the possibility of decomposing a main complex task in many simple subtasks by exploiting the decentralized implementation of the so called \emph{Null Space Behavioral} paradigm. This approach to the problem of merging different subtasks with assigned priority is slightly modified in order to handle critical situations that can be detected when robots are moving through an unknown environment. In fact, issues can occur when one or more robots got stuck in local minima: a smart strategy to avoid deadlock situations is provided by the author and the algorithm is validated by simulative analysis. The second problem deals with the use of concepts borrowed from \emph{graph theory} to control a group differential wheel robots by exploiting the Laplacian solution of the consensus problem. Constraints on the swarm communication topology have been introduced by the use of a range and bearing platform developed at the Distributed Intelligent Systems and Algorithms Laboratory (DISAL), EPFL (Lausanne, CH) where part of author's work has been carried out. The control algorithm is validated by demonstration and simulation analysis and, later, is performed by a team of four robots engaged in a formation mission. To conclude, the capabilities of the algorithm based on the local solution of the consensus problem for differential wheel robots are demonstrated with an application scenario, where nine robots are engaged in a hunting task.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Alzheimer's disease (AD) and cancer represent two of the main causes of death worldwide. They are complex multifactorial diseases and several biochemical targets have been recognized to play a fundamental role in their development. Basing on their complex nature, a promising therapeutical approach could be represented by the so-called "Multi-Target-Directed Ligand" approach. This new strategy is based on the assumption that a single molecule could hit several targets responsible for the onset and/or progression of the pathology. In particular in AD, most currently prescribed drugs aim to increase the level of acetylcholine in the brain by inhibiting the enzyme acetylcholinesterase (AChE). However, clinical experience shows that AChE inhibition is a palliative treatment, and the simple modulation of a single target does not address AD aetiology. Research into newer and more potent anti-AD agents is thus focused on compounds whose properties go beyond AChE inhibition (such as inhibition of the enzyme β-secretase and inhibition of the aggregation of beta-amyloid). Therefore, the MTDL strategy seems a more appropriate approach for addressing the complexity of AD and may provide new drugs for tackling its multifactorial nature. In this thesis, it is described the design of new MTDLs able to tackle the multifactorial nature of AD. Such new MTDLs designed are less flexible analogues of Caproctamine, one of the first MTDL owing biological properties useful for the AD treatment. These new compounds are able to inhibit the enzymes AChE, beta-secretase and to inhibit both AChE-induced and self-induced beta-amyloid aggregation. In particular, the most potent compound of the series is able to inhibit AChE in subnanomolar range, to inhibit β-secretase in micromolar concentration and to inhibit both AChE-induced and self-induced beta-amyloid aggregation in micromolar concentration. Cancer, as AD, is a very complex pathology and many different therapeutical approaches are currently use for the treatment of such pathology. However, due to its multifactorial nature the MTDL approach could be, in principle, apply also to this pathology. Aim of this thesis has been the development of new molecules owing different structural motifs able to simultaneously interact with some of the multitude of targets responsible for the pathology. The designed compounds displayed cytotoxic activity in different cancer cell lines. In particular, the most potent compounds of the series have been further evaluated and they were able to bind DNA resulting 100-fold more potent than the reference compound Mitonafide. Furthermore, these compounds were able to trigger apoptosis through caspases activation and to inhibit PIN1 (preliminary result). This last protein is a very promising target because it is overexpressed in many human cancers, it functions as critical catalyst for multiple oncogenic pathways and in several cancer cell lines depletion of PIN1 determines arrest of mitosis followed by apoptosis induction. In conclusion, this study may represent a promising starting pint for the development of new MTDLs hopefully useful for cancer and AD treatment.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

In this thesis we study three combinatorial optimization problems belonging to the classes of Network Design and Vehicle Routing problems that are strongly linked in the context of the design and management of transportation networks: the Non-Bifurcated Capacitated Network Design Problem (NBP), the Period Vehicle Routing Problem (PVRP) and the Pickup and Delivery Problem with Time Windows (PDPTW). These problems are NP-hard and contain as special cases some well known difficult problems such as the Traveling Salesman Problem and the Steiner Tree Problem. Moreover, they model the core structure of many practical problems arising in logistics and telecommunications. The NBP is the problem of designing the optimum network to satisfy a given set of traffic demands. Given a set of nodes, a set of potential links and a set of point-to-point demands called commodities, the objective is to select the links to install and dimension their capacities so that all the demands can be routed between their respective endpoints, and the sum of link fixed costs and commodity routing costs is minimized. The problem is called non- bifurcated because the solution network must allow each demand to follow a single path, i.e., the flow of each demand cannot be splitted. Although this is the case in many real applications, the NBP has received significantly less attention in the literature than other capacitated network design problems that allow bifurcation. We describe an exact algorithm for the NBP that is based on solving by an integer programming solver a formulation of the problem strengthened by simple valid inequalities and four new heuristic algorithms. One of these heuristics is an adaptive memory metaheuristic, based on partial enumeration, that could be applied to a wider class of structured combinatorial optimization problems. In the PVRP a fleet of vehicles of identical capacity must be used to service a set of customers over a planning period of several days. Each customer specifies a service frequency, a set of allowable day-combinations and a quantity of product that the customer must receive every time he is visited. For example, a customer may require to be visited twice during a 5-day period imposing that these visits take place on Monday-Thursday or Monday-Friday or Tuesday-Friday. The problem consists in simultaneously assigning a day- combination to each customer and in designing the vehicle routes for each day so that each customer is visited the required number of times, the number of routes on each day does not exceed the number of vehicles available, and the total cost of the routes over the period is minimized. We also consider a tactical variant of this problem, called Tactical Planning Vehicle Routing Problem, where customers require to be visited on a specific day of the period but a penalty cost, called service cost, can be paid to postpone the visit to a later day than that required. At our knowledge all the algorithms proposed in the literature for the PVRP are heuristics. In this thesis we present for the first time an exact algorithm for the PVRP that is based on different relaxations of a set partitioning-like formulation. The effectiveness of the proposed algorithm is tested on a set of instances from the literature and on a new set of instances. Finally, the PDPTW is to service a set of transportation requests using a fleet of identical vehicles of limited capacity located at a central depot. Each request specifies a pickup location and a delivery location and requires that a given quantity of load is transported from the pickup location to the delivery location. Moreover, each location can be visited only within an associated time window. Each vehicle can perform at most one route and the problem is to satisfy all the requests using the available vehicles so that each request is serviced by a single vehicle, the load on each vehicle does not exceed the capacity, and all locations are visited according to their time window. We formulate the PDPTW as a set partitioning-like problem with additional cuts and we propose an exact algorithm based on different relaxations of the mathematical formulation and a branch-and-cut-and-price algorithm. The new algorithm is tested on two classes of problems from the literature and compared with a recent branch-and-cut-and-price algorithm from the literature.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Deutsch:In der vorliegenden Arbeit konnten neue Methoden zur Synthese anorganischer Materialien mit neuartiger Architektur im Mikrometer und Nanometer Maßstab beschrieben werden. Die zentrale Rolle der Formgebung basiert dabei auf der templatinduzierten Abscheidung der anorganischen Materialien auf selbstorganisierten Monoschichten. Als Substrate eignen sich goldbedampfte Glasträger und Goldkolloide, die eine Mittelstellung in der Welt der Atome bzw. Moleküle und der makroskopischen Welt der ausgedehnten Festkörper einnehmen. Auf diesen Substraten lassen sich Thiole zu einer monomolekularen Schicht adsorbieren und damit die Oberflächeneigenschaften des Substrates ändern. Ein besonderer Schwerpunkt bei dieser Arbeit stellt die Synthese speziell auf die Bedürfnisse der jeweiligen Anwendung ausgerichteten Thiole dar.Im ersten Teil der Arbeit wurden goldbedampfte Glasoberflächen als Template verwendet. Die Abscheidung von Calciumcarbonat wurde in Abhängigkeit der Schichtdicke der adsorbierten Monolage untersucht. Aragonit, eine der drei Hauptphasen des Calciumcarbonat Systems, wurde auf polyaromatischen Amid - Oberflächen mit Schichtdicken von 5 - 400 nm Dicke unter milden Bedingung abgeschieden. Die einstellbaren Parameter waren dabei die Kettenlänge des Polymers, der w-Substituent, die Bindung an die Goldoberfläche über Verwendung verschiedener Aminothiole und die Kristallisationstemperatur. Die Schichtdickeneinstellung der Polymerfilme erfolgte hierbei über einen automatisierten Synthesezyklus.Titanoxid Filme konnten auf Oberflächen strukturiert werden. Dabei kam ein speziell synthetisiertes Thiol zum Einsatz, das die Funktionalität einer Styroleinheit an der Oberflächen Grenze als auch eine Möglichkeit zur späteren Entfernung von der Oberfläche in sich vereinte. Die PDMS Stempeltechnik erzeugte dabei Mikrostrukturen auf der Goldoberfläche im Bereich von 5 bis 10 µm, die ihrerseits über die Polymerisation und Abscheidung des Polymers in den Titanoxid Film überführt werden konnten. Drei dimensionale Strukturen wurden über Goldkolloid Template erhalten. Tetraethylenglykol konnte mit einer Thiolgruppe im Austausch zu einer Hydroxylgruppe monofunktionalisiert werden. Das erhaltene Molekül wurde auf kolloidalem Gold selbstorganisiert; es entstand dabei ein wasserlösliches Goldkolloid. Die Darstellung erfolgte dabei in einer Einphasenreaktion. Die so erhaltenen Goldkolloide wurden als Krstallisationstemplate für die drei dimensionale Abscheidung von Calciumcarbonat verwendet. Es zeigte sich, dass Glykol die Kristallisation bzw. den Habitus des krsitalls bei niedrigem pH Wert modifiziert. Bei erhöhtem pH Wert (pH = 12) jedoch agieren die Glykol belegten Goldkolloide als Template und führen zu sphärisch Aggregaten. Werden Goldkolloide langkettigen Dithiolen ausgesetzt, so führt dies zu einer Aggregation und Ausfällung der Kolloide aufgrund der Vernetzung mehrer Goldkolloide mit den Thiolgruppen der Alkyldithiole. Zur Vermeidung konnte in dieser Arbeit ein halbseitig geschütztes Dithiol synthetisiert werden, mit dessen Hilfe die Aggregation unterbunden werden konnte. Das nachfolgende Entschützten der Thiolfunktion führte zu Goldkolloiden, deren Oberfläche Thiol funktionalisiert werden konnte. Die thiolaktiven Goldkolloide fungierten als template für die Abscheidung von Bleisulfid aus organisch/wässriger Lösung. Die Funktionsweise der Schutzgruppe und die Entschützung konnte mittels Plasmonenresonanz Spektroskopie verdeutlicht werden. Titanoxid / Gold / Polystyrol Komposite in Röhrenform konnten synthetisiert werden. Dazu wurde ein menschliches Haar als biologisches Templat für die Formgebung gewählt.. Durch Bedampfung des Haares mit Gold, Assemblierung eines Stryrolmonomers, welches zusätzlich eine Thiolfunktionalität trug, Polymerisation auf der Oberfläche, Abscheidung des Titanoxid Films und anschließendem Auflösen des biologischen Templates konnte eine Röhrenstruktur im Mikrometer Bereich dargestellt werden. Goldkolloide fungierten in dieser Arbeit nicht nur als Kristallisationstemplate und Formgeber, auch sie selbst wurden dahingehend modifiziert, dass sie drahtförmige Agglormerate im Nanometerbereich ausbilden. Dazu wurden Template aus Siliziumdioxid benutzt. Zum einen konnten Nanoröhren aus amorphen SiO2 in einer Sol Gel Methode dargestellt werden, zum anderen bediente sich diese Arbeit biologischer Siliziumoxid Hohlnadeln aus marinen Schwämmen isoliert. Goldkolloide wurden in die Hohlstrukturen eingebettet und die Struktur durch Ausbildung von Kolloid - Thiol Netzwerken mittels Dithiol Zugabe gefestigt. Die Gold-Nanodrähte im Bereich von 100 bis 500 nm wurden durch Auflösen des SiO2 - Templates freigelegt.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

A path integral simulation algorithm which includes a higher-order Trotter approximation (HOA)is analyzed and compared to an approach which includes the correct quantum mechanical pair interaction (effective Propagator (EPr)). It is found that the HOA algorithmconverges to the quantum limit with increasing Trotter number P as P^{-4}, while the EPr algorithm converges as P^{-2}.The convergence rate of the HOA algorithm is analyzed for various physical systemssuch as a harmonic chain,a particle in a double-well potential, gaseous argon, gaseous helium and crystalline argon. A new expression for the estimator for the pair correlation function in the HOA algorithm is derived. A new path integral algorithm, the hybrid algorithm, is developed.It combines an exact treatment of the quadratic part of the Hamiltonian and thehigher-order Trotter expansion techniques.For the discrete quantum sine-Gordon chain (DQSGC), it is shown that this algorithm works more efficiently than all other improved path integral algorithms discussed in this work. The new simulation techniques developed in this work allow the analysis of theDQSGC and disordered model systems in the highly quantum mechanical regime using path integral molecular dynamics (PIMD)and adiabatic centroid path integral molecular dynamics (ACPIMD).The ground state phonon dispersion relation is calculated for the DQSGC by the ACPIMD method.It is found that the excitation gap at zero wave vector is reduced by quantum fluctuations. Two different phases exist: One phase with a finite excitation gap at zero wave vector, and a gapless phase where the excitation gap vanishes.The reaction of the DQSGC to an external driving force is analyzed at T=0.In the gapless phase the system creeps if a small force is applied, and in the phase with a gap the system is pinned. At a critical force, the systems undergo a depinning transition in both phases and flow is induced. The analysis of the DQSGC is extended to models with disordered substrate potentials. Three different cases are analyzed: Disordered substrate potentials with roughness exponent H=0, H=1/2,and a model with disordered bond length. For all models, the ground state phonon dispersion relation is calculated.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Die vorliegende Dissertation untersucht die biogeochemischen Vorgänge in der Vegetationsschicht (Bestand) und die Rückkopplungen zwischen physiologischen und physikalischen Umweltprozessen, die das Klima und die Chemie der unteren Atmosphäre beeinflussen. Ein besondere Schwerpunkt ist die Verwendung theoretischer Ansätze zur Quantifizierung des vertikalen Austauschs von Energie und Spurengasen (Vertikalfluss) unter besonderer Berücksichtigung der Wechselwirkungen der beteiligten Prozesse. Es wird ein differenziertes Mehrschicht-Modell der Vegetation hergeleitet, implementiert, für den amazonischen Regenwald parametrisiert und auf einen Standort in Rondonia (Südwest Amazonien) angewendet, welches die gekoppelten Gleichungen zur Energiebilanz der Oberfläche und CO2-Assimilation auf der Blattskala mit einer Lagrange-Beschreibung des Vertikaltransports auf der Bestandesskala kombiniert. Die hergeleiteten Parametrisierungen beinhalten die vertikale Dichteverteilung der Blattfläche, ein normalisiertes Profil der horizontalen Windgeschwindigkeit, die Lichtakklimatisierung der Photosynthesekapazität und den Austausch von CO2 und Wärme an der Bodenoberfläche. Desweiteren werden die Berechnungen zur Photosynthese, stomatären Leitfähigkeit und der Strahlungsabschwächung im Bestand mithilfe von Feldmessungen evaluiert. Das Teilmodell zum Vertikaltransport wird im Detail unter Verwendung von 222-Radon-Messungen evaluiert. Die ``Vorwärtslösung'' und der ``inverse Ansatz'' des Lagrangeschen Dispersionsmodells werden durch den Vergleich von beobachteten und vorhergesagten Konzentrationsprofilen bzw. Bodenflüssen bewertet. Ein neuer Ansatz wird hergeleitet, um die Unsicherheiten des inversen Ansatzes aus denjenigen des Eingabekonzentrationsprofils zu quantifizieren. Für nächtliche Bedingungen wird eine modifizierte Parametrisierung der Turbulenz vorgeschlagen, welche die freie Konvektion während der Nacht im unteren Bestand berücksichtigt und im Vergleich zu früheren Abschätzungen zu deutlich kürzeren Aufenthaltszeiten im Bestand führt. Die vorhergesagte Stratifizierung des Bestandes am Tage und in der Nacht steht im Einklang mit Beobachtungen in dichter Vegetation. Die Tagesgänge der vorhergesagten Flüsse und skalaren Profile von Temperatur, H2O, CO2, Isopren und O3 während der späten Regen- und Trockenzeit am Rondonia-Standort stimmen gut mit Beobachtungen überein. Die Ergebnisse weisen auf saisonale physiologische Änderungen hin, die sich durch höhere stomatäre Leitfähigkeiten bzw. niedrigere Photosyntheseraten während der Regen- und Trockenzeit manifestieren. Die beobachteten Depositionsgeschwindigkeiten für Ozon während der Regenzeit überschreiten diejenigen der Trockenzeit um 150-250%. Dies kann nicht durch realistische physiologische Änderungen erklärt werden, jedoch durch einen zusätzlichen cuticulären Aufnahmemechanismus, möglicherweise an feuchten Oberflächen. Der Vergleich von beobachteten und vorhergesagten Isoprenkonzentrationen im Bestand weist auf eine reduzierte Isoprenemissionskapazität schattenadaptierter Blätter und zusätzlich auf eine Isoprenaufnahme des Bodens hin, wodurch sich die globale Schätzung für den tropischen Regenwald um 30% reduzieren würde. In einer detaillierten Sensitivitätsstudie wird die VOC Emission von amazonischen Baumarten unter Verwendung eines neuronalen Ansatzes in Beziehung zu physiologischen und abiotischen Faktoren gesetzt. Die Güte einzelner Parameterkombinationen bezüglich der Vorhersage der VOC Emission wird mit den Vorhersagen eines Modells verglichen, das quasi als Standardemissionsalgorithmus für Isopren dient und Licht sowie Temperatur als Eingabeparameter verwendet. Der Standardalgorithmus und das neuronale Netz unter Verwendung von Licht und Temperatur als Eingabeparameter schneiden sehr gut bei einzelnen Datensätzen ab, scheitern jedoch bei der Vorhersage beobachteter VOC Emissionen, wenn Datensätze von verschiedenen Perioden (Regen/Trockenzeit), Blattentwicklungsstadien, oder gar unterschiedlichen Spezies zusammengeführt werden. Wenn dem Netzwerk Informationen über die Temperatur-Historie hinzugefügt werden, reduziert sich die nicht erklärte Varianz teilweise. Eine noch bessere Leistung wird jedoch mit physiologischen Parameterkombinationen erzielt. Dies verdeutlicht die starke Kopplung zwischen VOC Emission und Blattphysiologie.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

We present a non linear technique to invert strong motion records with the aim of obtaining the final slip and rupture velocity distributions on the fault plane. In this thesis, the ground motion simulation is obtained evaluating the representation integral in the frequency. The Green’s tractions are computed using the discrete wave-number integration technique that provides the full wave-field in a 1D layered propagation medium. The representation integral is computed through a finite elements technique, based on a Delaunay’s triangulation on the fault plane. The rupture velocity is defined on a coarser regular grid and rupture times are computed by integration of the eikonal equation. For the inversion, the slip distribution is parameterized by 2D overlapping Gaussian functions, which can easily relate the spectrum of the possible solutions with the minimum resolvable wavelength, related to source-station distribution and data processing. The inverse problem is solved by a two-step procedure aimed at separating the computation of the rupture velocity from the evaluation of the slip distribution, the latter being a linear problem, when the rupture velocity is fixed. The non-linear step is solved by optimization of an L2 misfit function between synthetic and real seismograms, and solution is searched by the use of the Neighbourhood Algorithm. The conjugate gradient method is used to solve the linear step instead. The developed methodology has been applied to the M7.2, Iwate Nairiku Miyagi, Japan, earthquake. The estimated magnitude seismic moment is 2.6326 dyne∙cm that corresponds to a moment magnitude MW 6.9 while the mean the rupture velocity is 2.0 km/s. A large slip patch extends from the hypocenter to the southern shallow part of the fault plane. A second relatively large slip patch is found in the northern shallow part. Finally, we gave a quantitative estimation of errors associates with the parameters.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

This study examines the case of Vietnam and uses the method of process tracing to explore the sources of foreign policy choice and change. Foreign policy is derived from grand strategy, which refers to the full package of a state’s domestic and foreign policies. I argue that a state’s grand strategy results from the interaction of four factors—its society’s historical experience, social motivation, international power, and political contest among domestic groups. Grand strategies emerge as a response to perceived shifts in the balance of international economic, political, and military power. However, this is not to say that international pressures and incentives are translated into foreign policy. Rather, pressures and incentives are given meaning by worldviews, which reflect a society’s historical experiences of its place in the international system at traumatic junctures of its encounter with the outside world. Strategic changes in foreign policy follow what I call the “strategic algorithm,” which incorporates four major mechanisms—balancing against threat, bandwagoning with power, learning, and survival by transformation. This case study generates hypotheses for a theory of strategic choice, a theory of foreign policy transformation, and a theory of grand strategy emergence.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Among the scientific objectives addressed by the Radio Science Experiment hosted on board the ESA mission BepiColombo is the retrieval of the rotational state of planet Mercury. In fact, the estimation of the obliquity and the librations amplitude were proven to be fundamental for constraining the interior composition of Mercury. This is accomplished by the Mercury Orbiter Radio science Experiment (MORE) via a strict interaction among different payloads thus making the experiment particularly challenging. The underlying idea consists in capturing images of the same landmark on the surface of the planet in different epochs in order to observe a displacement of the identified features with respect to a nominal rotation which allows to estimate the rotational parameters. Observations must be planned accurately in order to obtain image pairs carrying the highest information content for the following estimation process. This is not a trivial task especially in light of the several dynamical constraints involved. Another delicate issue is represented by the pattern matching process between image pairs for which the lowest correlation errors are desired. The research activity was conducted in the frame of the MORE rotation experiment and addressed the design and implementation of an end-to-end simulator of the experiment with the final objective of establishing an optimal science planning of the observations. In the thesis, the implementation of the singular modules forming the simulator is illustrated along with the simulations performed. The results obtained from the preliminary release of the optimization algorithm are finally presented although the software implemented is only at a preliminary release and will be improved and refined in the future also taking into account the developments of the mission.