970 resultados para computer algorithm


Relevância:

20.00% 20.00%

Publicador:

Resumo:

Please consult the paper edition of this thesis to read. It is available on the 5th Floor of the Library at Call Number: Z 9999 E38 D56 1992

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This thesis focuses on developing an evolutionary art system using genetic programming. The main goal is to produce new forms of evolutionary art that filter existing images into new non-photorealistic (NPR) styles, by obtaining images that look like traditional media such as watercolor or pencil, as well as brand new effects. The approach permits GP to generate creative forms of NPR results. The GP language is extended with different techniques and methods inspired from NPR research such as colour mixing expressions, image processing filters and painting algorithm. Colour mixing is a major new contribution, as it enables many familiar and innovative NPR effects to arise. Another major innovation is that many GP functions process the canvas (rendered image), while is dynamically changing. Automatic fitness scoring uses aesthetic evaluation models and statistical analysis, and multi-objective fitness evaluation is used. Results showed a variety of NPR effects, as well as new, creative possibilities.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Complex networks have recently attracted a significant amount of research attention due to their ability to model real world phenomena. One important problem often encountered is to limit diffusive processes spread over the network, for example mitigating pandemic disease or computer virus spread. A number of problem formulations have been proposed that aim to solve such problems based on desired network characteristics, such as maintaining the largest network component after node removal. The recently formulated critical node detection problem aims to remove a small subset of vertices from the network such that the residual network has minimum pairwise connectivity. Unfortunately, the problem is NP-hard and also the number of constraints is cubic in number of vertices, making very large scale problems impossible to solve with traditional mathematical programming techniques. Even many approximation algorithm strategies such as dynamic programming, evolutionary algorithms, etc. all are unusable for networks that contain thousands to millions of vertices. A computationally efficient and simple approach is required in such circumstances, but none currently exist. In this thesis, such an algorithm is proposed. The methodology is based on a depth-first search traversal of the network, and a specially designed ranking function that considers information local to each vertex. Due to the variety of network structures, a number of characteristics must be taken into consideration and combined into a single rank that measures the utility of removing each vertex. Since removing a vertex in sequential fashion impacts the network structure, an efficient post-processing algorithm is also proposed to quickly re-rank vertices. Experiments on a range of common complex network models with varying number of vertices are considered, in addition to real world networks. The proposed algorithm, DFSH, is shown to be highly competitive and often outperforms existing strategies such as Google PageRank for minimizing pairwise connectivity.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Hub Location Problems play vital economic roles in transportation and telecommunication networks where goods or people must be efficiently transferred from an origin to a destination point whilst direct origin-destination links are impractical. This work investigates the single allocation hub location problem, and proposes a genetic algorithm (GA) approach for it. The effectiveness of using a single-objective criterion measure for the problem is first explored. Next, a multi-objective GA employing various fitness evaluation strategies such as Pareto ranking, sum of ranks, and weighted sum strategies is presented. The effectiveness of the multi-objective GA is shown by comparison with an Integer Programming strategy, the only other multi-objective approach found in the literature for this problem. Lastly, two new crossover operators are proposed and an empirical study is done using small to large problem instances of the Civil Aeronautics Board (CAB) and Australian Post (AP) data sets.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The KCube interconnection topology was rst introduced in 2010. The KCube graph is a compound graph of a Kautz digraph and hypercubes. Compared with the at- tractive Kautz digraph and well known hypercube graph, the KCube graph could accommodate as many nodes as possible for a given indegree (and outdegree) and the diameter of interconnection networks. However, there are few algorithms designed for the KCube graph. In this thesis, we will concentrate on nding graph theoretical properties of the KCube graph and designing parallel algorithms that run on this network. We will explore several topological properties, such as bipartiteness, Hamiltonianicity, and symmetry property. These properties for the KCube graph are very useful to develop efficient algorithms on this network. We will then study the KCube network from the algorithmic point of view, and will give an improved routing algorithm. In addition, we will present two optimal broadcasting algorithms. They are fundamental algorithms to many applications. A literature review of the state of the art network designs in relation to the KCube network as well as some open problems in this field will also be given.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

While the influence of computer technology has been widely studied in a variety of contexts, the drawing teaching studio is a particularly interesting context because of the juxtaposition of traditional medium and computer technology. For this study, 5 Canadian postsecondary teachers engaged in a 2-round Delphi interview process to discuss their responses to computer technology on their drawing pedagogy. Data sources included transcribed interviews. Findings indicated that artist teachers are both cautious to embrace and curious to explore appropriate use of computer technology on their drawing pedagogy. Artist teachers are both critical and optimistic about the influence of computer technology.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Experimental Extended X-ray Absorption Fine Structure (EXAFS) spectra carry information about the chemical structure of metal protein complexes. However, pre- dicting the structure of such complexes from EXAFS spectra is not a simple task. Currently methods such as Monte Carlo optimization or simulated annealing are used in structure refinement of EXAFS. These methods have proven somewhat successful in structure refinement but have not been successful in finding the global minima. Multiple population based algorithms, including a genetic algorithm, a restarting ge- netic algorithm, differential evolution, and particle swarm optimization, are studied for their effectiveness in structure refinement of EXAFS. The oxygen-evolving com- plex in S1 is used as a benchmark for comparing the algorithms. These algorithms were successful in finding new atomic structures that produced improved calculated EXAFS spectra over atomic structures previously found.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Population-based metaheuristics, such as particle swarm optimization (PSO), have been employed to solve many real-world optimization problems. Although it is of- ten sufficient to find a single solution to these problems, there does exist those cases where identifying multiple, diverse solutions can be beneficial or even required. Some of these problems are further complicated by a change in their objective function over time. This type of optimization is referred to as dynamic, multi-modal optimization. Algorithms which exploit multiple optima in a search space are identified as niching algorithms. Although numerous dynamic, niching algorithms have been developed, their performance is often measured solely on their ability to find a single, global optimum. Furthermore, the comparisons often use synthetic benchmarks whose landscape characteristics are generally limited and unknown. This thesis provides a landscape analysis of the dynamic benchmark functions commonly developed for multi-modal optimization. The benchmark analysis results reveal that the mechanisms responsible for dynamism in the current dynamic bench- marks do not significantly affect landscape features, thus suggesting a lack of representation for problems whose landscape features vary over time. This analysis is used in a comparison of current niching algorithms to identify the effects that specific landscape features have on niching performance. Two performance metrics are proposed to measure both the scalability and accuracy of the niching algorithms. The algorithm comparison results demonstrate the algorithms best suited for a variety of dynamic environments. This comparison also examines each of the algorithms in terms of their niching behaviours and analyzing the range and trade-off between scalability and accuracy when tuning the algorithms respective parameters. These results contribute to the understanding of current niching techniques as well as the problem features that ultimately dictate their success.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this thesis we are going to analyze the dictionary graphs and some other kinds of graphs using the PagerRank algorithm. We calculated the correlation between the degree and PageRank of all nodes for a graph obtained from Merriam-Webster dictionary, a French dictionary and WordNet hypernym and synonym dictionaries. Our conclusion was that PageRank can be a good tool to compare the quality of dictionaries. We studied some artificial social and random graphs. We found that when we omitted some random nodes from each of the graphs, we have not noticed any significant changes in the ranking of the nodes according to their PageRank. We also discovered that some social graphs selected for our study were less resistant to the changes of PageRank.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The goal of most clustering algorithms is to find the optimal number of clusters (i.e. fewest number of clusters). However, analysis of molecular conformations of biological macromolecules obtained from computer simulations may benefit from a larger array of clusters. The Self-Organizing Map (SOM) clustering method has the advantage of generating large numbers of clusters, but often gives ambiguous results. In this work, SOMs have been shown to be reproducible when the same conformational dataset is independently clustered multiple times (~100), with the help of the Cramérs V-index (C_v). The ability of C_v to determine which SOMs are reproduced is generalizable across different SOM source codes. The conformational ensembles produced from MD (molecular dynamics) and REMD (replica exchange molecular dynamics) simulations of the penta peptide Met-enkephalin (MET) and the 34 amino acid protein human Parathyroid Hormone (hPTH) were used to evaluate SOM reproducibility. The training length for the SOM has a huge impact on the reproducibility. Analysis of MET conformational data definitively determined that toroidal SOMs cluster data better than bordered maps due to the fact that toroidal maps do not have an edge effect. For the source code from MATLAB, it was determined that the learning rate function should be LINEAR with an initial learning rate factor of 0.05 and the SOM should be trained by a sequential algorithm. The trained SOMs can be used as a supervised classification for another dataset. The toroidal 10×10 hexagonal SOMs produced from the MATLAB program for hPTH conformational data produced three sets of reproducible clusters (27%, 15%, and 13% of 100 independent runs) which find similar partitionings to those of smaller 6×6 SOMs. The χ^2 values produced as part of the C_v calculation were used to locate clusters with identical conformational memberships on independently trained SOMs, even those with different dimensions. The χ^2 values could relate the different SOM partitionings to each other.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Feature selection plays an important role in knowledge discovery and data mining nowadays. In traditional rough set theory, feature selection using reduct - the minimal discerning set of attributes - is an important area. Nevertheless, the original definition of a reduct is restrictive, so in one of the previous research it was proposed to take into account not only the horizontal reduction of information by feature selection, but also a vertical reduction considering suitable subsets of the original set of objects. Following the work mentioned above, a new approach to generate bireducts using a multi--objective genetic algorithm was proposed. Although the genetic algorithms were used to calculate reduct in some previous works, we did not find any work where genetic algorithms were adopted to calculate bireducts. Compared to the works done before in this area, the proposed method has less randomness in generating bireducts. The genetic algorithm system estimated a quality of each bireduct by values of two objective functions as evolution progresses, so consequently a set of bireducts with optimized values of these objectives was obtained. Different fitness evaluation methods and genetic operators, such as crossover and mutation, were applied and the prediction accuracies were compared. Five datasets were used to test the proposed method and two datasets were used to perform a comparison study. Statistical analysis using the one-way ANOVA test was performed to determine the significant difference between the results. The experiment showed that the proposed method was able to reduce the number of bireducts necessary in order to receive a good prediction accuracy. Also, the influence of different genetic operators and fitness evaluation strategies on the prediction accuracy was analyzed. It was shown that the prediction accuracies of the proposed method are comparable with the best results in machine learning literature, and some of them outperformed it.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The KCube interconnection network was first introduced in 2010 in order to exploit the good characteristics of two well-known interconnection networks, the hypercube and the Kautz graph. KCube links up multiple processors in a communication network with high density for a fixed degree. Since the KCube network is newly proposed, much study is required to demonstrate its potential properties and algorithms that can be designed to solve parallel computation problems. In this thesis we introduce a new methodology to construct the KCube graph. Also, with regard to this new approach, we will prove its Hamiltonicity in the general KC(m; k). Moreover, we will find its connectivity followed by an optimal broadcasting scheme in which a source node containing a message is to communicate it with all other processors. In addition to KCube networks, we have studied a version of the routing problem in the traditional hypercube, investigating this problem: whether there exists a shortest path in a Qn between two nodes 0n and 1n, when the network is experiencing failed components. We first conditionally discuss this problem when there is a constraint on the number of faulty nodes, and subsequently introduce an algorithm to tackle the problem without restrictions on the number of nodes.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Un résumé en français est également disponible.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

En salle d’opération, les tâches de l’anesthésiste sont nombreuses. Alors que l’utilisation de nouveaux outils technologiques l’informe plus fidèlement sur ce qui se passe pour son patient, ces outils font que ses tâches deviennent plus exigeantes. En vue de diminuer cette charge de travail, nous avons considérer l’administration automatique d’agents anesthésiques en se servant de contrôle en boucle fermée. À cette fin, nous avons développé un système d’administration d’un agent anesthésique (le propofol) visant à maintenir à un niveau optimal la perte de conscience du patient pendant toute la durée d’une chirurgie. Le système comprend un ordinateur, un moniteur d’anesthésie et une pompe de perfusion. L’ordinateur est doté d’un algorithme de contrôle qui, à partir d’un indice (Bispectral IndexTM ou BIS) fournit par le moniteur d’anesthésie détermine le taux d’infusion de l’agent anesthésiant. Au départ, l’anesthésiste choisit une valeur cible pour la variable de contrôle BIS et l’algorithme, basé sur système expert, calcule les doses de perfusion de propofol de sorte que la valeur mesurée de BIS se rapproche le plus possible de la valeur cible établie. Comme interface-utilisateur pour un nouveau moniteur d’anesthésie, quatre sortes d’affichage ont été considérés: purement numérique, purement graphique, un mélange entre graphique et numérique et un affichage graphique intégré (soit bidimensionnel). À partir de 20 scenarios différents où des paramètres normaux et anormaux en anesthésie étaient présentés à des anesthésistes et des résidents, l’étude des temps de réaction, de l’exactitude des réponses et de la convivialité (évaluée par le NASA-TLX) a montré qu’un affichage qui combine des éléments graphiques et numériques était le meilleur choix comme interface du système. Une étude clinique a été réalisée pour comparer le comportement du système d’administration de propofol en boucle fermée comparativement à une anesthésie contrôlée de façon manuelle et conventionnelle où le BIS était aussi utilisé. Suite à l’approbation du comité d’éthique et le consentement de personnes ayant à subir des chirurgies générales et orthopédiques, 40 patients ont été distribués également et aléatoirement soit dans le Groupe contrôle, soit dans le Groupe boucle fermée. Après l’induction manuelle de propofol (1.5 mg/kg), le contrôle en boucle fermée a été déclenché pour maintenir l’anesthésie à une cible de BIS fixée à 45. Dans l’autre groupe, le propofol a été administré à l’aide d’une pompe de perfusion et l’anesthésiste avait aussi à garder manuellement l’indice BIS le plus proche possible de 45. En fonction du BIS mesuré, la performance du contrôle exercé a été définie comme excellente pendant les moments où la valeur du BIS mesurée se situait à ±10% de la valeur cible, bonne si comprise de ±10% à ±20%, faible si comprise de ±20% à ±30% ou inadéquate lorsque >±30%. Dans le Groupe boucle fermée, le système a montré un contrôle excellent durant 55% du temps total de l’intervention, un bon contrôle durant 29% du temps et faible que pendant 9% du temps. Le temps depuis l’arrêt de la perfusion jusqu’à l’extubation est de 9 ± 3.7 min. Dans le Groupe contrôle, un contrôle excellent, bon, et faible a été enregistré durant 33%, 33% et 15% du temps respectivement et les doses ont été changées manuellement par l’anesthésiste en moyenne 9.5±4 fois par h. L’extubation a été accomplie après 11.9 ± 3.3 min de l’arrêt de la perfusion. Dans le Groupe boucle fermée, un contrôle excellent a été obtenu plus longtemps au cours des interventions (P<0.0001) et un contrôle inadéquat moins longtemps (P=0.001) que dans le Groupe contrôle. Le système en boucle fermée d’administration de propofol permet donc de maintenir plus facilement l’anesthésie au voisinage d’une cible choisie que l’administration manuelle.