902 resultados para Dynamic search fireworks algorithm with covariance mutation


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Imagery registration is a fundamental step, which greatly affects later processes in image mosaic, multi-spectral image fusion, digital surface modelling, etc., where the final solution needs blending of pixel information from more than one images. It is highly desired to find a way to identify registration regions among input stereo image pairs with high accuracy, particularly in remote sensing applications in which ground control points (GCPs) are not always available, such as in selecting a landing zone on an outer space planet. In this paper, a framework for localization in image registration is developed. It strengthened the local registration accuracy from two aspects: less reprojection error and better feature point distribution. Affine scale-invariant feature transform (ASIFT) was used for acquiring feature points and correspondences on the input images. Then, a homography matrix was estimated as the transformation model by an improved random sample consensus (IM-RANSAC) algorithm. In order to identify a registration region with a better spatial distribution of feature points, the Euclidean distance between the feature points is applied (named the S criterion). Finally, the parameters of the homography matrix were optimized by the Levenberg–Marquardt (LM) algorithm with selective feature points from the chosen registration region. In the experiment section, the Chang’E-2 satellite remote sensing imagery was used for evaluating the performance of the proposed method. The experiment result demonstrates that the proposed method can automatically locate a specific region with high registration accuracy between input images by achieving lower root mean square error (RMSE) and better distribution of feature points.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

P>Estimates of effective elastic thickness (T(e)) for the western portion of the South American Plate using, independently, forward flexural modelling and coherence analysis, suggest different thermomechanical properties for the same continental lithosphere. We present a review of these T(e) estimates and carry out a critical reappraisal using a common methodology of 3-D finite element method to solve a differential equation for the bending of a thin elastic plate. The finite element flexural model incorporates lateral variations of T(e) and the Andes topography as the load. Three T(e) maps for the entire Andes were analysed: Stewart & Watts (1997), Tassara et al. (2007) and Perez-Gussinye et al. (2007). The predicted flexural deformation obtained for each T(e) map was compared with the depth to the base of the foreland basin sequence. Likewise, the gravity effect of flexurally induced crust-mantle deformation was compared with the observed Bouguer gravity. T(e) estimates using forward flexural modelling by Stewart & Watts (1997) better predict the geological and gravity data for most of the Andean system, particularly in the Central Andes, where T(e) ranges from greater than 70 km in the sub-Andes to less than 15 km under the Andes Cordillera. The misfit between the calculated and observed foreland basin subsidence and the gravity anomaly for the Maranon basin in Peru and the Bermejo basin in Argentina, regardless of the assumed T(e) map, may be due to a dynamic topography component associated with the shallow subduction of the Nazca Plate beneath the Andes at these latitudes.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

We describe three patients with a comparable deletion encompassing SLC25A43, SLC25A5, CXorf56, UBE2A, NKRF, and two non-coding RNA genes, U1 and LOC100303728. Moderate to severe intellectual disability (ID), psychomotor retardation, severely impaired/absent speech, seizures, and urogenital anomalies were present in all three patients. Facial dysmorphisms include ocular hypertelorism, synophrys, and a depressed nasal bridge. These clinical features overlap with those described in two patients from a family with a similar deletion at Xq24 that also includes UBE2A, and in several patients of Brazilian and Polish families with point mutations in UBE2A. Notably, all five patients with an Xq24 deletion have ventricular septal defects that are not present inpatients with a point mutation, which might be attributed to the deletion of SLC25A5. Taken together, the UBE2A deficiency syndrome in male patients with a mutation in or a deletion of UBE2A is characterized by ID, absent speech, seizures, urogenital anomalies, frequently including a small penis, and skin abnormalities, which include generalized hirsutism, low posterior hairline, myxedematous appearance, widely spaced nipples, and hair whorls. Facial dysmorphisms include a wide face, a depressed nasal bridge, a large mouth with downturned corners, thin vermilion, and a short, broad neck. (C) 2010 Wiley-Liss, Inc.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In this paper, we present an algorithm for cluster analysis that integrates aspects from cluster ensemble and multi-objective clustering. The algorithm is based on a Pareto-based multi-objective genetic algorithm, with a special crossover operator, which uses clustering validation measures as objective functions. The algorithm proposed can deal with data sets presenting different types of clusters, without the need of expertise in cluster analysis. its result is a concise set of partitions representing alternative trade-offs among the objective functions. We compare the results obtained with our algorithm, in the context of gene expression data sets, to those achieved with multi-objective Clustering with automatic K-determination (MOCK). the algorithm most closely related to ours. (C) 2009 Elsevier B.V. All rights reserved.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

For a fixed family F of graphs, an F-packing in a graph G is a set of pairwise vertex-disjoint subgraphs of G, each isomorphic to an element of F. Finding an F-packing that maximizes the number of covered edges is a natural generalization of the maximum matching problem, which is just F = {K(2)}. In this paper we provide new approximation algorithms and hardness results for the K(r)-packing problem where K(r) = {K(2), K(3,) . . . , K(r)}. We show that already for r = 3 the K(r)-packing problem is APX-complete, and, in fact, we show that it remains so even for graphs with maximum degree 4. On the positive side, we give an approximation algorithm with approximation ratio at most 2 for every fixed r. For r = 3, 4, 5 we obtain better approximations. For r = 3 we obtain a simple 3/2-approximation, achieving a known ratio that follows from a more involved algorithm of Halldorsson. For r = 4, we obtain a (3/2 + epsilon)-approximation, and for r = 5 we obtain a (25/14 + epsilon)-approximation. (C) 2008 Elsevier B.V. All rights reserved.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

We consider the problems of finding the maximum number of vertex-disjoint triangles (VTP) and edge-disjoint triangles (ETP) in a simple graph. Both problems are NP-hard. The algorithm with the best approximation ratio known so far for these problems has ratio 3/2 + epsilon, a result that follows from a more general algorithm for set packing obtained by Hurkens and Schrijver [On the size of systems of sets every t of which have an SDR, with an application to the worst-case ratio of heuristics for packing problems, SIAM J. Discrete Math. 2(1) (1989) 68-72]. We present improvements on the approximation ratio for restricted cases of VTP and ETP that are known to be APX-hard: we give an approximation algorithm for VTP on graphs with maximum degree 4 with ratio slightly less than 1.2, and for ETP on graphs with maximum degree 5 with ratio 4/3. We also present an exact linear-time algorithm for VTP on the class of indifference graphs. (C) 2007 Elsevier B.V. All rights reserved.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Cell division in bacteria is carried out by an elaborate molecular machine composed of more than a dozen proteins and known as the divisome. Here we describe the characterization of a new divisome protein in Bacillus subtilis called YpsB. Sequence comparisons and phylogentic analysis demonstrated that YpsB is a paralog of the division site selection protein DivIVA. YpsB is present in several gram-positive bacteria and likely originated from the duplication of a DivIVA-like gene in the last common ancestor of bacteria of the orders Bacillales and Lactobacillales. We used green fluorescent protein microscopy to determine that YpsB localizes to the divisome. Similarly to that for DivIVA, the recruitment of YpsB to the divisome requires late division proteins and occurs significantly after Z-ring formation. In contrast to DivIVA, however, YpsB is not retained at the newly formed cell poles after septation. Deletion analysis suggests that the N terminus of YpsB is required to target the protein to the divisome. The high similarity between the N termini of YpsB and DivIVA suggests that the same region is involved in the targeting of DivIVA. YpsB is not essential for septum formation and does not appear to play a role in septum positioning. However, a ypsB deletion has a synthetic effect when combined with a mutation in the cell division gene ftsA. Thus, we conclude that YpsB is a novel B. subtilis cell division protein whose function has diverged from that of its paralog DivIVA.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In order to achieve the high performance, we need to have an efficient scheduling of a parallelprogram onto the processors in multiprocessor systems that minimizes the entire executiontime. This problem of multiprocessor scheduling can be stated as finding a schedule for ageneral task graph to be executed on a multiprocessor system so that the schedule length can be minimize [10]. This scheduling problem is known to be NP- Hard.In multi processor task scheduling, we have a number of CPU’s on which a number of tasksare to be scheduled that the program’s execution time is minimized. According to [10], thetasks scheduling problem is a key factor for a parallel multiprocessor system to gain betterperformance. A task can be partitioned into a group of subtasks and represented as a DAG(Directed Acyclic Graph), so the problem can be stated as finding a schedule for a DAG to beexecuted in a parallel multiprocessor system so that the schedule can be minimized. Thishelps to reduce processing time and increase processor utilization. The aim of this thesis workis to check and compare the results obtained by Bee Colony algorithm with already generatedbest known results in multi processor task scheduling domain.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In this study a case study was conducted at the Transport Agency. The government agency is seeking better alternatives to its current ways of keeping track of important dates stated in legal documents. The agency wants to explore the possibility for an IT-solution for keeping track on those dates. In the case study change analysis was conducted on the vehicle type approval process. Qualitative interviews were conducted with a section manager, vehicle type and component approvers, and with a system administrator and a system manager at the Transport Agency in Borlänge. The study describes an information environment in an organization in which date-controlled conditions apply and challenges with such an environment.The study also provides a proposal for designing an IT-solution for such an environment. The result that emerged showed that the information environment in terms of the legal documents and the institutions’ (EU and UNECE) websites where they are distributed is challenging for keeping track on dates. The challenges of the websites are that they are limited and complicated in terms of information search. The challenges with the legal documents are that date-controlled conditions apply. In other words, the dates themselves determine which dates and what applies and not. The legal documents are also many, retroactive, refer to each other and contain regulations for different kinds of areas, which makes difficult to find the relevant dates. The design proposal that emerged was made based on the existing IT environment at the Transport Agency. It was proposed that important key concepts are identified, categorized and realized in a database with searchability on the most important common concepts as denominators, and that it the information conforms with E-message. The developed database was based on date, date type, vehicle category, legal document, minor version and other info. A web interface was created in ASP.NET and C # for access to the database, with the message that a webbbased IT-solution should be considered.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The subgradient optimization method is a simple and flexible linear programming iterative algorithm. It is much simpler than Newton's method and can be applied to a wider variety of problems. It also converges when the objective function is non-differentiable. Since an efficient algorithm will not only produce a good solution but also take less computing time, we always prefer a simpler algorithm with high quality. In this study a series of step size parameters in the subgradient equation is studied. The performance is compared for a general piecewise function and a specific p-median problem. We examine how the quality of solution changes by setting five forms of step size parameter.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The rapid development of data transfer through internet made it easier to send the data accurate and faster to the destination. There are many transmission media to transfer the data to destination like e-mails; at the same time it is may be easier to modify and misuse the valuable information through hacking. So, in order to transfer the data securely to the destination without any modifications, there are many approaches like cryptography and steganography. This paper deals with the image steganography as well as with the different security issues, general overview of cryptography, steganography and digital watermarking approaches.  The problem of copyright violation of multimedia data has increased due to the enormous growth of computer networks that provides fast and error free transmission of any unauthorized duplicate and possibly manipulated copy of multimedia information. In order to be effective for copyright protection, digital watermark must be robust which are difficult to remove from the object in which they are embedded despite a variety of possible attacks. The message to be send safe and secure, we use watermarking. We use invisible watermarking to embed the message using LSB (Least Significant Bit) steganographic technique. The standard LSB technique embed the message in every pixel, but my contribution for this proposed watermarking, works with the hint for embedding the message only on the image edges alone. If the hacker knows that the system uses LSB technique also, it cannot decrypt correct message. To make my system robust and secure, we added cryptography algorithm as Vigenere square. Whereas the message is transmitted in cipher text and its added advantage to the proposed system. The standard Vigenere square algorithm works with either lower case or upper case. The proposed cryptography algorithm is Vigenere square with extension of numbers also. We can keep the crypto key with combination of characters and numbers. So by using these modifications and updating in this existing algorithm and combination of cryptography and steganography method we develop a secure and strong watermarking method. Performance of this watermarking scheme has been analyzed by evaluating the robustness of the algorithm with PSNR (Peak Signal to Noise Ratio) and MSE (Mean Square Error) against the quality of the image for large amount of data. While coming to see results of the proposed encryption, higher value of 89dB of PSNR with small value of MSE is 0.0017. Then it seems the proposed watermarking system is secure and robust for hiding secure information in any digital system, because this system collect the properties of both steganography and cryptography sciences.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The p-median problem is often used to locate p service centers by minimizing their distances to a geographically distributed demand (n). The optimal locations are sensitive to geographical context such as road network and demand points especially when they are asymmetrically distributed in the plane. Most studies focus on evaluating performances of the p-median model when p and n vary. To our knowledge this is not a very well-studied problem when the road network is alternated especially when it is applied in a real world context. The aim in this study is to analyze how the optimal location solutions vary, using the p-median model, when the density in the road network is alternated. The investigation is conducted by the means of a case study in a region in Sweden with an asymmetrically distributed population (15,000 weighted demand points), Dalecarlia. To locate 5 to 50 service centers we use the national transport administrations official road network (NVDB). The road network consists of 1.5 million nodes. To find the optimal location we start with 500 candidate nodes in the network and increase the number of candidate nodes in steps up to 67,000. To find the optimal solution we use a simulated annealing algorithm with adaptive tuning of the temperature. The results show that there is a limited improvement in the optimal solutions when nodes in the road network increase and p is low. When p is high the improvements are larger. The results also show that choice of the best network depends on p. The larger p the larger density of the network is needed. 

Relevância:

100.00% 100.00%

Publicador:

Resumo:

To have good data quality with high complexity is often seen to be important. Intuition says that the higher accuracy and complexity the data have the better the analytic solutions becomes if it is possible to handle the increasing computing time. However, for most of the practical computational problems, high complexity data means that computational times become too long or that heuristics used to solve the problem have difficulties to reach good solutions. This is even further stressed when the size of the combinatorial problem increases. Consequently, we often need a simplified data to deal with complex combinatorial problems. In this study we stress the question of how the complexity and accuracy in a network affect the quality of the heuristic solutions for different sizes of the combinatorial problem. We evaluate this question by applying the commonly used p-median model, which is used to find optimal locations in a network of p supply points that serve n demand points. To evaluate this, we vary both the accuracy (the number of nodes) of the network and the size of the combinatorial problem (p). The investigation is conducted by the means of a case study in a region in Sweden with an asymmetrically distributed population (15,000 weighted demand points), Dalecarlia. To locate 5 to 50 supply points we use the national transport administrations official road network (NVDB). The road network consists of 1.5 million nodes. To find the optimal location we start with 500 candidate nodes in the network and increase the number of candidate nodes in steps up to 67,000 (which is aggregated from the 1.5 million nodes). To find the optimal solution we use a simulated annealing algorithm with adaptive tuning of the temperature. The results show that there is a limited improvement in the optimal solutions when the accuracy in the road network increase and the combinatorial problem (low p) is simple. When the combinatorial problem is complex (large p) the improvements of increasing the accuracy in the road network are much larger. The results also show that choice of the best accuracy of the network depends on the complexity of the combinatorial (varying p) problem.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

o presente trabalho é um estudo exploratório a respeito da síntese de filmes de diamante via deposiçãoquímica a vapor (CVD) sobre alguns substratos cerâmicos: diboreto de titânio (TiB2), ítria (Y20a), zircão (ZrSi04), zircônia parcialmente e totalmente estabilizada com ítria (Zr02), pirofilita ( Al2Si4OlO(OHh), .alumina (Al2Oa) e nitreto de boro hexagonal (h-BN). Estes substratos foram produzidos, em sua maioria, a partir da sinterização de pós micrométricos em altas temperaturas. Além do estudo em relação a possíveis candidatos alternativos ao tradicional silício para o crescimento de filmes auto-sustentáveis, procuramos encontrar substratos onde o filme aderisse bem e cujas propriedades tribológicas pudessem ser melhoradas com o recobrimento com filme de diamante.Dentre os materiais selecionados, constatamos que a topografia da superfície relacionada à densidade de contornos de grão, desempenha um papel relevante na nucleação do diamante. Além disso, os materiais que favorecem a formação de carbonetos conduziram a melhores resultados na nucleação e crescimento do filme, indicando que a ação da atmosfera reativa do CVD com o substrato também contribui decisivamente para o processo de nucleação. A partir dos resultados obtidos, concluímos que a aderência do filme de diamante ao zircão é excelente, assim como a qualidade do filme, o que pode serexplorado convenientemente caso as propriedades mecânicas do sinterizado de zircão sejam adequadas. No caso da zircônia parcialmente estabilizada, os resultados obtidos foram surpreendentes e este material poderia substituir o convencional substrato de silício para a deposição de filmes auto-sustentados de diamante, com inúmeras vantagens, dentre elas o fato de ser reutilizável e de não ser necessário ataque com ácidos para remoção do substrato, o que evita a geração de resíduos químicos.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This article investigates the causes in the reduction of labor force participation of the old. We argue that the changes in social security policy, in technology and in demography may account for most of the changes in retirement over the second part of the last century in the U.S. economy. We develop a dynamic general equilibrium model with endogenous retirement that embeds social security legislation. The model is able to match very closely the increase in the retirement rate of males aged 65 and older. It also quanti es the isolated impact on retirement and on the solvency of the social security system of the di¤erent factors. The model suggests that technological and demographic changes had a strong in uence on retirement, so that it would have increased signi cantly even if the social security rules had not changed. However, as the latter became much more generous in the past, changes in social security policy can account not only for a sizeable part of the expansion of retirement, but also for the most of the observed increase in the social security expenses as a share of GDP.