965 resultados para Force-based finite elements


Relevância:

40.00% 40.00%

Publicador:

Resumo:

La multiplication dans le corps de Galois à 2^m éléments (i.e. GF(2^m)) est une opérations très importante pour les applications de la théorie des correcteurs et de la cryptographie. Dans ce mémoire, nous nous intéressons aux réalisations parallèles de multiplicateurs dans GF(2^m) lorsque ce dernier est généré par des trinômes irréductibles. Notre point de départ est le multiplicateur de Montgomery qui calcule A(x)B(x)x^(-u) efficacement, étant donné A(x), B(x) in GF(2^m) pour u choisi judicieusement. Nous étudions ensuite l'algorithme diviser pour régner PCHS qui permet de partitionner les multiplicandes d'un produit dans GF(2^m) lorsque m est impair. Nous l'appliquons pour la partitionnement de A(x) et de B(x) dans la multiplication de Montgomery A(x)B(x)x^(-u) pour GF(2^m) même si m est pair. Basé sur cette nouvelle approche, nous construisons un multiplicateur dans GF(2^m) généré par des trinôme irréductibles. Une nouvelle astuce de réutilisation des résultats intermédiaires nous permet d'éliminer plusieurs portes XOR redondantes. Les complexités de temps (i.e. le délais) et d'espace (i.e. le nombre de portes logiques) du nouveau multiplicateur sont ensuite analysées: 1. Le nouveau multiplicateur demande environ 25% moins de portes logiques que les multiplicateurs de Montgomery et de Mastrovito lorsque GF(2^m) est généré par des trinômes irréductible et m est suffisamment grand. Le nombre de portes du nouveau multiplicateur est presque identique à celui du multiplicateur de Karatsuba proposé par Elia. 2. Le délai de calcul du nouveau multiplicateur excède celui des meilleurs multiplicateurs d'au plus deux évaluations de portes XOR. 3. Nous determinons le délai et le nombre de portes logiques du nouveau multiplicateur sur les deux corps de Galois recommandés par le National Institute of Standards and Technology (NIST). Nous montrons que notre multiplicateurs contient 15% moins de portes logiques que les multiplicateurs de Montgomery et de Mastrovito au coût d'un délai d'au plus une porte XOR supplémentaire. De plus, notre multiplicateur a un délai d'une porte XOR moindre que celui du multiplicateur d'Elia au coût d'une augmentation de moins de 1% du nombre total de portes logiques.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Le cancer du sein est le cancer le plus fréquent chez la femme. Il demeure la cause de mortalité la plus importante chez les femmes âgées entre 35 et 55 ans. Au Canada, plus de 20 000 nouveaux cas sont diagnostiqués chaque année. Les études scientifiques démontrent que l'espérance de vie est étroitement liée à la précocité du diagnostic. Les moyens de diagnostic actuels comme la mammographie, l'échographie et la biopsie comportent certaines limitations. Par exemple, la mammographie permet de diagnostiquer la présence d’une masse suspecte dans le sein, mais ne peut en déterminer la nature (bénigne ou maligne). Les techniques d’imagerie complémentaires comme l'échographie ou l'imagerie par résonance magnétique (IRM) sont alors utilisées en complément, mais elles sont limitées quant à la sensibilité et la spécificité de leur diagnostic, principalement chez les jeunes femmes (< 50 ans) ou celles ayant un parenchyme dense. Par conséquent, nombreuses sont celles qui doivent subir une biopsie alors que leur lésions sont bénignes. Quelques voies de recherche sont privilégiées depuis peu pour réduire l`incertitude du diagnostic par imagerie ultrasonore. Dans ce contexte, l’élastographie dynamique est prometteuse. Cette technique est inspirée du geste médical de palpation et est basée sur la détermination de la rigidité des tissus, sachant que les lésions en général sont plus rigides que le tissu sain environnant. Le principe de cette technique est de générer des ondes de cisaillement et d'en étudier la propagation de ces ondes afin de remonter aux propriétés mécaniques du milieu via un problème inverse préétabli. Cette thèse vise le développement d'une nouvelle méthode d'élastographie dynamique pour le dépistage précoce des lésions mammaires. L'un des principaux problèmes des techniques d'élastographie dynamiques en utilisant la force de radiation est la forte atténuation des ondes de cisaillement. Après quelques longueurs d'onde de propagation, les amplitudes de déplacement diminuent considérablement et leur suivi devient difficile voir impossible. Ce problème affecte grandement la caractérisation des tissus biologiques. En outre, ces techniques ne donnent que l'information sur l'élasticité tandis que des études récentes montrent que certaines lésions bénignes ont les mêmes élasticités que des lésions malignes ce qui affecte la spécificité de ces techniques et motive la quantification de d'autres paramètres mécaniques (e.g.la viscosité). Le premier objectif de cette thèse consiste à optimiser la pression de radiation acoustique afin de rehausser l'amplitude des déplacements générés. Pour ce faire, un modèle analytique de prédiction de la fréquence de génération de la force de radiation a été développé. Une fois validé in vitro, ce modèle a servi pour la prédiction des fréquences optimales pour la génération de la force de radiation dans d'autres expérimentations in vitro et ex vivo sur des échantillons de tissu mammaire obtenus après mastectomie totale. Dans la continuité de ces travaux, un prototype de sonde ultrasonore conçu pour la génération d'un type spécifique d'ondes de cisaillement appelé ''onde de torsion'' a été développé. Le but est d'utiliser la force de radiation optimisée afin de générer des ondes de cisaillement adaptatives, et de monter leur utilité dans l'amélioration de l'amplitude des déplacements. Contrairement aux techniques élastographiques classiques, ce prototype permet la génération des ondes de cisaillement selon des parcours adaptatifs (e.g. circulaire, elliptique,…etc.) dépendamment de la forme de la lésion. L’optimisation des dépôts énergétiques induit une meilleure réponse mécanique du tissu et améliore le rapport signal sur bruit pour une meilleure quantification des paramètres viscoélastiques. Il est aussi question de consolider davantage les travaux de recherches antérieurs par un appui expérimental, et de prouver que ce type particulier d'onde de torsion peut mettre en résonance des structures. Ce phénomène de résonance des structures permet de rehausser davantage le contraste de déplacement entre les masses suspectes et le milieu environnant pour une meilleure détection. Enfin, dans le cadre de la quantification des paramètres viscoélastiques des tissus, la dernière étape consiste à développer un modèle inverse basé sur la propagation des ondes de cisaillement adaptatives pour l'estimation des paramètres viscoélastiques. L'estimation des paramètres viscoélastiques se fait via la résolution d'un problème inverse intégré dans un modèle numérique éléments finis. La robustesse de ce modèle a été étudiée afin de déterminer ces limites d'utilisation. Les résultats obtenus par ce modèle sont comparés à d'autres résultats (mêmes échantillons) obtenus par des méthodes de référence (e.g. Rheospectris) afin d'estimer la précision de la méthode développée. La quantification des paramètres mécaniques des lésions permet d'améliorer la sensibilité et la spécificité du diagnostic. La caractérisation tissulaire permet aussi une meilleure identification du type de lésion (malin ou bénin) ainsi que son évolution. Cette technique aide grandement les cliniciens dans le choix et la planification d'une prise en charge adaptée.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Sharing of information with those in need of it has always been an idealistic goal of networked environments. With the proliferation of computer networks, information is so widely distributed among systems, that it is imperative to have well-organized schemes for retrieval and also discovery. This thesis attempts to investigate the problems associated with such schemes and suggests a software architecture, which is aimed towards achieving a meaningful discovery. Usage of information elements as a modelling base for efficient information discovery in distributed systems is demonstrated with the aid of a novel conceptual entity called infotron.The investigations are focused on distributed systems and their associated problems. The study was directed towards identifying suitable software architecture and incorporating the same in an environment where information growth is phenomenal and a proper mechanism for carrying out information discovery becomes feasible. An empirical study undertaken with the aid of an election database of constituencies distributed geographically, provided the insights required. This is manifested in the Election Counting and Reporting Software (ECRS) System. ECRS system is a software system, which is essentially distributed in nature designed to prepare reports to district administrators about the election counting process and to generate other miscellaneous statutory reports.Most of the distributed systems of the nature of ECRS normally will possess a "fragile architecture" which would make them amenable to collapse, with the occurrence of minor faults. This is resolved with the help of the penta-tier architecture proposed, that contained five different technologies at different tiers of the architecture.The results of experiment conducted and its analysis show that such an architecture would help to maintain different components of the software intact in an impermeable manner from any internal or external faults. The architecture thus evolved needed a mechanism to support information processing and discovery. This necessitated the introduction of the noveI concept of infotrons. Further, when a computing machine has to perform any meaningful extraction of information, it is guided by what is termed an infotron dictionary.The other empirical study was to find out which of the two prominent markup languages namely HTML and XML, is best suited for the incorporation of infotrons. A comparative study of 200 documents in HTML and XML was undertaken. The result was in favor ofXML.The concept of infotron and that of infotron dictionary, which were developed, was applied to implement an Information Discovery System (IDS). IDS is essentially, a system, that starts with the infotron(s) supplied as clue(s), and results in brewing the information required to satisfy the need of the information discoverer by utilizing the documents available at its disposal (as information space). The various components of the system and their interaction follows the penta-tier architectural model and therefore can be considered fault-tolerant. IDS is generic in nature and therefore the characteristics and the specifications were drawn up accordingly. Many subsystems interacted with multiple infotron dictionaries that were maintained in the system.In order to demonstrate the working of the IDS and to discover the information without modification of a typical Library Information System (LIS), an Information Discovery in Library Information System (lDLIS) application was developed. IDLIS is essentially a wrapper for the LIS, which maintains all the databases of the library. The purpose was to demonstrate that the functionality of a legacy system could be enhanced with the augmentation of IDS leading to information discovery service. IDLIS demonstrates IDS in action. IDLIS proves that any legacy system could be augmented with IDS effectively to provide the additional functionality of information discovery service.Possible applications of IDS and scope for further research in the field are covered.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Magnetism and magnetic materials have been an ever-attractive subject area for engineers and scientists alike because of its versatility in finding applications in useful devices. They find applications in a host of devices ranging from rudimentary devices like loud speakers to sophisticated gadgets like waveguides and Magnetic Random Access Memories (MRAM).The one and only material in the realm of magnetism that has been at the centre stage of applications is ferrites and in that spinel ferrites received the lions share as far as practical applications are concerned.It has been the endeavour of scientists and engineers to remove obsolescence and improve upon the existing so as to save energy and integrate in to various other systems. This has been the hallmark of material scientists and this has led to new materials and new technologies.In the field of ferrites too there has been considerable interest to devise new materials based on iron oxides and other compounds. This means synthesising ultra fine particles and tuning its properties to device new materials. There are various preparation techniques ranging from top- down to bottom-up approaches. This includes synthesising at molecular level, self assembling,gas based condensation. Iow temperature eo-precipitation, solgel process and high energy ball milling. Among these methods sol-gel process allows good control of the properties of ceramic materials. The advantage of this method includes processing at low temperature. mixing at the molecular level and fabrication of novel materials for various devices.Composites are materials. which combine the good qualities of one or more components. They can be prepared in situ or by mechanical means by the incorporation of fine particles in appropriate matrixes. The size of the magnetic powders as well as the nature of matrix affect the processability and other physical properties of the final product. These plastic/rubber magnets can in turn be useful for various applications in different devices. In applications involving ferrites at high frequencies, it is essential that the material possesses an appropriate dielectric permittivity and suitable magnetic permeability. This can be achieved by synthesizing rubber ferrite composites (RFC's). RFCs are very useful materials for microwave absorptions. Hence the synthesis of ferrites in the nanoregirne.investigations on their size effects on the structural, magnetic, and electrical properties and the incorporation of these ferrites into polymer matrixes assume significance.In the present study, nano particles of NiFe204, Li(!5Fe2S04 and Col-e-O, are prepared by sol gel method. By appropriate heat treatments, particles of different grain sizes are obtained. The structural, magnetic and electrical measurements are evaluated as a function of grain size and temperature. NiFel04 prepared in the ultrafine regime are then incorporated in nitrile rubber matrix. The incorporation was carried out according to a specific recipe and for various loadings of magnetic fillers. The cure characteristics, magnetic properties, electrical properties and mechanical properties of these elastomer blends are carried out. The electrical permittivity of all the rubber samples in the X - band are also conducted.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

This thesis attempts to investigate the problems associated with such schemes and suggests a software architecture, which is aimed towards achieving a meaningful discovery. Usage of information elements as a modelling base for efficient information discovery in distributed systems is demonstrated with the aid of a novel conceptual entity called infotron. The investigations are focused on distributed systems and their associated problems. The study was directed towards identifying suitable software architecture and incorporating the same in an environment where information growth is phenomenal and a proper mechanism for carrying out information discovery becomes feasible. An empirical study undertaken with the aid of an election database of constituencies distributed geographically, provided the insights required. This is manifested in the Election Counting and Reporting Software (ECRS) System. ECRS system is a software system, which is essentially distributed in nature designed to prepare reports to district administrators about the election counting process and to generate other miscellaneous statutory reports.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

In the present study, nano particles of NiFe3O4, I_.l()5Feg5O4 and CoFegO4 are prepared by sol gel method. By appropriate heat treatments, particles of different grain sizes are obtained. The structural, magnetic and electrical measurements are evaluated as a function of grain size and temperature. NiFe3O4 prepared in the ultrafine regime are then incorporated in nitrile rubber matrix. The incorporation was carried out according to a specific recipe and for various loadings of magnetic fillers. The cure characteristics, magnetic properties, electrical properties and mechanical properties of these elastomer blends are carried out. The electrical permittivity of all the rubber samples in the X — band are also conducted

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Der Vielelektronen Aspekt wird in einteilchenartigen Formulierungen berücksichtigt, entweder in Hartree-Fock Näherung oder unter dem Einschluß der Elektron-Elektron Korrelationen durch die Dichtefunktional Theorie. Da die Physik elektronischer Systeme (Atome, Moleküle, Cluster, Kondensierte Materie, Plasmen) relativistisch ist, habe ich von Anfang an die relativistische 4 Spinor Dirac Theorie eingesetzt, in jüngster Zeit aber, und das wird der hauptfortschritt in den relativistischen Beschreibung durch meine Promotionsarbeit werden, eine ebenfalls voll relativistische, auf dem sogenannten Minimax Prinzip beruhende 2-Spinor Theorie umgesetzt. Im folgenden ist eine kurze Beschreibung meiner Dissertation: Ein wesentlicher Effizienzgewinn in der relativistischen 4-Spinor Dirac Rechnungen konnte durch neuartige singuläre Koordinatentransformationen erreicht werden, so daß sich auch noch für das superschwere Th2 179+ hächste Lösungsgenauigkeiten mit moderatem Computer Aufwand ergaben, und zu zwei weiteren interessanten Veröffentlichungen führten (Publikationsliste). Trotz der damit bereits ermöglichten sehr viel effizienteren relativistischen Berechnung von Molekülen und Clustern blieben diese Rechnungen Größenordnungen aufwendiger als entsprechende nicht-relativistische. Diese behandeln das tatsächliche (relativitische) Verhalten elektronischer Systeme nur näherungsweise richtig, um so besser jedoch, je leichter die beteiligten Atome sind (kleine Kernladungszahl Z). Deshalb habe ich nach einem neuen Formalismus gesucht, der dem möglichst gut Rechnung trägt und trotzdem die Physik richtig relativistisch beschreibt. Dies gelingt durch ein 2-Spinor basierendes Minimax Prinzip: Systeme mit leichten Atomen sind voll relativistisch nunmehr nahezu ähnlich effizient beschrieben wie nicht-relativistisch, was natürlich große Hoffnungen für genaue (d.h. relativistische) Berechnungen weckt. Es ergab sich eine erste grundlegende Veröffentlichung (Publikationsliste). Die Genauigkeit in stark relativistischen Systemen wie Th2 179+ ist ähnlich oder leicht besser als in 4-Spinor Dirac-Formulierung. Die Vorteile der neuen Formulierung gehen aber entscheidend weiter: A. Die neue Minimax Formulierung der Dirac-Gl. ist frei von spuriosen Zuständen und hat keine positronischen Kontaminationen. B. Der Aufwand ist weit reduziert, da nur ein 1/3 der Matrix Elemente gegenüber 4-Spinor noch zu berechnen ist, und alle Matrixdimensionen Faktor 2 kleiner sind. C. Numerisch verhält sich die neue Formulierung ähnlilch gut wie die nichtrelativistische Schrödinger Gleichung (Obwohl es eine exakte Formulierung und keine Näherung der Dirac-Gl. ist), und hat damit bessere Konvergenzeigenschaften als 4-Spinor. Insbesondere die Fehlerwichtung (singulärer und glatter Anteil) ist in 2-Spinor anders, und diese zeigt die guten Extrapolationseigenschaften wie bei der nichtrelativistischen Schrödinger Gleichung. Die Ausweitung des Anwendungsbereichs von (relativistischen) 2-Spinor ist bereits in FEM Dirac-Fock-Slater, mit zwei Beispielen CO und N2, erfolgreich gemacht. Weitere Erweiterungen sind nahezu möglich. Siehe Minmax LCAO Nährung.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Analytical potential energy functions which are valid at all dissociation limits have been derived for the ground states of SO2 and O3. The procedure involves minimizing the errors between the observed vibrational spectra and spectra calculated by a variational procedure. Good agreement is obtained between the observed and calculated spectra for both molecules. Comparisons are made between anharmonic force fields, previously determined from the spectral data, and the force fields obtained by differentiating the derived analytical functions at the equilibrium configurations.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

This work investigates the optimum decision delay and tap-length of the finite-length decision feedback equalizer. First we show that, if the feedback filter (FBF) length Nb is equal to or larger than the channel memory v and the decision delay Δ is smaller than the feedforward filter (FFF) length Nf, then only the first Δ+1 elements of the FFF can be nonzero. Based on this result we prove that the maximum effective FBF length is equal to the channel memory v, and if Nb ≥ v and Nf is long enough, the optimum decision delay that minimizes the MMSE is Nf-1.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

though discrete cell-based frameworks are now commonly used to simulate a whole range of biological phenomena, it is typically not obvious how the numerous different types of model are related to one another, nor which one is most appropriate in a given context. Here we demonstrate how individual cell movement on the discrete scale modeled using nonlinear force laws can be described by nonlinear diffusion coefficients on the continuum scale. A general relationship between nonlinear force laws and their respective diffusion coefficients is derived in one spatial dimension and, subsequently, a range of particular examples is considered. For each case excellent agreement is observed between numerical solutions of the discrete and corresponding continuum models. Three case studies are considered in which we demonstrate how the derived nonlinear diffusion coefficients can be used to (a) relate different discrete models of cell behavior; (b) derive discrete, intercell force laws from previously posed diffusion coefficients, and (c) describe aggregative behavior in discrete simulations.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

We propose a discontinuous-Galerkin-based immersed boundary method for elasticity problems. The resulting numerical scheme does not require boundary fitting meshes and avoids boundary locking by switching the elements intersected by the boundary to a discontinuous Galerkin approximation. Special emphasis is placed on the construction of a method that retains an optimal convergence rate in the presence of non-homogeneous essential and natural boundary conditions. The role of each one of the approximations introduced is illustrated by analyzing an analog problem in one spatial dimension. Finally, extensive two- and three-dimensional numerical experiments on linear and nonlinear elasticity problems verify that the proposed method leads to optimal convergence rates under combinations of essential and natural boundary conditions. (C) 2009 Elsevier B.V. All rights reserved.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

A numerical method to approximate partial differential equations on meshes that do not conform to the domain boundaries is introduced. The proposed method is conceptually simple and free of user-defined parameters. Starting with a conforming finite element mesh, the key ingredient is to switch those elements intersected by the Dirichlet boundary to a discontinuous-Galerkin approximation and impose the Dirichlet boundary conditions strongly. By virtue of relaxing the continuity constraint at those elements. boundary locking is avoided and optimal-order convergence is achieved. This is shown through numerical experiments in reaction-diffusion problems. Copyright (c) 2008 John Wiley & Sons, Ltd.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Felsic microgranular enclaves with structures indicating that they interacted in a plastic state with their chemically similar host granite are abundant in the Maua Pluton, SE Brazil. Larger plagioclase xenocrysts are in textural disequilibrium with the enclave groundmass and show complex zoning patterns with partially resorbed An-rich cores (locally with patchy textures) surrounded by more sodic rims. In situ laser ablation-(multi-collector) inductively coupled plasma mass spectrometry trace element and Sr isotopic analyses performed on the plagioclase xenocrysts indicate open-system crystallization; however, no evidence of derivation from more primitive basic melts is observed. The An-rich cores have more radiogenic initial Sr isotopic ratios that decrease towards the outermost part of the rims, which are in isotopic equilibrium with the matrix plagioclase. These profiles may have been produced by either (1) diffusional re-equilibration after rim crystallization from the enclave-forming magma, as indicated by relatively short calculated residence times, or (2) episodic contamination with a decrease of the contaminant ratio proportional to the extent to which the country rocks were isolated by the crystallization front. Profiles of trace elements with high diffusion coefficients would require unrealistically long residence times, and can be modeled in terms of fractional crystallization. A combination of trace element and Sr isotope data suggests that the felsic microgranular enclaves from the Maua Pluton are the products of interaction between end-member magmas that had similar compositions, thus recording `self-mixing` events.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)