957 resultados para Quasi-3D mechanics model
Resumo:
The widespread implementation of GIS-based 3D topographical models has been a great aid in the development and testing of archaeological hypotheses. In this paper, a topographical reconstruction of the ancient city of Tarraco, the Roman capital of the Tarraconensis province, is presented. This model is based on topographical data obtained through archaeological excavations, old photographic documentation, georeferenced archive maps depicting the pre-modern city topography, modern detailed topographical maps and differential GPS measurements. The addition of the Roman urban architectural features to the model offers the possibility to test hypotheses concerning the ideological background manifested in the city shape. This is accomplished mainly through the use of 3D views from the main city accesses. These techniques ultimately demonstrate the ‘theatre-shaped’ layout of the city (to quote Vitrubius) as well as its southwest oriented architecture, whose monumental character was conceived to present a striking aspect to visitors, particularly those arriving from the sea.
Resumo:
Previous studies assessed the outcome of ankle arthrodesis (AA) and total ankle replacement (TAR) surgeries; however, the extent of postoperative recovery towards bilateral gait mechanics (BGM) is unknown. We evaluated the outcome of the two surgeries at least 2 years post rehabilitation, focusing on BGM. 36 participants, including 12 AA patients, 12 TAR patients, and 12 controls were included. Gait assessment over 50 m distance was performed utilizing pressure insoles and 3D inertial sensors, following which an intraindividual comparison was performed. Most spatiotemporal and kinematic parameters in the TAR group were indicative of good gait symmetry, while the AA group presented significant differences. Plantar pressure symmetry among the AA group was also significantly distorted. Abnormality in biomechanical behavior of the AA unoperated, contralateral foot was observed. In summary, our results indicate an altered BGM in AA patients, whereas a relatively fully recovered BGM is observed in TAR patients, despite the quantitative differences in several parameters when compared to a healthy population. Our study supports a biomechanical assessment and rehabilitation of both operated and unoperated sides after major surgeries for ankle osteoarthrosis.
Resumo:
Un système efficace de sismique tridimensionnelle (3-D) haute-résolution adapté à des cibles lacustres de petite échelle a été développé. Dans le Lac Léman, près de la ville de Lausanne, en Suisse, des investigations récentes en deux dimension (2-D) ont mis en évidence une zone de faille complexe qui a été choisie pour tester notre système. Les structures observées incluent une couche mince (<40 m) de sédiments quaternaires sub-horizontaux, discordants sur des couches tertiaires de molasse pentées vers le sud-est. On observe aussi la zone de faille de « La Paudèze » qui sépare les unités de la Molasse du Plateau de la Molasse Subalpine. Deux campagnes 3-D complètes, d?environ d?un kilomètre carré, ont été réalisées sur ce site de test. La campagne pilote (campagne I), effectuée en 1999 pendant 8 jours, a couvert 80 profils en utilisant une seule flûte. Pendant la campagne II (9 jours en 2001), le nouveau système trois-flûtes, bien paramétrés pour notre objectif, a permis l?acquisition de données de très haute qualité sur 180 lignes CMP. Les améliorations principales incluent un système de navigation et de déclenchement de tirs grâce à un nouveau logiciel. Celui-ci comprend un contrôle qualité de la navigation du bateau en temps réel utilisant un GPS différentiel (dGPS) à bord et une station de référence près du bord du lac. De cette façon, les tirs peuvent être déclenchés tous les 5 mètres avec une erreur maximale non-cumulative de 25 centimètres. Tandis que pour la campagne I la position des récepteurs de la flûte 48-traces a dû être déduite à partir des positions du bateau, pour la campagne II elle ont pu être calculées précisément (erreur <20 cm) grâce aux trois antennes dGPS supplémentaires placées sur des flotteurs attachés à l?extrémité de chaque flûte 24-traces. Il est maintenant possible de déterminer la dérive éventuelle de l?extrémité des flûtes (75 m) causée par des courants latéraux ou de petites variations de trajet du bateau. De plus, la construction de deux bras télescopiques maintenant les trois flûtes à une distance de 7.5 m les uns des autres, qui est la même distance que celle entre les lignes naviguées de la campagne II. En combinaison avec un espacement de récepteurs de 2.5 m, la dimension de chaque «bin» de données 3-D de la campagne II est de 1.25 m en ligne et 3.75 m latéralement. L?espacement plus grand en direction « in-line » par rapport à la direction «cross-line» est justifié par l?orientation structurale de la zone de faille perpendiculaire à la direction «in-line». L?incertitude sur la navigation et le positionnement pendant la campagne I et le «binning» imprécis qui en résulte, se retrouve dans les données sous forme d?une certaine discontinuité des réflecteurs. L?utilisation d?un canon à air à doublechambre (qui permet d?atténuer l?effet bulle) a pu réduire l?aliasing observé dans les sections migrées en 3-D. Celui-ci était dû à la combinaison du contenu relativement haute fréquence (<2000 Hz) du canon à eau (utilisé à 140 bars et à 0.3 m de profondeur) et d?un pas d?échantillonnage latéral insuffisant. Le Mini G.I 15/15 a été utilisé à 80 bars et à 1 m de profondeur, est mieux adapté à la complexité de la cible, une zone faillée ayant des réflecteurs pentés jusqu?à 30°. Bien que ses fréquences ne dépassent pas les 650 Hz, cette source combine une pénétration du signal non-aliasé jusqu?à 300 m dans le sol (par rapport au 145 m pour le canon à eau) pour une résolution verticale maximale de 1.1 m. Tandis que la campagne I a été acquise par groupes de plusieurs lignes de directions alternées, l?optimisation du temps d?acquisition du nouveau système à trois flûtes permet l?acquisition en géométrie parallèle, ce qui est préférable lorsqu?on utilise une configuration asymétrique (une source et un dispositif de récepteurs). Si on ne procède pas ainsi, les stacks sont différents selon la direction. Toutefois, la configuration de flûtes, plus courtes que pour la compagne I, a réduit la couverture nominale, la ramenant de 12 à 6. Une séquence classique de traitement 3-D a été adaptée à l?échantillonnage à haute fréquence et elle a été complétée par deux programmes qui transforment le format non-conventionnel de nos données de navigation en un format standard de l?industrie. Dans l?ordre, le traitement comprend l?incorporation de la géométrie, suivi de l?édition des traces, de l?harmonisation des «bins» (pour compenser l?inhomogénéité de la couverture due à la dérive du bateau et de la flûte), de la correction de la divergence sphérique, du filtrage passe-bande, de l?analyse de vitesse, de la correction DMO en 3-D, du stack et enfin de la migration 3-D en temps. D?analyses de vitesse détaillées ont été effectuées sur les données de couverture 12, une ligne sur deux et tous les 50 CMP, soit un nombre total de 600 spectres de semblance. Selon cette analyse, les vitesses d?intervalles varient de 1450-1650 m/s dans les sédiments non-consolidés et de 1650-3000 m/s dans les sédiments consolidés. Le fait que l?on puisse interpréter plusieurs horizons et surfaces de faille dans le cube, montre le potentiel de cette technique pour une interprétation tectonique et géologique à petite échelle en trois dimensions. On distingue cinq faciès sismiques principaux et leurs géométries 3-D détaillées sur des sections verticales et horizontales: les sédiments lacustres (Holocène), les sédiments glacio-lacustres (Pléistocène), la Molasse du Plateau, la Molasse Subalpine de la zone de faille (chevauchement) et la Molasse Subalpine au sud de cette zone. Les couches de la Molasse du Plateau et de la Molasse Subalpine ont respectivement un pendage de ~8° et ~20°. La zone de faille comprend de nombreuses structures très déformées de pendage d?environ 30°. Des tests préliminaires avec un algorithme de migration 3-D en profondeur avant sommation et à amplitudes préservées démontrent que la qualité excellente des données de la campagne II permet l?application de telles techniques à des campagnes haute-résolution. La méthode de sismique marine 3-D était utilisée jusqu?à présent quasi-exclusivement par l?industrie pétrolière. Son adaptation à une échelle plus petite géographiquement mais aussi financièrement a ouvert la voie d?appliquer cette technique à des objectifs d?environnement et du génie civil.<br/><br/>An efficient high-resolution three-dimensional (3-D) seismic reflection system for small-scale targets in lacustrine settings was developed. In Lake Geneva, near the city of Lausanne, Switzerland, past high-resolution two-dimensional (2-D) investigations revealed a complex fault zone (the Paudèze thrust zone), which was subsequently chosen for testing our system. Observed structures include a thin (<40 m) layer of subhorizontal Quaternary sediments that unconformably overlie southeast-dipping Tertiary Molasse beds and the Paudèze thrust zone, which separates Plateau and Subalpine Molasse units. Two complete 3-D surveys have been conducted over this same test site, covering an area of about 1 km2. In 1999, a pilot survey (Survey I), comprising 80 profiles, was carried out in 8 days with a single-streamer configuration. In 2001, a second survey (Survey II) used a newly developed three-streamer system with optimized design parameters, which provided an exceptionally high-quality data set of 180 common midpoint (CMP) lines in 9 days. The main improvements include a navigation and shot-triggering system with in-house navigation software that automatically fires the gun in combination with real-time control on navigation quality using differential GPS (dGPS) onboard and a reference base near the lake shore. Shots were triggered at 5-m intervals with a maximum non-cumulative error of 25 cm. Whereas the single 48-channel streamer system of Survey I requires extrapolation of receiver positions from the boat position, for Survey II they could be accurately calculated (error <20 cm) with the aid of three additional dGPS antennas mounted on rafts attached to the end of each of the 24- channel streamers. Towed at a distance of 75 m behind the vessel, they allow the determination of feathering due to cross-line currents or small course variations. Furthermore, two retractable booms hold the three streamers at a distance of 7.5 m from each other, which is the same distance as the sail line interval for Survey I. With a receiver spacing of 2.5 m, the bin dimension of the 3-D data of Survey II is 1.25 m in in-line direction and 3.75 m in cross-line direction. The greater cross-line versus in-line spacing is justified by the known structural trend of the fault zone perpendicular to the in-line direction. The data from Survey I showed some reflection discontinuity as a result of insufficiently accurate navigation and positioning and subsequent binning errors. Observed aliasing in the 3-D migration was due to insufficient lateral sampling combined with the relatively high frequency (<2000 Hz) content of the water gun source (operated at 140 bars and 0.3 m depth). These results motivated the use of a double-chamber bubble-canceling air gun for Survey II. A 15 / 15 Mini G.I air gun operated at 80 bars and 1 m depth, proved to be better adapted for imaging the complexly faulted target area, which has reflectors dipping up to 30°. Although its frequencies do not exceed 650 Hz, this air gun combines a penetration of non-aliased signal to depths of 300 m below the water bottom (versus 145 m for the water gun) with a maximum vertical resolution of 1.1 m. While Survey I was shot in patches of alternating directions, the optimized surveying time of the new threestreamer system allowed acquisition in parallel geometry, which is preferable when using an asymmetric configuration (single source and receiver array). Otherwise, resulting stacks are different for the opposite directions. However, the shorter streamer configuration of Survey II reduced the nominal fold from 12 to 6. A 3-D conventional processing flow was adapted to the high sampling rates and was complemented by two computer programs that format the unconventional navigation data to industry standards. Processing included trace editing, geometry assignment, bin harmonization (to compensate for uneven fold due to boat/streamer drift), spherical divergence correction, bandpass filtering, velocity analysis, 3-D DMO correction, stack and 3-D time migration. A detailed semblance velocity analysis was performed on the 12-fold data set for every second in-line and every 50th CMP, i.e. on a total of 600 spectra. According to this velocity analysis, interval velocities range from 1450-1650 m/s for the unconsolidated sediments and from 1650-3000 m/s for the consolidated sediments. Delineation of several horizons and fault surfaces reveal the potential for small-scale geologic and tectonic interpretation in three dimensions. Five major seismic facies and their detailed 3-D geometries can be distinguished in vertical and horizontal sections: lacustrine sediments (Holocene) , glaciolacustrine sediments (Pleistocene), Plateau Molasse, Subalpine Molasse and its thrust fault zone. Dips of beds within Plateau and Subalpine Molasse are ~8° and ~20°, respectively. Within the fault zone, many highly deformed structures with dips around 30° are visible. Preliminary tests with 3-D preserved-amplitude prestack depth migration demonstrate that the excellent data quality of Survey II allows application of such sophisticated techniques even to high-resolution seismic surveys. In general, the adaptation of the 3-D marine seismic reflection method, which to date has almost exclusively been used by the oil exploration industry, to a smaller geographical as well as financial scale has helped pave the way for applying this technique to environmental and engineering purposes.<br/><br/>La sismique réflexion est une méthode d?investigation du sous-sol avec un très grand pouvoir de résolution. Elle consiste à envoyer des vibrations dans le sol et à recueillir les ondes qui se réfléchissent sur les discontinuités géologiques à différentes profondeurs et remontent ensuite à la surface où elles sont enregistrées. Les signaux ainsi recueillis donnent non seulement des informations sur la nature des couches en présence et leur géométrie, mais ils permettent aussi de faire une interprétation géologique du sous-sol. Par exemple, dans le cas de roches sédimentaires, les profils de sismique réflexion permettent de déterminer leur mode de dépôt, leurs éventuelles déformations ou cassures et donc leur histoire tectonique. La sismique réflexion est la méthode principale de l?exploration pétrolière. Pendant longtemps on a réalisé des profils de sismique réflexion le long de profils qui fournissent une image du sous-sol en deux dimensions. Les images ainsi obtenues ne sont que partiellement exactes, puisqu?elles ne tiennent pas compte de l?aspect tridimensionnel des structures géologiques. Depuis quelques dizaines d?années, la sismique en trois dimensions (3-D) a apporté un souffle nouveau à l?étude du sous-sol. Si elle est aujourd?hui parfaitement maîtrisée pour l?imagerie des grandes structures géologiques tant dans le domaine terrestre que le domaine océanique, son adaptation à l?échelle lacustre ou fluviale n?a encore fait l?objet que de rares études. Ce travail de thèse a consisté à développer un système d?acquisition sismique similaire à celui utilisé pour la prospection pétrolière en mer, mais adapté aux lacs. Il est donc de dimension moindre, de mise en oeuvre plus légère et surtout d?une résolution des images finales beaucoup plus élevée. Alors que l?industrie pétrolière se limite souvent à une résolution de l?ordre de la dizaine de mètres, l?instrument qui a été mis au point dans le cadre de ce travail permet de voir des détails de l?ordre du mètre. Le nouveau système repose sur la possibilité d?enregistrer simultanément les réflexions sismiques sur trois câbles sismiques (ou flûtes) de 24 traces chacun. Pour obtenir des données 3-D, il est essentiel de positionner les instruments sur l?eau (source et récepteurs des ondes sismiques) avec une grande précision. Un logiciel a été spécialement développé pour le contrôle de la navigation et le déclenchement des tirs de la source sismique en utilisant des récepteurs GPS différentiel (dGPS) sur le bateau et à l?extrémité de chaque flûte. Ceci permet de positionner les instruments avec une précision de l?ordre de 20 cm. Pour tester notre système, nous avons choisi une zone sur le Lac Léman, près de la ville de Lausanne, où passe la faille de « La Paudèze » qui sépare les unités de la Molasse du Plateau et de la Molasse Subalpine. Deux campagnes de mesures de sismique 3-D y ont été réalisées sur une zone d?environ 1 km2. Les enregistrements sismiques ont ensuite été traités pour les transformer en images interprétables. Nous avons appliqué une séquence de traitement 3-D spécialement adaptée à nos données, notamment en ce qui concerne le positionnement. Après traitement, les données font apparaître différents faciès sismiques principaux correspondant notamment aux sédiments lacustres (Holocène), aux sédiments glacio-lacustres (Pléistocène), à la Molasse du Plateau, à la Molasse Subalpine de la zone de faille et la Molasse Subalpine au sud de cette zone. La géométrie 3-D détaillée des failles est visible sur les sections sismiques verticales et horizontales. L?excellente qualité des données et l?interprétation de plusieurs horizons et surfaces de faille montrent le potentiel de cette technique pour les investigations à petite échelle en trois dimensions ce qui ouvre des voies à son application dans les domaines de l?environnement et du génie civil.
Resumo:
In order that the radius and thus ununiform structure of the teeth and otherelectrical and magnetic parts of the machine may be taken into consideration the calculation of an axial flux permanent magnet machine is, conventionally, doneby means of 3D FEM-methods. This calculation procedure, however, requires a lotof time and computer recourses. This study proves that also analytical methods can be applied to perform the calculation successfully. The procedure of the analytical calculation can be summarized into following steps: first the magnet is divided into slices, which makes the calculation for each section individually, and then the parts are submitted to calculation of the final results. It is obvious that using this method can save a lot of designing and calculating time. Thecalculation program is designed to model the magnetic and electrical circuits of surface mounted axial flux permanent magnet synchronous machines in such a way, that it takes into account possible magnetic saturation of the iron parts. Theresult of the calculation is the torque of the motor including the vibrations. The motor geometry and the materials and either the torque or pole angle are defined and the motor can be fed with an arbitrary shape and amplitude of three-phase currents. There are no limits for the size and number of the pole pairs nor for many other factors. The calculation steps and the number of different sections of the magnet are selectable, but the calculation time is strongly depending on this. The results are compared to the measurements of real prototypes. The permanent magnet creates part of the flux in the magnetic circuit. The form and amplitude of the flux density in the air-gap depends on the geometry and material of the magnetic circuit, on the length of the air-gap and remanence flux density of the magnet. Slotting is taken into account by using the Carter factor in the slot opening area. The calculation is simple and fast if the shape of the magnetis a square and has no skew in relation to the stator slots. With a more complicated magnet shape the calculation has to be done in several sections. It is clear that according to the increasing number of sections also the result will become more accurate. In a radial flux motor all sections of the magnets create force with a same radius. In the case of an axial flux motor, each radial section creates force with a different radius and the torque is the sum of these. The magnetic circuit of the motor, consisting of the stator iron, rotor iron, air-gap, magnet and the slot, is modelled with a reluctance net, which considers the saturation of the iron. This means, that several iterations, in which the permeability is updated, has to be done in order to get final results. The motor torque is calculated using the instantaneous linkage flux and stator currents. Flux linkage is called the part of the flux that is created by the permanent magnets and the stator currents passing through the coils in stator teeth. The angle between this flux and the phase currents define the torque created by the magnetic circuit. Due to the winding structure of the stator and in order to limit the leakage flux the slot openings of the stator are normally not made of ferromagnetic material even though, in some cases, semimagnetic slot wedges are used. In the slot opening faces the flux enters the iron almost normally (tangentially with respect to the rotor flux) creating tangential forces in the rotor. This phenomenon iscalled cogging. The flux in the slot opening area on the different sides of theopening and in the different slot openings is not equal and so these forces do not compensate each other. In the calculation it is assumed that the flux entering the left side of the opening is the component left from the geometrical centre of the slot. This torque component together with the torque component calculated using the Lorenz force make the total torque of the motor. It is easy to assume that when all the magnet edges, where the derivative component of the magnet flux density is at its highest, enter the slot openings at the same time, this will have as a result a considerable cogging torque. To reduce the cogging torquethe magnet edges can be shaped so that they are not parallel to the stator slots, which is the common way to solve the problem. In doing so, the edge may be spread along the whole slot pitch and thus also the high derivative component willbe spread to occur equally along the rotation. Besides forming the magnets theymay also be placed somewhat asymmetric on the rotor surface. The asymmetric distribution can be made in many different ways. All the magnets may have a different deflection of the symmetrical centre point or they can be for example shiftedin pairs. There are some factors that limit the deflection. The first is that the magnets cannot overlap. The magnet shape and the relative width compared to the pole define the deflection in this case. The other factor is that a shifting of the poles limits the maximum torque of the motor. If the edges of adjacent magnets are very close to each other the leakage flux from one pole to the other increases reducing thus the air-gap magnetization. The asymmetric model needs some assumptions and simplifications in order to limit the size of the model and calculation time. The reluctance net is made for symmetric distribution. If the magnets are distributed asymmetrically the flux in the different pole pairs will not be exactly the same. Therefore, the assumption that the flux flows from the edges of the model to the next pole pairs, in the calculation model from one edgeto the other, is not correct. If it were wished for that this fact should be considered in multi-pole pair machines, this would mean that all the poles, in other words the whole machine, should be modelled in reluctance net. The error resulting from this wrong assumption is, nevertheless, irrelevant.
Resumo:
PURPOSE: Statistical shape and appearance models play an important role in reducing the segmentation processing time of a vertebra and in improving results for 3D model development. Here, we describe the different steps in generating a statistical shape model (SSM) of the second cervical vertebra (C2) and provide the shape model for general use by the scientific community. The main difficulties in its construction are the morphological complexity of the C2 and its variability in the population. METHODS: The input dataset is composed of manually segmented anonymized patient computerized tomography (CT) scans. The alignment of the different datasets is done with the procrustes alignment on surface models, and then, the registration is cast as a model-fitting problem using a Gaussian process. A principal component analysis (PCA)-based model is generated which includes the variability of the C2. RESULTS: The SSM was generated using 92 CT scans. The resulting SSM was evaluated for specificity, compactness and generalization ability. The SSM of the C2 is freely available to the scientific community in Slicer (an open source software for image analysis and scientific visualization) with a module created to visualize the SSM using Statismo, a framework for statistical shape modeling. CONCLUSION: The SSM of the vertebra allows the shape variability of the C2 to be represented. Moreover, the SSM will enable semi-automatic segmentation and 3D model generation of the vertebra, which would greatly benefit surgery planning.
Resumo:
Les syndromes de déficiences cérébrales en créatine (CCDS) sont dus à des mutations dans les gènes GATM et G AMT (codant pour les enzymes AGAT et G AMT de la voie de synthèse de créatine) ainsi que SLC6A8 (transporteur de créatine), et génèrent une absence ou une très forte baisse de créatine (Cr) dans le cerveau, mesurée par spectroscopic de résonance magnétique. Les patients CCDS développent des handicaps neurologiques sévères. Les patients AGAT et GAMT peuvent être traités avec des doses importantes de Cr, mais gardent dans la plupart des cas des séquelles neurologiques irréversibles. Aucun traitement efficace n'existe à ce jour pour la déficience en SLC6A8. Bien que de nombreux modèles aient été développés pour comprendre la Cr cérébrale en conditions physiologiques, les pathomécanismes des CCDS ne sont pas encore compris. Des souris transgéniques pour les gènes Gatm, Gamt et Slc6a8 ont été générées, mais elles ne miment que partiellement la pathologie humaine. Parmi les CCDS, la déficience en GAMT est la plus sévère, en raison de l'accumulation cérébrale de l'intermédiaire guanidinoacétate (GAA). Alors que la toxicité cérébrale du GAA a été étudiée par exposition directe au GAA d'animaux adultes sains, les mécanismes de la toxicité du GAA en condition de déficience en GAMT dans le cerveau en développement sont encore inconnus. Le but de ce projet était donc de développer un modèle de déficience en GAMT dans des cultures 3D primaires de cellules nerveuses de rat en agrégats par knock-down du gène GAMT, en utilisant un virus adéno-associé (AAV) induisant le mécanisme d'interférence à l'ARN (RNAi). Le virus scAAV2, à la multiplicité d'infection de 1000, s'est révélé le plus efficace pour transduire tous les types de cellules nerveuses des cultures (neurones, astrocytes, oligodendrocytes), et générer un knock-down maximal de la protéine GAMT de 85% (jour in vitro 18). Cette déficience partielle en GAMT s'est révélée insuffisante pour générer une déficience en Cr, mais a causé l'accumulation attendue de GAA, à des doses comparables aux niveaux observés dans le LCR des patients GAMT. Le GAA a induit une croissance axonale anarchique accompagnée d'une baisse de l'apoptose naturelle, suivis par une induction tardive de mort cellulaire non-apoptotique. Le co-traitement par la Cr a prévenu tous les effets toxiques du GAA. Ce travail montre que l'accumulation de GAA en absence de déficience en Cr est suffisante pour affecter le développement du tissu nerveux, et suggère que des formes de déficiences en GAMT supplémentaires, ne présentant pas de déficiences en Cr, pourraient être découvertes par mesure du GAA, en particulier à travers les programmes récemment proposés de dépistage néonatal de la déficience en GAMT. -- Cerebral creatine deficiency syndromes (CCDS) are caused by mutations in the genes GATM and GAMT (respectively coding for the two enzymes of the creatine synthetic pathway, AGAT and GAMT) as well as SLC6A8 (creatine transporter), and lead to the absence or very strong decrease of creatine (Cr) in the brain when measured by magnetic resonance spectroscopy. Affected patients show severe neurological impairments. While AGAT and GAMT deficient patients can be treated with high dosages of Cr, most remain with irreversible brain sequelae. No treatment has been successful so far for SLC6A8 deficiency. While many models have helped understanding the cerebral Cr pathways in physiological conditions, the pathomechanisms underlying CCDS are yet to be elucidated. Transgenic mice carrying mutations in the Gatm, Gamt and Slc6a8 genes have been developed, but only partially mimic the human pathology. Among CCDS, GAMT deficiency is the most severe, due to the CNS accumulation of the guanidinoacetate (GAA) intermediate. While brain toxicity of GAA has been explored through direct GAA exposure of adult healthy animals, the mechanisms underlying GAA toxicity in GAMT deficiency conditions on the developing CNS are yet unknown. The aim of this project was thus to develop and characterize a GAMT deficiency model in developing brain cells by gene knockdown, by adeno-associated virus (AAV)-driven RNA interference (RNAi) in rat 3D organotypic primary brain cell cultures in aggregates. scAAV2 with a multiplicity of infection of 1000 was shown as the most efficient serotype, was able to transduce all brain cell types (neurons, astrocytes, oligodendrocytes) and to induce a maximal GAMT protein knockdown of 85% (day in vitro 18). Metabolite analysis showed that partial GAMT knockdown was insufficient to induce Cr deficiency but generated the awaited GAA accumulation at concentrations comparable to the levels observed in cerebrospinal fluid of GAMT-deficient patients. Accumulated GAA induced axonal hypersprouting paralleled with inhibition of natural apoptosis, followed by a later induction in non-apoptotic cell death. Cr supplementation led to the prevention of all GAA-induced toxic effects. This work shows that GAA accumulation without Cr deficiency is sufficient to affect CNS development, and suggests that additional partial GAMT deficiencies, which may not show the classical brain Cr deficiency, may be discovered through GAA measurement including by recently proposed neonatal screening programs for GAMT deficiency.
Resumo:
Modelling the shoulder's musculature is challenging given its mechanical and geometric complexity. The use of the ideal fibre model to represent a muscle's line of action cannot always faithfully represent the mechanical effect of each muscle, leading to considerable differences between model-estimated and in vivo measured muscle activity. While the musculo-tendon force coordination problem has been extensively analysed in terms of the cost function, only few works have investigated the existence and sensitivity of solutions to fibre topology. The goal of this paper is to present an analysis of the solution set using the concepts of torque-feasible space (TFS) and wrench-feasible space (WFS) from cable-driven robotics. A shoulder model is presented and a simple musculo-tendon force coordination problem is defined. The ideal fibre model for representing muscles is reviewed and the TFS and WFS are defined, leading to the necessary and sufficient conditions for the existence of a solution. The shoulder model's TFS is analysed to explain the lack of anterior deltoid (DLTa) activity. Based on the analysis, a modification of the model's muscle fibre geometry is proposed. The performance with and without the modification is assessed by solving the musculo-tendon force coordination problem for quasi-static abduction in the scapular plane. After the proposed modification, the DLTa reaches 20% of activation.
Resumo:
The goal of this thesis is to implement software for creating 3D models from point clouds. Point clouds are acquired with stereo cameras, monocular systems or laser scanners. The created 3D models are triangular models or NURBS (Non-Uniform Rational B-Splines) models. Triangular models are constructed from selected areas from the point clouds and resulted triangular models are translated into a set of quads. The quads are further translated into an estimated grid structure and used for NURBS surface approximation. Finally, we have a set of NURBS surfaces which represent the whole model. The problem wasn’t so easy to solve. The selected triangular surface reconstruction algorithm did not deal well with noise in point clouds. To handle this problem, a clustering method is introduced for simplificating the model and removing noise. As we had better results with the smaller point clouds produced by clustering, we used points in clusters to better estimate the grids for NURBS models. The overall results were good when the point cloud did not have much noise. The point clouds with small amount of error had good results as the triangular model was solid. NURBS surface reconstruction performed well on solid models.
Resumo:
A new model for the H2 antagonists binding site is postulated based on adsorption coefficient values of sixteen antagonists, in the affinities constants of the primary and secondary binding sites, and in the chemical characterization of these sites by 3D-QSAR. All study compounds are in the extended conformation and deprotonated form. The lateral validation of the QSARs, CoMFA analysis, affinity constants and chemical similarity data suggest that the antagonists block the proton pump in the H2 receptor interacting with two tyrosines - one in the helix 5, and other in the helix 6.
Resumo:
A continuum damage model for the prediction of damage onset and structural collapse of structures manufactured in fiber-reinforced plastic laminates is proposed. The principal damage mechanisms occurring in the longitudinal and transverse directions of a ply are represented by a damage tensor that is fixed in space. Crack closure under load reversal effects are taken into account using damage variables established as a function of the sign of the components of the stress tensor. Damage activation functions based on the LaRC04 failure criteria are used to predict the different damage mechanisms occurring at the ply level. The constitutive damage model is implemented in a finite element code. The objectivity of the numerical model is assured by regularizing the dissipated energy at a material point using Bazant’s Crack Band Model. To verify the accuracy of the approach, analyses ofcoupon specimens were performed, and the numerical predictions were compared with experimental data
Resumo:
A thermodynamically consistent damage model for the simulation of progressive delamination under variable mode ratio is presented. The model is formulated in the context of the Damage Mechanics. The constitutive equation that results from the definition of the free energy as a function of a damage variable is used to model the initiation and propagation of delamination. A new delamination initiation criterion is developed to assure that the formulation can account for changes in the loading mode in a thermodynamically consistent way. The formulation proposed accounts for crack closure effets avoiding interfacial penetration of two adjacent layers aftercomplete decohesion. The model is implemented in a finite element formulation. The numerical predictions given by the model are compared with experimental results
Resumo:
Alzheimer's disease (AD) is considered the main cause of cognitive decline in adults. The available therapies for AD treatment seek to maintain the activity of cholinergic system through the inhibition of the enzyme acetylcholinesterase. However, butyrylcholinesterase (BuChE) can be considered an alternative target for AD treatment. Aiming at developing new BuChE inhibitors, robust QSAR 3D models with high predictive power were developed. The best model presents a good fit (r²=0.82, q²=0.76, with two PCs) and high predictive power (r²predict=0.88). Analysis of regression vector shows that steric properties have considerable importance to the inhibition of the BuChE.
Resumo:
Solid-state silicon detectors have replaced conventional ones in almost all recent high-energy physics experiments. Pixel silicon sensors don't have any alternative in the area near the interaction point because of their high resolution and fast operation speed. However, present detectors hardly withstand high radiation doses. Forthcoming upgrade of the LHC in 2014 requires development of a new generation of pixel detectors which will be able to operate under ten times increased luminosity. A planar fabrication technique has some physical limitations; an improvement of the radiation hardness will reduce sensitivity of a detector. In that case a 3D pixel detector seems to be the most promising device which can overcome these difficulties. The objective of this work was to model a structure of the 3D stripixel detector and to simulate electrical characteristics of the device. Silvaco Atlas software has been used for these purposes. The structures of single and double sided dual column detectors with active edges were described using special command language. Simulations of these detectors have shown that electric field inside an active area has more uniform distribution in comparison to the planar structure. A smaller interelectrode space leads to a stronger field and also decreases the collection time. This makes the new type of detectors more radiation resistant. Other discovered advantages are the lower full depletion voltage and increased charge collection efficiency. So the 3D stripixel detectors have demonstrated improved characteristics and will be a suitable replacement for the planar ones.
Resumo:
The developing energy markets and rising energy system costs have sparked the need to find new forms of energy production and increase the self-sufficiency of energy production. One alternative is gasification, whose principles have been known for decades, but it is only recently when the technology has become a true alternative. However, in order to meet the requirements of modern energy production methods, it is necessary to study the phenomenon thoroughly. In order to understand the gasification process better and optimize it from the viewpoint of ecology and energy efficiency, it is necessary to develop effective and reliable modeling tools for gasifiers. The main aims of this work have been to understand gasification as a process and furthermore to develop an existing three-dimensional circulating fluidized bed modeling tool for modeling of gasification. The model is applied to two gasification processes of 12 and 50 MWth. The results of modeling and measurements have been compared and subsequently reviewed. The work was done in co-operation with Lappeenranta University of Technology and Foster Wheeler Energia Oy.
Resumo:
Previous studies on pencil grip have typically dealt with the developmental aspects in young children while handwriting research is mainly concerned with speed and legibility. Studies linking these areas are few. Evaluation of the existing pencil grip studies is hampered by methodological inconsistencies. The operational definitions of pencil grip arerational but tend to be oversimplified while detailed descriptors tend to be impractical due to their multiplicity. The present study introduces a descriptive two-dimensional model for the categorisation of pencil grip suitable for research applications in a classroom setting. The model is used in four empirical studies of children during the first six years of writing instruction. Study 1 describes the pencil grips observed in a large group of pupils in Finland (n = 504). The results indicate that in Finland the majority of grips resemble the traditional dynamic tripod grip. Significant genderrelated differences in pencil grip were observed. Study 2 is a longitudinal exploration of grip stability vs. change (n = 117). Both expected and unexpected changes were observed in about 25 per cent of the children's grips over four years. A new finding emerged using the present model for categorisation: whereas pencil grips would change, either in terms of ease of grip manipulation or grip configuration, no instances were found where a grip would have changed concurrently on both dimensions. Study 3 is a cross-cultural comparison of grips observed in Finland and the USA (n = 793). The distribution of the pencil grips observed in the American pupils was significantly different from those found in Finland. The cross-cultural disparity is most likely related to the differences in the onset of writing instruction. The differences between the boys' and girls' grips in the American group were non-significant.An implication of Studies 2 and 3 is that the initial pencil grip is of foremost importance since pencil grips are largely stable over time. Study 4 connects the pencil grips to assessment of the mechanics of writing (n = 61). It seems that certain previously not recommended pencil grips might nevertheless be includedamong those accepted since they did not appear to hamper either fluency or legibility.