880 resultados para one-to-many mapping
Resumo:
Frequent advances in medical technologies have brought fonh many innovative treatments that allow medical teams to treal many patients with grave illness and serious trauma who would have died only a few years earlier. These changes have given some patients a second chance at life, but for others. these new treatments have merely prolonged their dying. Instead of dying relatively painlessly, these unfortunate patients often suffer from painful tenninal illnesses or exist in a comatose state that robs them of their dignity, since they cannot survive without advanced and often dehumanizing forms of treatment. Due to many of these concerns, euthanasia has become a central issue in medical ethics. Additionally, the debate is impacted by those who believe that patients have the right make choices about the method and timing of their deaths. Euthanasia is defined as a deliberate act by a physician to hasten the death of a patient, whether through active methods such as an injection of morphine, or through the withdrawal of advanced forms of medical care, for reasons of mercy because of a medical condition that they have. This study explores the question of whether euthanasia is an ethical practice and, as determined by ethical theories and professional codes of ethics, whether the physician is allowed to provide the means to give the patient a path to a "good death," rather than one filled with physical and mental suffering. The paper also asks if there is a relevant moral difference between the active and passive forms of euthanasia and seeks to define requirements to ensure fully voluntary decision making through an evaluation of the factors necessary to produce fully informed consent. Additionally, the proper treatments for patients who suffer from painful terminal illnesses, those who exist in persistent vegetative states and infants born with many diverse medical problems are examined. The ultimate conclusions that are reached in the paper are that euthanasia is an ethical practice in certain specific circumstances for patients who have a very low quality of life due to pain, illness or serious mental deficits as a result of irreversible coma, persistent vegetative state or end-stage clinical dementia. This is defended by the fact that the rights of the patient to determine his or her own fate and to autonomously decide the way that he or she dies are paramount to all other factors in decisions of life and death. There are also circumstances where decisions can be made by health care teams in conjunction with the family to hasten the deaths of incompetent patients when continued existence is clearly not in their best interest, as is the case of infants who are born with serious physical anomalies, who are either 'born dying' or have no prospect for a life that is of a reasonable quality. I have rejected the distinction between active and passive methods of euthanasia and have instead chosen to focus on the intentions of the treating physician and the voluntary nature of the patient's request. When applied in equivalent circumstances, active and passive methods of euthanasia produce the same effects, and if the choice to hasten the death of the patient is ethical, then the use of either method can be accepted. The use of active methods of euthanasia and active forms of withdrawal of life support, such as the removal of a respirator are both conscious decisions to end the life of the patient and both bring death within a short period of time. It is false to maintain a distinction that believes that one is active killing. whereas the other form only allows nature to take it's course. Both are conscious choices to hasten the patient's death and should be evaluated as such. Additionally, through an examination of the Hippocratic Oath, and statements made by the American Medical Association and the American College of physicians, it can be shown that the ideals that the medical profession maintains and the respect for the interests of the patient that it holds allows the physician to give aid to patients who wish to choose death as an alternative to continued suffering. The physician is also allowed to and in some circumstances, is morally required, to help dying patients whether through active or passive forms of euthanasia or through assisted suicide. Euthanasia is a difficult topic to think about, but in the end, we should support the choice that respects the patient's autonomous choice or clear best interest and the respect that we have for their dignity and personal worth.
Resumo:
Worldwide, families Carangidae and Rachycentridae represent one of the groups most important commercial fish, used for food, and great potential for marine aquaculture. However, the genetic bases that can underpin the future cultivation of these species, cytogenetic between these aspects are very weak. The chromosomal patterns have provided basic data for the exploration of biotechnological processes aimed at handling chromosomal genetic improvement, such as induction of polyploidy, androgenesis and ginogenesis, as well as obtaining monosex stocks and interspecific hybridizations. This paper presents a comprehensive cytogenetic survey in 10 species, seven of the family Carangidae and the monotypic family Rachycentridae. Classical cytogenetic analysis and in situ mapping of multigene sequences were employed, and additionally for the genus Selene and morphotypes of Caranx lugubris, comparisons were made using geometric morphometrics. In general, conservative species exhibit a marked chromosome number (2n=48). Although present in large part, different karyotypic form, retain many characteristics typical of chromosomal Order Perciformes, the high number of elements monobrachyal, Ag-NORs/18S rDNA sites and heterochromatin simply reduced, preferably centromeric. The main mechanisms involved in karyotypic diversification are the pericentric inversions, with secondary action of centric fusions. In addition to physical mapping and chromosome detail for the species are presented and discussed patterns of intra-and interspecific diversity, cytotaxonomic markers. This data set provides a better understanding of these patterns caryoevolutyonary groups and conditions for the development of protocols based on Biotechnology for chromosomal manipulation Atlantic these species
Resumo:
This paper presents the results of electrical resistivity methods in the area delineation that was potentially contaminated by liquefaction products, which are also called putrefactive liquids in Vila Rezende municipal cemetery, Piracicaba, So Paulo, Brazil. The results indicate a depth of water table between 3.1 and 5.1 m, with two groundwater direction flows, one to the SW and another to the SE. Due to the contamination plumes, which have the same groundwater direction flow, as well the conductive anomalies observed in the geoelectric sections, the contamination suspicions in the area were confirmed. The probable plume to the SE extends beyond the limits of the cemetery. The location of the conductive anomalies and the probable contamination plumes showed that the contamination is linked with the depth of the water table and the burial time. Mapping using the geostatistical method of ordinary kriging applied to the work drew structural characteristics of the regional phenomenon and spatial behavior of the electrical resistivity data, resulting in continued surfaces. Thus, this method has proved to be an important tool for mapping contamination plumes in cemeteries.
Resumo:
Conselho Nacional de Desenvolvimento Científico e Tecnológico (CNPq)
Resumo:
Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)
Resumo:
Coordenação de Aperfeiçoamento de Pessoal de Nível Superior (CAPES)
Resumo:
This paper presents the results of electrical resistivity methods in the area delineation that was potentially contaminated by liquefaction products, which are also called putrefactive liquids in Vila Rezende municipal cemetery, Piracicaba, So Paulo, Brazil. The results indicate a depth of water table between 3.1 and 5.1 m, with two groundwater direction flows, one to the SW and another to the SE. Due to the contamination plumes, which have the same groundwater direction flow, as well the conductive anomalies observed in the geoelectric sections, the contamination suspicions in the area were confirmed. The probable plume to the SE extends beyond the limits of the cemetery. The location of the conductive anomalies and the probable contamination plumes showed that the contamination is linked with the depth of the water table and the burial time. Mapping using the geostatistical method of ordinary kriging applied to the work drew structural characteristics of the regional phenomenon and spatial behavior of the electrical resistivity data, resulting in continued surfaces. Thus, this method has proved to be an important tool for mapping contamination plumes in cemeteries.
Resumo:
Research in art conservation has been developed from the early 1950s, giving a significant contribution to the conservation-restoration of cultural heritage artefacts. In fact, only through a profound knowledge about the nature and conditions of constituent materials, suitable decisions on the conservation and restoration measures can thus be adopted and preservation practices enhanced. The study of ancient artworks is particularly challenging as they can be considered as heterogeneous and multilayered systems where numerous interactions between the different components as well as degradation and ageing phenomena take place. However, difficulties to physically separate the different layers due to their thickness (1-200 µm) can result in the inaccurate attribution of the identified compounds to a specific layer. Therefore, details can only be analysed when the sample preparation method leaves the layer structure intact, as for example the preparation of embedding cross sections in synthetic resins. Hence, spatially resolved analytical techniques are required not only to exactly characterize the nature of the compounds but also to obtain precise chemical and physical information about ongoing changes. This thesis focuses on the application of FTIR microspectroscopic techniques for cultural heritage materials. The first section is aimed at introducing the use of FTIR microscopy in conservation science with a particular attention to the sampling criteria and sample preparation methods. The second section is aimed at evaluating and validating the use of different FTIR microscopic analytical methods applied to the study of different art conservation issues which may be encountered dealing with cultural heritage artefacts: the characterisation of the artistic execution technique (chapter II-1), the studies on degradation phenomena (chapter II-2) and finally the evaluation of protective treatments (chapter II-3). The third and last section is divided into three chapters which underline recent developments in FTIR spectroscopy for the characterisation of paint cross sections and in particular thin organic layers: a newly developed preparation method with embedding systems in infrared transparent salts (chapter III-1), the new opportunities offered by macro-ATR imaging spectroscopy (chapter III-2) and the possibilities achieved with the different FTIR microspectroscopic techniques nowadays available (chapter III-3). In chapter II-1, FTIR microspectroscopy as molecular analysis, is presented in an integrated approach with other analytical techniques. The proposed sequence is optimized in function of the limited quantity of sample available and this methodology permits to identify the painting materials and characterise the adopted execution technique and state of conservation. Chapter II-2 describes the characterisation of the degradation products with FTIR microscopy since the investigation on the ageing processes encountered in old artefacts represents one of the most important issues in conservation research. Metal carboxylates resulting from the interaction between pigments and binding media are characterized using synthesised metal palmitates and their production is detected on copper-, zinc-, manganese- and lead- (associated with lead carbonate) based pigments dispersed either in oil or egg tempera. Moreover, significant effects seem to be obtained with iron and cobalt (acceleration of the triglycerides hydrolysis). For the first time on sienna and umber paints, manganese carboxylates are also observed. Finally in chapter II-3, FTIR microscopy is combined with further elemental analyses to characterise and estimate the performances and stability of newly developed treatments, which should better fit conservation-restoration problems. In the second part, in chapter III-1, an innovative embedding system in potassium bromide is reported focusing on the characterisation and localisation of organic substances in cross sections. Not only the identification but also the distribution of proteinaceous, lipidic or resinaceous materials, are evidenced directly on different paint cross sections, especially in thin layers of the order of 10 µm. Chapter III-2 describes the use of a conventional diamond ATR accessory coupled with a focal plane array to obtain chemical images of multi-layered paint cross sections. A rapid and simple identification of the different compounds is achieved without the use of any infrared microscope objectives. Finally, the latest FTIR techniques available are highlighted in chapter III-3 in a comparative study for the characterisation of paint cross sections. Results in terms of spatial resolution, data quality and chemical information obtained are presented and in particular, a new FTIR microscope equipped with a linear array detector, which permits reducing the spatial resolution limit to approximately 5 µm, provides very promising results and may represent a good alternative to either mapping or imaging systems.
Resumo:
A one-dimensional multi-component reactive fluid transport algorithm, 1DREACT (Steefel, 1993) was used to investigate different fluid-rock interaction systems. A major short coming of mass transport calculations which include mineral reactions is that solid solutions occurring in many minerals are not treated adequately. Since many thermodynamic models of solid solutions are highly non-linear, this can seriously impact on the stability and efficiency of the solution algorithms used. Phase petrology community saw itself faced with a similar predicament 10 years ago. To improve performance and reliability, phase equilibrium calculations have been using pseudo compounds. The same approach is used here in the first, using the complex plagioclase solid solution as an example. Thermodynamic properties of a varying number of intermediate plagioclase phases were calculated using ideal molecular, Al-avoidance, and non-ideal mixing models. These different mixing models can easily be incorporated into the simulations without modification of the transport code. Simulation results show that as few as nine intermediate compositions are sufficient to characterize the diffusional profile between albite and anorthite. Hence this approach is very efficient, and can be used with little effort. A subsequent chapter reports the results of reactive fluid transport modeling designed to constrain the hydrothermal alteration of Paleoproterozoic sediments of the Southern Lake Superior region. Field observations reveal that quartz-pyrophyllite (or kaolinite) bearing assemblages have been transformed into muscovite-pyrophyllite-diaspore bearing assemblages due to action of fluids migrating along permeable flow channels. Fluid-rock interaction modeling with an initial qtz-prl assemblage and a K-rich fluid simulates the formation of observed mineralogical transformation. The bulk composition of the system evolves from an SiO2-rich one to an Al2O3+K2O-rich one. Simulations show that the fluid flow was up-temperature (e.g. recharge) and that fluid was K-rich. Pseudo compound approach to include solid solutions in reactive transport models was tested in modeling hydrothermal alteration of Icelandic basalts. Solid solutions of chlorites, amphiboles and plagioclase were included as the secondary mineral phases. Saline and fresh water compositions of geothermal fluids were used to investigate the effect of salinity on alteration. Fluid-rock interaction simulations produce the observed mineral transformations. They show that roughly the same alteration minerals are formed due to reactions with both types of fluid which is in agreement with the field observations. A final application is directed towards the remediation of nitrate rich groundwaters. Removal of excess nitrate from groundwater by pyrite oxidation was modeled using the reactive fluid transport algorithm. Model results show that, when a pyrite-bearing, permeable zone is placed in the flow path, nitrate concentration in infiltrating water can be significantly lowered, in agreement with proposals from the literature. This is due to nitrogen reduction. Several simulations investigate the efficiency of systems with different mineral reactive surface areas, reactive barrier zone widths, and flow rates to identify the optimum setup.
Resumo:
1. Teil: Bekannte Konstruktionen. Die vorliegende Arbeit gibt zunächst einen ausführlichen Überblick über die bisherigen Entwicklungen auf dem klassischen Gebiet der Hyperflächen mit vielen Singularitäten. Die maximale Anzahl mu^n(d) von Singularitäten auf einer Hyperfläche vom Grad d im P^n(C) ist nur in sehr wenigen Fällen bekannt, im P^3(C) beispielsweise nur für d<=6. Abgesehen von solchen Ausnahmen existieren nur obere und untere Schranken. 2. Teil: Neue Konstruktionen. Für kleine Grade d ist es oft möglich, bessere Resultate zu erhalten als jene, die durch allgemeine Schranken gegeben sind. In dieser Arbeit beschreiben wir einige algorithmische Ansätze hierfür, von denen einer Computer Algebra in Charakteristik 0 benutzt. Unsere anderen algorithmischen Methoden basieren auf einer Suche über endlichen Körpern. Das Liften der so experimentell gefundenen Hyperflächen durch Ausnutzung ihrer Geometrie oder Arithmetik liefert beispielsweise eine Fläche vom Grad 7 mit $99$ reellen gewöhnlichen Doppelpunkten und eine Fläche vom Grad 9 mit 226 gewöhnlichen Doppelpunkten. Diese Konstruktionen liefern die ersten unteren Schranken für mu^3(d) für ungeraden Grad d>5, die die allgemeine Schranke übertreffen. Unser Algorithmus hat außerdem das Potential, auf viele weitere Probleme der algebraischen Geometrie angewendet zu werden. Neben diesen algorithmischen Methoden beschreiben wir eine Konstruktion von Hyperflächen vom Grad d im P^n mit vielen A_j-Singularitäten, j>=2. Diese Beispiele, deren Existenz wir mit Hilfe der Theorie der Dessins d'Enfants beweisen, übertreffen die bekannten unteren Schranken in den meisten Fällen und ergeben insbesondere neue asymptotische untere Schranken für j>=2, n>=3. 3. Teil: Visualisierung. Wir beschließen unsere Arbeit mit einer Anwendung unserer neuen Visualisierungs-Software surfex, die die Stärken mehrerer existierender Programme bündelt, auf die Konstruktion affiner Gleichungen aller 45 topologischen Typen reeller kubischer Flächen.
Resumo:
In this work, we discuss some theoretical topics related to many-body physics in ultracold atomic and molecular gases. First, we present a comparison between experimental data and theoretical predictions in the context of quantum emulator of quantum field theories, finding good results which supports the efficiency of such simulators. In the second and third parts, we investigate several many-body properties of atomic and molecular gases confined in one dimension.
Resumo:
Zeitreihen sind allgegenwärtig. Die Erfassung und Verarbeitung kontinuierlich gemessener Daten ist in allen Bereichen der Naturwissenschaften, Medizin und Finanzwelt vertreten. Das enorme Anwachsen aufgezeichneter Datenmengen, sei es durch automatisierte Monitoring-Systeme oder integrierte Sensoren, bedarf außerordentlich schneller Algorithmen in Theorie und Praxis. Infolgedessen beschäftigt sich diese Arbeit mit der effizienten Berechnung von Teilsequenzalignments. Komplexe Algorithmen wie z.B. Anomaliedetektion, Motivfabfrage oder die unüberwachte Extraktion von prototypischen Bausteinen in Zeitreihen machen exzessiven Gebrauch von diesen Alignments. Darin begründet sich der Bedarf nach schnellen Implementierungen. Diese Arbeit untergliedert sich in drei Ansätze, die sich dieser Herausforderung widmen. Das umfasst vier Alignierungsalgorithmen und ihre Parallelisierung auf CUDA-fähiger Hardware, einen Algorithmus zur Segmentierung von Datenströmen und eine einheitliche Behandlung von Liegruppen-wertigen Zeitreihen.rnrnDer erste Beitrag ist eine vollständige CUDA-Portierung der UCR-Suite, die weltführende Implementierung von Teilsequenzalignierung. Das umfasst ein neues Berechnungsschema zur Ermittlung lokaler Alignierungsgüten unter Verwendung z-normierten euklidischen Abstands, welches auf jeder parallelen Hardware mit Unterstützung für schnelle Fouriertransformation einsetzbar ist. Des Weiteren geben wir eine SIMT-verträgliche Umsetzung der Lower-Bound-Kaskade der UCR-Suite zur effizienten Berechnung lokaler Alignierungsgüten unter Dynamic Time Warping an. Beide CUDA-Implementierungen ermöglichen eine um ein bis zwei Größenordnungen schnellere Berechnung als etablierte Methoden.rnrnAls zweites untersuchen wir zwei Linearzeit-Approximierungen für das elastische Alignment von Teilsequenzen. Auf der einen Seite behandeln wir ein SIMT-verträgliches Relaxierungschema für Greedy DTW und seine effiziente CUDA-Parallelisierung. Auf der anderen Seite führen wir ein neues lokales Abstandsmaß ein, den Gliding Elastic Match (GEM), welches mit der gleichen asymptotischen Zeitkomplexität wie Greedy DTW berechnet werden kann, jedoch eine vollständige Relaxierung der Penalty-Matrix bietet. Weitere Verbesserungen umfassen Invarianz gegen Trends auf der Messachse und uniforme Skalierung auf der Zeitachse. Des Weiteren wird eine Erweiterung von GEM zur Multi-Shape-Segmentierung diskutiert und auf Bewegungsdaten evaluiert. Beide CUDA-Parallelisierung verzeichnen Laufzeitverbesserungen um bis zu zwei Größenordnungen.rnrnDie Behandlung von Zeitreihen beschränkt sich in der Literatur in der Regel auf reellwertige Messdaten. Der dritte Beitrag umfasst eine einheitliche Methode zur Behandlung von Liegruppen-wertigen Zeitreihen. Darauf aufbauend werden Distanzmaße auf der Rotationsgruppe SO(3) und auf der euklidischen Gruppe SE(3) behandelt. Des Weiteren werden speichereffiziente Darstellungen und gruppenkompatible Erweiterungen elastischer Maße diskutiert.
Resumo:
Nowadays we live in densely populated regions and this leads to many environmental issues. Among all pollutants that human activities originate, metals are relevant because they can be potentially toxic for most of living beings. We studied the fate of Cd, Cr, Cu, Fe, Mn, Ni, Pb and Zn in a vineyard environment analysing samples of plant, wine and soil. Sites were chosen considering the type of wine produced, the type of cultivation (both organic and conventional agriculture) and the geographic location. We took vineyards that cultivate the same grape variety, the Trebbiano). We investigated 5 vineyards located in the Ravenna district (Italy): two on the Lamone Valley slopes, one in the area of river-bank deposits near Ravenna city, then a farm near Lugo and one near Bagnacavallo in interfluve regions. We carried out a very detailed characterization of soils in the sites, including the analysis of: pH, electric conductivity, texture, total carbonate and extimated content of dolomite, active carbonate, iron from ammonium oxalate, Iron Deficiency Chlorosis Index (IDCI), total nitrogen and organic carbon, available phosphorous, available potassium and Cation Exchange Capacity (CEC). Then we made the analysis of the bulk chemical composition and a DTPA extraction to determine the available fraction of elements in soils. All the sites have proper ground to cultivate, with already a good amount of nutrients, such as not needing strong fertilisations, but a vineyard on hills suffers from iron deficiency chlorosis due to the high level of active carbonate. We found some soils with much silica and little calcium oxide that confirm the marly sandstone substratum, while other soils have more calcium oxide and more aluminium oxide that confirm the argillaceous marlstone substratum. We found some critical situations, such as high concentrations of Chromium, especially in the farm near Lugo, and we noticed differences between organic vineyards and conventional ones: the conventional ones have a higher enrichment in soils of some metals (Copper and Zinc). Each metal accumulates differently in every single part of grapevines. We found differences between hill plants and lowland ones: behaviors of plants in metal accumulations seems to have patterns. Metals are more abundant in barks, then in leaves or sometimes in roots. Plants seem trying to remove excesses of metal storing them in bark. Two wines have excess of acetic acid and one conventional farm produces wine with content of Zinc over the Italian law limit. We already found evidence of high values relating them with uncontaminated environments, but more investigations are suggested to link those values to their anthropogenic supplies.
Resumo:
The difficulties of applying the Hartree-Fock method to many body problems is illustrated by treating Helium's electrons up to the point where tractability vanishes. Second, the problem of applying Hartree-Fock methods to the helium atom's electrons, when they are constrained to remain on a sphere, is revisited. The 6-dimensional total energy operator is reduced to a 2-dimensional one, and the application of that 2-dimensional operator in the Hartree-Fock mode is discussed.
Resumo:
Los hipergrafos dirigidos se han empleado en problemas relacionados con lógica proposicional, bases de datos relacionales, linguística computacional y aprendizaje automático. Los hipergrafos dirigidos han sido también utilizados como alternativa a los grafos (bipartitos) dirigidos para facilitar el estudio de las interacciones entre componentes de sistemas complejos que no pueden ser fácilmente modelados usando exclusivamente relaciones binarias. En este contexto, este tipo de representación es conocida como hiper-redes. Un hipergrafo dirigido es una generalización de un grafo dirigido especialmente adecuado para la representación de relaciones de muchos a muchos. Mientras que una arista en un grafo dirigido define una relación entre dos de sus nodos, una hiperarista en un hipergrafo dirigido define una relación entre dos conjuntos de sus nodos. La conexión fuerte es una relación de equivalencia que divide el conjunto de nodos de un hipergrafo dirigido en particiones y cada partición define una clase de equivalencia conocida como componente fuertemente conexo. El estudio de los componentes fuertemente conexos de un hipergrafo dirigido puede ayudar a conseguir una mejor comprensión de la estructura de este tipo de hipergrafos cuando su tamaño es considerable. En el caso de grafo dirigidos, existen algoritmos muy eficientes para el cálculo de los componentes fuertemente conexos en grafos de gran tamaño. Gracias a estos algoritmos, se ha podido averiguar que la estructura de la WWW tiene forma de “pajarita”, donde más del 70% del los nodos están distribuidos en tres grandes conjuntos y uno de ellos es un componente fuertemente conexo. Este tipo de estructura ha sido también observada en redes complejas en otras áreas como la biología. Estudios de naturaleza similar no han podido ser realizados en hipergrafos dirigidos porque no existe algoritmos capaces de calcular los componentes fuertemente conexos de este tipo de hipergrafos. En esta tesis doctoral, hemos investigado como calcular los componentes fuertemente conexos de un hipergrafo dirigido. En concreto, hemos desarrollado dos algoritmos para este problema y hemos determinado que son correctos y cuál es su complejidad computacional. Ambos algoritmos han sido evaluados empíricamente para comparar sus tiempos de ejecución. Para la evaluación, hemos producido una selección de hipergrafos dirigidos generados de forma aleatoria inspirados en modelos muy conocidos de grafos aleatorios como Erdos-Renyi, Newman-Watts-Strogatz and Barabasi-Albert. Varias optimizaciones para ambos algoritmos han sido implementadas y analizadas en la tesis. En concreto, colapsar los componentes fuertemente conexos del grafo dirigido que se puede construir eliminando ciertas hiperaristas complejas del hipergrafo dirigido original, mejora notablemente los tiempos de ejecucion de los algoritmos para varios de los hipergrafos utilizados en la evaluación. Aparte de los ejemplos de aplicación mencionados anteriormente, los hipergrafos dirigidos han sido también empleados en el área de representación de conocimiento. En concreto, este tipo de hipergrafos se han usado para el cálculo de módulos de ontologías. Una ontología puede ser definida como un conjunto de axiomas que especifican formalmente un conjunto de símbolos y sus relaciones, mientras que un modulo puede ser entendido como un subconjunto de axiomas de la ontología que recoge todo el conocimiento que almacena la ontología sobre un conjunto especifico de símbolos y sus relaciones. En la tesis nos hemos centrado solamente en módulos que han sido calculados usando la técnica de localidad sintáctica. Debido a que las ontologías pueden ser muy grandes, el cálculo de módulos puede facilitar las tareas de re-utilización y mantenimiento de dichas ontologías. Sin embargo, analizar todos los posibles módulos de una ontología es, en general, muy costoso porque el numero de módulos crece de forma exponencial con respecto al número de símbolos y de axiomas de la ontología. Afortunadamente, los axiomas de una ontología pueden ser divididos en particiones conocidas como átomos. Cada átomo representa un conjunto máximo de axiomas que siempre aparecen juntos en un modulo. La decomposición atómica de una ontología es definida como un grafo dirigido de tal forma que cada nodo del grafo corresponde con un átomo y cada arista define una dependencia entre una pareja de átomos. En esta tesis introducimos el concepto de“axiom dependency hypergraph” que generaliza el concepto de descomposición atómica de una ontología. Un modulo en una ontología correspondería con un componente conexo en este tipo de hipergrafos y un átomo de una ontología con un componente fuertemente conexo. Hemos adaptado la implementación de nuestros algoritmos para que funcionen también con axiom dependency hypergraphs y poder de esa forma calcular los átomos de una ontología. Para demostrar la viabilidad de esta idea, hemos incorporado nuestros algoritmos en una aplicación que hemos desarrollado para la extracción de módulos y la descomposición atómica de ontologías. A la aplicación la hemos llamado HyS y hemos estudiado sus tiempos de ejecución usando una selección de ontologías muy conocidas del área biomédica, la mayoría disponibles en el portal de Internet NCBO. Los resultados de la evaluación muestran que los tiempos de ejecución de HyS son mucho mejores que las aplicaciones más rápidas conocidas. ABSTRACT Directed hypergraphs are an intuitive modelling formalism that have been used in problems related to propositional logic, relational databases, computational linguistic and machine learning. Directed hypergraphs are also presented as an alternative to directed (bipartite) graphs to facilitate the study of the interactions between components of complex systems that cannot naturally be modelled as binary relations. In this context, they are known as hyper-networks. A directed hypergraph is a generalization of a directed graph suitable for representing many-to-many relationships. While an edge in a directed graph defines a relation between two nodes of the graph, a hyperedge in a directed hypergraph defines a relation between two sets of nodes. Strong-connectivity is an equivalence relation that induces a partition of the set of nodes of a directed hypergraph into strongly-connected components. These components can be collapsed into single nodes. As result, the size of the original hypergraph can significantly be reduced if the strongly-connected components have many nodes. This approach might contribute to better understand how the nodes of a hypergraph are connected, in particular when the hypergraphs are large. In the case of directed graphs, there are efficient algorithms that can be used to compute the strongly-connected components of large graphs. For instance, it has been shown that the macroscopic structure of the World Wide Web can be represented as a “bow-tie” diagram where more than 70% of the nodes are distributed into three large sets and one of these sets is a large strongly-connected component. This particular structure has been also observed in complex networks in other fields such as, e.g., biology. Similar studies cannot be conducted in a directed hypergraph because there does not exist any algorithm for computing the strongly-connected components of the hypergraph. In this thesis, we investigate ways to compute the strongly-connected components of directed hypergraphs. We present two new algorithms and we show their correctness and computational complexity. One of these algorithms is inspired by Tarjan’s algorithm for directed graphs. The second algorithm follows a simple approach to compute the stronglyconnected components. This approach is based on the fact that two nodes of a graph that are strongly-connected can also reach the same nodes. In other words, the connected component of each node is the same. Both algorithms are empirically evaluated to compare their performances. To this end, we have produced a selection of random directed hypergraphs inspired by existent and well-known random graphs models like Erd˝os-Renyi and Newman-Watts-Strogatz. Besides the application examples that we mentioned earlier, directed hypergraphs have also been employed in the field of knowledge representation. In particular, they have been used to compute the modules of an ontology. An ontology is defined as a collection of axioms that provides a formal specification of a set of terms and their relationships; and a module is a subset of an ontology that completely captures the meaning of certain terms as defined in the ontology. In particular, we focus on the modules computed using the notion of syntactic locality. As ontologies can be very large, the computation of modules facilitates the reuse and maintenance of these ontologies. Analysing all modules of an ontology, however, is in general not feasible as the number of modules grows exponentially in the number of terms and axioms of the ontology. Nevertheless, the modules can succinctly be represented using the Atomic Decomposition of an ontology. Using this representation, an ontology can be partitioned into atoms, which are maximal sets of axioms that co-occur in every module. The Atomic Decomposition is then defined as a directed graph such that each node correspond to an atom and each edge represents a dependency relation between two atoms. In this thesis, we introduce the notion of an axiom dependency hypergraph which is a generalization of the atomic decomposition of an ontology. A module in the ontology corresponds to a connected component in the hypergraph, and the atoms of the ontology to the strongly-connected components. We apply our algorithms for directed hypergraphs to axiom dependency hypergraphs and in this manner, we compute the atoms of an ontology. To demonstrate the viability of this approach, we have implemented the algorithms in the application HyS which computes the modules of ontologies and calculate their atomic decomposition. In the thesis, we provide an experimental evaluation of HyS with a selection of large and prominent biomedical ontologies, most of which are available in the NCBO Bioportal. HyS outperforms state-of-the-art implementations in the tasks of extracting modules and computing the atomic decomposition of these ontologies.