798 resultados para Data-Intensive Science


Relevância:

80.00% 80.00%

Publicador:

Resumo:

Con la mayor capacidad de los nodos de procesamiento en relación a la potencia de cómputo, cada vez más aplicaciones intensivas de datos como las aplicaciones de la bioinformática, se llevarán a ejecutar en clusters no dedicados. Los clusters no dedicados se caracterizan por su capacidad de combinar la ejecución de aplicaciones de usuarios locales con aplicaciones, científicas o comerciales, ejecutadas en paralelo. Saber qué efecto las aplicaciones con acceso intensivo a dados producen respecto a la mezcla de otro tipo (batch, interativa, SRT, etc) en los entornos no-dedicados permite el desarrollo de políticas de planificación más eficientes. Algunas de las aplicaciones intensivas de E/S se basan en el paradigma MapReduce donde los entornos que las utilizan, como Hadoop, se ocupan de la localidad de los datos, balanceo de carga de forma automática y trabajan con sistemas de archivos distribuidos. El rendimiento de Hadoop se puede mejorar sin aumentar los costos de hardware, al sintonizar varios parámetros de configuración claves para las especificaciones del cluster, para el tamaño de los datos de entrada y para el procesamiento complejo. La sincronización de estos parámetros de sincronización puede ser demasiado compleja para el usuario y/o administrador pero procura garantizar prestaciones más adecuadas. Este trabajo propone la evaluación del impacto de las aplicaciones intensivas de E/S en la planificación de trabajos en clusters no-dedicados bajo los paradigmas MPI y Mapreduce.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

L’èxit del Projecte Genoma Humà (PGH) l’any 2000 va fer de la “medicina personalitzada” una realitat més propera. Els descobriments del PGH han simplificat les tècniques de seqüenciació de tal manera que actualment qualsevol persona pot aconseguir la seva seqüència d’ADN complerta. La tecnologia de Read Mapping destaca en aquest tipus de tècniques i es caracteritza per manegar una gran quantitat de dades. Hadoop, el framework d’Apache per aplicacions intensives de dades sota el paradigma Map Reduce, resulta un aliat perfecte per aquest tipus de tecnologia i ha sigut l’opció escollida per a realitzar aquest projecte. Durant tot el treball es realitza l’estudi, l’anàlisi i les experimentacions necessàries per aconseguir un Algorisme Genètic innovador que utilitzi tot el potencial de Hadoop.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Background: To enhance our understanding of complex biological systems like diseases we need to put all of the available data into context and use this to detect relations, pattern and rules which allow predictive hypotheses to be defined. Life science has become a data rich science with information about the behaviour of millions of entities like genes, chemical compounds, diseases, cell types and organs, which are organised in many different databases and/or spread throughout the literature. Existing knowledge such as genotype - phenotype relations or signal transduction pathways must be semantically integrated and dynamically organised into structured networks that are connected with clinical and experimental data. Different approaches to this challenge exist but so far none has proven entirely satisfactory. Results: To address this challenge we previously developed a generic knowledge management framework, BioXM™, which allows the dynamic, graphic generation of domain specific knowledge representation models based on specific objects and their relations supporting annotations and ontologies. Here we demonstrate the utility of BioXM for knowledge management in systems biology as part of the EU FP6 BioBridge project on translational approaches to chronic diseases. From clinical and experimental data, text-mining results and public databases we generate a chronic obstructive pulmonary disease (COPD) knowledge base and demonstrate its use by mining specific molecular networks together with integrated clinical and experimental data. Conclusions: We generate the first semantically integrated COPD specific public knowledge base and find that for the integration of clinical and experimental data with pre-existing knowledge the configuration based set-up enabled by BioXM reduced implementation time and effort for the knowledge base compared to similar systems implemented as classical software development projects. The knowledgebase enables the retrieval of sub-networks including protein-protein interaction, pathway, gene - disease and gene - compound data which are used for subsequent data analysis, modelling and simulation. Pre-structured queries and reports enhance usability; establishing their use in everyday clinical settings requires further simplification with a browser based interface which is currently under development.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Multiprocessor system-on-chip (MPSoC) designs utilize the available technology and communication architectures to meet the requirements of the upcoming applications. In MPSoC, the communication platform is both the key enabler, as well as the key differentiator for realizing efficient MPSoCs. It provides product differentiation to meet a diverse, multi-dimensional set of design constraints, including performance, power, energy, reconfigurability, scalability, cost, reliability and time-to-market. The communication resources of a single interconnection platform cannot be fully utilized by all kind of applications, such as the availability of higher communication bandwidth for computation but not data intensive applications is often unfeasible in the practical implementation. This thesis aims to perform the architecture-level design space exploration towards efficient and scalable resource utilization for MPSoC communication architecture. In order to meet the performance requirements within the design constraints, careful selection of MPSoC communication platform, resource aware partitioning and mapping of the application play important role. To enhance the utilization of communication resources, variety of techniques such as resource sharing, multicast to avoid re-transmission of identical data, and adaptive routing can be used. For implementation, these techniques should be customized according to the platform architecture. To address the resource utilization of MPSoC communication platforms, variety of architectures with different design parameters and performance levels, namely Segmented bus (SegBus), Network-on-Chip (NoC) and Three-Dimensional NoC (3D-NoC), are selected. Average packet latency and power consumption are the evaluation parameters for the proposed techniques. In conventional computing architectures, fault on a component makes the connected fault-free components inoperative. Resource sharing approach can utilize the fault-free components to retain the system performance by reducing the impact of faults. Design space exploration also guides to narrow down the selection of MPSoC architecture, which can meet the performance requirements with design constraints.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Mothers represent the natural caring. Natural caring is the object of caring science and of research interest because it establishes the central core of professional caring. In this study, we encounter patients who are mothers in need of care in a psychiatric context. Motherhood involves taking responsibility that extends beyond one's own life, because the child represents possibilities in a yet unknown future. Understanding and knowledge about the mothers' struggle in health and suffering are of crucial importance to enable clinical practice to make provisions for and adapt to the individual patient. The overall purpose of this dissertation is to illuminate how the innermost essence of caring emerges in health and suffering in patients who are mothers in psychiatric care. The purpose of the study in a clinical sense is to seek to understand and illuminate the patient's inner world in health and suffering in terms of contextual, existential, ontological and ethical dimensions. The dissertation is exploratory and descriptive in nature and encompasses induction, deduction and abduction as logics tools of reasoning. A theoretical model of natural caring and a universal theoretical model of the innermost essence of caring is developed as seen from the patient's world in a psychiatric context. The dissertation is anchored in human science's view of the human being and the world and in caring science's perspective. Caring science's view of the human being as a unity comprising body, soul and spirit is central in the study's concept of the patient. This multi-dimensional conception of the human being encompasses the dissertation's basic values and is decisive for choice of methodology. Hermeneutic epistemology guided the interpretation of the empirical data, the paradigmatic theses and assumptions. The dialectical movement in interpretation moves back and forth between empirical data, caring science theory and philosophical theory and reveals deeper insight into meaningful content in the clinical context. The interpretation process comprises four levels of abstraction: rational, contextual, existential and ontological. Hermeneutic philosophy guides the inductive and deductive approach to interpretation, as well as the movement between the clinical context and the caring science paradigm. In this encounter between the visible and invisible reality, the image of natural caring – motherliness emerged. The dissertation consists of four studies. The first study is a systematic review of nineteen research articles. The three other studies are hermeneutical interpretations based on text materials from open interviews. Fifteen participants were interviewed, all of whom are mothers of children between 0 and 18 years of age. All were outpatients in the psychiatric specialist health service. In the interpretation process, the mothers' struggle in health and suffering emerges as a struggle between the inner and outer world. Being a mother and patient in health and suffering in a psychiatric context means to struggle to be oneself, to create oneself, to live and realize one's good deeds as a mother and human being. To be oneself, to possess oneself as a mother is not only a question of tending, playing and learning in order to master a practical situation or to survive. It involves constituting a deep, inner desire to courageously create oneself so that the child is able to realize his or her potential in health and suffering. Motherliness manifests itself in caring as a call to ministering humanity and life. The voice of motherliness is understood as the voice of life—the eternal, inner call of love and freedom. The inner call craves fulfilment. Motherliness in natural caring does not retreat. Motherliness defines the Other as freedom and proceeds without regard for all other exterior requirements to realizing wellbeing. The inner essence of caring is attentive, aware and heeds the call of the heart. The innermost essence of caring is to be and to make oneself responsible for the Other. Responsibility cannot be relinquished; free choice consists in whether or not to follow the call. To renounce the inner call to responsibility is to deny oneself and one's dignity as a human being. The theoretical models provide clinical and systematic caring science with knowledge and understanding based on the natural caring spirit inherent in the human being. The study elucidates and strengthens the ontological basic assumptions about the human being as a unity of body, soul and spirit, the sanctity of the human being and the core of caring, ethos. The results of the dissertation will provide clinical practice with knowledge about the inner movements of the mothers' souls in relation to their responsibility as mothers and human beings. Being able to understand the basic conditions for responsibility is crucial for developing care that encompasses mother and child and the mutual relationship between them. This is basic knowledge for developing attitudes and actions that meet and provide for the needs of the patient as mother and as a whole, suffering human being.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The quality and quantity of dissolved organic matter (DOM) exported by Arctic rivers is known to vary with hydrology and this exported material plays a fundamental role in the biogeochemical cycling of carbon at high latitudes. We highlight the potential of optical measurements to examine DOM quality across the hydrograph in Arctic rivers. Furthermore, we establish chromophoric DOM (CDOM) relationships to dissolved organic carbon (DOC) and lignin phenols in the Yukon River and model DOC and lignin loads from CDOM measurements, the former in excellent agreement with long-term DOC monitoring data. Intensive sampling across the historically under-sampled spring flush period highlights the importance of this time for total export of DOC and particularly lignin. Calculated riverine DOC loads to the Arctic Ocean show an increase from previous estimates, especially when new higher discharge data are incorporated. Increased DOC loads indicate decreased residence times for terrigenous DOM in the Arctic Ocean with important implications for the reactivity and export of this material to the Atlantic Ocean.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Mainframes, corporate and central servers are becoming information servers. The requirement for more powerful information servers is the best opportunity to exploit the potential of parallelism. ICL recognized the opportunity of the 'knowledge spectrum' namely to convert raw data into information and then into high grade knowledge. Parallel Processing and Data Management Its response to this and to the underlying search problems was to introduce the CAFS retrieval engine. The CAFS product demonstrates that it is possible to move functionality within an established architecture, introduce a different technology mix and exploit parallelism to achieve radically new levels of performance. CAFS also demonstrates the benefit of achieving this transparently behind existing interfaces. ICL is now working with Bull and Siemens to develop the information servers of the future by exploiting new technologies as available. The objective of the joint Esprit II European Declarative System project is to develop a smoothly scalable, highly parallel computer system, EDS. EDS will in the main be an SQL server and an information server. It will support the many data-intensive applications which the companies foresee; it will also support application-intensive and logic-intensive systems.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

A conceptual problem that appears in different contexts of clustering analysis is that of measuring the degree of compatibility between two sequences of numbers. This problem is usually addressed by means of numerical indexes referred to as sequence correlation indexes. This paper elaborates on why some specific sequence correlation indexes may not be good choices depending on the application scenario in hand. A variant of the Product-Moment correlation coefficient and a weighted formulation for the Goodman-Kruskal and Kendall`s indexes are derived that may be more appropriate for some particular application scenarios. The proposed and existing indexes are analyzed from different perspectives, such as their sensitivity to the ranks and magnitudes of the sequences under evaluation, among other relevant aspects of the problem. The results help suggesting scenarios within the context of clustering analysis that are possibly more appropriate for the application of each index. (C) 2008 Elsevier Inc. All rights reserved.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Point placement strategies aim at mapping data points represented in higher dimensions to bi-dimensional spaces and are frequently used to visualize relationships amongst data instances. They have been valuable tools for analysis and exploration of data sets of various kinds. Many conventional techniques, however, do not behave well when the number of dimensions is high, such as in the case of documents collections. Later approaches handle that shortcoming, but may cause too much clutter to allow flexible exploration to take place. In this work we present a novel hierarchical point placement technique that is capable of dealing with these problems. While good grouping and separation of data with high similarity is maintained without increasing computation cost, its hierarchical structure lends itself both to exploration in various levels of detail and to handling data in subsets, improving analysis capability and also allowing manipulation of larger data sets.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

We present a variable time step, fully adaptive in space, hybrid method for the accurate simulation of incompressible two-phase flows in the presence of surface tension in two dimensions. The method is based on the hybrid level set/front-tracking approach proposed in [H. D. Ceniceros and A. M. Roma, J. Comput. Phys., 205, 391400, 2005]. Geometric, interfacial quantities are computed from front-tracking via the immersed-boundary setting while the signed distance (level set) function, which is evaluated fast and to machine precision, is used as a fluid indicator. The surface tension force is obtained by employing the mixed Eulerian/Lagrangian representation introduced in [S. Shin, S. I. Abdel-Khalik, V. Daru and D. Juric, J. Comput. Phys., 203, 493-516, 2005] whose success for greatly reducing parasitic currents has been demonstrated. The use of our accurate fluid indicator together with effective Lagrangian marker control enhance this parasitic current reduction by several orders of magnitude. To resolve accurately and efficiently sharp gradients and salient flow features we employ dynamic, adaptive mesh refinements. This spatial adaption is used in concert with a dynamic control of the distribution of the Lagrangian nodes along the fluid interface and a variable time step, linearly implicit time integration scheme. We present numerical examples designed to test the capabilities and performance of the proposed approach as well as three applications: the long-time evolution of a fluid interface undergoing Rayleigh-Taylor instability, an example of bubble ascending dynamics, and a drop impacting on a free interface whose dynamics we compare with both existing numerical and experimental data.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Data-intensive Grid applications require huge data transfers between grid computing nodes. These computing nodes, where computing jobs are executed, are usually geographically separated. A grid network that employs optical wavelength division multiplexing (WDM) technology and optical switches to interconnect computing resources with dynamically provisioned multi-gigabit rate bandwidth lightpath is called a Lambda Grid network. A computing task may be executed on any one of several computing nodes which possesses the necessary resources. In order to reflect the reality in job scheduling, allocation of network resources for data transfer should be taken into consideration. However, few scheduling methods consider the communication contention on Lambda Grids. In this paper, we investigate the joint scheduling problem while considering both optical network and computing resources in a Lambda Grid network. The objective of our work is to maximize the total number of jobs that can be scheduled in a Lambda Grid network. An adaptive routing algorithm is proposed and implemented for accomplishing the communication tasks for every job submitted in the network. Four heuristics (FIFO, ESTF, LJF, RS) are implemented for job scheduling of the computational tasks. Simulation results prove the feasibility and efficiency of the proposed solution.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Data-intensive Grid applications require huge data transfers between grid computing nodes. These computing nodes, where computing jobs are executed, are usually geographically separated. A grid network that employs optical wavelength division multiplexing (WDM) technology and optical switches to interconnect computing resources with dynamically provisioned multi-gigabit rate bandwidth lightpath is called a Lambda Grid network. A computing task may be executed on any one of several computing nodes which possesses the necessary resources. In order to reflect the reality in job scheduling, allocation of network resources for data transfer should be taken into consideration. However, few scheduling methods consider the communication contention on Lambda Grids. In this paper, we investigate the joint scheduling problem while considering both optical network and computing resources in a Lambda Grid network. The objective of our work is to maximize the total number of jobs that can be scheduled in a Lambda Grid network. An adaptive routing algorithm is proposed and implemented for accomplishing the communication tasks for every job submitted in the network. Four heuristics (FIFO, ESTF, LJF, RS) are implemented for job scheduling of the computational tasks. Simulation results prove the feasibility and efficiency of the proposed solution.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Background: The hypothalamus plays a pivotal role in numerous mechanisms highly relevant to the maintenance of body homeostasis, such as the control of food intake and energy expenditure. Impairment of these mechanisms has been associated with the metabolic disturbances involved in the pathogenesis of obesity. Since rodent species constitute important models for metabolism studies and the rat hypothalamus is poorly characterized by proteomic strategies, we performed experiments aimed at constructing a two-dimensional gel electrophoresis (2-DE) profile of rat hypothalamus proteins. Results: As a first step, we established the best conditions for tissue collection and protein extraction, quantification and separation. The extraction buffer composition selected for proteome characterization of rat hypothalamus was urea 7 M, thiourea 2 M, CHAPS 4%, Triton X-100 0.5%, followed by a precipitation step with chloroform/methanol. Two-dimensional (2-D) gels of hypothalamic extracts from four-month-old rats were analyzed; the protein spots were digested and identified by using tandem mass spectrometry and database query using the protein search engine MASCOT. Eighty-six hypothalamic proteins were identified, the majority of which were classified as participating in metabolic processes, consistent with the finding of a large number of proteins with catalytic activity. Genes encoding proteins identified in this study have been related to obesity development. Conclusion: The present results indicate that the 2-DE technique will be useful for nutritional studies focusing on hypothalamic proteins. The data presented herein will serve as a reference database for studies testing the effects of dietary manipulations on hypothalamic proteome. We trust that these experiments will lead to important knowledge on protein targets of nutritional variables potentially able to affect the complex central nervous system control of energy homeostasis.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This thesis deals with an investigation of combinatorial and robust optimisation models to solve railway problems. Railway applications represent a challenging area for operations research. In fact, most problems in this context can be modelled as combinatorial optimisation problems, in which the number of feasible solutions is finite. Yet, despite the astonishing success in the field of combinatorial optimisation, the current state of algorithmic research faces severe difficulties with highly-complex and data-intensive applications such as those dealing with optimisation issues in large-scale transportation networks. One of the main issues concerns imperfect information. The idea of Robust Optimisation, as a way to represent and handle mathematically systems with not precisely known data, dates back to 1970s. Unfortunately, none of those techniques proved to be successfully applicable in one of the most complex and largest in scale (transportation) settings: that of railway systems. Railway optimisation deals with planning and scheduling problems over several time horizons. Disturbances are inevitable and severely affect the planning process. Here we focus on two compelling aspects of planning: robust planning and online (real-time) planning.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Vergleich von Datensätzen,die mit Hilfe von den Programmen "Volume©","Pulmo©", "Yacta©" und PulmoFUNC(ILab) erstellt wurden. Dabei wurden jeweils die Lungen- und Emphysemvolumina verglichen, die mit den 4 Programmen ermittelt wurden. Außerdem wurde noch die mittlere Lungendichte als Mittelwert aller Lungenvoxel bestimmt. Zusätzlich wurde noch der Emphysemindex als Quotient aus Emphysem- und Lungenvolumina errechnet. Die Programme waren unterschiedlich benutzerfreundlich in der Bearbeitung: Die weitestgehend manuell zu bearbeitenden Programme Volume© und Pulmo© benötigten zur Bearbeitung deutlich mehr Zeit als die überwiegend automatisch arbeitenden Programme Yacta und PulmoFUNC(ILab).