972 resultados para european projects


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Besides core project partners, the SCI-BUS project also supported several external user communities in developing and setting up customized science gateways. The focus was on large communities typically represented by other European research projects. However, smaller local efforts with the potential of generalizing the solution to wider communities were also supported. This chapter gives an overview of support activities related to user communities external to the SCI-BUS project. A generic overview of such activities is provided followed by the detailed description of three gateways developed in collaboration with European projects: the agINFRA Science Gateway for Workflows for agricultural research, the VERCE Science Gateway for seismology, and the DRIHM Science Gateway for weather research and forecasting.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The aging process is characterized by the progressive fitness decline experienced at all the levels of physiological organization, from single molecules up to the whole organism. Studies confirmed inflammaging, a chronic low-level inflammation, as a deeply intertwined partner of the aging process, which may provide the “common soil” upon which age-related diseases develop and flourish. Thus, albeit inflammation per se represents a physiological process, it can rapidly become detrimental if it goes out of control causing an excess of local and systemic inflammatory response, a striking risk factor for the elderly population. Developing interventions to counteract the establishment of this state is thus a top priority. Diet, among other factors, represents a good candidate to regulate inflammation. Building on top of this consideration, the EU project NU-AGE is now trying to assess if a Mediterranean diet, fortified for the elderly population needs, may help in modulating inflammaging. To do so, NU-AGE enrolled a total of 1250 subjects, half of which followed a 1-year long diet, and characterized them by mean of the most advanced –omics and non –omics analyses. The aim of this thesis was the development of a solid data management pipeline able to efficiently cope with the results of these assays, which are now flowing inside a centralized database, ready to be used to test the most disparate scientific hypotheses. At the same time, the work hereby described encompasses the data analysis of the GEHA project, which was focused on identifying the genetic determinants of longevity, with a particular focus on developing and applying a method for detecting epistatic interactions in human mtDNA. Eventually, in an effort to propel the adoption of NGS technologies in everyday pipeline, we developed a NGS variant calling pipeline devoted to solve all the sequencing-related issues of the mtDNA.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Congenital anomalies (CA) are the paradigm example of rare diseases liable to primary prevention actions due to the multifactorial etiology of many of them, involving a number of environmental factors together with genetic predispositions. Yet despite the preventive potential, lack of attention to an integrated preventive strategy has led to the prevalence of CA remaining relatively stable in recent decades. The 2 European projects, EUROCAT and EUROPLAN, have joined efforts to provide the first science-based and comprehensive set of recommendations for the primary prevention of CA in the European Union. The resulting EUROCAT-EUROPLAN 'Recommendations on Policies to Be Considered for the Primary Prevention of Congenital Anomalies in National Plans and Strategies on Rare Diseases' were issued in 2012 and endorsed by EUCERD (European Union Committee of Experts on Rare Diseases) in 2013. The recommendations exploit interdisciplinary expertise encompassing drugs, diet, lifestyles, maternal health status, and the environment. The recommendations include evidence-based actions aimed at reducing risk factors and at increasing protective factors and behaviors at both individual and population level. Moreover, consideration is given to topics specifically related to CA (e.g. folate status, teratogens) as well as of broad public health impact (e.g. obesity, smoking) which call for specific attention to their relevance in the pre- and periconceptional period. The recommendations, reported entirely in this paper, are a comprehensive tool to implement primary prevention into national policies on rare diseases in Europe.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

Urban mobility in Europe is always a responsibility of the municipalities which propose measures to reduce CO2 emissions in terms of mobility aimed at reducing individual private transport (car). The European Commission's Action Plan on Urban Mobility calls for an increase in the take-up of Sustainable Urban Mobility Plans in Europe. SUMPs aim to create a sustainable urban transport system. Europe has got some long term initiatives and has been using some evaluation procedures, many of them through European projects. Nevertheless, the weak point with the SUMPs in Spain, has been the lack of concern about the evaluation and the effectiveness of the measures implemented in a SUMP. For this reason, it is difficult to know exactly whether or not the SUMPs have positively influenced in the modal split of the cities, and its contribution to reduce CO2 levels. The case of the City of Burgos is a very illustrative example as it developed a CiViTAS project during the years 2005-2009, with a total investment of 6M?. The results have been considered as ?very successful? even at European level. The modal split has changed considerably for better, The cost-effectiveness ratio of the SUMP in the city can be measured with the CO2 ton saved, specifically 36 ? per CO2 ton saved, which is fully satisfactory and in line with calculations from other European researchers. Additionally, the authors propose a single formula to measure the effectiveness of the activities developed under the umbrella of a SUMP.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

Part 1: Introduction

Relevância:

60.00% 60.00%

Publicador:

Resumo:

O processo do neoliberalismo revela, no seu momento mais intenso, a submissão de todos os níveis da vida humana a transações mercantis, a inclinação capitalista à mercantilização mundial. Desse modo, as propostas neoliberais objetivam, sobretudo, a criação de um sentimento único para garantir o sucesso de seus ideais de globalização, de livre-economia e de Estado mínimo, não interventor nas políticas sociais, mas servidor do mercado. Dentre os artifícios utilizados pelos propositores neoliberais encontram-se as Parcerias Público-Privadas (PPPs) com as quais este artigo se ocupa. A partir da conceituação e da caracterização do instrumento Parcerias Público-Privadas (PPPs), foram analisados projetos nacionais e europeus desenvolvidos através das PPPs.Os temas desses projetos envolvem Tecnologia de Informática e Comunicação e eLearning (ensino + tecnologia + a distância). As reflexões evidenciaram que as parcerias entre governos e empresas multinacionais podem levar a uma perda de controle por parte do Estado sobre a formação educacional dos cidadãos e da perda de identidade de seus currículos. Por outro lado, essas parcerias trazem ganhos políticos aos Estados e econômicos às empresas.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

This volume is a collection of the work done in a three years-lasting PhD, focused in the analysis of Central and Southern Adriatic marine sediments, deriving from the collection of a borehole and many cores, achieved thanks to the good seismic-stratigraphic knowledge of the study area. The work was made out within European projects EC-EURODELTA (coordinated by Fabio Trincardi, ISMAR-CNR), EC-EUROSTRATAFORM (coordinated by Phil P. E. Weaver, NOC, UK), and PROMESS1 (coordinated by Serge Bernè, IFREMER, France). The analysed sedimentary successions presented highly expanded stratigraphic intervals, particularly for the last 400 kyr, 60 kyr and 6 kyr BP. These three different time-intervals resulted in a tri-partition of the PhD thesis. The study consisted of the analysis of planktic and benthic foraminifers’ assemblages (more than 560 samples analysed), as well as in preparing the material for oxygen and carbon stable isotope analyses, and interpreting and discussing the obtained dataset. The chronologic framework of the last 400 kyr was achieved for borehole PRAD1-2 (within the work-package WP6 of PROMESS1 project), collected in 186.5 m water depth. The proposed chronology derives from a multi-disciplinary approach, consisting of the integration of numerous and independent proxies, some of which analysed by other specialists within the project. The final framework based on: micropaleontology (calcareous nannofossils and foraminifers’ bioevents), climatic cyclicity (foraminifers’ assemblages), geochemistry (oxygen stable isotope, made out on planktic and benthic records), paleomagnetism, radiometric ages (14C AMS), teprhochronology, identification of sapropel-equivalent levels (Se). It’s worth to note the good consistency between the oxygen stable isotope curve obtained for borehole PRAD1-2 and other deeper Mediterranean records. The studied proxies allowed the recognition of all the isotopic intervals from MIS10 to MIS1 in PRAD1-2 record, and the base of the borehole has been ascribed to the early MIS11. Glacial and interglacial intervals identified in the Central Adriatic record have been analysed in detail for the paleo-environmental reconstruction, as well. For instance, glacial stages MIS6, MIS8 and MIS10 present peculiar foraminifers’ assemblages, composed by benthic species typical of polar regions and no longer living in the Central Adriatic nowadays. Moreover, a deepening trend in the paleo-bathymetry during glacial intervals was observed, from MIS10 (inner-shelf environment) to MIS4 (mid-shelf environment).Ten sapropel-equivalent levels have been recognised in PRAD1-2 Central Adriatic record. They showed different planktic foraminifers’ assemblages, which allowed the first distinction of events occurred during warm-climate (Se5, Se7), cold-climate (Se4, Se6 and Se8) and temperate-intermediate-climate (Se1, Se3, Se9, Se’, Se10) conditions, consistently with literature. Cold-climate sapropel equivalents are characterised by the absence of an oligotrophic phase, whereas warm-temeprate-climate sapropel equivalents present both the oligotrophic and the eutrophic phases (except for Se1). Sea floor conditions vary, according to benthic foraminifers’ assemblages, from relatively well oxygenated (Se1, Se3), to dysoxic (Se9, Se’, Se10), to highly dysoxic (Se4, Se6, Se8) to events during which benthic foraminifers are absent (Se5, Se7). These two latter levels are also characterised by the lamination of the sediment, feature never observed in literature in such shallow records. The enhanced stratification of the water column during the events Se8, Se7, Se6, Se5, Se4, and the concurring strong dilution of shallow water, pointed out by the isotope record, lead to the hypothesis of a period of intense precipitation in the Central Adriatic region, possibly due to a northward shift of the African Monsoon. Finally, the expression of Central Adriatic PRAD1-2 Se5 equivalent was compared with the same event, as registered in other Eastern Mediterranean areas. The sequence of substantially the same planktic foraminifers’ bioevents has been consistently recognised, indicating a similar evolution of the water column all over the Eastern Mediterranean; yet, the synchronism of these events cannot be demonstrated. A high resolution analysis of late Holocene (last 6000 years BP) climate change was carried out for the Adriatic area, through the recognition of planktic and benthic foraminifers’ bioevents. In particular, peaks of planktic Globigerinoides sacculifer (four during the last 5500 years BP in the most expanded core) have been interpreted, based on the ecological requirements of this species, as warm-climate, arid intervals, correspondent to periods of relative climatic optimum, such as, for instance, the Medieval Warm Period, the Roman Age, the Late Bronze Age and the Copper Age. Consequently, the minima in the abundance of this biomarker could correspond to relatively cooler and more rainy periods. These conclusions are in good agreement with the isotopic and the pollen data. The Last Occurrence (LO) of G. sacculifer has been dated in this work at an average age of 550 years BP, and it is the best bioevent approximating the base of the Little Ice Age in the Adriatic. Recent literature reports the same bioevent in the Levantine Basin, showing a rather consistent age. Therefore, the LO of G. sacculifer has the potential to be extended to all the Eastern Mediterranean. Within the Little Ice Age, benthic foraminifer V. complanata shows two distinct peaks in the shallower Adriatic cores analysed, collected hundred kilometres apart, inside the mud belt environment. Based on the ecological requirements of this species, these two peaks have been interpreted as the more intense (cold and rainy) oscillations inside the LIA. The chronologic framework of the analysed cores is robust, being based on several range-finding 14C AMS ages, on estimates of the secular variation of the magnetic field, on geochemical estimates of the activity depth of 210Pb short-lived radionuclide (for the core-top ages), and is in good agreement with tephrochronologic, pollen and foraminiferal data. The intra-holocenic climate oscillations find out in the Adriatic have been compared with those pointed out in literature from other records of the Northern Hemisphere, and the chronologic constraint seems quite good. Finally, the sedimentary successions analysed allowed the review and the update of the foraminifers’ ecobiostratigraphy available from literature for the Adriatic region, thanks to the achievement of 16 ecobiozones for the last 60 kyr BP. Some bioevents are restricted to the Central Adriatic (for instance the LO of benthic Hyalinea balthica , approximating the MIS3/MIS2 boundary), others occur all over the Adriatic basin (for instance the LO of planktic Globorotalia inflata during MIS3, individuating Dansgaard-Oeschger cycle 8 (Denekamp)).

Relevância:

60.00% 60.00%

Publicador:

Resumo:

OntoTag - A Linguistic and Ontological Annotation Model Suitable for the Semantic Web 1. INTRODUCTION. LINGUISTIC TOOLS AND ANNOTATIONS: THEIR LIGHTS AND SHADOWS Computational Linguistics is already a consolidated research area. It builds upon the results of other two major ones, namely Linguistics and Computer Science and Engineering, and it aims at developing computational models of human language (or natural language, as it is termed in this area). Possibly, its most well-known applications are the different tools developed so far for processing human language, such as machine translation systems and speech recognizers or dictation programs. These tools for processing human language are commonly referred to as linguistic tools. Apart from the examples mentioned above, there are also other types of linguistic tools that perhaps are not so well-known, but on which most of the other applications of Computational Linguistics are built. These other types of linguistic tools comprise POS taggers, natural language parsers and semantic taggers, amongst others. All of them can be termed linguistic annotation tools. Linguistic annotation tools are important assets. In fact, POS and semantic taggers (and, to a lesser extent, also natural language parsers) have become critical resources for the computer applications that process natural language. Hence, any computer application that has to analyse a text automatically and ‘intelligently’ will include at least a module for POS tagging. The more an application needs to ‘understand’ the meaning of the text it processes, the more linguistic tools and/or modules it will incorporate and integrate. However, linguistic annotation tools have still some limitations, which can be summarised as follows: 1. Normally, they perform annotations only at a certain linguistic level (that is, Morphology, Syntax, Semantics, etc.). 2. They usually introduce a certain rate of errors and ambiguities when tagging. This error rate ranges from 10 percent up to 50 percent of the units annotated for unrestricted, general texts. 3. Their annotations are most frequently formulated in terms of an annotation schema designed and implemented ad hoc. A priori, it seems that the interoperation and the integration of several linguistic tools into an appropriate software architecture could most likely solve the limitations stated in (1). Besides, integrating several linguistic annotation tools and making them interoperate could also minimise the limitation stated in (2). Nevertheless, in the latter case, all these tools should produce annotations for a common level, which would have to be combined in order to correct their corresponding errors and inaccuracies. Yet, the limitation stated in (3) prevents both types of integration and interoperation from being easily achieved. In addition, most high-level annotation tools rely on other lower-level annotation tools and their outputs to generate their own ones. For example, sense-tagging tools (operating at the semantic level) often use POS taggers (operating at a lower level, i.e., the morphosyntactic) to identify the grammatical category of the word or lexical unit they are annotating. Accordingly, if a faulty or inaccurate low-level annotation tool is to be used by other higher-level one in its process, the errors and inaccuracies of the former should be minimised in advance. Otherwise, these errors and inaccuracies would be transferred to (and even magnified in) the annotations of the high-level annotation tool. Therefore, it would be quite useful to find a way to (i) correct or, at least, reduce the errors and the inaccuracies of lower-level linguistic tools; (ii) unify the annotation schemas of different linguistic annotation tools or, more generally speaking, make these tools (as well as their annotations) interoperate. Clearly, solving (i) and (ii) should ease the automatic annotation of web pages by means of linguistic tools, and their transformation into Semantic Web pages (Berners-Lee, Hendler and Lassila, 2001). Yet, as stated above, (ii) is a type of interoperability problem. There again, ontologies (Gruber, 1993; Borst, 1997) have been successfully applied thus far to solve several interoperability problems. Hence, ontologies should help solve also the problems and limitations of linguistic annotation tools aforementioned. Thus, to summarise, the main aim of the present work was to combine somehow these separated approaches, mechanisms and tools for annotation from Linguistics and Ontological Engineering (and the Semantic Web) in a sort of hybrid (linguistic and ontological) annotation model, suitable for both areas. This hybrid (semantic) annotation model should (a) benefit from the advances, models, techniques, mechanisms and tools of these two areas; (b) minimise (and even solve, when possible) some of the problems found in each of them; and (c) be suitable for the Semantic Web. The concrete goals that helped attain this aim are presented in the following section. 2. GOALS OF THE PRESENT WORK As mentioned above, the main goal of this work was to specify a hybrid (that is, linguistically-motivated and ontology-based) model of annotation suitable for the Semantic Web (i.e. it had to produce a semantic annotation of web page contents). This entailed that the tags included in the annotations of the model had to (1) represent linguistic concepts (or linguistic categories, as they are termed in ISO/DCR (2008)), in order for this model to be linguistically-motivated; (2) be ontological terms (i.e., use an ontological vocabulary), in order for the model to be ontology-based; and (3) be structured (linked) as a collection of ontology-based triples, as in the usual Semantic Web languages (namely RDF(S) and OWL), in order for the model to be considered suitable for the Semantic Web. Besides, to be useful for the Semantic Web, this model should provide a way to automate the annotation of web pages. As for the present work, this requirement involved reusing the linguistic annotation tools purchased by the OEG research group (http://www.oeg-upm.net), but solving beforehand (or, at least, minimising) some of their limitations. Therefore, this model had to minimise these limitations by means of the integration of several linguistic annotation tools into a common architecture. Since this integration required the interoperation of tools and their annotations, ontologies were proposed as the main technological component to make them effectively interoperate. From the very beginning, it seemed that the formalisation of the elements and the knowledge underlying linguistic annotations within an appropriate set of ontologies would be a great step forward towards the formulation of such a model (henceforth referred to as OntoTag). Obviously, first, to combine the results of the linguistic annotation tools that operated at the same level, their annotation schemas had to be unified (or, preferably, standardised) in advance. This entailed the unification (id. standardisation) of their tags (both their representation and their meaning), and their format or syntax. Second, to merge the results of the linguistic annotation tools operating at different levels, their respective annotation schemas had to be (a) made interoperable and (b) integrated. And third, in order for the resulting annotations to suit the Semantic Web, they had to be specified by means of an ontology-based vocabulary, and structured by means of ontology-based triples, as hinted above. Therefore, a new annotation scheme had to be devised, based both on ontologies and on this type of triples, which allowed for the combination and the integration of the annotations of any set of linguistic annotation tools. This annotation scheme was considered a fundamental part of the model proposed here, and its development was, accordingly, another major objective of the present work. All these goals, aims and objectives could be re-stated more clearly as follows: Goal 1: Development of a set of ontologies for the formalisation of the linguistic knowledge relating linguistic annotation. Sub-goal 1.1: Ontological formalisation of the EAGLES (1996a; 1996b) de facto standards for morphosyntactic and syntactic annotation, in a way that helps respect the triple structure recommended for annotations in these works (which is isomorphic to the triple structures used in the context of the Semantic Web). Sub-goal 1.2: Incorporation into this preliminary ontological formalisation of other existing standards and standard proposals relating the levels mentioned above, such as those currently under development within ISO/TC 37 (the ISO Technical Committee dealing with Terminology, which deals also with linguistic resources and annotations). Sub-goal 1.3: Generalisation and extension of the recommendations in EAGLES (1996a; 1996b) and ISO/TC 37 to the semantic level, for which no ISO/TC 37 standards have been developed yet. Sub-goal 1.4: Ontological formalisation of the generalisations and/or extensions obtained in the previous sub-goal as generalisations and/or extensions of the corresponding ontology (or ontologies). Sub-goal 1.5: Ontological formalisation of the knowledge required to link, combine and unite the knowledge represented in the previously developed ontology (or ontologies). Goal 2: Development of OntoTag’s annotation scheme, a standard-based abstract scheme for the hybrid (linguistically-motivated and ontological-based) annotation of texts. Sub-goal 2.1: Development of the standard-based morphosyntactic annotation level of OntoTag’s scheme. This level should include, and possibly extend, the recommendations of EAGLES (1996a) and also the recommendations included in the ISO/MAF (2008) standard draft. Sub-goal 2.2: Development of the standard-based syntactic annotation level of the hybrid abstract scheme. This level should include, and possibly extend, the recommendations of EAGLES (1996b) and the ISO/SynAF (2010) standard draft. Sub-goal 2.3: Development of the standard-based semantic annotation level of OntoTag’s (abstract) scheme. Sub-goal 2.4: Development of the mechanisms for a convenient integration of the three annotation levels already mentioned. These mechanisms should take into account the recommendations included in the ISO/LAF (2009) standard draft. Goal 3: Design of OntoTag’s (abstract) annotation architecture, an abstract architecture for the hybrid (semantic) annotation of texts (i) that facilitates the integration and interoperation of different linguistic annotation tools, and (ii) whose results comply with OntoTag’s annotation scheme. Sub-goal 3.1: Specification of the decanting processes that allow for the classification and separation, according to their corresponding levels, of the results of the linguistic tools annotating at several different levels. Sub-goal 3.2: Specification of the standardisation processes that allow (a) complying with the standardisation requirements of OntoTag’s annotation scheme, as well as (b) combining the results of those linguistic tools that share some level of annotation. Sub-goal 3.3: Specification of the merging processes that allow for the combination of the output annotations and the interoperation of those linguistic tools that share some level of annotation. Sub-goal 3.4: Specification of the merge processes that allow for the integration of the results and the interoperation of those tools performing their annotations at different levels. Goal 4: Generation of OntoTagger’s schema, a concrete instance of OntoTag’s abstract scheme for a concrete set of linguistic annotations. These linguistic annotations result from the tools and the resources available in the research group, namely • Bitext’s DataLexica (http://www.bitext.com/EN/datalexica.asp), • LACELL’s (POS) tagger (http://www.um.es/grupos/grupo-lacell/quees.php), • Connexor’s FDG (http://www.connexor.eu/technology/machinese/glossary/fdg/), and • EuroWordNet (Vossen et al., 1998). This schema should help evaluate OntoTag’s underlying hypotheses, stated below. Consequently, it should implement, at least, those levels of the abstract scheme dealing with the annotations of the set of tools considered in this implementation. This includes the morphosyntactic, the syntactic and the semantic levels. Goal 5: Implementation of OntoTagger’s configuration, a concrete instance of OntoTag’s abstract architecture for this set of linguistic tools and annotations. This configuration (1) had to use the schema generated in the previous goal; and (2) should help support or refute the hypotheses of this work as well (see the next section). Sub-goal 5.1: Implementation of the decanting processes that facilitate the classification and separation of the results of those linguistic resources that provide annotations at several different levels (on the one hand, LACELL’s tagger operates at the morphosyntactic level and, minimally, also at the semantic level; on the other hand, FDG operates at the morphosyntactic and the syntactic levels and, minimally, at the semantic level as well). Sub-goal 5.2: Implementation of the standardisation processes that allow (i) specifying the results of those linguistic tools that share some level of annotation according to the requirements of OntoTagger’s schema, as well as (ii) combining these shared level results. In particular, all the tools selected perform morphosyntactic annotations and they had to be conveniently combined by means of these processes. Sub-goal 5.3: Implementation of the merging processes that allow for the combination (and possibly the improvement) of the annotations and the interoperation of the tools that share some level of annotation (in particular, those relating the morphosyntactic level, as in the previous sub-goal). Sub-goal 5.4: Implementation of the merging processes that allow for the integration of the different standardised and combined annotations aforementioned, relating all the levels considered. Sub-goal 5.5: Improvement of the semantic level of this configuration by adding a named entity recognition, (sub-)classification and annotation subsystem, which also uses the named entities annotated to populate a domain ontology, in order to provide a concrete application of the present work in the two areas involved (the Semantic Web and Corpus Linguistics). 3. MAIN RESULTS: ASSESSMENT OF ONTOTAG’S UNDERLYING HYPOTHESES The model developed in the present thesis tries to shed some light on (i) whether linguistic annotation tools can effectively interoperate; (ii) whether their results can be combined and integrated; and, if they can, (iii) how they can, respectively, interoperate and be combined and integrated. Accordingly, several hypotheses had to be supported (or rejected) by the development of the OntoTag model and OntoTagger (its implementation). The hypotheses underlying OntoTag are surveyed below. Only one of the hypotheses (H.6) was rejected; the other five could be confirmed. H.1 The annotations of different levels (or layers) can be integrated into a sort of overall, comprehensive, multilayer and multilevel annotation, so that their elements can complement and refer to each other. • CONFIRMED by the development of: o OntoTag’s annotation scheme, o OntoTag’s annotation architecture, o OntoTagger’s (XML, RDF, OWL) annotation schemas, o OntoTagger’s configuration. H.2 Tool-dependent annotations can be mapped onto a sort of tool-independent annotations and, thus, can be standardised. • CONFIRMED by means of the standardisation phase incorporated into OntoTag and OntoTagger for the annotations yielded by the tools. H.3 Standardisation should ease: H.3.1: The interoperation of linguistic tools. H.3.2: The comparison, combination (at the same level and layer) and integration (at different levels or layers) of annotations. • H.3 was CONFIRMED by means of the development of OntoTagger’s ontology-based configuration: o Interoperation, comparison, combination and integration of the annotations of three different linguistic tools (Connexor’s FDG, Bitext’s DataLexica and LACELL’s tagger); o Integration of EuroWordNet-based, domain-ontology-based and named entity annotations at the semantic level. o Integration of morphosyntactic, syntactic and semantic annotations. H.4 Ontologies and Semantic Web technologies (can) play a crucial role in the standardisation of linguistic annotations, by providing consensual vocabularies and standardised formats for annotation (e.g., RDF triples). • CONFIRMED by means of the development of OntoTagger’s RDF-triple-based annotation schemas. H.5 The rate of errors introduced by a linguistic tool at a given level, when annotating, can be reduced automatically by contrasting and combining its results with the ones coming from other tools, operating at the same level. However, these other tools might be built following a different technological (stochastic vs. rule-based, for example) or theoretical (dependency vs. HPS-grammar-based, for instance) approach. • CONFIRMED by the results yielded by the evaluation of OntoTagger. H.6 Each linguistic level can be managed and annotated independently. • REJECTED: OntoTagger’s experiments and the dependencies observed among the morphosyntactic annotations, and between them and the syntactic annotations. In fact, Hypothesis H.6 was already rejected when OntoTag’s ontologies were developed. We observed then that several linguistic units stand on an interface between levels, belonging thereby to both of them (such as morphosyntactic units, which belong to both the morphological level and the syntactic level). Therefore, the annotations of these levels overlap and cannot be handled independently when merged into a unique multileveled annotation. 4. OTHER MAIN RESULTS AND CONTRIBUTIONS First, interoperability is a hot topic for both the linguistic annotation community and the whole Computer Science field. The specification (and implementation) of OntoTag’s architecture for the combination and integration of linguistic (annotation) tools and annotations by means of ontologies shows a way to make these different linguistic annotation tools and annotations interoperate in practice. Second, as mentioned above, the elements involved in linguistic annotation were formalised in a set (or network) of ontologies (OntoTag’s linguistic ontologies). • On the one hand, OntoTag’s network of ontologies consists of − The Linguistic Unit Ontology (LUO), which includes a mostly hierarchical formalisation of the different types of linguistic elements (i.e., units) identifiable in a written text; − The Linguistic Attribute Ontology (LAO), which includes also a mostly hierarchical formalisation of the different types of features that characterise the linguistic units included in the LUO; − The Linguistic Value Ontology (LVO), which includes the corresponding formalisation of the different values that the attributes in the LAO can take; − The OIO (OntoTag’s Integration Ontology), which  Includes the knowledge required to link, combine and unite the knowledge represented in the LUO, the LAO and the LVO;  Can be viewed as a knowledge representation ontology that describes the most elementary vocabulary used in the area of annotation. • On the other hand, OntoTag’s ontologies incorporate the knowledge included in the different standards and recommendations for linguistic annotation released so far, such as those developed within the EAGLES and the SIMPLE European projects or by the ISO/TC 37 committee: − As far as morphosyntactic annotations are concerned, OntoTag’s ontologies formalise the terms in the EAGLES (1996a) recommendations and their corresponding terms within the ISO Morphosyntactic Annotation Framework (ISO/MAF, 2008) standard; − As for syntactic annotations, OntoTag’s ontologies incorporate the terms in the EAGLES (1996b) recommendations and their corresponding terms within the ISO Syntactic Annotation Framework (ISO/SynAF, 2010) standard draft; − Regarding semantic annotations, OntoTag’s ontologies generalise and extend the recommendations in EAGLES (1996a; 1996b) and, since no stable standards or standard drafts have been released for semantic annotation by ISO/TC 37 yet, they incorporate the terms in SIMPLE (2000) instead; − The terms coming from all these recommendations and standards were supplemented by those within the ISO Data Category Registry (ISO/DCR, 2008) and also of the ISO Linguistic Annotation Framework (ISO/LAF, 2009) standard draft when developing OntoTag’s ontologies. Third, we showed that the combination of the results of tools annotating at the same level can yield better results (both in precision and in recall) than each tool separately. In particular, 1. OntoTagger clearly outperformed two of the tools integrated into its configuration, namely DataLexica and FDG in all the combination sub-phases in which they overlapped (i.e. POS tagging, lemma annotation and morphological feature annotation). As far as the remaining tool is concerned, i.e. LACELL’s tagger, it was also outperformed by OntoTagger in POS tagging and lemma annotation, and it did not behave better than OntoTagger in the morphological feature annotation layer. 2. As an immediate result, this implies that a) This type of combination architecture configurations can be applied in order to improve significantly the accuracy of linguistic annotations; and b) Concerning the morphosyntactic level, this could be regarded as a way of constructing more robust and more accurate POS tagging systems. Fourth, Semantic Web annotations are usually performed by humans or else by machine learning systems. Both of them leave much to be desired: the former, with respect to their annotation rate; the latter, with respect to their (average) precision and recall. In this work, we showed how linguistic tools can be wrapped in order to annotate automatically Semantic Web pages using ontologies. This entails their fast, robust and accurate semantic annotation. As a way of example, as mentioned in Sub-goal 5.5, we developed a particular OntoTagger module for the recognition, classification and labelling of named entities, according to the MUC and ACE tagsets (Chinchor, 1997; Doddington et al., 2004). These tagsets were further specified by means of a domain ontology, namely the Cinema Named Entities Ontology (CNEO). This module was applied to the automatic annotation of ten different web pages containing cinema reviews (that is, around 5000 words). In addition, the named entities annotated with this module were also labelled as instances (or individuals) of the classes included in the CNEO and, then, were used to populate this domain ontology. • The statistical results obtained from the evaluation of this particular module of OntoTagger can be summarised as follows. On the one hand, as far as recall (R) is concerned, (R.1) the lowest value was 76,40% (for file 7); (R.2) the highest value was 97, 50% (for file 3); and (R.3) the average value was 88,73%. On the other hand, as far as the precision rate (P) is concerned, (P.1) its minimum was 93,75% (for file 4); (R.2) its maximum was 100% (for files 1, 5, 7, 8, 9, and 10); and (R.3) its average value was 98,99%. • These results, which apply to the tasks of named entity annotation and ontology population, are extraordinary good for both of them. They can be explained on the basis of the high accuracy of the annotations provided by OntoTagger at the lower levels (mainly at the morphosyntactic level). However, they should be conveniently qualified, since they might be too domain- and/or language-dependent. It should be further experimented how our approach works in a different domain or a different language, such as French, English, or German. • In any case, the results of this application of Human Language Technologies to Ontology Population (and, accordingly, to Ontological Engineering) seem very promising and encouraging in order for these two areas to collaborate and complement each other in the area of semantic annotation. Fifth, as shown in the State of the Art of this work, there are different approaches and models for the semantic annotation of texts, but all of them focus on a particular view of the semantic level. Clearly, all these approaches and models should be integrated in order to bear a coherent and joint semantic annotation level. OntoTag shows how (i) these semantic annotation layers could be integrated together; and (ii) they could be integrated with the annotations associated to other annotation levels. Sixth, we identified some recommendations, best practices and lessons learned for annotation standardisation, interoperation and merge. They show how standardisation (via ontologies, in this case) enables the combination, integration and interoperation of different linguistic tools and their annotations into a multilayered (or multileveled) linguistic annotation, which is one of the hot topics in the area of Linguistic Annotation. And last but not least, OntoTag’s annotation scheme and OntoTagger’s annotation schemas show a way to formalise and annotate coherently and uniformly the different units and features associated to the different levels and layers of linguistic annotation. This is a great scientific step ahead towards the global standardisation of this area, which is the aim of ISO/TC 37 (in particular, Subcommittee 4, dealing with the standardisation of linguistic annotations and resources).

Relevância:

60.00% 60.00%

Publicador:

Resumo:

With the success of Web 2.0 we are witnessing a growing number of services and APIs exposed by Telecom, IT and content providers. Targeting the Web community and, in particular, Web application developers, service providers expose capabilities of their infrastructures and applications in order to open new markets and to reach new customer groups. However, due to the complexity of the underlying technologies, the last step, i.e., the consumption and integration of the offered services, is a non-trivial and time-consuming task that is still a prerogative of expert developers. Although many approaches to lower the entry barriers for end users exist, little success has been achieved so far. In this paper, we introduce the OMELETTE project and show how it addresses end-user-oriented telco mashup development. We present the goals of the project, describe its contributions, summarize current results, and describe current and future work.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

La Unión Europea acaba de lanzar una iniciativa para fomentar Participaciones Público Privadas (PPPs) mediante bonos de proyecto más atractivos a inversores institucionales para promover proyectos transeuropeos. Esto se logra a través de mecanismos de mejora crediticia como garantías de liquidez o tramos de deuda subordinada facilitados por el Banco Europeo de Inversiones. Esta iniciativa pretende evitar los problemas de liquidez experimentados actualmente por bancos comerciales en Europa para financiar megaproyectos. En este artículo exploramos las ventajas e inconvenientes de esta iniciativa para promover redes de infraestructuras transnacionales en Europa, y analizamos su aplicabilidad a otras áreas como Latino-América. The European Union recently launched an initiative to foster Public Private Partnerships (PPPs) for delivering Trans-European projects by making long-term project-bonds more appealing to institutional investors. This is achieved through credit-enhancement mechanisms such as partial stand-by liquidity guarantees, or layers of subordinated debt provided by the European Investment Bank. This initiative intends to circumvent the liquidity problems currently endured by commercial banks in Europe to fund megaprojects. In this paper we explore the advantages and drawbacks of this initiative for promoting transnational infrastructure networks in Europe, and analyse its applicability to other economic areas such as Latin America.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

El sector porcino ibérico es un sector típicamente español. La imagen que asocia el consumidor a los productos ibéricos es la correspondiente a los productos elaborados a partir de cerdos ibéricos en explotaciones extensivas, (dehesas), y cebados en montanera. Sin embargo, según los datos publicados por el Ministerio de Agricultura, Alimentación y Medio Ambiente, más de un 70 por ciento de los cerdos ibéricos que se producen en España son criados en granjas intensivas. El sector porcino ibérico está en pleno proceso de cambio, tanto por las variaciones que se derivan de la crisis que sufre el sector, como por las modificaciones que se esperan en la norma de calidad, recientemente acordadas. En este entorno en el que existe una reducción importante de los márgenes comerciales en todos los eslabones de la cadena de valor del jamón ibérico de bellota, adquiere importancia el estudio de su estructura y funcionamiento, la determinación de los aspectos que mejoran la eficiencia económica a lo largo de la misma y su concordancia a lo largo de ella. Así mismo, un sector de los consumidores comienza a buscar productos de calidad, se interesa por productos con determinados etiquetados que garantizan la calidad o unas determinadas prácticas de producción y elaboración, como sucede con los productos ecológicos. Existen numerosos estudios sobre distintos aspectos del sector porcino ibérico, realizados por distintas instituciones, como universidades, instituciones privadas y proyectos europeos. En general, se centran en la mejora de la producción de los cerdos ibéricos o en la mejora de la elaboración de los jamones. La novedad de la investigación realizada en la presente Tesis Doctoral reside en el estudio de los aspectos que determinan la mayor eficiencia económica de la cadena de valor del jamón ibérico de bellota de la denominación de origen de Guijuelo, y en las características que aporta al jamón ibérico de bellota la producción ecológica frente a la convencional. Resumen Los objetivos planteados en esta investigación son la mejora de la cadena de valor del jamón ibérico de bellota de la denominación de origen de Guijuelo, mediante el estudio de la problemática estructural y de funcionamiento de la misma. Como objetivos específicos se proponen la mejora de la eficiencia económica en las relaciones comerciales entre los distintos eslabones de la cadena de valor, y el impulso de la oferta de los productos ecológicos de ibérico, mediante el estudio comparativo de las producciones convencionales y ecológicas. Para alcanzar estos objetivos la metodología utilizada es la cadena de valor agroalimentaria, utilizando como herramientas de análisis el método Delphi y la realización de un análisis DAFO del sector porcino ibérico. Las fuentes de información primarias utilizadas proceden por una parte de la colaboración realizada en el proyecto europeo Q-Porkchains, coordinado por el Dr. Jacques H. Trienekens, y por otra parte de los cuestionarios especialmente diseñados para el análisis Delphi realizado. Las fuentes de información secundarias proceden de artículos académicos publicados, de artículos de revistas especializadas en el sector y de informes y estadísticas publicados por el Ministerio de Agricultura, Alimentación y Medio Ambiente. Las conclusiones de la investigación son las siguientes. Las relaciones comerciales de la cadena de valor estudiada están gestionadas, ya que se otorga gran valor a las relaciones comerciales basadas en la confianza, tanto en el aprovisionamiento de productos como en la venta de los mismos en cada eslabón. Sin embargo, no se gestiona adecuadamente la gama de productos que se oferta a lo largo de la cadena de valor, puesto que los eslabones de la misma obtienen diferente eficiencia económica para el mismo producto. Por otra parte, aunque la denominación de origen de Guijuelo es una marca conocida por el consumidor, el esfuerzo por tener una marca colectiva de calidad no ha sido totalmente satisfactorio, en especial para el eslabón de la distribución. Por esta razón, es importante resaltar que es recomendable la creación de una marca propia que cuide la calidad con especial atención, modificando incluso las empresas integrantes de la cadena de valor si el objetivo de calidad no coincide. Es importante garantizar la calidad del producto al consumidor, mediante la trasmisión la información generada por los informes de inspección y certificación realizados por las entidades correspondientes en relación a la norma de calidad de los productos de cerdo ibérico y por el Consejo Regulador de Guijuelo. Se requiere un estudio detallado de los posibles nichos de mercado para llegar a los clientes que buscan un producto de calidad diferenciada como es el jamón ibérico de denominación de origen de Guijuelo. Dentro de los posibles clientes debe tenerse en consideración al colectivo de personas alérgicas, ya que no se encuentran con facilidad en el mercado productos cárnicos de cerdo libres de alérgenos. Por último, la innovación es muy importante en este sector, y pueden ofertarse nuevos productos que ayuden al acortamiento de la cadena de valor, como podría ser la carne fresca de cerdo ibérico de bellota. La producción ecológica de porcino ibérico, en la actualidad no es significativa. El manejo que se realiza de los cerdos ibéricos convencionales de bellota es muy similar al manejo que se realiza de los cerdos ecológicos, salvo en la alimentación que reciben los cerdos ibéricos antes del cebo en montanera y en los medicamentos permitidos en ambas producciones. La producción de porcino ecológico de bellota puede ser interesante para la exportación. Los jamones ibéricos de bellota dispondrían de un sello que garantiza un manejo tradicional –como se realiza en muchas explotaciones extensivas de cerdos ibéricos de bellota-, y que es apreciado por los consumidores de otros países, dispuestos a pagar por productos de calidad diferenciada. Este tipo de producción quizá podría solventar la limitación al ámbito nacional de la norma de calidad de los productos de cerdo ibérico. La falta de claridad y unanimidad en el sector sobre la calidad de los productos de ibéricos que se percibe en los cambios previstos en la legislación, y el interés del eslabón de la industria elaboradora de jamones en calidades no ligadas a producciones extensivas sino intensivas, por su mayor volumen de ventas y sus mejores rendimientos, mueven a los productores de “ibérico puro de bellota” a desmarcarse del sector “ibérico”, donde el peso del producto ibérico cruzado de pienso tiene mucho peso en las decisiones que se toman. The Iberian pork industry is a typically Spanish industry. The associated image by the consumer to these products corresponds to products made from Iberian pigs in extensive farms (meadows), and fattened in open range. However, according to data released by the Ministry of Agriculture, Food and Environment, more than 70 percent of the Iberian pigs produced in Spain are raised on factory farms. The Iberian pork industry is in process of change, both by modifications resulting from the crisis in the sector, and the expected ones in the quality standard, recently agreed. In this environment where there is a significant reduction in trade margins in all links of the value chain of Iberian ham, becomes important to study its structure and operation, identifying the aspects that improve economic efficiency along the value chain and their concordance along it. Likewise, part of the consumers begins searching for quality products, they are interested in certain products that guarantee some sort of quality or production and processing practices, like organic products. There are several studies on various aspects of the Iberian pork industry, made by different institutions, such as universities, private institutions and some European projects. In general, their goal is focused on improving the Iberian hog production or on improving the elaboration of hams. The novelty of the research conducted in this PhD thesis lies in the study of the aspects that determine the economic efficiency of the value chain of Iberian ham from Guijuelo designation of origin, and the comparison between the features that organic or conventional production of Iberian ham brings to it. The objectives propound in this research are the improvement of the value chain of Iberian ham from Guijuelo designation of origin, through the study of its structural and operational problematic. The proposed specific objectives are the improvement of the economic efficiency Resumen in trade relations between the different links in the value chain, and the promotion of the supply of Iberian organic pork products, through the comparative study of conventional and organic productions. In order to achieve these objectives, the used methodology is the agrifood value chain, using as analysis tools the Delphi method, and a SWOT analysis of the Iberian pork industry. The primary sources of information come partly from the collaboration in the European project Q-Porkchains, coordinated by Dr. Jacques Trienekens, and partly from the questionnaires specially designed for the Delphi analysis performed. The secondary sources come from published academic papers, specialized journal articles and reports published by the Ministry of Agriculture, Food and Environment. The research conclusions are as follows. Trade relations in the studied value chain are managed, as it is granted a great value to business relationships based on trust, both in the supply of products and in the sale in each link. However, the portfolio that is offered throughout the value chain is not properly managed, since the economic efficiency obtained for the same product is different in each link. Moreover, although the designation of origin of Guijuelo is a well-known brand, the effort made to acquire a collective quality trademark has not been entirely satisfactory, especially in the distribution link. Due to it, it is important to highlight that it is recommended to create a brand focused on quality, and indeed to change the firms integrating the vale chain, if the goal is not shared by all the companies. It is essential to ensure the quality of the product to the consumer, through the transmission of the available information related to the quality standard of Iberian pork products and of the Guijuelo Regulating Council. A detailed study of potential niche markets is required, in order to reach customers looking for a differentiated quality product as Iberian ham of Guijuelo designation of origin. It should be taken into consideration the group of allergy sufferers as potential customers, as they do not easily find pork products in the market free of allergens. Innovation is very important in this industry, and new products could be commercialized that help to shorten the value chain, as fresh acorn Iberian pork. Nowadays, the organic production of Iberian pork products is not significant. The operation is very similar in both conventional and organic production of Iberian pork products, except in the food received by Iberian hogs and the sort of medicinal treatments allowed in each production. The organic acorn Iberian pork production could be interesting for export. Acorn Iberian hams would have a well-known hallmark, which guarantees a traditional production, -as it is done in many extensive holdings of Iberian acorn-pigs-, and that is appreciated by consumers in other countries This niche of consumers could pay for differentiated quality products. Perhaps, this sort of production could solve the limitation to national scope of the quality standard of Iberian pork products. The lack of clarity and unanimity in the industry on the quality of Iberian pork products has been perceived in the agreed changes of the legislation as well as the interest of the processing industry in hams of quality categories which are not linked to extensive but intensive productions, due to their higher sales volume and better yields. This situation pushed "acorn pure Iberian " pigs farmers to split from the Iberian industry, where the importance of cross Iberian pig from intensive farms is growing so high that they can influence the negotiations and decisions taken inside and outside the industry.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Los años cincuenta y sesenta son los años de la incorporación definitiva de la arquitectura española al panorama internacional. Entre los arquitectos que protagonizan ese salto sin retorno, se encuentra el grupo de aquellos que unos años más tarde serán denominados por Juan Daniel Fullaondo como Escuela de Madrid. Carlos Flores, en su libro Arquitectura Española Contemporánea 1880-1950, se refiere a esos arquitectos como aquellos que se aplicaban a la difícil tarea de restablecer en España un tipo de arquitectura que conectaba con las teorías, soluciones y lenguajes establecidos por Europa durante las primeras décadas del siglo XX. Sigfried Giedion plantea en Espacio, Tiempo y Arquitectura el origen de una nueva tradición, surgida a partir de la revolución óptica de principios de siglo. Con tradición se refiere a una nueva cultura, que abarca la interrelación de las diferentes actividades del hombre: la similitud de los métodos que se usan en la arquitectura, la construcción, la pintura, el urbanismo o la ciencia. Esa novedad, fundamentada en su independencia y desvinculación con el periodo anterior, se inscribe dentro del esquema evolutivo que Thomas Kuhn plantea en su texto La Estructura de la Revoluciones Científicas, conforme a periodos no acumulativos. Kuhn habla del surgimiento de anomalías en cada periodo, origen de las crisis de pensamiento cuya explicación precisará un necesario cambio paradigmático. En la ciencia, en el campo de la óptica Thomas Young demuestra a principios del siglo XIX la naturaleza ondulatoria de la luz con su experimento de doble rendija; en el electromagnetismo se produce el salto conceptual que supone la postulación de la existencia del campo eléctrico por parte de Michael Faraday, y en termodinámica la consideración apuntada por Planck de que la radiación de la energía de produce de forma discreta, a través de cuantos. En las artes plásticas, paralelamente, Gleizes y Metzinger, en su recopilación de logros cubistas recogida en Sobre el Cubismo, hablan de la evolución sufrida durante el siglo XIX por la pintura: desde el idealismo de principios de siglo, para pasando por el realismo y la representación impresionista de la realidad, concluir prescindiendo de la perspectiva clásica. También la matemática, una vez desarrolladas por Gauss o Lobachevsky y Bolyai geometrías coherentes que incumplen el quinto postulado de Euclides, terminará dando validez a través de Riemann a los espacios ambiente en los que habitan dichas geometrías, desvinculando la relación directa entre espacio geométrico –el espacio ambiente al que da lugar un tipo de geometría- y el espacio físico. Capi Corrales refleja en su libro Contando el Espacio, cómo hasta la teoría de la relatividad y el cubismo, las geometrías no euclídeas no se hicieron notorias también fuera del campo de las matemáticas. El origen de la nueva tradición con la que Giedion se refiere a la nueva cultura de la modernidad coincide con los saltos paradigmáticos que suponen la teoría de la relatividad en las ciencias y el cubismo en las artes plásticas. Ambas se prolongan durante las primeras décadas hasta la teoría cuántica y la abstracción absoluta, barreras que los dos principales precursores de la relatividad y el cubismo, Einstein y Picasso, nunca llegan a franquear. En ese sentido Giedion habla también, además del origen, de su desarrollo, e incorpora las aportaciones periféricas en la arquitectura de Brasil, Japón o Finlandia, incluyendo por tanto la revisión orgánica propugnada por Zevi como parte de esa nueva tradición, quedando abierta a la incorporación tardía de nuevas aportaciones al desarrollo de esa cultura de la modernidad. Eliminado el concepto de la estética trascendental de Kant del tiempo como una referencia absoluta, y asumido el valor constante de la velocidad de la luz, para la teoría de la relatividad no existe una simultaneidad auténtica. Queda así fijada la velocidad de la luz como uno de los límites del universo, y la equivalencia entre masa y energía. En el cubismo la simultaneidad espacial viene motivada por la eliminación del punto de vista preferente, cuyo resultado es la multiplicidad descriptiva de la realidad, que se visualiza en la descomposición en planos, tanto del objeto como del espacio, y la consecuente continuidad entre fondo y figura que en arquitectura se refleja en la continuidad entre edificio y territorio. Sin la consideración de un punto de vista absoluto, no existe una forma auténtica. El cubismo, y su posterior desarrollo por las vanguardias plásticas, hacen uso de la geometría como mecanismo de recomposición de la figura y el espacio, adoptando mecanismos de penetración, superposición y transparencia. Gyorgy Kepes indica en El Lenguaje de la Visión que la descomposición cubista del objeto implica la sucesiva autonomía de los planos, hasta convertirse en elementos constituyentes. Algo que refleja las axonometrías arquitectónicas de Van Doesburg y que culmina con los espacios propuestos por Mies van der Rohe en sus primeros proyectos europeos. Estos mecanismos, encuentran eco en los primeros planteamientos de Javier Carvajal: en la ampliación del Panteón de españoles del cementerio de Campo Verano, un recinto virtual reconstruido mentalmente a partir del uso de tres únicos planos; o en el Pabellón de Nueva York, que organiza su planta baja desde el recorrido, introduciendo el parámetro temporal como una dimensión más. Al uso diferenciado del plano como elemento constituyente, Carvajal incorpora su plegado y su disposición conformando envolventes como mecanismo de cualificación espacial y formal, potenciando la prolongación entre arquitectura y territorio. Una continuidad que quedará culminada en las dos viviendas unifamiliares construidas en Somosaguas. La descomposición volumétrica conduce a unos niveles de abstracción que hace precisa la incorporación de elementos de la memoria -fuentes, patios, celosías…- a modo de red de señales, como las que Picasso y Braque introducen en sus cuadros para permitir su interpretación. Braque insiste en el interés por el espacio que rodea a los objetos. Una búsqueda de la tactilidad del espacio contraria a la perspectiva que aleja el objeto del observador, y que en los jardines de las viviendas de Somosaguas parece emanar de su propia materialidad. Un espacio táctil alejado del espacio geométrico y que Braque identifica con el espacio representativo en el que Poincaré, en La Ciencia y la Hipótesis, ubica nuestras sensaciones. Desdibujar los límites del objeto prolonga el espacio indefinidamente. Con el paso en el arte griego del mito al logos, se abre paso a la matemática como herramienta de comprensión de la naturaleza hasta el siglo XIX. Leon Lederman, en Simetría y la Belleza del Universo, apunta a que una de las mayores contribuciones de la teoría de Einstein es hacer cambiar el modo de pensar la naturaleza, orientándolo hacia la búsqueda de los principios de simetría que subyacen bajo las leyes físicas. Considerando que la simetría es la invariancia de un objeto o un sistema frente a una transformación y que las leyes físicas son las mismas en cualquier punto del espacio, el espacio de nuestro universo posee una simetría traslacional continua. En la ocupación del espacio de las primeras propuestas de Corrales y Molezún aparecen estructuras subyacentes que responden a enlosetados: paralelogramos sometidos a transformaciones continuas, que la naturaleza identifica tridimensionalmente con los grupos cristalográficos. Las plantas del museo de Arte Contemporáneo de la Castellana, la residencia de Miraflores, el pabellón de Bruselas o la torre Peugeot pertenecen a este grupo. La arquitectura como proceso de ocupación continua del territorio y de su trasposición al plano de cubierta, se materializa en líneas estructurales coincidentes con la estructura matemática de sus simetrías de traslación cuya posibilidad de prolongación infinita queda potenciada por el uso de la envolvente transparente. Junto a esta transparencia literal, inherente al material, Colin Rowe y Robert Slutzky nos alertan sobre otra transparencia inherente a la estructura: la transparencia fenomenal, ilustrada por los cuadros de Juan Gris, y cuya intuición aparece reflejada en la casa Huarte en Puerta de Hierro de Madrid. Corrales y Molezún insisten en una lectura de su volumetría alejada de la frontalidad, en la que los contornos de sus cubiertas inclinadas y las visuales tangenciales sugeridas por la organización de sus recorridos introducen una estructura diagonal que se superpone al entendimiento ortogonal de su planta, dibujando una intrincada red de líneas quebradas que permiten al espacio fluctuar entre las secuencia volumétrica propuesta. Los datos relativos al contenido energético de la luz y el concepto de átomo parten de la consideración de la emisión de energía en cuantos realizada por Planck, y concluyen con una circunstancia paradójica: la doble naturaleza de la luz -demostrada por la explicación de Einstein del efecto fotoeléctrico- y la doble naturaleza de la materia -asumida por Bohr y demostrada por el efecto Compton-. Schrödinger y Heisenberg formularán finalmente la ecuación universal del movimiento que rige en las ondas de materia, y cuya representación matemática es lo que se conoce como función de onda. El objeto es así identificado con su función de onda. Su ondulatoriedad expresará la probabilidad de encontrarse en un lugar determinado. Gyorgy Kepes subraya la necesidad de simplificar el lenguaje para pasar de la objetividad que aún permanece en la pintura cubista a la abstracción total del espacio. Y es así como los artistas plásticos reducen los objetos a simples formas geométricas, haciendo aflorar a la vez, las fuerzas plásticas que los tensionan o equilibran, en un proceso que acaba por eliminar cualquier atisbo de materia. Robert Rosenblum en La Pintura Moderna y la Tradición del Romanticismo Nórdico habla de cómo ese rechazo de la materia en favor de un vacío casi impalpable, campos luminosos de color denso que difunden un sereno resplandor y parecen engendrar las energías elementales de la luz natural, está directamente vinculado a la relación con la naturaleza que establece el romanticismo nórdico. La expresión de la energía de la naturaleza concentrada en un vacío que ya había sido motivo de reflexión para Michael Faraday en su postulación del concepto de campo eléctrico. Sáenz de Oíza incide en la expresión de la condición material de la energía en su propuesta junto a José Luis Romany para la capilla en el Camino de Santiago. La evocación de diferentes fuerzas electromagnéticas, las únicas junto a las gravitatorias susceptibles de ser experimentadas por el hombre, aparecerán visualizadas también en el carácter emergente de algunas de sus obras: el Santuario de Aránzazu o Torres Blancas; pero también en la naturaleza fluyente de sus contornos, la dispersión perimetral de los espacios -el umbral como centro del universoo la configuración del límite como respuesta a las tensiones germinales de la naturaleza. Miguel Fisac, a la vuelta de su viaje a los países nórdicos, aborda una simplificación lingüística orientada hacia la adecuación funcional de los espacios. En el Instituto de Daimiel, el Instituto de formación del profesorado o los complejos para los Padres Dominicos en Valladolid o Alcobendas, organiza progresivamente la arquitectura en diferentes volúmenes funcionales, incidiendo de un modo paralelo en la manifestación de los vínculos que se establecen entre dichos volúmenes como una visualización de las fuerzas que los tensionan y equilibran. En ellos la prolongación de la realidad física más allá de los límites de la envolvente ya es algo más que una simple intuición. Un proceso en el que el tratamiento de la luz como un material de construcción más, tendrá un especial protagonismo. En la iglesia de la Coronación, la iluminación del muro curvo escenifica la condición ondulatoria de la luz, manifestándose como si de un patrón de interferencia se tratara. Frente a la disolución de lo material, el espacio se manifiesta aquí como un medio denso, alejado de la tradicional noción de vacío. Una doble naturaleza, onda y partícula, que será intuido también por Fisac en la materia a través de su uso comprometido del hormigón como único material de construcción. Richard Feynmann nos alerta de la ocupación del espacio por multitud de fuerzas electromagnéticas que, al igual que la luz, precisan de receptores específicos para captar su presencia. Sus célebres diagramas suponen además la visualización definitiva de los procesos subatómicos. Al igual que la abstracción absoluta en las artes plásticas, esas representaciones diagramáticas no son asimilables a imágenes obtenidas de nuestra experiencia. Una intuición plasmada en el uso del diagrama, que irán adquiriendo progresivamente los dibujos de Alejandro de la Sota. La sección del gimnasio Maravillas recoge los trazos de sus principales elementos constructivos: estructura, cerramientos, compartimentaciones…, pero también, y con la misma intensidad, los de las fuerzas que generan su espacio, considerando así su condición de elementos constituyentes. El vacío, nos deja claro Sota, es el lugar donde habitan dichas tensiones. La posterior simplificación de las formas acompañadas de la obsesión por su aligeramiento, la casi desaparición de la envolvente, incide en aquella idea con la que Paul Klee define la actividad del artista en su Teoría del Arte Moderno, y en la que se transmite el distanciamiento hacia lo aparente: No se trata de reproducir lo visible, se trata de volver visible. Así, en Bankunión y Aviaco, como en tantos otros proyectos, frente al objetivo de la forma, Sota plantea el límite como la acotación de un ámbito de actuación. Su propia representación aséptica y diagramática transmite la renuncia a una especificidad espacial. Gilles Deleuze expresa ese posicionamiento en Pintura, el Concepto de Diagrama: el diagrama como la posibilidad de cuadros infinitos, o la posibilidad infinita de cuadros. Aparece así una concepción probabilística del espacio en la que frente a la renuncia por la forma, la tendencia al aligeramiento, y lo difuso de su definición – ideas claras, definición borrosa, en palabras de Llinás referidas al modo de operar de Sota-, la insistente atención a algunos elementos como escaleras, protecciones o miradores parece trasmitir la idea de que la arquitectura queda condensada en aquellos acontecimientos que delatan su condición dinámica, transitoria. Primando la relación frente al objeto, el vínculo frente a lo tangible. English summary. The fifties and sixties were the years of the final incorporation of Spanish architecture to the international scene. Among the architects who star that no return leap, is the group of those who a few years later will be named by Juan Daniel Fullaondo as Escuela de Madrid. Carlos Flores, in his book Arquitectura Española Contemporánea 1880-1950, refers to those architects as those that applied to the difficult task of restoring in Spain an architecture that connected with theories, solutions and established languages in Europe during the first decades of the twentieth century. Sigfried Giedion proposes in Space, Time and Architecture, the origin of a new tradition, arising from the optical revolution at the beginning of the century. With tradition he refers to a new culture, covering the interplay of different human activities: the similarity of the methods used in architecture, building, painting, urban planning or science. This new feature, based on its independence and detachment from the previous period, is part of the evolutionary scheme that Thomas Kuhn proposes in his text The Structure of Scientific Revolutions, according to non-accumulative periods. Kuhn talks about the emergence of anomalies in each period, origin of thought crisis whose explanation will require a paradigm shift needed. In science, in the field of optical Thomas Young demonstrates at the early nineteenth century the wave nature of light with its double-slit experiment , in electromagnetism the postulation of the existence of the electric field by Michael Faraday involves a conceptual leap, and in thermodynamic, the consideration pointed by Planck about quantum energy radiation. In the arts, in a parallel process, Gleizes and Metzinger , in his collection of cubism achievements on their book Du Cubisme, speak of evolution occurring during the nineteenth century by the painting: from the idealism of beginning of the century, going for realism and impressionist representation of reality, and finishing regardless of the classical perspective . Mathematics also, once developed by Gauss and Lobachevsky and Bolyai consistent geometries that violate Euclid's fifth postulate , will end validating Riemann’s ambient spaces in which these geometries inhabit, decoupling the direct relationship between geometric space -the space environment that results in a type of geometry- , and physical space. Capi Corrales reflectes in his book Contando el Espacio, that non-Euclidean geometries were not noticeable outside the field of mathematics until the theory of relativity and cubism. The origin of the new tradition that Giedion relates to the new culture of modernity coincides with paradigmatic leaps pointed by the theory of relativity in science and Cubism in the visual arts. Both are extended during the first decades until quantum theory and absolute abstraction, barriers that the two main precursors of relativity and cubism, Einstein and Picasso never overcome. In that sense Giedion speaks about the origin, but also the development, and incorporates peripheral inputs from Brazil, Japan and Finland architecture, thus including organic revision advocated by Zevi as part of this new tradition, being open to the late addition of new contributions to the development of that culture of modernity. Removed the concept of Kant's transcendental aesthetics, of time as an absolute reference, and assumed the constant value of the speed of light, theory of relativity says there is no authentic concurrency. It is thus fixed the speed of light as one of the limits of the universe, and the equivalence of mass and energy. In cubism, spatial simultaneity results from the elimination of preferential points of view, resulting in the multiplicity descriptive of reality, which is displayed in decomposition levels, both the object and the space, and the resulting continuity between figure and background that architecture is reflected in the continuity between building and land. Without the consideration of an absolute point of view, there isn’t an authentic shape. Cubism, and its subsequent development by the vanguard arts, make use of geometry as a means of rebuilding the figure and space, taking penetration mechanisms, overlapping and transparency. Gyorgy Kepes suggest in Languaje of Vision, that cubist decomposition of the object involves successive planes autonomy, to become constituent elements. Something that reflects the Van Doesburg’s architectural axonometrics and culminates with the spaces proposed by Mies van der Rohe in his first European projects. These mechanisms are reflected in the first approaches by Javier Carvajal: the extension of Spanish Pantheon in Campo Verano Cemetery, virtual enclosure mentally reconstructed from 24 the use of only three planes, or in the Spanish Pavilion of New York, which organizes its ground floor from the tour, introducing the time parameter as an additional dimension. Carvajal adds to the differential use of the plane as a constituent, Carvajal incorporates its folding and forming enclosures available as a mechanism for spatial and formal qualification, promoting the extension between architecture and territory. A continuity that will be completed in the two houses built in Somosaguas. Volumetric decomposition, as the fragmentation achieved in the last cubist experiences, needs the incorporation of elements of memory - fountains, patios, shutters...- as a network of signals, such as those introduced by Picasso and Braque in their paintings to allow their interpretation. Braque insists in his interest in the space surrounding the objects. A search of the tactility of space contrary to the perspective, which moves the observer away from the object, and that in the gardens of Somosaguas seems to emanate from its own materiality. A tactile space away from the geometric space and Braque identified with the representative space in which Poincaré in La Science et l´hypothèse, located our feelings. To blur those boundaries of the object extends the space indefinitely. With the passage in Greek art from myth to logos, it opens up to mathematics as a tool for understanding the nature until the nineteenth century. Leon Lederman, in Symmetry and beautiful Universe, suggests that one of the greatest contributions of Einstein's theory is to change the mindset of nature, namely the search for symmetry principles that underlie physical laws. Considering that symmetry is the invariance of an object or system from a transformation and that physical laws are the same at any point in space, the space of our universe has a continuous translational symmetry. In the space occupation of the first proposals by Corrales and Molezún underlying structures appear that match enlosetados: parallelograms under continuous transformations, which nature identifies tridimensionally with the crystallographic groups. Plants in the Contemporary Art Museum in La Castellana, the residence in Miraflores, the Brussels pavilion or the Peugeot tower belong to this group. The architecture as a process of continuous occupation of the territory and of its transposition to the deck, embodied in structural lines coincide with the mathematical structure of the translational symmetry and infinite extension whose possibility is enhanced by the use of the transparent cover. Alongside this literal transparency inherent to the material, Colin Rowe and Robert Slutzky alert us another transparency inherent in the structure: phenomenal transparency, illustrated by the Juan Gris’ works, and whose intuition is reflected in the Huarte’s house in Puerta de Hierro in Madrid. Corrales and Molezún insist on a reading of its volume away from the frontal, in which the outline of their inclined roofs and tangential visual suggested by the organization of his circulations introduce a diagonal structure which overlaps the orthogonal understanding of its plant, drawing an intricate web of broken lines that allow the space fluctuate between the volumetric sequence proposal. Information concerning to the energy mean of light and the concept of atom start from the consideration by Plank about the energy emission, and conclude with a paradoxical situation: the dual nature of light - demonstrated by the explanation of Einstein's photoelectric effect-, and the dual nature of matter -assumed by Bohr and demonstrated by the Compton effect-. Finally, Schrödinger and Heisenberg will formulate the universal movement equation governing in undulatory matter, whose mathematical representation is what is known as a wave function. The object is thus identified with its wave function. Its undulatory expression speaks about the probability of being found in a certain place. Gyorgy Kepes emphasizess the need to simplify the language to move from the objectivity that still remains in the cubist painting to the total abstraction of the space. And this is how artists reduced the objects to simple geometric shapes, making emerge at a time, the plastic forces that tense or balance them, in a process that eventually eliminate any trace of matter. Robert Rosenblum in Modern Painting and the Northern Romantic Tradition. Friedrich to Rothko talks about how this rejection of matter in an almost impalpable vacuum: dense color light fields that broadcast a serene glow and seem to generate the elemental energies of natural light is directly linked to the relationship with nature that sets the northern romanticism. An expression of the power of nature concentrated in a vacuum which had been reason for thought by Michael Faraday in his application of the concept of electric field. Saenz de Oíza touches upon the material expression of the energy in its proposal with Jose Luis Romany to the chapel on the Camino de Santiago. The presence of electromagnetic forces, the only ones with the gravitational one capable of being experienced by the man will also visualize in the emerging nature of some of his works: the sanctuary of Aránzazu or Torres Blancas, but also in the flowing nature of its contours, and the inclusion of interest in the realization of space fluctuating boundary: the threshold as the center of the universe. Miguel Fisac, back from his trip to the Northern Countries, starts on a linguistic simplification oriented to the functional adequacy of spaces. In the Daimiel Institute, in the Institute to Teacher Formation or in the complex to the Dominican Fathers in Valladolid or Alcobendas, progressively organized into different functional volumes architecture, focusing in a parallel way in the manifestation of the links established between these volumes as a visualization of the forces that tense and balance them. The prolongation of the physical reality beyond the limits of the envelope is already something more than a simple intuition. A process in which the treatment of light as a construction material, have a special role. In the Coronation church, curved wall lighting dramatizes the undulatory condition of the light, manifesting as if an interference pattern is involved. Versus the dissolution of the material, the space is expressed here as a dense atmosphere, away from the traditional notion of the vacuum. A dual nature, wave and particle, which is also sensed by Fisac in his committed use of concrete as a unique construction material. Richard Feynman alerts us to the occupation of space by many electromagnetic forces, which like the light, require specific receptors to capture their presence. His famous diagrams also involve the final visualization of atomic processes. As absolute abstraction in the visual arts, these representations are not assimilated to images obtained from our experience. A diagrammatic nature, abstracted from figuration, which will obtein the pictures of Alejandro de la Sota. The section of Maravillas gym collects traces of its main building blocks: structure, enclosures... but also, and with the same intensity, of the forces that generate their space as constituent elements. Sota makes it clear: the vacuum is where inhabit these tensions. The subsequent simplification of forms, accompanied by the obsession with his lightening, the near disappearance of the envelope, touches upon that idea which Paul Klee defines the activity of the artist in his Modern Art Theory, the spacing out to the apparent: it is not to reproduce the visible, it is to turn visible. Thus, in Bankunión and Aviaco, as in many other projects, against the shape, raises the limit as the dimension of a scope. His own aseptic and diagrammatic representation transmits waiver to a spatial specificity that Gilles Deleuze clearly expressed in Painting. The Concept Diagram: The diagram as the possibility of infinite pictures, or infinite possibility of the picture. Thus appears the probabilistic concept of space in which, opposite to the diffuse of its definition -clear ideas, diffuse definition, as Llinas said- the insistent attention to some elements like stairs, guards or lookouts seems to concentrate the architecture in its dynamic condition, transitional. The relationship opposite the object, the link opposite the tangible.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

This paper tackles the optimization of applications in multi-provider hybrid cloud scenarios from an economic point of view. In these scenarios the great majority of solutions offer the automatic allocation of resources on different cloud providers based on their current prices. However our approach is intended to introduce a novel solution by making maximum use of divide and rule. This paper describes a methodology to create cost aware cloud applications that can be broken down into the three most important components in cloud infrastructures: computation, network and storage. A real videoconference system has been modified in order to evaluate this idea with both theoretical and empirical experiments. This system has become a widely used tool in several national and European projects for e-learning and collaboration purposes.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

El trabajo se enmarca dentro de los proyecto INTEGRATE y EURECA, cuyo objetivo es el desarrollo de una capa de interoperabilidad semántica que permita la integración de datos e investigación clínica, proporcionando una plataforma común que pueda ser integrada en diferentes instituciones clínicas y que facilite el intercambio de información entre las mismas. De esta manera se promueve la mejora de la práctica clínica a través de la cooperación entre instituciones de investigación con objetivos comunes. En los proyectos se hace uso de estándares y vocabularios clínicos ya existentes, como pueden ser HL7 o SNOMED, adaptándolos a las necesidades particulares de los datos con los que se trabaja en INTEGRATE y EURECA. Los datos clínicos se representan de manera que cada concepto utilizado sea único, evitando ambigüedades y apoyando la idea de plataforma común. El alumno ha formado parte de un equipo de trabajo perteneciente al Grupo de Informática de la UPM, que a su vez trabaja como uno de los socios de los proyectos europeos nombrados anteriormente. La herramienta desarrollada, tiene como objetivo realizar tareas de homogenización de la información almacenada en las bases de datos de los proyectos haciendo uso de los mecanismos de normalización proporcionados por el vocabulario médico SNOMED-CT. Las bases de datos normalizadas serán las utilizadas para llevar a cabo consultas por medio de servicios proporcionados en la capa de interoperabilidad, ya que contendrán información más precisa y completa que las bases de datos sin normalizar. El trabajo ha sido realizado entre el día 12 de Septiembre del año 2014, donde comienza la etapa de formación y recopilación de información, y el día 5 de Enero del año 2015, en el cuál se termina la redacción de la memoria. El ciclo de vida utilizado ha sido el de desarrollo en cascada, en el que las tareas no comienzan hasta que la etapa inmediatamente anterior haya sido finalizada y validada. Sin embargo, no todas las tareas han seguido este modelo, ya que la realización de la memoria del trabajo se ha llevado a cabo de manera paralela con el resto de tareas. El número total de horas dedicadas al Trabajo de Fin de Grado es 324. Las tareas realizadas y el tiempo de dedicación de cada una de ellas se detallan a continuación:  Formación. Etapa de recopilación de información necesaria para implementar la herramienta y estudio de la misma [30 horas.  Especificación de requisitos. Se documentan los diferentes requisitos que ha de cumplir la herramienta [20 horas].  Diseño. En esta etapa se toman las decisiones de diseño de la herramienta [35 horas].  Implementación. Desarrollo del código de la herramienta [80 horas].  Pruebas. Etapa de validación de la herramienta, tanto de manera independiente como integrada en los proyectos INTEGRATE y EURECA [70 horas].  Depuración. Corrección de errores e introducción de mejoras de la herramienta [45 horas].  Realización de la memoria. Redacción de la memoria final del trabajo [44 horas].---ABSTRACT---This project belongs to the semantic interoperability layer developed in the European projects INTEGRATE and EURECA, which aims to provide a platform to promote interchange of medical information from clinical trials to clinical institutions. Thus, research institutions may cooperate to enhance clinical practice. Different health standards and clinical terminologies has been used in both INTEGRATE and EURECA projects, e.g. HL7 or SNOMED-CT. These tools have been adapted to the projects data requirements. Clinical data are represented by unique concepts, avoiding ambiguity problems. The student has been working in the Biomedical Informatics Group from UPM, partner of the INTEGRATE and EURECA projects. The tool developed aims to perform homogenization tasks over information stored in databases of the project, through normalized representation provided by the SNOMED-CT terminology. The data query is executed against the normalized version of the databases, since the information retrieved will be more informative than non-normalized databases. The project has been performed from September 12th of 2014, when initiation stage began, to January 5th of 2015, when the final report was finished. The waterfall model for software development was followed during the working process. Therefore, a phase may not start before the previous one finishes and has been validated, except from the final report redaction, which has been carried out in parallel with the others phases. The tasks that have been developed and time for each one are detailed as follows:  Training. Gathering the necessary information to develop the tool [30 hours].  Software requirement specification. Requirements the tool must accomplish [20 hours].  Design. Decisions on the design of the tool [35 hours].  Implementation. Tool development [80 hours].  Testing. Tool evaluation within the framework of the INTEGRATE and EURECA projects [70 hours].  Debugging. Improve efficiency and correct errors [45 hours].  Documenting. Final report elaboration [44 hours].