30 resultados para higher level collective agrrement
em Universidad Politécnica de Madrid
Resumo:
A new formalism, called Hiord, for defining type-free higherorder logic programming languages with predicate abstraction is introduced. A model theory, based on partial combinatory algebras, is presented, with respect to which the formalism is shown sound. A programming language built on a subset of Hiord, and its implementation are discussed. A new proposal for defining modules in this framework is considered, along with several examples.
Resumo:
The aim of this study was to compare the race characteristics of the start and turn segments of national and regional level swimmers. In the study, 100 and 200-m events were analysed during the finals session of the Open Comunidad de Madrid (Spain) tournament. The “individualized-distance” method with two-dimensional direct linear transformation algorithm was used to perform race analyses. National level swimmers obtained faster velocities in all race segments and stroke comparisons,although significant inter-level differences in start velocity were only obtained in half (8 out of 16) of the analysed events. Higher level swimmers also travelled for longer start and turn distances but only in the race segments where the gain of speed was high. This was observed in the turn segments, in the backstroke and butterfly strokes and during the 200-m breaststroke event, but not in any of the freestyle events. Time improvements due to the appropriate extension of the underwater subsections appeared to be critical for the end race result and should be carefully evaluated by the “individualized-distance” method.
Resumo:
In the last two decades, there has been an important increase in research on speech technology in Spain, mainly due to a higher level of funding from European, Spanish and local institutions and also due to a growing interest in these technologies for developing new services and applications. This paper provides a review of the main areas of speech technology addressed by research groups in Spain, their main contributions in the recent years and the main focus of interest these days. This description is classified in five main areas: audio processing including speech, speaker characterization, speech and language processing, text to speech conversion and spoken language applications. This paper also introduces the Spanish Network of Speech Technologies (RTTH. Red Temática en Tecnologías del Habla) as the research network that includes almost all the researchers working in this area, presenting some figures, its objectives and its main activities developed in the last years.
Resumo:
This paper describes a novel architecture to introduce automatic annotation and processing of semantic sensor data within context-aware applications. Based on the well-known state-charts technologies, and represented using W3C SCXML language combined with Semantic Web technologies, our architecture is able to provide enriched higher-level semantic representations of user’s context. This capability to detect and model relevant user situations allows a seamless modeling of the actual interaction situation, which can be integrated during the design of multimodal user interfaces (also based on SCXML) for them to be adequately adapted. Therefore, the final result of this contribution can be described as a flexible context-aware SCXML-based architecture, suitable for both designing a wide range of multimodal context-aware user interfaces, and implementing the automatic enrichment of sensor data, making it available to the entire Semantic Sensor Web
Resumo:
OntoTag - A Linguistic and Ontological Annotation Model Suitable for the Semantic Web
1. INTRODUCTION. LINGUISTIC TOOLS AND ANNOTATIONS: THEIR LIGHTS AND SHADOWS
Computational Linguistics is already a consolidated research area. It builds upon the results of other two major ones, namely Linguistics and Computer Science and Engineering, and it aims at developing computational models of human language (or natural language, as it is termed in this area). Possibly, its most well-known applications are the different tools developed so far for processing human language, such as machine translation systems and speech recognizers or dictation programs.
These tools for processing human language are commonly referred to as linguistic tools. Apart from the examples mentioned above, there are also other types of linguistic tools that perhaps are not so well-known, but on which most of the other applications of Computational Linguistics are built. These other types of linguistic tools comprise POS taggers, natural language parsers and semantic taggers, amongst others. All of them can be termed linguistic annotation tools.
Linguistic annotation tools are important assets. In fact, POS and semantic taggers (and, to a lesser extent, also natural language parsers) have become critical resources for the computer applications that process natural language. Hence, any computer application that has to analyse a text automatically and ‘intelligently’ will include at least a module for POS tagging. The more an application needs to ‘understand’ the meaning of the text it processes, the more linguistic tools and/or modules it will incorporate and integrate.
However, linguistic annotation tools have still some limitations, which can be summarised as follows:
1. Normally, they perform annotations only at a certain linguistic level (that is, Morphology, Syntax, Semantics, etc.).
2. They usually introduce a certain rate of errors and ambiguities when tagging. This error rate ranges from 10 percent up to 50 percent of the units annotated for unrestricted, general texts.
3. Their annotations are most frequently formulated in terms of an annotation schema designed and implemented ad hoc.
A priori, it seems that the interoperation and the integration of several linguistic tools into an appropriate software architecture could most likely solve the limitations stated in (1). Besides, integrating several linguistic annotation tools and making them interoperate could also minimise the limitation stated in (2). Nevertheless, in the latter case, all these tools should produce annotations for a common level, which would have to be combined in order to correct their corresponding errors and inaccuracies. Yet, the limitation stated in (3) prevents both types of integration and interoperation from being easily achieved.
In addition, most high-level annotation tools rely on other lower-level annotation tools and their outputs to generate their own ones. For example, sense-tagging tools (operating at the semantic level) often use POS taggers (operating at a lower level, i.e., the morphosyntactic) to identify the grammatical category of the word or lexical unit they are annotating. Accordingly, if a faulty or inaccurate low-level annotation tool is to be used by other higher-level one in its process, the errors and inaccuracies of the former should be minimised in advance. Otherwise, these errors and inaccuracies would be transferred to (and even magnified in) the annotations of the high-level annotation tool.
Therefore, it would be quite useful to find a way to
(i) correct or, at least, reduce the errors and the inaccuracies of lower-level linguistic tools;
(ii) unify the annotation schemas of different linguistic annotation tools or, more generally speaking, make these tools (as well as their annotations) interoperate.
Clearly, solving (i) and (ii) should ease the automatic annotation of web pages by means of linguistic tools, and their transformation into Semantic Web pages (Berners-Lee, Hendler and Lassila, 2001). Yet, as stated above, (ii) is a type of interoperability problem. There again, ontologies (Gruber, 1993; Borst, 1997) have been successfully applied thus far to solve several interoperability problems. Hence, ontologies should help solve also the problems and limitations of linguistic annotation tools aforementioned.
Thus, to summarise, the main aim of the present work was to combine somehow these separated approaches, mechanisms and tools for annotation from Linguistics and Ontological Engineering (and the Semantic Web) in a sort of hybrid (linguistic and ontological) annotation model, suitable for both areas. This hybrid (semantic) annotation model should (a) benefit from the advances, models, techniques, mechanisms and tools of these two areas; (b) minimise (and even solve, when possible) some of the problems found in each of them; and (c) be suitable for the Semantic Web. The concrete goals that helped attain this aim are presented in the following section.
2. GOALS OF THE PRESENT WORK
As mentioned above, the main goal of this work was to specify a hybrid (that is, linguistically-motivated and ontology-based) model of annotation suitable for the Semantic Web (i.e. it had to produce a semantic annotation of web page contents). This entailed that the tags included in the annotations of the model had to (1) represent linguistic concepts (or linguistic categories, as they are termed in ISO/DCR (2008)), in order for this model to be linguistically-motivated; (2) be ontological terms (i.e., use an ontological vocabulary), in order for the model to be ontology-based; and (3) be structured (linked) as a collection of ontology-based
Resumo:
Competitive abstract machines for Prolog are usually large, intricate, and incorpórate sophisticated optimizations. This makes them difñcult to code, optimize, and, especially, maintain and extend. This is partly due to the fact that efñciency considerations make it necessary to use low-level languages in their implementation. Writing the abstract machine (and ancillary code) in a higher-level language can help harness this inherent complexity. In this paper we show how the semantics of basic components of an efficient virtual machine for Prolog can be described using (a variant of) Prolog which retains much of its semantics. These descriptions are then compiled to C and assembled to build a complete bytecode emulator. Thanks to the high level of the language used and its closeness to Prolog the abstract machine descriptions can be manipulated using standard Prolog compilation and optimization techniques with relative ease. We also show how, by applying program transformations selectively, we obtain abstract machine implementations whose performance can match and even exceed that of highly-tuned, hand-crafted emulators.
Resumo:
La mayor parte de los entornos diseñados por el hombre presentan características geométricas específicas. En ellos es frecuente encontrar formas poligonales, rectangulares, circulares . . . con una serie de relaciones típicas entre distintos elementos del entorno. Introducir este tipo de conocimiento en el proceso de construcción de mapas de un robot móvil puede mejorar notablemente la calidad y la precisión de los mapas resultantes. También puede hacerlos más útiles de cara a un razonamiento de más alto nivel. Cuando la construcción de mapas se formula en un marco probabilístico Bayesiano, una especificación completa del problema requiere considerar cierta información a priori sobre el tipo de entorno. El conocimiento previo puede aplicarse de varias maneras, en esta tesis se presentan dos marcos diferentes: uno basado en el uso de primitivas geométricas y otro que emplea un método de representación cercano al espacio de las medidas brutas. Un enfoque basado en características geométricas supone implícitamente imponer un cierto modelo a priori para el entorno. En este sentido, el desarrollo de una solución al problema SLAM mediante la optimización de un grafo de características geométricas constituye un primer paso hacia nuevos métodos de construcción de mapas en entornos estructurados. En el primero de los dos marcos propuestos, el sistema deduce la información a priori a aplicar en cada caso en base a una extensa colección de posibles modelos geométricos genéricos, siguiendo un método de Maximización de la Esperanza para hallar la estructura y el mapa más probables. La representación de la estructura del entorno se basa en un enfoque jerárquico, con diferentes niveles de abstracción para los distintos elementos geométricos que puedan describirlo. Se llevaron a cabo diversos experimentos para mostrar la versatilidad y el buen funcionamiento del método propuesto. En el segundo marco, el usuario puede definir diferentes modelos de estructura para el entorno mediante grupos de restricciones y energías locales entre puntos vecinos de un conjunto de datos del mismo. El grupo de restricciones que se aplica a cada grupo de puntos depende de la topología, que es inferida por el propio sistema. De este modo, se pueden incorporar nuevos modelos genéricos de estructura para el entorno con gran flexibilidad y facilidad. Se realizaron distintos experimentos para demostrar la flexibilidad y los buenos resultados del enfoque propuesto. Abstract Most human designed environments present specific geometrical characteristics. In them, it is easy to find polygonal, rectangular and circular shapes, with a series of typical relations between different elements of the environment. Introducing this kind of knowledge in the mapping process of mobile robots can notably improve the quality and accuracy of the resulting maps. It can also make them more suitable for higher level reasoning applications. When mapping is formulated in a Bayesian probabilistic framework, a complete specification of the problem requires considering a prior for the environment. The prior over the structure of the environment can be applied in several ways; this dissertation presents two different frameworks, one using a feature based approach and another one employing a dense representation close to the measurements space. A feature based approach implicitly imposes a prior for the environment. In this sense, feature based graph SLAM was a first step towards a new mapping solution for structured scenarios. In the first framework, the prior is inferred by the system from a wide collection of feature based priors, following an Expectation-Maximization approach to obtain the most probable structure and the most probable map. The representation of the structure of the environment is based on a hierarchical model with different levels of abstraction for the geometrical elements describing it. Various experiments were conducted to show the versatility and the good performance of the proposed method. In the second framework, different priors can be defined by the user as sets of local constraints and energies for consecutive points in a range scan from a given environment. The set of constraints applied to each group of points depends on the topology, which is inferred by the system. This way, flexible and generic priors can be incorporated very easily. Several tests were carried out to demonstrate the flexibility and the good results of the proposed approach.
Resumo:
Service compositions put together loosely-coupled component services to perform more complex, higher level, or cross-organizational tasks in a platform-independent manner. Quality-of-Service (QoS) properties, such as execution time, availability, or cost, are critical for their usability, and permissible boundaries for their values are defined in Service Level Agreements (SLAs). We propose a method whereby constraints that model SLA conformance and violation are derived at any given point of the execution of a service composition. These constraints are generated using the structure of the composition and properties of the component services, which can be either known or empirically measured. Violation of these constraints means that the corresponding scenario is unfeasible, while satisfaction gives values for the constrained variables (start / end times for activities, or number of loop iterations) which make the scenario possible. These results can be used to perform optimized service matching or trigger preventive adaptation or healing.
Resumo:
We present a new free library for Constraint Logic Programming over Finite Domains, included with the Ciao Prolog system. The library is entirely written in Prolog, leveraging on Ciao's module system and code transformation capabilities in order to achieve a highly modular design without compromising performance. We describe the interface, implementation, and design rationale of each modular component. The library meets several design goals: a high level of modularity, allowing the individual components to be replaced by different versions; highefficiency, being competitive with other TT> implementations; a glass-box approach, so the user can specify new constraints at different levels; and a Prolog implementation, in order to ease the integration with Ciao's code analysis components. The core is built upon two small libraries which implement integer ranges and closures. On top of that, a finite domain variable datatype is defined, taking care of constraint reexecution depending on range changes. These three libraries form what we call the TT> kernel of the library. This TT> kernel is used in turn to implement several higher-level finite domain constraints, specified using indexicals. Together with a labeling module this layer forms what we name the TT> solver. A final level integrates the CLP (J7©) paradigm with our TT> solver. This is achieved using attributed variables and a compiler from the CLP (J7©) language to the set of constraints provided by the solver. It should be noted that the user of the library is encouraged to work in any of those levels as seen convenient: from writing a new range module to enriching the set of TT> constraints by writing new indexicals.
Resumo:
El principal objetivo de la presente investigación fue el conocer el perfil de rendimiento técnico de los triatletas, desde un punto de vista biomecánica, en el segmento carrera a pie durante la competición en triatlón. Asimismo, como el genero y el nivel deportivo del triatleta podrían influir en su respuesta motriz durante la competicion. Para ello, se necesitaba desarrollar y validar una técnica experimental que fuera lo suficientemente precisa (validez interna), con una alta fiabilidad y con una gran validez externa (ecologica) debido al entorno de la competicion. La muestra la formaron un total de 64 deportistas: 32 triatletas participantes en la Copa del Mundo de Triatlon de Madrid-2008 (16 hombres y 16 mujeres) y 32 triatletas participantes en el Clasificatorio del Campeonato de Espana Elite (16 hombres y 16 mujeres). El análisis de la técnica de carrera de los deportistas se realizo mediante un sistema fotogramétrico en 2d que permitió calcular las coordenadas (x,y) de los centros articulares con un error de 1.66% en el eje x y de un 2.10% en el eje y. Las imágenes fueron obtenidas por una cámara que filmaba el movimiento en un plano antero-posterior del triatleta. Algoritmos basados en la DLT (Abdel-Aziz & Karara, 1971) permitieron conocer las coordenadas reales a partir de las coordenadas digitalizadas en el plano y posteriormente las distintas variables analizadas. El análisis biomecánica de la carrera se realizo en 4 ocasiones diferentes durante la competición, correspondiendo con cada una de las vueltas de 2,5 km, que el triatleta tenía que realizar. La velocidad de carrera resulto estar íntimamente ligada al nivel deportivo del triatleta. Del mismo modo, 3 de los 4 grupos analizados presentaron valores inferiores a 3 minutos 30 segundos por kilometro recorrido, poniendo de manifiesto el altísimo nivel de los sujetos analizados. Del mismo modo parece que las chicos consiguen una mayor velocidad gracias a una mayor longitud de ciclo en relación a las chicas, ya que estas muestran valores mayores en cuanto a frecuencia de zancada. La frecuencia de zancada presento los valores más altos en la primera vuelta en todos los deportistas analizados. Asimismo, los triatletas de nivel internacional y las chicas fueron los que mostraron los mayores valores. La longitud de zancada presento distintas tendencias en función del nivel y el género del deportista. Así pues, en los deportistas internacionales y en los chicos los mayores valores se encontraron en la primera vuelta mientras que la tendencia fue al descenso, siendo probablemente la fatiga acumulada la causante de dicha tendencia. En cambio, aquellos deportistas de nivel nacional y las chicas mostraron valores mayores en la segunda vuelta que en la primera, evidenciando que además de la fatiga, el ciclismo previo tiene una incidencia directa sobre su rendimiento. Los tiempos de vuelo permanecieron constantes durante toda la carrera, encontrando cierta evolución en los tiempos de apoyo, la cual provoca una modificación en los porcentajes relativos en los tiempos de vuelo. Los tiempos de apoyo más bajos se encontraron en la primera vuelta. Del mismo modo, los deportistas de nivel internacional y los chicos mostraron valores inferiores. También, estos grupos fueron más constantes en sus valores a lo largo de las vueltas. Por el contrario, se encontraron tendencias al aumento en los triatletas de nivel nacional y en las chicas, los cuales no fueron capaces de mantener el mismo rendimiento debido seguramente a su menor nivel deportivo. La oscilación vertical de la cadera se mostro constante en los triatletas de mayor nivel, encontrándose tendencias al aumento en los de menor nivel. Del mismo modo, los valores más altos correspondieron a las chicas y a los deportistas de nivel nacional. La distancia de la cadera al apoyo permaneció constante a lo largo de las vueltas en todos los grupos, obteniéndose valores mayores en los triatletas de nivel internacional y en los chicos. El ángulo de la rodilla apoyada en el momento del despegue no mostro una tendencia clara. Los deportistas de nivel internacional y los chicos presentaron los valores más bajos. El ángulo de la rodilla libre en el momento del despegue mostro una correlación muy alta con la velocidad de carrera. Del mismo modo, los ángulos más pequeños se encontraron en los triatletas internacionales y en los chicos, debido seguramente a los mayores valores de velocidad registrados por ambos grupos. Los ángulos de los tobillos no mostraron ninguna tendencia clara durante la competición analizada. Los cuatro grupos de población presentaron valores similares, por lo que parece que no representan una variable que pueda incidir sobre el rendimiento biomecánica del triatleta. Los resultados obtenidos en el presente estudio de investigación avalan la utilización de la fotogrametría-video en 2d para el análisis de la técnica de carrera durante la competición en triatlón. Su aplicación en una competición de máximo nivel internacional ha posibilitado conocer el perfil técnico que presentan los triatletas a lo largo del segmento de carrera a pie. Del mismo modo, se ha podido demostrar como los estudios realizados en laboratorio no reflejan la realidad competitiva de un triatlón de máximo nivel. The aim of this research was to determine the running technique profile during a triathlon competition from a biomechanical perspective. Also, to analyze the triathlete gender’s and level of performance’s influence on this profile in competition. An accurate (internal validity) and reliable methodology with a high external validity (ecological) had to be developed to get those aims in competition. Sixty-four triathletes were analyzed. 32 (16 males, 16 females) took part in the Madrid 2008 Triathlon World Cup and 32 (16 males and 16 females) took part in the Spanish Triathlon National Championships. The biomechanical analyses were carried out by a photogrammetric system that allow to calculate the landmarks coordinates (x,y) with a 1.66% error in x axis, and a 2.10% error in y axis. The frames were obtained with a camera situated perpendicular to the triathletes’ trajectory, filming the saggittal plane. DLT based algorithms (Abdel-Aziz & Karara, 1971) were used to calculate the real coordinates from the digitalized ones and the final variables afterwards. The biomechanical analisys itself was performed in four different moments during the competition, according to each 2.5 km lap the triathletes had to do. Running speed was highly related to performance level. Also, 3 of the 4 analyzed groups showed speed values under the 3 minutes and 30 seconds per kilometer. It demonstrated the very high performance level of the analized triathletes. Furthermore, it seems that men get higher speeds because their longer stride length, while women shows higher stride frequency values. The highest stride frequency values were found in the first lap. Women and the international level triathletes showed the highest values. Stride length showed different tendencies according to the gender and level of performance. Men and international level triathletes showed the highest level in the first lap and a decreasing tendency after that. The accumulated fatigue was probably the reason of this tendency. On the other hand, higher values than in first lap were found in the second one in women and national level triathletes. It demonstrated the previous cycling can affect to those groups in terms of biomechanics. Flight times remained constant during the running part, while the contact times showed an increasing tendency that caused a variation in flight times percents. The lowest contact times were found in the first lap and in men and international triathletes’ values. Also, these two groups were more consistent during the whole running. On the other hand, increasing tendencies were found in women and national level triathletes, who were not able to maintain the same values probably due to their lower level of performance. Higher level triathletes showed more consistent hip vertical oscillation values than lower level triathletes, who presented increasing tendencies. The highest values were found in women and national level triathletes. The horizontal distance hip-toe cap remained constant among the laps in all the groups. Men and international level triathletes showed the highest values. The support knee angle at toe-off did not show a clear tendency. The lowest values were found in men and international level triathletes. A high correlation was found between the non-support knee angle and the running speed. Furthermore, men and international level triathletes showed the smallest values, due to the higher velocities reached by these two groups. Ankles angles did not show any tendency during the running part. Similar values were found in the four analyzed groups, so this variable does not seem to represent an important one within the triathlete’s performance. The results obtained in the present research support the use of the bidimensional photogrammetric video-system to analyze the running technique during a triathlon competition. Its application in international triathlon meetings has allowed determining the triathletes’ technique profile during the running part. Also, it has been demonstrated the laboratory-based studies does not reproduce a top-level competition.
Resumo:
Fe–Cr based alloys are the leading structural material candidates in the design of next generation reactors due to their high resistance to swelling and corrosion. Despite these good properties there are others, such as embrittlement, which require a higher level of understanding in order to improve aspects such as safety or lifetime of the reactors. The addition of Cr improves the behavior of the steels under irradiation, but not in a monotonic way. Therefore, understanding the changes in the Fe–Cr based alloys microstructure induced by irradiation and the role played by the alloying element (Cr) is needed in order to predict the response of these materials under the extreme conditions they are going to support. In this work we perform a study of the effect of Cr concentration in a bcc Fe–Cr matrix on formation and binding energies of vacancy clusters up to 5 units. The dependence of the calculated formation and binding energy is investigated with two empirical interatomic potentials specially developed to study radiation damage in Fe–Cr alloys. Results are very similar for both potentials showing an increase of the defect stability with the cluster size and no real dependence on Cr concentration for the binding energy.
Resumo:
Look-up tables are collected and analysed for 12 European National Travel Surveys (NTS) in a harmonized way covering the age group 13-84 year. Travel behaviour measured as kilometres, time use and trips per traveller is compared. Trips per traveller are very similar over the countries whereas kilometres differ most, from minus 28% for Spain to plus 19% and 14% for Sweden and Finland. It is shown that two main factors for differences are GDP per capita and density in the urban areas. The latter is the main reason for the low level in Spain. Mode share is except for Spain with a very high level of walking trips rather similar with a higher level of cycling in the Netherlands, more public transport in Switzerland, and more air traffic in Sweden. Normally kilometres per respondent/inhabitant is used for national planning purpose and this is very affected by the share of mobile travellers. The immobile share is varying between 8 and 28% with 6 NTS at a 15-17% level. These differences are analysed and discussed and it is concluded that the immobile share should be a little less than 15-17% because it is assessed that some short trips might have been forgotten in these 6 countries. The share has a downward tendency with higher density. The resulting immobile share is very dependent on data collection methodology, sampling method, quality of interviewer felt-work etc. The paper shows other possibilities to improve local surveys based on comparison with other countries.
Resumo:
La computación basada en servicios (Service-Oriented Computing, SOC) se estableció como un paradigma ampliamente aceptado para el desarollo de sistemas de software flexibles, distribuidos y adaptables, donde las composiciones de los servicios realizan las tareas más complejas o de nivel más alto, frecuentemente tareas inter-organizativas usando los servicios atómicos u otras composiciones de servicios. En tales sistemas, las propriedades de la calidad de servicio (Quality of Service, QoS), como la rapídez de procesamiento, coste, disponibilidad o seguridad, son críticas para la usabilidad de los servicios o sus composiciones en cualquier aplicación concreta. El análisis de estas propriedades se puede realizarse de una forma más precisa y rica en información si se utilizan las técnicas de análisis de programas, como el análisis de complejidad o de compartición de datos, que son capables de analizar simultáneamente tanto las estructuras de control como las de datos, dependencias y operaciones en una composición. El análisis de coste computacional para la composicion de servicios puede ayudar a una monitorización predictiva así como a una adaptación proactiva a través de una inferencia automática de coste computacional, usando los limites altos y bajos como funciones del valor o del tamaño de los mensajes de entrada. Tales funciones de coste se pueden usar para adaptación en la forma de selección de los candidatos entre los servicios que minimizan el coste total de la composición, basado en los datos reales que se pasan al servicio. Las funciones de coste también pueden ser combinadas con los parámetros extraídos empíricamente desde la infraestructura, para producir las funciones de los límites de QoS sobre los datos de entrada, cuales se pueden usar para previsar, en el momento de invocación, las violaciones de los compromisos al nivel de servicios (Service Level Agreements, SLA) potenciales or inminentes. En las composiciones críticas, una previsión continua de QoS bastante eficaz y precisa se puede basar en el modelado con restricciones de QoS desde la estructura de la composition, datos empiricos en tiempo de ejecución y (cuando estén disponibles) los resultados del análisis de complejidad. Este enfoque se puede aplicar a las orquestaciones de servicios con un control centralizado del flujo, así como a las coreografías con participantes multiples, siguiendo unas interacciones complejas que modifican su estado. El análisis del compartición de datos puede servir de apoyo para acciones de adaptación, como la paralelización, fragmentación y selección de los componentes, las cuales son basadas en dependencias funcionales y en el contenido de información en los mensajes, datos internos y las actividades de la composición, cuando se usan construcciones de control complejas, como bucles, bifurcaciones y flujos anidados. Tanto las dependencias funcionales como el contenido de información (descrito a través de algunos atributos definidos por el usuario) se pueden expresar usando una representación basada en la lógica de primer orden (claúsulas de Horn), y los resultados del análisis se pueden interpretar como modelos conceptuales basados en retículos. ABSTRACT Service-Oriented Computing (SOC) is a widely accepted paradigm for development of flexible, distributed and adaptable software systems, in which service compositions perform more complex, higher-level, often cross-organizational tasks using atomic services or other service compositions. In such systems, Quality of Service (QoS) properties, such as the performance, cost, availability or security, are critical for the usability of services and their compositions in concrete applications. Analysis of these properties can become more precise and richer in information, if it employs program analysis techniques, such as the complexity and sharing analyses, which are able to simultaneously take into account both the control and the data structures, dependencies, and operations in a composition. Computation cost analysis for service composition can support predictive monitoring and proactive adaptation by automatically inferring computation cost using the upper and lower bound functions of value or size of input messages. These cost functions can be used for adaptation by selecting service candidates that minimize total cost of the composition, based on the actual data that is passed to them. The cost functions can also be combined with the empirically collected infrastructural parameters to produce QoS bounds functions of input data that can be used to predict potential or imminent Service Level Agreement (SLA) violations at the moment of invocation. In mission-critical applications, an effective and accurate continuous QoS prediction, based on continuations, can be achieved by constraint modeling of composition QoS based on its structure, known data at runtime, and (when available) the results of complexity analysis. This approach can be applied to service orchestrations with centralized flow control, and choreographies with multiple participants with complex stateful interactions. Sharing analysis can support adaptation actions, such as parallelization, fragmentation, and component selection, which are based on functional dependencies and information content of the composition messages, internal data, and activities, in presence of complex control constructs, such as loops, branches, and sub-workflows. Both the functional dependencies and the information content (described using user-defined attributes) can be expressed using a first-order logic (Horn clause) representation, and the analysis results can be interpreted as a lattice-based conceptual models.
Resumo:
Esta tesis se suma a los intentos teóricos de clarificar la impronta del dominio digital en lo arquitectónico. Propone una cartografía crítica para reconstruir el proceso de tal convergencia considerando aquellos acontecimientos reveladores que lo han pautado. La integración de la extensión digital propia las tecnologías de la información y la comunicación en el contexto tradicional arquitectónico ha supuesto el advenimiento de un ecosistema artificial complejo. A esta realidad o proceso concurrente se la denomina el Entorno Aumentado. La línea principal de investigación explora el desarrollo de la interacción hombre-máquina, en sendas trayectorias sincrónicas convergentes. El análisis se aborda por tanto desde la consideración de esa naturaleza dual, atendiendo simultáneamente a la humanización del dominio digital o cómo la computación se adapta a la condición natural de ser humano, y a la digitalización del ser humano o cómo éste asume el imperativo digital. El análisis resulta vertebrado desde la condición panóptica del punto de observación del acontecimiento: la cuarta pared, entendida como pantalla y punto de inflexión que estructura la convergencia de los entornos físico y digital. La reflexión acometida sobre la construcción del Entorno Aumentado procura la verificación de la tesis, que es central en esta investigación: la multiplicación dimensional del lugar físico mediante su extensión con un campo informacional procedente del dominio digital, y sus efectos en la construcción de la nueva ecología digital propia del Entorno Aumentado. Esta circunstancia se produce tras la eclosión de la Revolución Digital en la segunda mitad del siglo XX, el consecuente incremento de la interacción entre los entornos digital y físico, y el alcance de un nivel superior de comunicación en los procesos arquitectónicos. Los síntomas del Entorno Aumentado se hacen notar en nuestra contemporaneidad; en ese sentido, la tesis alcanza un diagnóstico del cambio. La fractura y obsolescencia del límite espacio-temporal establecido por las dicotomías históricas privado-público, casa-ciudad, trabajo-ocio,…etc., o la vigencia del proyecto procedimental son algunas de sus consecuencias en el modo de abordar la disciplina arquitectónica. Abstract This dissertation aims to complete the theoretical attempts to clarify the digital domain imprint on the architectural realm. It constructs a critical cartography to reconstruct the process of such convergence, considering those principal events who have scheduled it. The integration of TIC’s digital extension through the traditional architectural context has meant the advent of an artificial complex ecosystem. This reality or concurrent process is called The Augmented Environment. The key research attempt explores man-machine interaction process in both synchronous converging trajectories. The analysis therefore addresses from the consideration of this dual nature, focusing simultaneously in humanizing digital domain process or how the computer fits the natural condition of human beings, and in digitalizing human beings or how it affords the digital imperative. The analysis is structured from the panoptic condition of the event scope: the fourth wall as screen that structures the convergence of physical and digital environments. The examination of The Augmented Environment’s assembly pretends the verification of the central point of this research: the dimensional multiplication of physical space by extending informational fields from the digital domain, and its effects on the construction of the Augmented Environment new digital ecology. This circumstance occurs after the arrival of the Digital Revolution in the second half of the twentieth century, with the consequent increase in the interaction between digital and physical environments, and the reach of a higher level of communication in architectural processes. The Augmented Environment signs are alive nowadays; in that sense, the thesis reaches a diagnostic of the changes. The fracture and obsolescence of the time-space limit established by historic dichotomies as private-public, home-city, working-leisure...etc., or the validity of the procedural design are some of its consequences on the architectural discipline.
Case study on mobile applications UX: effect of the usage of a crosss-platform development framework
Resumo:
Cross-platform development frameworks for mobile applications promise important advantages in cost cuttings and easy maintenance, posing as a very good option for organizations interested in the design of mobile applications for several platforms. Given that platform conventions are especially important for the User eXperience (UX) of mobile applications, the usage of framework where the same code defines the behavior of the app in different platforms could have negative impact in the UX. The objetive of this study is comparing the cross-platform and the native approach for being able to determine if the selected development approach has any impact on the users in terms of UX. To be able to set a base line under this subject, study on cross-platform frameworks was performed to select the most appropriate one from a UX point of view. In order to achieve the objectives of this work, two development teams have developed two versions of the same application; one using framework that generates Android and iOS versions automatically, and another team developing native versions of the same application. The alternative versions for each platform have been evaluated with 37 users with a combination of a laboratory usability test and a longitudinal study. The results show that differences are minimal in the Android version, but in iOS, even if a reasonable good UX can be obtained with the usage of this framework by an UX-conscious design team, a higher level of UX can be obtained directly developing in native code.