21 resultados para Jacobian Unification
em Universidad Politécnica de Madrid
Resumo:
Global linear instability theory is concerned with the temporal or spatial development of small-amplitude perturbations superposed upon laminar steady or time-periodic threedimensional flows, which are inhomogeneous in two (and periodic in one) or all three spatial directions.1 The theory addresses flows developing in complex geometries, in which the parallel or weakly nonparallel basic flow approximation invoked by classic linear stability theory does not hold. As such, global linear theory is called to fill the gap in research into stability and transition in flows over or through complex geometries. Historically, global linear instability has been (and still is) concerned with solution of multi-dimensional eigenvalue problems; the maturing of non-modal linear instability ideas in simple parallel flows during the last decade of last century2–4 has given rise to investigation of transient growth scenarios in an ever increasing variety of complex flows. After a brief exposition of the theory, connections are sought with established approaches for structure identification in flows, such as the proper orthogonal decomposition and topology theory in the laminar regime and the open areas for future research, mainly concerning turbulent and three-dimensional flows, are highlighted. Recent results obtained in our group are reported in both the time-stepping and the matrix-forming approaches to global linear theory. In the first context, progress has been made in implementing a Jacobian-Free Newton Krylov method into a standard finite-volume aerodynamic code, such that global linear instability results may now be obtained in compressible flows of aeronautical interest. In the second context a new stable very high-order finite difference method is implemented for the spatial discretization of the operators describing the spatial BiGlobal EVP, PSE-3D and the TriGlobal EVP; combined with sparse matrix treatment, all these problems may now be solved on standard desktop computers.
Resumo:
OntoTag - A Linguistic and Ontological Annotation Model Suitable for the Semantic Web
1. INTRODUCTION. LINGUISTIC TOOLS AND ANNOTATIONS: THEIR LIGHTS AND SHADOWS
Computational Linguistics is already a consolidated research area. It builds upon the results of other two major ones, namely Linguistics and Computer Science and Engineering, and it aims at developing computational models of human language (or natural language, as it is termed in this area). Possibly, its most well-known applications are the different tools developed so far for processing human language, such as machine translation systems and speech recognizers or dictation programs.
These tools for processing human language are commonly referred to as linguistic tools. Apart from the examples mentioned above, there are also other types of linguistic tools that perhaps are not so well-known, but on which most of the other applications of Computational Linguistics are built. These other types of linguistic tools comprise POS taggers, natural language parsers and semantic taggers, amongst others. All of them can be termed linguistic annotation tools.
Linguistic annotation tools are important assets. In fact, POS and semantic taggers (and, to a lesser extent, also natural language parsers) have become critical resources for the computer applications that process natural language. Hence, any computer application that has to analyse a text automatically and ‘intelligently’ will include at least a module for POS tagging. The more an application needs to ‘understand’ the meaning of the text it processes, the more linguistic tools and/or modules it will incorporate and integrate.
However, linguistic annotation tools have still some limitations, which can be summarised as follows:
1. Normally, they perform annotations only at a certain linguistic level (that is, Morphology, Syntax, Semantics, etc.).
2. They usually introduce a certain rate of errors and ambiguities when tagging. This error rate ranges from 10 percent up to 50 percent of the units annotated for unrestricted, general texts.
3. Their annotations are most frequently formulated in terms of an annotation schema designed and implemented ad hoc.
A priori, it seems that the interoperation and the integration of several linguistic tools into an appropriate software architecture could most likely solve the limitations stated in (1). Besides, integrating several linguistic annotation tools and making them interoperate could also minimise the limitation stated in (2). Nevertheless, in the latter case, all these tools should produce annotations for a common level, which would have to be combined in order to correct their corresponding errors and inaccuracies. Yet, the limitation stated in (3) prevents both types of integration and interoperation from being easily achieved.
In addition, most high-level annotation tools rely on other lower-level annotation tools and their outputs to generate their own ones. For example, sense-tagging tools (operating at the semantic level) often use POS taggers (operating at a lower level, i.e., the morphosyntactic) to identify the grammatical category of the word or lexical unit they are annotating. Accordingly, if a faulty or inaccurate low-level annotation tool is to be used by other higher-level one in its process, the errors and inaccuracies of the former should be minimised in advance. Otherwise, these errors and inaccuracies would be transferred to (and even magnified in) the annotations of the high-level annotation tool.
Therefore, it would be quite useful to find a way to
(i) correct or, at least, reduce the errors and the inaccuracies of lower-level linguistic tools;
(ii) unify the annotation schemas of different linguistic annotation tools or, more generally speaking, make these tools (as well as their annotations) interoperate.
Clearly, solving (i) and (ii) should ease the automatic annotation of web pages by means of linguistic tools, and their transformation into Semantic Web pages (Berners-Lee, Hendler and Lassila, 2001). Yet, as stated above, (ii) is a type of interoperability problem. There again, ontologies (Gruber, 1993; Borst, 1997) have been successfully applied thus far to solve several interoperability problems. Hence, ontologies should help solve also the problems and limitations of linguistic annotation tools aforementioned.
Thus, to summarise, the main aim of the present work was to combine somehow these separated approaches, mechanisms and tools for annotation from Linguistics and Ontological Engineering (and the Semantic Web) in a sort of hybrid (linguistic and ontological) annotation model, suitable for both areas. This hybrid (semantic) annotation model should (a) benefit from the advances, models, techniques, mechanisms and tools of these two areas; (b) minimise (and even solve, when possible) some of the problems found in each of them; and (c) be suitable for the Semantic Web. The concrete goals that helped attain this aim are presented in the following section.
2. GOALS OF THE PRESENT WORK
As mentioned above, the main goal of this work was to specify a hybrid (that is, linguistically-motivated and ontology-based) model of annotation suitable for the Semantic Web (i.e. it had to produce a semantic annotation of web page contents). This entailed that the tags included in the annotations of the model had to (1) represent linguistic concepts (or linguistic categories, as they are termed in ISO/DCR (2008)), in order for this model to be linguistically-motivated; (2) be ontological terms (i.e., use an ontological vocabulary), in order for the model to be ontology-based; and (3) be structured (linked) as a collection of ontology-based
Resumo:
It is generally recognized that information about the runtime cost of computations can be useful for a variety of applications, including program transformation, granularity control during parallel execution, and query optimization in deductive databases. Most of the work to date on compile-time cost estimation of logic programs has focused on the estimation of upper bounds on costs. However, in many applications, such as parallel implementations on distributed-memory machines, one would prefer to work with lower bounds instead. The problem with estimating lower bounds is that in general, it is necessary to account for the possibility of failure of head unification, leading to a trivial lower bound of 0. In this paper, we show how, given type and mode information about procedures in a logic program, it is possible to (semi-automatically) derive nontrivial lower bounds on their computational costs. We also discuss the cost analysis for the special and frequent case of divide-and-conquer programs and show how —as a pragmatic short-term solution —it may be possible to obtain useful results simply by identifying and treating divide-and-conquer programs specially.
Resumo:
We present a parallel graph narrowing machine, which is used to implement a functional logic language on a shared memory multiprocessor. It is an extensión of an abstract machine for a purely functional language. The result is a programmed graph reduction machine which integrates the mechanisms of unification, backtracking, and independent and-parallelism. In the machine, the subexpressions of an expression can run in parallel. In the case of backtracking, the structure of an expression is used to avoid the reevaluation of subexpressions as far as possible. Deterministic computations are detected. Their results are maintained and need not be reevaluated after backtracking.
Resumo:
In this paper, abstract interpretation algorithms are described for computing the sharmg as well as the freeness information about the run-time instantiations of program variables. An abstract domain is proposed which accurately and concisely represents combined freeness and sharing information for program variables. Abstract unification and all other domain-specific functions for an abstract interpreter working on this domain are presented. These functions are illustrated with an example. The importance of inferring freeness is stressed by showing (1) the central role it plays in non-strict goal independence, and (2) the improved accuracy it brings to the analysis of sharing information when both are computed together. Conversely, it is shown that keeping accurate track of sharing allows more precise inference of freeness, thus resulting in an overall much more powerful abstract interpreter.
Resumo:
A mathematical formulation for finite strain elasto plastic consolidation of fully saturated soil media is presented. Strong and weak forms of the boundary-value problem are derived using both the material and spatial descriptions. The algorithmic treatment of finite strain elastoplasticity for the solid phase is based on multiplicative decomposition and is coupled with the algorithm for fluid flow via the Kirchhoff pore water pressure. Balance laws are written for the soil-water mixture following the motion of the soil matrix alone. It is shown that the motion of the fluid phase only affects the Jacobian of the solid phase motion, and therefore can be characterized completely by the motion of the soil matrix. Furthermore, it is shown from energy balance consideration that the effective, or intergranular, stress is the appropriate measure of stress for describing the constitutive response of the soil skeleton since it absorbs all the strain energy generated in the saturated soil-water mixture. Finally, it is shown that the mathematical model is amenable to consistent linearization, and that explicit expressions for the consistent tangent operators can be derived for use in numerical solutions such as those based on the finite element method.
Resumo:
This paper presents improved unification algorithms, an implementation, and an analysis of the effectiveness of an abstract interpreter based on the sharing + freeness domain presented in a previous paper, which was designed to accurately and concisely represent combined freeness and sharing information for program variables. We first briefly review this domain and the unification algorithms previously proposed. We then improve these algorithms and correct them to deal with some cases which were not well analyzed previously, illustrating the improvement with an example. We then present the implementation of the improved algorithm and evaluate its performance by comparing the effectiveness of the information inferred to that of other interpreters available to us for an application (program parallelization) that is common to all these interpreters. All these systems have been embedded in a real parallelizing compiler. Effectiveness of the analysis is measured in terms of actual final performance of the system: i.e. in terms of the actual speedups obtained. The results show good performance for the combined domain in that it improves the accuracy of both types of information and also in that the analyzer using the combined domain is more effective in the application than any of the other analyzers it is compared to.
Resumo:
The performance indices are important tools for motion planning and design of robot manipulators. In this paper we present a collection of some of the performance indices that have generated interest in the robotics community. These indices are four different types: kinetostatic performance indices, dynamic performance indices, indices of joint limits, and finally global performance indices. In addition, we review the strategies that have been proposed to solve the problems that occur when the units of the Jacobian matrix elements are not homogeneous. At the end of this paper, we propose a set of global performance indices that can be useful in the design of robot manipulators.
Resumo:
Global linear instability theory is concerned with the temporal or spatial development of small-amplitude perturbations superposed upon laminar steady or time-periodic three-dimensional flows, which are inhomogeneous in two(and periodic in one)or all three spatial directions.After a brief exposition of the theory,some recent advances are reported. First, results are presented on the implementation of a Jacobian-free Newton–Krylov time-stepping method into a standard finite-volume aerodynamic code to obtain global linear instability results in flows of industrial interest. Second, connections are sought between established and more-modern approaches for structure identification in flows, such as proper orthogonal decomposition and Koopman modes analysis (dynamic mode decomposition), and the possibility to connect solutions of the eigenvalue problem obtained by matrix formation or time-stepping with those delivered by dynamic mode decomposition, residual algorithm, and proper orthogonal decomposition analysis is highlighted in the laminar regime; turbulent and three-dimensional flows are identified as open areas for future research. Finally, a new stable very-high-order finite-difference method is implemented for the spatial discretization of the operators describing the spatial biglobal eigenvalue problem, parabolized stability equation three-dimensional analysis, and the triglobal eigenvalue problem; it is shown that, combined with sparse matrix treatment, all these problems may now be solved on standard desktop computers
Resumo:
En la interacción con el entorno que nos rodea durante nuestra vida diaria (utilizar un cepillo de dientes, abrir puertas, utilizar el teléfono móvil, etc.) y en situaciones profesionales (intervenciones médicas, procesos de producción, etc.), típicamente realizamos manipulaciones avanzadas que incluyen la utilización de los dedos de ambas manos. De esta forma el desarrollo de métodos de interacción háptica multi-dedo dan lugar a interfaces hombre-máquina más naturales y realistas. No obstante, la mayoría de interfaces hápticas disponibles en el mercado están basadas en interacciones con un solo punto de contacto; esto puede ser suficiente para la exploración o palpación del entorno pero no permite la realización de tareas más avanzadas como agarres. En esta tesis, se investiga el diseño mecánico, control y aplicaciones de dispositivos hápticos modulares con capacidad de reflexión de fuerzas en los dedos índice, corazón y pulgar del usuario. El diseño mecánico de la interfaz diseñada, ha sido optimizado con funciones multi-objetivo para conseguir una baja inercia, un amplio espacio de trabajo, alta manipulabilidad y reflexión de fuerzas superiores a 3 N en el espacio de trabajo. El ancho de banda y la rigidez del dispositivo se han evaluado mediante simulación y experimentación real. Una de las áreas más importantes en el diseño de estos dispositivos es el efector final, ya que es la parte que está en contacto con el usuario. Durante este trabajo se ha diseñado un dedal de bajo peso, adaptable a diferentes usuarios que, mediante la incorporación de sensores de contacto, permite estimar fuerzas normales y tangenciales durante la interacción con entornos reales y virtuales. Para el diseño de la arquitectura de control, se estudiaron los principales requisitos para estos dispositivos. Entre estos, cabe destacar la adquisición, procesado e intercambio a través de internet de numerosas señales de control e instrumentación; la computación de equaciones matemáticas incluyendo la cinemática directa e inversa, jacobiana, algoritmos de detección de agarres, etc. Todos estos componentes deben calcularse en tiempo real garantizando una frecuencia mínima de 1 KHz. Además, se describen sistemas para manipulación de precisión virtual y remota; así como el diseño de un método denominado "desacoplo cinemático iterativo" para computar la cinemática inversa de robots y la comparación con otros métodos actuales. Para entender la importancia de la interacción multimodal, se ha llevado a cabo un estudio para comprobar qué estímulos sensoriales se correlacionan con tiempos de respuesta más rápidos y de mayor precisión. Estos experimentos se desarrollaron en colaboración con neurocientíficos del instituto Technion Israel Institute of Technology. Comparando los tiempos de respuesta en la interacción unimodal (auditiva, visual y háptica) con combinaciones bimodales y trimodales de los mismos, se demuestra que el movimiento sincronizado de los dedos para generar respuestas de agarre se basa principalmente en la percepción háptica. La ventaja en el tiempo de procesamiento de los estímulos hápticos, sugiere que los entornos virtuales que incluyen esta componente sensorial generan mejores contingencias motoras y mejoran la credibilidad de los eventos. Se concluye que, los sistemas que incluyen percepción háptica dotan a los usuarios de más tiempo en las etapas cognitivas para rellenar información de forma creativa y formar una experiencia más rica. Una aplicación interesante de los dispositivos hápticos es el diseño de nuevos simuladores que permitan entrenar habilidades manuales en el sector médico. En colaboración con fisioterapeutas de Griffith University en Australia, se desarrolló un simulador que permite realizar ejercicios de rehabilitación de la mano. Las propiedades de rigidez no lineales de la articulación metacarpofalange del dedo índice se estimaron mediante la utilización del efector final diseñado. Estos parámetros, se han implementado en un escenario que simula el comportamiento de la mano humana y que permite la interacción háptica a través de esta interfaz. Las aplicaciones potenciales de este simulador están relacionadas con entrenamiento y educación de estudiantes de fisioterapia. En esta tesis, se han desarrollado nuevos métodos que permiten el control simultáneo de robots y manos robóticas en la interacción con entornos reales. El espacio de trabajo alcanzable por el dispositivo háptico, se extiende mediante el cambio de modo de control automático entre posición y velocidad. Además, estos métodos permiten reconocer el gesto del usuario durante las primeras etapas de aproximación al objeto para su agarre. Mediante experimentos de manipulación avanzada de objetos con un manipulador y diferentes manos robóticas, se muestra que el tiempo en realizar una tarea se reduce y que el sistema permite la realización de la tarea con precisión. Este trabajo, es el resultado de una colaboración con investigadores de Harvard BioRobotics Laboratory. ABSTRACT When we interact with the environment in our daily life (using a toothbrush, opening doors, using cell-phones, etc.), or in professional situations (medical interventions, manufacturing processes, etc.) we typically perform dexterous manipulations that involve multiple fingers and palm for both hands. Therefore, multi-Finger haptic methods can provide a realistic and natural human-machine interface to enhance immersion when interacting with simulated or remote environments. Most commercial devices allow haptic interaction with only one contact point, which may be sufficient for some exploration or palpation tasks but are not enough to perform advanced object manipulations such as grasping. In this thesis, I investigate the mechanical design, control and applications of a modular haptic device that can provide force feedback to the index, thumb and middle fingers of the user. The designed mechanical device is optimized with a multi-objective design function to achieve a low inertia, a large workspace, manipulability, and force-feedback of up to 3 N within the workspace; the bandwidth and rigidity for the device is assessed through simulation and real experimentation. One of the most important areas when designing haptic devices is the end-effector, since it is in contact with the user. In this thesis the design and evaluation of a thimble-like, lightweight, user-adaptable, and cost-effective device that incorporates four contact force sensors is described. This design allows estimation of the forces applied by a user during manipulation of virtual and real objects. The design of a real-time, modular control architecture for multi-finger haptic interaction is described. Requirements for control of multi-finger haptic devices are explored. Moreover, a large number of signals have to be acquired, processed, sent over the network and mathematical computations such as device direct and inverse kinematics, jacobian, grasp detection algorithms, etc. have to be calculated in Real Time to assure the required high fidelity for the haptic interaction. The Hardware control architecture has different modules and consists of an FPGA for the low-level controller and a RT controller for managing all the complex calculations (jacobian, kinematics, etc.); this provides a compact and scalable solution for the required high computation capabilities assuring a correct frequency rate for the control loop of 1 kHz. A set-up for dexterous virtual and real manipulation is described. Moreover, a new algorithm named the iterative kinematic decoupling method was implemented to solve the inverse kinematics of a robotic manipulator. In order to understand the importance of multi-modal interaction including haptics, a subject study was carried out to look for sensory stimuli that correlate with fast response time and enhanced accuracy. This experiment was carried out in collaboration with neuro-scientists from Technion Israel Institute of Technology. By comparing the grasping response times in unimodal (auditory, visual, and haptic) events with the response times in events with bimodal and trimodal combinations. It is concluded that in grasping tasks the synchronized motion of the fingers to generate the grasping response relies on haptic cues. This processing-speed advantage of haptic cues suggests that multimodalhaptic virtual environments are superior in generating motor contingencies, enhancing the plausibility of events. Applications that include haptics provide users with more time at the cognitive stages to fill in missing information creatively and form a richer experience. A major application of haptic devices is the design of new simulators to train manual skills for the medical sector. In collaboration with physical therapists from Griffith University in Australia, we developed a simulator to allow hand rehabilitation manipulations. First, the non-linear stiffness properties of the metacarpophalangeal joint of the index finger were estimated by using the designed end-effector; these parameters are implemented in a scenario that simulates the behavior of the human hand and that allows haptic interaction through the designed haptic device. The potential application of this work is related to educational and medical training purposes. In this thesis, new methods to simultaneously control the position and orientation of a robotic manipulator and the grasp of a robotic hand when interacting with large real environments are studied. The reachable workspace is extended by automatically switching between rate and position control modes. Moreover, the human hand gesture is recognized by reading the relative movements of the index, thumb and middle fingers of the user during the early stages of the approximation-to-the-object phase and then mapped to the robotic hand actuators. These methods are validated to perform dexterous manipulation of objects with a robotic manipulator, and different robotic hands. This work is the result of a research collaboration with researchers from the Harvard BioRobotics Laboratory. The developed experiments show that the overall task time is reduced and that the developed methods allow for full dexterity and correct completion of dexterous manipulations.
Resumo:
La tesis doctoral desarrolla una investigación original, dentro del marco disciplinario de la historia de la construcción, sobre los fundamentos constructivos de las fortificaciones bajomedievales fronterizas entre las Coronas de Castilla y Aragón en la actual provincia de Soria. En el título de la tesis ya queda expresado el objeto fundamental y fundacional, así como el ámbito temporal —desde la reconquista del oriente soriano por parte de Alfonso I el Batallador a principios del siglo XII hasta la unificación de las coronas hispánicas en el siglo XV bajo el común mandato de los Reyes Católicos— y la extensión territorial que delimita la investigación: aquéllas comarcas castellanas lindantes con Aragón pertenecientes a la actual provincia de Soria. Durante este período bajomedieval se produjeron una serie de enfrentamientos fronterizos que obligó a fortificar la frontera y las vías de comunicación entre ambas coronas. La falta de estudios de conjunto de estas fortificaciones entendiéndolas como participantes en un sistema fortificado ha constituido la justificación de la investigación, que se realiza en varios niveles de análisis: territorial, histórico, arquitectónico, poliorcético y constructivo. Así mismo, se ha detectado cierta falta de rigor acompañada de inexactitudes en las consideraciones constructivas publicadas sobre algunas de las fortificaciones del ámbito de estudio, lo que ha provocado errores en su datación al no más haber elementos de corte artístico o estilístico que marquen indudablemente la pertenencia a una época. En la tesis se ponen en duda las dataciones tradicionalmente aceptadas planteando la hipótesis que da pie a la investigación: ante la falta de elementos artísticos o estilísticos en unos sobrios edificios eminentemente funcionales es posible establecer con suficiente aproximación la fecha de construcción en base a criterios constructivos una vez formada una clasificación cronotipológica de cada técnica constructiva. La hipótesis, por lo tanto, plantea un objetivo principal —el estudio de la razón constructiva del sistema fortificado fronterizo— desarrollado en una serie de objetivos específicos cuya consecución programa los sucesivos niveles de análisis: - Conocer y detallar los elementos históricos que originaron los enfrentamientos entre las Coronas de Castilla y Aragón y su desarrollo mediante herramientas historiográficas y analizar las características naturales del territorio en litigio mediante instrumentos cartográficos. - Conocer y analizar los tipos arquitectónicos y las tradiciones constructivas empleadas en las construcciones castrenses en el ámbito temporal en que se enmarca la investigación. - Localizar, documentar y seleccionar para su análisis las fortalezas y construcciones militares erigidas durante dichas luchas fronterizas en la actual provincia de Soria a través del trabajo de campo y métodos cartográficos y bibliográficos. - Realizar un estudio general sobre el sistema fortificado a escala territorial - Investigar la tipología arquitectónica, poliorcética y constructiva del conjunto de estas fortificaciones bajomedievales fronterizas. - Analizar los fundamentos constructivos de los casos de estudio seleccionados entre estas construcciones y caracterizarlas en cuanto al material, elementos, sistemas y procesos constructivos. - Ordenar la información histórica dispersa y corregir errores para hacer una base sobre la que establecer un discurso histórico de cada caso de estudio. - Comparar y relacionar las técnicas constructivas empleadas en estas fortalezas con los utilizados en el mismo ámbito temporal. - Difundir para su debate los resultados de la investigación por los foros científicos habituales. El método empleado combina los trabajos de gabinete con una intensa labor de campo, en la que se han documentado cincuenta fortificaciones y se han redactado sus correspondientes fichas de toma de datos. La recopilación de datos se ha incluido en una base de datos que incluye aspectos generales, tipológicos, constructivos y bibliográficos básicos del conjunto, a modo de inventario, de fortificaciones de la provincia. Las fortificaciones seleccionadas se agrupan según una clasificación tipológica y constructiva que marca las líneas de estudio posteriores. Se desarrolla un capítulo de antecedentes en el que se estudia la historia de la construcción fortificada medieval tanto en Europa como en España analizando la evolución de los tipos arquitectónicos y las múltiples influencias culturales que surcaron el Mediterráneo desde el Oriente cruzado e islámico al Poniente donde se desarrollaba la empresa reconquistadora que mantuvo en estado de guerra continuo a la Península Ibérica durante ochocientos años. El análisis del territorio como contenedor del hecho fortificado revela que hay una relación íntima entre la ubicación de las fortificaciones y las formas naturales que definen las vías de comunicación entre los valles del Duero, del Ebro y del Tajo. En efecto, el ámbito de estudio ha supuesto desde la Antigüedad un territorio de paso fundamental en la articulación de las comunicaciones en la Península Ibérica. Este carácter de paso más que de frontera explica las inquietudes y la preocupación por su control tanto por Roma como por el califato cordobés como por los reinos cristianos medievales. El análisis de los elementos históricos se complementa con el estudio detallado de los enfrentamientos fronterizos entre Castilla y Aragón así como los aspectos sociales y políticos que provocaron la fortificación como sistema de definición de la frontera y de organización espacial, jurisdiccional, social y administrativa del territorio. La arquitectura fortificada es esencialmente funcional: su cometido es la defensa. En este sentido, tras un estudio morfológico de los castillos seleccionados se realiza un extenso análisis poliorcético de sus elementos, investigando su origen y aplicación para servir también de parámetros de datación. Siendo el objeto inaugural de la tesis el estudio de los fundamentos constructivos, se explican los distintos materiales de construcción empleados y se agrupan las fábricas de las fortificaciones seleccionadas en dos grandes grupos constructivos: las fábricas aparejadas y las fábricas encofradas. Se han destacado y estudiado la evolución histórica y la tipología y mensiología constructiva de tres técnicas destacadas: el uso del ladrillo, la tapia de cal y canto o mampostería encofrada y la tapia de tierra. Para el estudio de la componente histórica y de la dimensión constructiva de cada técnica ha sido necesario documentar numerosos casos tanto en el ámbito de estudio como en la Península Ibérica con el fin de establecer grupos cronotipológicos constructivos entre los que poder ubicar las fábricas de estas técnicas presentes en el ámbito de estudio. Se ha observado una evolución dimensional de las fábricas de tapia que es más evidente en las hispanomusulmanas al modularse en codos pero que también se advierte significativamente en las cristianas bajomedievales. De cada una de las técnicas analizadas se ha seleccionado un caso de estudio singular y representativo. El castillo de Arcos de Jalón es un ejemplo significativo del empleo de la fábrica mixta de mampostería con verdugadas de ladrillo, así como las murallas de la ciudad fortificada de Peñalcázar lo es de la fábrica de mampostería encofrada y el castillo de Serón de Nágima constituye un caso característico y principal de la utilización de la tapia de tierra en la arquitectura militar bajomedieval. Cada uno de estos tres casos de estudios se examina bajo los mismos cuatro niveles anteriormente mencionados: territorial, histórico, arquitectónico y defensivo y constructivo. El sistemático método de estudio ha facilitado el orden en la investigación y la obtención de unos resultados y conclusiones que verifican la hipótesis y cumplen los objetivos marcados al comienzo. Se ha revisado la datación en la construcción de las fortificaciones analizadas mediante el estudio cronotipológico de sus fábricas, pudiendo trasladarse el método a otros sistemas fortificados. La tesis abre, finalmente, dos vías principales de investigación encaminadas a completar el estudio del sistema fortificado fronterizo bajomedieval en la raya oriental soriana de Castilla: la caracterización y datación por métodos físico-químicos de las muestras de piezas de madera de construcción que se conservan embebidas en las fábricas y la búsqueda documental y archivística que pueda revelar nuevos datos respecto a la fundación, reparación, venta o cualquier aspecto económico, legislativo, organizativo o administrativo relativo a las fortificaciones en documentos coetáneos. ABSTRACT The doctoral thesis develops an original research, held in the field of the Construction History, about the constructive reason of the frontier fortifications in the Late Middle Age between the Crowns of Castile and Aragón in the actual province of Soria, Spain. In the title is expressed the main objective, and also the temporal scope —from the reconquest in the 12th Century by Alfonso the First of Aragón to the unification under the common kingdom of the Catholic Kings— and the territorial extension that the research delimits: those Castilian regions in the border with Aragón in the actual province of Soria. During this period, a series of border wars were been, and this is the reason for the fortification of the border line and the main roads between both Crowns. The lack of studies of these fortifications as participants in a fortified system is the justification of the research. There is several analysis levels: territorial, historical, architectonic, defensive and constructive. Likewise, there is a lack of strictness and inaccuracy in the constructive items in the publications about several fortifications of this study field. This aspect has caused mistakes in the dating because there is neither artistic nor stylistic elements which determines a epoch. The traditionally accepted datings are challenged. An hypothesis is formulated: in the absence of artistic or stylistic elements in a sober and functional buildings is possible to date the time of construction with sufficiently approximation based on construction criteria once formed a cronotypologic classification of each building technique. The hypothesis, therefore, propose a main aim: the study of the constructive reason of the fortified border system. This aim is developed in a series of specifically targets whose achievement programs the analysis levels: - To know and to detail the historical elements which started the wars between Castile and Aragon and its development using historiographical tools, and to analyze the natural characteristic of the territory through cartographical tools. - To understand and to analyze the different architectural types and the building traditions employed in the military buildings in the time researched. - To locate, to document, and to select for their analysis the fortresses and military constructions erected during these border wars in the actual province of Soria through fieldwork and bibliographical and cartographical methods. - To conduct a general study on the fortified system in territorial scale. - To research the architectural, constructive and defensive typology of the system of these border late medieval fortifications. - To analyze the construction logic of the selected case studies and to characterize in the items of material, elements, systems and construction processes. - To sort scattered historical information and to correct mistakes to make a base by which to establish a historic speech of each case study. - To compare and to relate the construction techniques employed in these fortresses with those used in the same time range . - To spread for discussion the research results in the usual scientists forums. The method combines the destock work with an intense fieldwork. Fifty fortifications have been documented and it has written their corresponding data collection card. Data collection has been included in a database that includes general aspects, typological, constructive and basic bibliographical data, as an inventory of fortifications in the province. The selected fortifications are grouped according to a typological and constructive classification which lead the lines of the later study. There is a chapter for the antecedents in which the history of the medieval fortified construction in Europe and in Spain is studied by analyzing the evolution of architectural types and the many cultural influences along the opposite seasides of the Mediterranean Sea, from the Islamic and Crusader East to the Iberian Peninsula in where there were a long and continuous war during eight hundred years. The territory is analyzed as a container of fortifications. This analysis reveals that there is an intimate relationship between the location of the fortifications and the natural forms that define the communication roads between the Duero, Ebro and Tajo valleys. Indeed, the study area has been a cross-territory from ancient times more than a frontierterritory. This communication character explains the concerns about its control both by Rome and by the Muslims of Córdoba as medieval Christian kingdoms. The analysis of historical elements is supplemented by detailed study of border war between Castile and Aragon and the social and political issues that led to the fortification as border definition system and spatial, jurisdictional, social and administrative planning. The fortified architecture is essentially functional: it is responsible for defense. In this sense, after a morphological study of selected castles is performed an extensive analysis of its defensive elements, investigating its origin and application. This analisis serves for the definition of parameters for dating. The purpose of the thesis is the study of the constructive logic. First, various building materials are explained. Then, masonry is grouped into two major constructive groups: rigged masonry and formwork masonry. The historical evolution and the constructive typology and mensiology are studied for each one of the three main techniques: the use of brickwork, the mortar wall and rammed-earth. Many case studies have been documented along the Iberian Peninsula and also in the study area. As conclusion, there is a dimensional evolution of the rammed-earth walls. This evolution is more evident into the Muslim masonry than in the late medieval walls: the reason is the use of the cubit as module. From each of the techniques discussed, a singular and representative case of study has been selected. The castle of Arcos de Jalon is a significant example of mix masonry of stone and brick rows. The walled city of Peñalcázar is built with masonry formwork. Serón de Nágima castle, at last, is a typical and main case of the use of the rammedearth wall of late medieval military architecture. Each of these three case studies were examined under the same four analysis levels above mentioned: territorial, historical, architectural and defensive and constructive. The systematic method of study has facilitated the order in the research and the obtaining of results and conclusions that verify the hypotheses and achieve the research objectives. Dating of the fortifications construction has been revised by studying the cronotypological issues of its masonry. The method can be transferred to the study of other fortified systems. Finally, the thesis describes two main research new ways aimed at completing the study of the late medieval fortified border of Castile in the actual province of Soria. The first of them is the characterization and datig by physicochemical methods the sample pieces of wood construction preserved embedded in the masonry. The second research way is the investigation of the documents in archives that may reveal new information about the foundation, repair, sale or any aspect to economic, legal, organizational or administrative concerning fortifications in contemporary documents.
Resumo:
Basándonos en la recopilación inicial de preposiciones, locuciones preposicionales, términos con preposición dependiente y phrasal verbs utilizados en el texto técnico realizada en otros proyectos anteriores del Departamento de Lingüística Aplicada a la Ciencia y a la Tecnología, el objetivo de este trabajo es completar, organizar, actualizar y dar visibilidad a esta información inicial. Tras realizar un proceso exhaustivo de verificación, unificación, clasificación y ampliación de la información existente, en caso necesario, el listado resultante se utiliza para elaborar un glosario de términos con preposición. El objetivo final de este proyecto es que este glosario esté a disposición de los usuarios, a través de una consulta on-line, en la página del ILLLab (http://illlab.euitt.upm.es/wordpress/), dependiente del Departamento de Lingüística Aplicada a la Ciencia y a la Tecnología. Para incluir en el glosario ejemplos actualizados de textos técnicos, se ha recopilado un corpus lingüístico de textos técnicos, tomando como base diferentes números de la revista IEEE Spectrum, en su edición digital, publicados entre los años 2009 y 2012. El objetivo de esta recopilación es la de ofrecer al consultante diferentes ejemplos de uso en el texto técnico de los distintos términos con preposición que componen el glosario, de manera que pueda acceder de manera rápida y sencilla a ejemplos de uso real de los términos que está buscando, con objeto de clarificar aspectos relacionados con su uso o, en su caso, facilitar su aprendizaje. Toda esta información, tanto el listado de términos con preposición como las frases pertenecientes al corpus recopilado, se incorpora a una base de datos, alojada dentro de la misma página web del ILLLab. A través de un formulario de consulta, a disposición del usuario en dicha página, se pueden obtener todos los términos recopilados que coincidan con los criterios de búsqueda introducidos. El usuario puede realizar dos tipos de búsqueda principales: por preposición o por término completo. Además, puede elegir una búsqueda global (entre todos los términos que integran el glosario) o parcial (en una sola de las categorías en las que se han dividido los diferentes términos, de acuerdo con su función gramatical). Por último, se presentan unas estadísticas de uso de los términos recopilados dentro de los diferentes textos que integran el corpus lingüístico, de manera que pueda establecerse una relación de los que aparecen con más frecuencia en el texto técnico. ABSTRACT. Based on the initial collection of prepositions, prepositional phrases, dependent prepositions and phrasal verbs used in technical texts collected on previous projects in the Department of Applied Linguistics to Science and Technology, the aim of this project is to improve, organize, update and provide visibility to this initial information. Following a process of verification, unification, classification and extension of existing information, if necessary, a glossary of terms with preposition is built. The ultimate objective of this project is to make this glossary available to users through an online consultation in the ILLLab webpage (http://illlab.euitt.upm.es/wordpress/). The administration of tis webpage depends of the Department of Applied Linguistics in Science and Technology. A linguistic corpus of technical texts has been compiled, based on different numbers of the IEEE Spectrum magazine, in its online edition, published between the years 2009 and 2012. The aim of this collection is to provide different examples of use in the technical text for the terms included in the glossary, so that examples of the actual use of the terms consulted can be easily and quickly accessed, in order to clarify doubts regarding their meaning or translation into Spanish and facilitate learning. All this information, both the list of terms with prepositional phrases as well as the corpus developed, is incorporated in a database. Through a searching form, the ILLLab's user may obtain all the terms matching the search criteria entered. The user can perform two types of main search: by preposition or by full term. Additionally, a global search can be selected (including all terms included in the glossary) or a partial one (including only one of the glossary's categories). Finally, some statistics of use are presented according to the various texts included in the corpus, so a relation of the most frequent prepositions in the technical text can be established.
Resumo:
A novel method for generating patient-specific high quality conforming hexahedral meshes is presented. The meshes are directly obtained from the segmentation of patient magnetic resonance (MR) images of abdominal aortic aneu-rysms (AAA). The MRI permits distinguishing between struc-tures of interest in soft tissue. Being so, the contours of the lumen, the aortic wall and the intraluminal thrombus (ILT) are available and thus the meshes represent the actual anato-my of the patient?s aneurysm, including the layered morpholo-gies of these structures. Most AAAs are located in the lower part of the aorta and the upper section of the iliac arteries, where the inherent tortuosity of the anatomy and the presence of the ILT makes the generation of high-quality elements at the bifurcation is a challenging task. In this work we propose a novel approach for building quadrilateral meshes for each surface of the sectioned geometry, and generating conforming hexahedral meshes by combining the quadrilateral meshes. Conforming hexahedral meshes are created for the wall and the ILT. The resulting elements are evaluated on four patients? datasets using the Scaled Jacobian metric. Hexahedral meshes of 25,000 elements with 94.8% of elements well-suited for FE analysis are generated.
Resumo:
In order to perform finite element (FE) analyses of patient-specific abdominal aortic aneurysms, geometries derived from medical images must be meshed with suitable elements. We propose a semi-automatic method for generating conforming hexahedral meshes directly from contours segmented from medical images. Magnetic resonance images are generated using a protocol developed to give the abdominal aorta high contrast against the surrounding soft tissue. These data allow us to distinguish between the different structures of interest. We build novel quadrilateral meshes for each surface of the sectioned geometry and generate conforming hexahedral meshes by combining the quadrilateral meshes. The three-layered morphology of both the arterial wall and thrombus is incorporated using parameters determined from experiments. We demonstrate the quality of our patient-specific meshes using the element Scaled Jacobian. The method efficiently generates high-quality elements suitable for FE analysis, even in the bifurcation region of the aorta into the iliac arteries. For example, hexahedral meshes of up to 125,000 elements are generated in less than 130 s, with 94.8 % of elements well suited for FE analysis. We provide novel input for simulations by independently meshing both the arterial wall and intraluminal thrombus of the aneurysm, and their respective layered morphologies.
Resumo:
La importancia de los sistemas de recomendación ha experimentado un crecimiento exponencial como consecuencia del auge de las redes sociales. En esta tesis doctoral presentaré una amplia visión sobre el estado del arte de los sistemas de recomendación. Incialmente, estos estaba basados en fitrado demográfico, basado en contendio o colaborativo. En la actualidad, estos sistemas incorporan alguna información social al proceso de recomendación. En el futuro utilizarán información implicita, local y personal proveniente del Internet de las cosas. Los sistemas de recomendación basados en filtrado colaborativo se pueden modificar con el fin de realizar recomendaciones a grupos de usuarios. Existen trabajos previos que han incluido estas modificaciones en diferentes etapas del algoritmo de filtrado colaborativo: búsqueda de los vecinos, predicción de las votaciones y elección de las recomendaciones. En esta tesis doctoral proporcionaré un nuevo método que realizar el proceso de unficación (pasar de varios usuarios a un grupo) en el primer paso del algoritmo de filtrado colaborativo: cálculo de la métrica de similaridad. Proporcionaré una formalización completa del método propuesto. Explicaré cómo obtener el conjunto de k vecinos del grupo de usuarios y mostraré cómo obtener recomendaciones usando dichos vecinos. Asimismo, incluiré un ejemplo detallando cada paso del método propuesto en un sistema de recomendación compuesto por 8 usuarios y 10 items. Las principales características del método propuesto son: (a) es más rápido (más eficiente) que las alternativas proporcionadas por otros autores, y (b) es al menos tan exacto y preciso como otras soluciones estudiadas. Para contrastar esta hipótesis realizaré varios experimentos que miden la precisión, la exactitud y el rendimiento del método. Los resultados obtenidos se compararán con los resultados de otras alternativas utilizadas en la recomendación de grupos. Los experimentos se realizarán con las bases de datos de MovieLens y Netflix. ABSTRACT The importance of recommender systems has grown exponentially with the advent of social networks. In this PhD thesis I will provide a wide vision about the state of the art of recommender systems. They were initially based on demographic, contentbased and collaborative filtering. Currently, these systems incorporate some social information to the recommendation process. In the future, they will use implicit, local and personal information from the Internet of Things. As we will see here, recommender systems based on collaborative filtering can be used to perform recommendations to group of users. Previous works have made this modification in different stages of the collaborative filtering algorithm: establishing the neighborhood, prediction phase and determination of recommended items. In this PhD thesis I will provide a new method that carry out the unification process (many users to one group) in the first stage of the collaborative filtering algorithm: similarity metric computation. I will provide a full formalization of the proposed method. I will explain how to obtain the k nearest neighbors of the group of users and I will show how to get recommendations using those users. I will also include a running example of a recommender system with 8 users and 10 items detailing all the steps of the method I will present. The main highlights of the proposed method are: (a) it will be faster (more efficient) that the alternatives provided by other authors, and (b) it will be at least as precise and accurate as other studied solutions. To check this hypothesis I will conduct several experiments measuring the accuracy, the precision and the performance of my method. I will compare these results with the results generated by other methods of group recommendation. The experiments will be carried out using MovieLens and Netflix datasets.