960 resultados para continuous inverse systems
Resumo:
The Tara Oceans Expedition (2009-2013) sampled the world oceans on board a 36 m long schooner, collecting environmental data and organisms from viruses to planktonic metazoans for later analyses using modern sequencing and state-of-the-art imaging technologies. Tara Oceans Data are particularly suited to study the genetic, morphological and functional diversity of plankton. The present data set provides continuous measurements of partial pressure of carbon dioxide (pCO2), using a ProOceanus CO2-Pro instrument mounted on the flowthrough system. This automatic sensor is fitted with an equilibrator made of gas permeable silicone membrane and an internal detection loop with a non-dispersive infrared detector of PPSystems SBA-4 CO2 analyzer. A zero-CO2 baseline is provided for the subsequent measurements circulating the internal gas through a CO2 absorption chamber containing soda lime or Ascarite. The frequency of this automatic zero point calibration was set to be 24 hours. All data recorded during zeroing processes were discarded with the 15-minute data after each calibration. The output of CO2-Pro is the mole fraction of CO2 in the measured water and the pCO2 is obtained using the measured total pressure of the internal wet gas. The fugacity of CO2 (fCO2) in the surface seawater, whose difference with the atmospheric CO2 fugacity is proportional to the air-sea CO2 fluxes, is obtained by correcting the pCO2 for non-ideal CO2 gas concentration according to Weiss (1974). The fCO2 computed using CO2-Pro measurements was corrected to the sea surface condition by considering the temperature effect on fCO2 (Takahashi et al., 1993). The surface seawater observations that were initially estimated with a 15 seconds frequency were averaged every 5-min cycle. The performance of CO2-Pro was adjusted by comparing the sensor outputs against the thermodynamic carbonate calculation of pCO2 using the carbonic system constants of Millero et al. (2006) from the determinations of total inorganic carbon (CT ) and total alkalinity (AT ) in discrete samples collected at sea surface. AT was determined using an automated open cell potentiometric titration (Haraldsson et al. 1997). CT was determined with an automated coulometric titration (Johnson et al. 1985; 1987), using the MIDSOMMA system (Mintrop, 2005). fCO2 data are flagged according to the WOCE guidelines following Pierrot et al. (2009) identifying recommended values and questionable measurements giving additional information about the reasons of the questionability.
Resumo:
In this paper, we describe our research on bio-inspired locomotion systems using deformable structures and smart materials, concretely shape memory alloys (SMAs). These types of materials allow us to explore the possibility of building motor-less and gear-less robots. A swimming underwater fish-like robot has been developed whose movements are generated using SMAs. These actuators are suitable for bending the continuous backbone of the fish, which in turn causes a change in the curvature of the body. This type of structural arrangement is inspired by fish red muscles, which are mainly recruited during steady swimming for the bending of a flexible but nearly incompressible structure such as the fishbone. This paper reviews the design process of these bio-inspired structures, from the motivations and physiological inspiration to the mechatronics design, control and simulations, leading to actual experimental trials and results. The focus of this work is to present the mechanisms by which standard swimming patterns can be reproduced with the proposed design. Moreover, the performance of the SMA-based actuators’ control in terms of actuation speed and position accuracy is also addressed.
Resumo:
This article describes a knowledge-based method for generating multimedia descriptions that summarize the behavior of dynamic systems. We designed this method for users who monitor the behavior of a dynamic system with the help of sensor networks and make decisions according to prefixed management goals. Our method generates presentations using different modes such as text in natural language, 2D graphics and 3D animations. The method uses a qualitative representation of the dynamic system based on hierarchies of components and causal influences. The method includes an abstraction generator that uses the system representation to find and aggregate relevant data at an appropriate level of abstraction. In addition, the method includes a hierarchical planner to generate a presentation using a model with dis- course patterns. Our method provides an efficient and flexible solution to generate concise and adapted multimedia presentations that summarize thousands of time series. It is general to be adapted to differ- ent dynamic systems with acceptable knowledge acquisition effort by reusing and adapting intuitive rep- resentations. We validated our method and evaluated its practical utility by developing several models for an application that worked in continuous real time operation for more than 1 year, summarizing sen- sor data of a national hydrologic information system in Spain.
Resumo:
La Organización Mundial de la Salud (OMS) prevé que para el año 2020, el Daño Cerebral Adquirido (DCA) estará entre las 10 causas más comunes de discapacidad. Estas lesiones, dadas sus consecuencias físicas, sensoriales, cognitivas, emocionales y socioeconómicas, cambian dramáticamente la vida de los pacientes y sus familias. Las nuevas técnicas de intervención precoz y el desarrollo de la medicina intensiva en la atención al DCA han mejorado notablemente la probabilidad de supervivencia. Sin embargo, hoy por hoy, las lesiones cerebrales no tienen ningún tratamiento quirúrgico que tenga por objetivo restablecer la funcionalidad perdida, sino que las terapias rehabilitadoras se dirigen hacia la compensación de los déficits producidos. Uno de los objetivos principales de la neurorrehabilitación es, por tanto, dotar al paciente de la capacidad necesaria para ejecutar las Actividades de Vida Diaria (AVDs) necesarias para desarrollar una vida independiente, siendo fundamentales aquellas en las que la Extremidad Superior (ES) está directamente implicada, dada su gran importancia a la hora de la manipulación de objetos. Con la incorporación de nuevas soluciones tecnológicas al proceso de neurorrehabilitación se pretende alcanzar un nuevo paradigma centrado en ofrecer una práctica personalizada, monitorizada y ubicua con una valoración continua de la eficacia y de la eficiencia de los procedimientos y con capacidad de generar conocimientos que impulsen la ruptura del paradigma de actual. Los nuevos objetivos consistirán en minimizar el impacto de las enfermedades que afectan a la capacidad funcional de las personas, disminuir el tiempo de incapacidad y permitir una gestión más eficiente de los recursos. Estos objetivos clínicos, de gran impacto socio-económico, sólo pueden alcanzarse desde una apuesta decidida en nuevas tecnologías, metodologías y algoritmos capaces de ocasionar la ruptura tecnológica necesaria que permita superar las barreras que hasta el momento han impedido la penetración tecnológica en el campo de la rehabilitación de manera universal. De esta forma, los trabajos y resultados alcanzados en la Tesis son los siguientes: 1. Modelado de AVDs: como paso previo a la incorporación de ayudas tecnológicas al proceso rehabilitador, se hace necesaria una primera fase de modelado y formalización del conocimiento asociado a la ejecución de las actividades que se realizan como parte de la terapia. En particular, las tareas más complejas y a su vez con mayor repercusión terapéutica son las AVDs, cuya formalización permitirá disponer de modelos de movimiento sanos que actuarán de referencia para futuros desarrollos tecnológicos dirigidos a personas con DCA. Siguiendo una metodología basada en diagramas de estados UML se han modelado las AVDs 'servir agua de una jarra' y 'coger un botella' 2. Monitorización ubícua del movimiento de la ES: se ha diseñado, desarrollado y validado un sistema de adquisición de movimiento basado en tecnología inercial que mejora las limitaciones de los dispositivos comerciales actuales (coste muy elevado e incapacidad para trabajar en entornos no controlados); los altos coeficientes de correlación y los bajos niveles de error obtenidos en los corregistros llevados a cabo con el sistema comercial BTS SMART-D demuestran la alta precisión del sistema. También se ha realizado un trabajo de investigación exploratorio de un sistema de captura de movimiento de coste muy reducido basado en visión estereoscópica, habiéndose detectado los puntos clave donde se hace necesario incidir desde un punto de vista tecnológico para su incorporación en un entorno real 3. Resolución del Problema Cinemático Inverso (PCI): se ha diseñado, desarrollado y validado una solución al PCI cuando el manipulador se corresponde con una ES humana estudiándose 2 posibles alternativas, una basada en la utilización de un Perceptrón Multicapa (PMC) y otra basada en sistemas Artificial Neuro-Fuzzy Inference Systems (ANFIS). La validación, llevada a cabo utilizando información relativa a los modelos disponibles de AVDs, indica que una solución basada en un PMC con 3 neuronas en la capa de entrada, una capa oculta también de 3 neuronas y una capa de salida con tantas neuronas como Grados de Libertad (GdLs) tenga el modelo de la ES, proporciona resultados, tanto de precisión como de tiempo de cálculo, que la hacen idónea para trabajar en sistemas con requisitos de tiempo real 4. Control inteligente assisted-as-needed: se ha diseñado, desarrollado y validado un algoritmo de control assisted-as-needed para una ortesis robótica con capacidades de actuación anticipatoria de la que existe un prototipo implementado en la actualidad. Los resultados obtenidos demuestran cómo el sistema es capaz de adaptarse al perfil disfuncional del paciente activando la ayuda en instantes anteriores a la ocurrencia de movimientos incorrectos. Esta estrategia implica un aumento en la participación del paciente y, por tanto, en su actividad muscular, fomentándose los procesos la plasticidad cerebral responsables del reaprendizaje o readaptación motora 5. Simuladores robóticos para planificación: se propone la utilización de un simulador robótico assisted-as-needed como herramienta de planificación de sesiones de rehabilitación personalizadas y con un objetivo clínico marcado en las que interviene una ortesis robotizada. Los resultados obtenidos evidencian como, tras la ejecución de ciertos algoritmos sencillos, es posible seleccionar automáticamente una configuración para el algoritmo de control assisted-as-needed que consigue que la ortesis se adapte a los criterios establecidos desde un punto de vista clínico en función del paciente estudiado. Estos resultados invitan a profundizar en el desarrollo de algoritmos más avanzados de selección de parámetros a partir de baterías de simulaciones Estos trabajos han servido para corroborar las hipótesis de investigación planteadas al inicio de la misma, permitiendo, asimismo, la apertura de nuevas líneas de investigación. Summary The World Health Organization (WHO) predicts that by the year 2020, Acquired Brain Injury (ABI) will be among the ten most common ailments. These injuries dramatically change the life of the patients and their families due to their physical, sensory, cognitive, emotional and socio-economic consequences. New techniques of early intervention and the development of intensive ABI care have noticeably improved the survival rate. However, in spite of these advances, brain injuries still have no surgical or pharmacological treatment to re-establish the lost functions. Neurorehabilitation therapies address this problem by restoring, minimizing or compensating the functional alterations in a person disabled because of a nervous system injury. One of the main objectives of Neurorehabilitation is to provide patients with the capacity to perform specific Activities of the Daily Life (ADL) required for an independent life, especially those in which the Upper Limb (UL) is directly involved due to its great importance in manipulating objects within the patients' environment. The incorporation of new technological aids to the neurorehabilitation process tries to reach a new paradigm focused on offering a personalized, monitored and ubiquitous practise with continuous assessment of both the efficacy and the efficiency of the procedures and with the capacity of generating new knowledge. New targets will be to minimize the impact of the sicknesses affecting the functional capabilitiies of the subjects, to decrease the time of the physical handicap and to allow a more efficient resources handling. These targets, of a great socio-economic impact, can only be achieved by means of new technologies and algorithms able to provoke the technological break needed to beat the barriers that are stopping the universal penetration of the technology in the field of rehabilitation. In this way, this PhD Thesis has achieved the following results: 1. ADL Modeling: as a previous step to the incorporation of technological aids to the neurorehabilitation process, it is necessary a first modelling and formalization phase of the knowledge associated to the execution of the activities that are performed as a part of the therapy. In particular, the most complex and therapeutically relevant tasks are the ADLs, whose formalization will produce healthy motion models to be used as a reference for future technological developments. Following a methodology based on UML state-chart diagrams, the ADLs 'serving water from a jar' and 'picking up a bottle' have been modelled 2. Ubiquitous monitoring of the UL movement: it has been designed, developed and validated a motion acquisition system based on inertial technology that improves the limitations of the current devices (high monetary cost and inability of working within uncontrolled environments); the high correlation coefficients and the low error levels obtained throughout several co-registration sessions with the commercial sys- tem BTS SMART-D show the high precision of the system. Besides an exploration of a very low cost stereoscopic vision-based motion capture system has been carried out and the key points where it is necessary to insist from a technological point of view have been detected 3. Inverse Kinematics (IK) problem solving: a solution to the IK problem has been proposed for a manipulator that corresponds to a human UL. This solution has been faced by means of two different alternatives, one based on a Mulilayer Perceptron (MLP) and another based on Artificial Neuro-Fuzzy Inference Systems (ANFIS). The validation of these solutions, carried out using the information regarding the previously generated motion models, indicate that a MLP-based solution, with an architecture consisting in 3 neurons in the input layer, one hidden layer of 3 neurons and an output layer with as many neurons as the number of Degrees of Freedom (DoFs) that the UL model has, is the one that provides the best results both in terms of precission and in terms of processing time, making in idoneous to be integrated within a system with real time restrictions 4. Assisted-as-needed intelligent control: an assisted-as-needed control algorithm with anticipatory actuation capabilities has been designed, developed and validated for a robotic orthosis of which there is an already implemented prototype. Obtained results demonstrate that the control system is able to adapt to the dysfunctional profile of the patient by triggering the assistance right before an incorrect movement is going to take place. This strategy implies an increase in the participation of the patients and in his or her muscle activity, encouraging the neural plasticity processes in charge of the motor learning 5. Planification with a robotic simulator: in this work a robotic simulator is proposed as a planification tool for personalized rehabilitation sessions under a certain clinical criterium. Obtained results indicate that, after the execution of simple parameter selection algorithms, it is possible to automatically choose a specific configuration that makes the assisted-as-needed control algorithm to adapt both to the clinical criteria and to the patient. These results invite researchers to work in the development of more complex parameter selection algorithms departing from simulation batteries Obtained results have been useful to corroborate the hypotheses set out at the beginning of this PhD Thesis. Besides, they have allowed the creation of new research lines in all the studied application fields.
Resumo:
Objective: This research is focused in the creation and validation of a solution to the inverse kinematics problem for a 6 degrees of freedom human upper limb. This system is intended to work within a realtime dysfunctional motion prediction system that allows anticipatory actuation in physical Neurorehabilitation under the assisted-as-needed paradigm. For this purpose, a multilayer perceptron-based and an ANFIS-based solution to the inverse kinematics problem are evaluated. Materials and methods: Both the multilayer perceptron-based and the ANFIS-based inverse kinematics methods have been trained with three-dimensional Cartesian positions corresponding to the end-effector of healthy human upper limbs that execute two different activities of the daily life: "serving water from a jar" and "picking up a bottle". Validation of the proposed methodologies has been performed by a 10 fold cross-validation procedure. Results: Once trained, the systems are able to map 3D positions of the end-effector to the corresponding healthy biomechanical configurations. A high mean correlation coefficient and a low root mean squared error have been found for both the multilayer perceptron and ANFIS-based methods. Conclusions: The obtained results indicate that both systems effectively solve the inverse kinematics problem, but, due to its low computational load, crucial in real-time applications, along with its high performance, a multilayer perceptron-based solution, consisting in 3 input neurons, 1 hidden layer with 3 neurons and 6 output neurons has been considered the most appropriated for the target application.
Resumo:
Agro-areas of Arroyos Menores (La Colacha) west and south of Rand south of R?o Cuarto (Prov. of Cordoba, Argentina) basins are very fertile but have high soil loses. Extreme rain events, inundations and other severe erosions forming gullies demand urgently actions in this area to avoid soil degradation and erosion supporting good levels of agro production. The authors first improved hydrologic data on La Colacha, evaluated the systems of soil uses and actions that could be recommended considering the relevant aspects of the study area and applied decision support systems (DSS) with mathematic tools for planning of defences and uses of soils in these areas. These were conducted here using multi-criteria models, in multi-criteria decision making (MCDM); first of discrete MCDM to chose among global types of use of soils, and then of continuous MCDM to evaluate and optimize combined actions, including repartition of soil use and the necessary levels of works for soil conservation and for hydraulic management to conserve against erosion these basins. Relatively global solutions for La Colacha area have been defined and were optimised by Linear Programming in Goal Programming forms that are presented as Weighted or Lexicographic Goal Programming and as Compromise Programming. The decision methods used are described, indicating algorithms used, and examples for some representative scenarios on La Colacha area are given.
Resumo:
INTRODUCTION: Objective assessment of motor skills has become an important challenge in minimally invasive surgery (MIS) training.Currently, there is no gold standard defining and determining the residents' surgical competence.To aid in the decision process, we analyze the validity of a supervised classifier to determine the degree of MIS competence based on assessment of psychomotor skills METHODOLOGY: The ANFIS is trained to classify performance in a box trainer peg transfer task performed by two groups (expert/non expert). There were 42 participants included in the study: the non-expert group consisted of 16 medical students and 8 residents (< 10 MIS procedures performed), whereas the expert group consisted of 14 residents (> 10 MIS procedures performed) and 4 experienced surgeons. Instrument movements were captured by means of the Endoscopic Video Analysis (EVA) tracking system. Nine motion analysis parameters (MAPs) were analyzed, including time, path length, depth, average speed, average acceleration, economy of area, economy of volume, idle time and motion smoothness. Data reduction was performed by means of principal component analysis, and then used to train the ANFIS net. Performance was measured by leave one out cross validation. RESULTS: The ANFIS presented an accuracy of 80.95%, where 13 experts and 21 non-experts were correctly classified. Total root mean square error was 0.88, while the area under the classifiers' ROC curve (AUC) was measured at 0.81. DISCUSSION: We have shown the usefulness of ANFIS for classification of MIS competence in a simple box trainer exercise. The main advantage of using ANFIS resides in its continuous output, which allows fine discrimination of surgical competence. There are, however, challenges that must be taken into account when considering use of ANFIS (e.g. training time, architecture modeling). Despite this, we have shown discriminative power of ANFIS for a low-difficulty box trainer task, regardless of the individual significances between MAPs. Future studies are required to confirm the findings, inclusion of new tasks, conditions and sample population.
Resumo:
La computación basada en servicios (Service-Oriented Computing, SOC) se estableció como un paradigma ampliamente aceptado para el desarollo de sistemas de software flexibles, distribuidos y adaptables, donde las composiciones de los servicios realizan las tareas más complejas o de nivel más alto, frecuentemente tareas inter-organizativas usando los servicios atómicos u otras composiciones de servicios. En tales sistemas, las propriedades de la calidad de servicio (Quality of Service, QoS), como la rapídez de procesamiento, coste, disponibilidad o seguridad, son críticas para la usabilidad de los servicios o sus composiciones en cualquier aplicación concreta. El análisis de estas propriedades se puede realizarse de una forma más precisa y rica en información si se utilizan las técnicas de análisis de programas, como el análisis de complejidad o de compartición de datos, que son capables de analizar simultáneamente tanto las estructuras de control como las de datos, dependencias y operaciones en una composición. El análisis de coste computacional para la composicion de servicios puede ayudar a una monitorización predictiva así como a una adaptación proactiva a través de una inferencia automática de coste computacional, usando los limites altos y bajos como funciones del valor o del tamaño de los mensajes de entrada. Tales funciones de coste se pueden usar para adaptación en la forma de selección de los candidatos entre los servicios que minimizan el coste total de la composición, basado en los datos reales que se pasan al servicio. Las funciones de coste también pueden ser combinadas con los parámetros extraídos empíricamente desde la infraestructura, para producir las funciones de los límites de QoS sobre los datos de entrada, cuales se pueden usar para previsar, en el momento de invocación, las violaciones de los compromisos al nivel de servicios (Service Level Agreements, SLA) potenciales or inminentes. En las composiciones críticas, una previsión continua de QoS bastante eficaz y precisa se puede basar en el modelado con restricciones de QoS desde la estructura de la composition, datos empiricos en tiempo de ejecución y (cuando estén disponibles) los resultados del análisis de complejidad. Este enfoque se puede aplicar a las orquestaciones de servicios con un control centralizado del flujo, así como a las coreografías con participantes multiples, siguiendo unas interacciones complejas que modifican su estado. El análisis del compartición de datos puede servir de apoyo para acciones de adaptación, como la paralelización, fragmentación y selección de los componentes, las cuales son basadas en dependencias funcionales y en el contenido de información en los mensajes, datos internos y las actividades de la composición, cuando se usan construcciones de control complejas, como bucles, bifurcaciones y flujos anidados. Tanto las dependencias funcionales como el contenido de información (descrito a través de algunos atributos definidos por el usuario) se pueden expresar usando una representación basada en la lógica de primer orden (claúsulas de Horn), y los resultados del análisis se pueden interpretar como modelos conceptuales basados en retículos. ABSTRACT Service-Oriented Computing (SOC) is a widely accepted paradigm for development of flexible, distributed and adaptable software systems, in which service compositions perform more complex, higher-level, often cross-organizational tasks using atomic services or other service compositions. In such systems, Quality of Service (QoS) properties, such as the performance, cost, availability or security, are critical for the usability of services and their compositions in concrete applications. Analysis of these properties can become more precise and richer in information, if it employs program analysis techniques, such as the complexity and sharing analyses, which are able to simultaneously take into account both the control and the data structures, dependencies, and operations in a composition. Computation cost analysis for service composition can support predictive monitoring and proactive adaptation by automatically inferring computation cost using the upper and lower bound functions of value or size of input messages. These cost functions can be used for adaptation by selecting service candidates that minimize total cost of the composition, based on the actual data that is passed to them. The cost functions can also be combined with the empirically collected infrastructural parameters to produce QoS bounds functions of input data that can be used to predict potential or imminent Service Level Agreement (SLA) violations at the moment of invocation. In mission-critical applications, an effective and accurate continuous QoS prediction, based on continuations, can be achieved by constraint modeling of composition QoS based on its structure, known data at runtime, and (when available) the results of complexity analysis. This approach can be applied to service orchestrations with centralized flow control, and choreographies with multiple participants with complex stateful interactions. Sharing analysis can support adaptation actions, such as parallelization, fragmentation, and component selection, which are based on functional dependencies and information content of the composition messages, internal data, and activities, in presence of complex control constructs, such as loops, branches, and sub-workflows. Both the functional dependencies and the information content (described using user-defined attributes) can be expressed using a first-order logic (Horn clause) representation, and the analysis results can be interpreted as a lattice-based conceptual models.
Resumo:
We report conditions on a switching signal that guarantee that solutions of a switched linear systems converge asymptotically to zero. These conditions are apply to continuous, discrete-time and hybrid switched linear systems, both those having stable subsystems and mixtures of stable and unstable subsystems.
Resumo:
The analysis of complex nonlinear systems is often carried out using simpler piecewise linear representations of them. A principled and practical technique is proposed to linearize and evaluate arbitrary continuous nonlinear functions using polygonal (continuous piecewise linear) models under the L1 norm. A thorough error analysis is developed to guide an optimal design of two kinds of polygonal approximations in the asymptotic case of a large budget of evaluation subintervals N. The method allows the user to obtain the level of linearization (N) for a target approximation error and vice versa. It is suitable for, but not limited to, an efficient implementation in modern Graphics Processing Units (GPUs), allowing real-time performance of computationally demanding applications. The quality and efficiency of the technique has been measured in detail on two nonlinear functions that are widely used in many areas of scientific computing and are expensive to evaluate.
Resumo:
With the advent of cloud computing model, distributed caches have become the cornerstone for building scalable applications. Popular systems like Facebook [1] or Twitter use Memcached [5], a highly scalable distributed object cache, to speed up applications by avoiding database accesses. Distributed object caches assign objects to cache instances based on a hashing function, and objects are not moved from a cache instance to another unless more instances are added to the cache and objects are redistributed. This may lead to situations where some cache instances are overloaded when some of the objects they store are frequently accessed, while other cache instances are less frequently used. In this paper we propose a multi-resource load balancing algorithm for distributed cache systems. The algorithm aims at balancing both CPU and Memory resources among cache instances by redistributing stored data. Considering the possible conflict of balancing multiple resources at the same time, we give CPU and Memory resources weighted priorities based on the runtime load distributions. A scarcer resource is given a higher weight than a less scarce resource when load balancing. The system imbalance degree is evaluated based on monitoring information, and the utility load of a node, a unit for resource consumption. Besides, since continuous rebalance of the system may affect the QoS of applications utilizing the cache system, our data selection policy ensures that each data migration minimizes the system imbalance degree and hence, the total reconfiguration cost can be minimized. An extensive simulation is conducted to compare our policy with other policies. Our policy shows a significant improvement in time efficiency and decrease in reconfiguration cost.
Resumo:
Los modelos de simulación de cultivos permiten analizar varias combinaciones de laboreo-rotación y explorar escenarios de manejo. El modelo DSSAT fue evaluado bajo condiciones de secano en un experimento de campo de 16 años en la semiárida España central. Se evaluó el efecto del sistema de laboreo y las rotaciones basadas en cereales de invierno, en el rendimiento del cultivo y la calidad del suelo. Los modelos CERES y CROPGRO se utilizaron para simular el crecimiento y rendimiento del cultivo, mientras que el modelo DSSAT CENTURY se utilizó en las simulaciones de SOC y SN. Tanto las observaciones de campo como las simulaciones con CERES-Barley, mostraron que el rendimiento en grano de la cebada era mas bajo para el cereal continuo (BB) que para las rotaciones de veza (VB) y barbecho (FB) en ambos sistemas de laboreo. El modelo predijo más nitrógeno disponible en el laboreo convencional (CT) que en el no laboreo (NT) conduciendo a un mayor rendimiento en el CT. El SOC y el SN en la capa superficial del suelo, fueron mayores en NT que en CT, y disminuyeron con la profundidad en los valores tanto observados como simulados. Las mejores combinaciones para las condiciones de secano estudiadas fueron CT-VB y CT-FB, pero CT presentó menor contenido en SN y SOC que NT. El efecto beneficioso del NT en SOC y SN bajo condiciones Mediterráneas semiáridas puede ser identificado por observaciones de campo y por simulaciones de modelos de cultivos. La simulación del balance de agua en sistemas de cultivo es una herramienta útil para estudiar como el agua puede ser utilizado eficientemente. La comparación del balance de agua de DSSAT , con una simple aproximación “tipping bucket”, con el modelo WAVE más mecanicista, el cual integra la ecuación de Richard , es un potente método para valorar el funcionamiento del modelo. Los parámetros de suelo fueron calibrados usando el método de optimización global Simulated Annealing (SA). Un lisímetro continuo de pesada en suelo desnudo suministró los valores observados de drenaje y evapotranspiración (ET) mientras que el contenido de agua en el suelo (SW) fue suministrado por sensores de capacitancia. Ambos modelos funcionaron bien después de la optimización de los parámetros de suelo con SA, simulando el balance de agua en el suelo para el período de calibración. Para el período de validación, los modelos optimizados predijeron bien el contenido de agua en el suelo y la evaporación del suelo a lo largo del tiempo. Sin embargo, el drenaje fue predicho mejor con WAVE que con DSSAT, el cual presentó mayores errores en los valores acumulados. Esto podría ser debido a la naturaleza mecanicista de WAVE frente a la naturaleza más funcional de DSSAT. Los buenos resultados de WAVE indican que, después de la calibración, este puede ser utilizado como "benchmark" para otros modelos para periodos en los que no haya medidas de campo del drenaje. El funcionamiento de DSSAT-CENTURY en la simulación de SOC y N depende fuertemente del proceso de inicialización. Se propuso como método alternativo (Met.2) la inicialización de las fracciones de SOC a partir de medidas de mineralización aparente del suelo (Napmin). El Met.2 se comparó con el método de inicialización de Basso et al. (2011) (Met.1), aplicando ambos métodos a un experimento de campo de 4 años en un área en regadío de España central. Nmin y Napmin fueron sobreestimados con el Met.1, ya que la fracción estable obtenida (SOC3) en las capas superficiales del suelo fue más baja que con Met.2. El N lixiviado simulado fue similar en los dos métodos, con buenos resultados en los tratamientos de barbecho y cebada. El Met.1 subestimó el SOC en la capa superficial del suelo cuando se comparó con una serie observada de 12 años. El crecimiento y rendimiento del cultivo fueron adecuadamente simulados con ambos métodos, pero el N en la parte aérea de la planta y en el grano fueron sobreestimados con el Met.1. Los resultados variaron significativamente con las fracciones iniciales de SOC, resaltando la importancia del método de inicialización. El Met.2 ofrece una alternativa para la inicialización del modelo CENTURY, mejorando la simulación de procesos de N en el suelo. La continua emergencia de nuevas variedades de híbridos modernos de maíz limita la aplicación de modelos de simulación de cultivos, ya que estos nuevos híbridos necesitan ser calibrados en el campo para ser adecuados para su uso en los modelos. El desarrollo de relaciones basadas en la duración del ciclo, simplificaría los requerimientos de calibración facilitando la rápida incorporación de nuevos cultivares en DSSAT. Seis híbridos de maiz (FAO 300 hasta FAO 700) fueron cultivados en un experimento de campo de dos años en un área semiárida de regadío en España central. Los coeficientes genéticos fueron obtenidos secuencialmente, comenzando con los parámetros de desarrollo fenológico (P1, P2, P5 and PHINT), seguido de los parámetros de crecimiento del cultivo (G2 and G3). Se continuó el procedimiento hasta que la salida de las simulaciones estuvo en concordancia con las observaciones fenológicas de campo. Después de la calibración, los parámetros simulados se ajustaron bien a los parámetros observados, con bajos RMSE en todos los casos. Los P1 y P5 calibrados, incrementaron con la duración del ciclo. P1 fue una función lineal del tiempo térmico (TT) desde emergencia hasta floración y P5 estuvo linealmente relacionada con el TT desde floración a madurez. No hubo diferencias significativas en PHINT entre híbridos de FAO-500 a 700 , ya que tuvieron un número de hojas similar. Como los coeficientes fenológicos estuvieron directamente relacionados con la duración del ciclo, sería posible desarrollar rangos y correlaciones que permitan estimar dichos coeficientes a partir de la clasificación del ciclo. ABSTRACT Crop simulation models allow analyzing various tillage-rotation combinations and exploring management scenarios. DSSAT model was tested under rainfed conditions in a 16-year field experiment in semiarid central Spain. The effect of tillage system and winter cereal-based rotations on the crop yield and soil quality was evaluated. The CERES and CROPGRO models were used to simulate crop growth and yield, while the DSSAT CENTURY was used in the SOC and SN simulations. Both field observations and CERES-Barley simulations, showed that barley grain yield was lower for continuous cereal (BB) than for vetch (VB) and fallow (FB) rotations for both tillage systems. The model predicted higher nitrogen availability in the conventional tillage (CT) than in the no tillage (NT) leading to a higher yield in the CT. The SOC and SN in the top layer, were higher in NT than in CT, and decreased with depth in both simulated and observed values. The best combinations for the dry land conditions studied were CT-VB and CT-FB, but CT presented lower SN and SOC content than NT. The beneficial effect of NT on SOC and SN under semiarid Mediterranean conditions can be identified by field observations and by crop model simulations. The simulation of the water balance in cropping systems is a useful tool to study how water can be used efficiently. The comparison of DSSAT soil water balance, with a simpler “tipping bucket” approach, with the more mechanistic WAVE model, which integrates Richard’s equation, is a powerful method to assess model performance. The soil parameters were calibrated by using the Simulated Annealing (SA) global optimizing method. A continuous weighing lysimeter in a bare fallow provided the observed values of drainage and evapotranspiration (ET) while soil water content (SW) was supplied by capacitance sensors. Both models performed well after optimizing soil parameters with SA, simulating the soil water balance components for the calibrated period. For the validation period, the optimized models predicted well soil water content and soil evaporation over time. However, drainage was predicted better by WAVE than by DSSAT, which presented larger errors in the cumulative values. That could be due to the mechanistic nature of WAVE against the more functional nature of DSSAT. The good results from WAVE indicate that, after calibration, it could be used as benchmark for other models for periods when no drainage field measurements are available. The performance of DSSAT-CENTURY when simulating SOC and N strongly depends on the initialization process. Initialization of the SOC pools from apparent soil N mineralization (Napmin) measurements was proposed as alternative method (Met.2). Method 2 was compared to the Basso et al. (2011) initialization method (Met.1), by applying both methods to a 4-year field experiment in a irrigated area of central Spain. Nmin and Napmin were overestimated by Met.1, since the obtained stable pool (SOC3) in the upper layers was lower than from Met.2. Simulated N leaching was similar for both methods, with good results in fallow and barley treatments. Method 1 underestimated topsoil SOC when compared with a 12-year observed serial. Crop growth and yield were properly simulated by both methods, but N in shoots and grain were overestimated by Met.1. Results varied significantly with the initial SOC pools, highlighting the importance of the initialization procedure. Method 2 offers an alternative to initialize the CENTURY model, enhancing the simulation of soil N processes. The continuous emergence of new varieties of modern maize hybrids limits the application of crop simulation models, since these new hybrids should be calibrated in the field to be suitable for model use. The development of relationships based on the cycle duration, would simplify the calibration requirements facilitating the rapid incorporation of new cultivars into DSSAT. Six maize hybrids (FAO 300 through FAO 700) were grown in a 2-year field experiment in a semiarid irrigated area of central Spain. Genetic coefficients were obtained sequentially, starting with the phenological development parameters (P1, P2, P5 and PHINT), followed by the crop growth parameters (G2 and G3). The procedure was continued until the simulated outputs were in good agreement with the field phenological observations. After calibration, simulated parameters matched observed parameters well, with low RMSE in most cases. The calibrated P1 and P5 increased with the duration of the cycle. P1 was a linear function of the thermal time (TT) from emergence to silking and P5 was linearly related with the TT from silking to maturity . There were no significant differences in PHINT between hybrids from FAO-500 to 700 , as they had similar leaf number. Since phenological coefficients were directly related with the cycle duration, it would be possible to develop ranges and correlations which allow to estimate such coefficients from the cycle classification.