960 resultados para Separating of variables
Resumo:
Una apropiada evaluación de los márgenes de seguridad de una instalación nuclear, por ejemplo, una central nuclear, tiene en cuenta todas las incertidumbres que afectan a los cálculos de diseño, funcionanmiento y respuesta ante accidentes de dicha instalación. Una fuente de incertidumbre son los datos nucleares, que afectan a los cálculos neutrónicos, de quemado de combustible o activación de materiales. Estos cálculos permiten la evaluación de las funciones respuesta esenciales para el funcionamiento correcto durante operación, y también durante accidente. Ejemplos de esas respuestas son el factor de multiplicación neutrónica o el calor residual después del disparo del reactor. Por tanto, es necesario evaluar el impacto de dichas incertidumbres en estos cálculos. Para poder realizar los cálculos de propagación de incertidumbres, es necesario implementar metodologías que sean capaces de evaluar el impacto de las incertidumbres de estos datos nucleares. Pero también es necesario conocer los datos de incertidumbres disponibles para ser capaces de manejarlos. Actualmente, se están invirtiendo grandes esfuerzos en mejorar la capacidad de analizar, manejar y producir datos de incertidumbres, en especial para isótopos importantes en reactores avanzados. A su vez, nuevos programas/códigos están siendo desarrollados e implementados para poder usar dichos datos y analizar su impacto. Todos estos puntos son parte de los objetivos del proyecto europeo ANDES, el cual ha dado el marco de trabajo para el desarrollo de esta tesis doctoral. Por tanto, primero se ha llevado a cabo una revisión del estado del arte de los datos nucleares y sus incertidumbres, centrándose en los tres tipos de datos: de decaimiento, de rendimientos de fisión y de secciones eficaces. A su vez, se ha realizado una revisión del estado del arte de las metodologías para la propagación de incertidumbre de estos datos nucleares. Dentro del Departamento de Ingeniería Nuclear (DIN) se propuso una metodología para la propagación de incertidumbres en cálculos de evolución isotópica, el Método Híbrido. Esta metodología se ha tomado como punto de partida para esta tesis, implementando y desarrollando dicha metodología, así como extendiendo sus capacidades. Se han analizado sus ventajas, inconvenientes y limitaciones. El Método Híbrido se utiliza en conjunto con el código de evolución isotópica ACAB, y se basa en el muestreo por Monte Carlo de los datos nucleares con incertidumbre. En esta metodología, se presentan diferentes aproximaciones según la estructura de grupos de energía de las secciones eficaces: en un grupo, en un grupo con muestreo correlacionado y en multigrupos. Se han desarrollado diferentes secuencias para usar distintas librerías de datos nucleares almacenadas en diferentes formatos: ENDF-6 (para las librerías evaluadas), COVERX (para las librerías en multigrupos de SCALE) y EAF (para las librerías de activación). Gracias a la revisión del estado del arte de los datos nucleares de los rendimientos de fisión se ha identificado la falta de una información sobre sus incertidumbres, en concreto, de matrices de covarianza completas. Además, visto el renovado interés por parte de la comunidad internacional, a través del grupo de trabajo internacional de cooperación para evaluación de datos nucleares (WPEC) dedicado a la evaluación de las necesidades de mejora de datos nucleares mediante el subgrupo 37 (SG37), se ha llevado a cabo una revisión de las metodologías para generar datos de covarianza. Se ha seleccionando la actualización Bayesiana/GLS para su implementación, y de esta forma, dar una respuesta a dicha falta de matrices completas para rendimientos de fisión. Una vez que el Método Híbrido ha sido implementado, desarrollado y extendido, junto con la capacidad de generar matrices de covarianza completas para los rendimientos de fisión, se han estudiado diferentes aplicaciones nucleares. Primero, se estudia el calor residual tras un pulso de fisión, debido a su importancia para cualquier evento después de la parada/disparo del reactor. Además, se trata de un ejercicio claro para ver la importancia de las incertidumbres de datos de decaimiento y de rendimientos de fisión junto con las nuevas matrices completas de covarianza. Se han estudiado dos ciclos de combustible de reactores avanzados: el de la instalación europea para transmutación industrial (EFIT) y el del reactor rápido de sodio europeo (ESFR), en los cuales se han analizado el impacto de las incertidumbres de los datos nucleares en la composición isotópica, calor residual y radiotoxicidad. Se han utilizado diferentes librerías de datos nucleares en los estudios antreriores, comparando de esta forma el impacto de sus incertidumbres. A su vez, mediante dichos estudios, se han comparando las distintas aproximaciones del Método Híbrido y otras metodologías para la porpagación de incertidumbres de datos nucleares: Total Monte Carlo (TMC), desarrollada en NRG por A.J. Koning y D. Rochman, y NUDUNA, desarrollada en AREVA GmbH por O. Buss y A. Hoefer. Estas comparaciones demostrarán las ventajas del Método Híbrido, además de revelar sus limitaciones y su rango de aplicación. ABSTRACT For an adequate assessment of safety margins of nuclear facilities, e.g. nuclear power plants, it is necessary to consider all possible uncertainties that affect their design, performance and possible accidents. Nuclear data are a source of uncertainty that are involved in neutronics, fuel depletion and activation calculations. These calculations can predict critical response functions during operation and in the event of accident, such as decay heat and neutron multiplication factor. Thus, the impact of nuclear data uncertainties on these response functions needs to be addressed for a proper evaluation of the safety margins. Methodologies for performing uncertainty propagation calculations need to be implemented in order to analyse the impact of nuclear data uncertainties. Nevertheless, it is necessary to understand the current status of nuclear data and their uncertainties, in order to be able to handle this type of data. Great eórts are underway to enhance the European capability to analyse/process/produce covariance data, especially for isotopes which are of importance for advanced reactors. At the same time, new methodologies/codes are being developed and implemented for using and evaluating the impact of uncertainty data. These were the objectives of the European ANDES (Accurate Nuclear Data for nuclear Energy Sustainability) project, which provided a framework for the development of this PhD Thesis. Accordingly, first a review of the state-of-the-art of nuclear data and their uncertainties is conducted, focusing on the three kinds of data: decay, fission yields and cross sections. A review of the current methodologies for propagating nuclear data uncertainties is also performed. The Nuclear Engineering Department of UPM has proposed a methodology for propagating uncertainties in depletion calculations, the Hybrid Method, which has been taken as the starting point of this thesis. This methodology has been implemented, developed and extended, and its advantages, drawbacks and limitations have been analysed. It is used in conjunction with the ACAB depletion code, and is based on Monte Carlo sampling of variables with uncertainties. Different approaches are presented depending on cross section energy-structure: one-group, one-group with correlated sampling and multi-group. Differences and applicability criteria are presented. Sequences have been developed for using different nuclear data libraries in different storing-formats: ENDF-6 (for evaluated libraries) and COVERX (for multi-group libraries of SCALE), as well as EAF format (for activation libraries). A revision of the state-of-the-art of fission yield data shows inconsistencies in uncertainty data, specifically with regard to complete covariance matrices. Furthermore, the international community has expressed a renewed interest in the issue through the Working Party on International Nuclear Data Evaluation Co-operation (WPEC) with the Subgroup (SG37), which is dedicated to assessing the need to have complete nuclear data. This gives rise to this review of the state-of-the-art of methodologies for generating covariance data for fission yields. Bayesian/generalised least square (GLS) updating sequence has been selected and implemented to answer to this need. Once the Hybrid Method has been implemented, developed and extended, along with fission yield covariance generation capability, different applications are studied. The Fission Pulse Decay Heat problem is tackled first because of its importance during events after shutdown and because it is a clean exercise for showing the impact and importance of decay and fission yield data uncertainties in conjunction with the new covariance data. Two fuel cycles of advanced reactors are studied: the European Facility for Industrial Transmutation (EFIT) and the European Sodium Fast Reactor (ESFR), and response function uncertainties such as isotopic composition, decay heat and radiotoxicity are addressed. Different nuclear data libraries are used and compared. These applications serve as frameworks for comparing the different approaches of the Hybrid Method, and also for comparing with other methodologies: Total Monte Carlo (TMC), developed at NRG by A.J. Koning and D. Rochman, and NUDUNA, developed at AREVA GmbH by O. Buss and A. Hoefer. These comparisons reveal the advantages, limitations and the range of application of the Hybrid Method.
Resumo:
El análisis del rendimiento en deportes juega un papel esencial en el fútbol profesional. Aunque el estudio del análisis del juego en fútbol se ha utilizado desde diferentes ámbitos y situaciones, todavía existen diferentes aspectos y componentes del juego que siguen sin estar estudiados. En este sentido existen diferentes aspectos que deben de superar los estudios previos centrados en el componente descriptivo tales como el uso de variables/ indicadores de rendimiento que no se han definido ni estudiado, la validez de los métodos observaciones que no han sido testados con los softwares específicos en fútbol, la aplicación y utilidad de los resultados, así como las limitaciones del estudio de las variables situacionales/contextuales. Con el objetivo de cubrir las citadas limitaciones se han diseñado 6 estudios independientes e inter-relacionados que tratan de estudiar los aspectos anteriormente referidos. El primer estudio evalua la fiabilidad inter-observadores de las estadísticas de juego de la empresa privada OPTA Sportsdata, estos datos son la muestra de estudio de la presente tesis doctoral. Dos grupos de observadores experimentados se requieren para analizar un partido de la liga española de manera independiente. Los resultados muestran que los eventos de equipos y porteros codificados por los inter-operadores alcanzan un acuerdo muy bueno (valores kappa entre 0.86 y 0.94). La validez inter-observadores de las acciones de juego y los datos de jugadores individuales se evaluó con elevados niveles de acuerdo (valores del coeficiente de correlación intraclase entre 0.88 hasta 1.00, el error típico estandarizado variaba entre 0.00 hasta 0.37). Los resultados sugieren que las estadísticas de juego registradas por los operadores de la empresa OPTA Sportsdata están bien entrenados y son fiables. El segundo, tercer y cuarto estudio se centran en resaltar la aplicabilidad del análisis de rendimiento en el fútbol así como para explicar en profundidad las influencias de las variables situacionales. Utilizando la técnica de los perfiles de rendimiento de jugadores y equipos de fútbol se puede evaluar y comparar de manera gráfica, fácil y visual. Así mismo, mediante esta técnica se puede controlar el efecto de las variables situacionales (localización del partido, nivel del equipo y del oponente, y el resultado final del partido). Los perfiles de rendimiento de porteros (n = 46 porteros, 744 observaciones) y jugadores de campo (n = 409 jugadores, 5288 observaciones) de la primera division professional de fútbol Española (La Liga, temporada 2012-13), los equipos (n = 496 partidos, 992 observaciones) de la UEFA Champions League (temporadas 2009-10 a 2012-13) fueron analizados registrando la media, desviación típica, mediana, cuartiles superior e inferior y el recuento de valores de cada indicador de rendimiento y evento, los cuales se presentaron en su forma tipificada y normalizada. Los valores medios de los porteros de los equipos de diferentes niveles de La Liga y de los equipos de diferente nivel de la UEFA Champions League cuando jugaban en diferentes contextos de juego y situaciones (variables situacionales) fueron comparados utilizando el ANOVA de un factor y la prueba t para muestras independientes (localización del partido, diferencias entre casa y fuera), y fueron establecidos en los perfiles de red después de unificar todos los registros en la misma escala derivada con valores estandarizados. Mientras que las diferencias de rendimiento entre los jugadores de los mejores equipos (Top3) y los peores (Bottom3) fueron comparados mediante el uso de diferencias en la magnitud del tamaño del efecto. El quinto y el sexto estudio analizaban el rendimiento del fútbol desde un punto de vista de predicción del rendimiento. El modelo linear general y el modelo lineal general mixto fue empleado para analizar la magnitud de las relaciones de los indicadores y estadísticas de juego con el resultado final del partido en función del tipo de partido (partidos ajustados o todos los partidos) en la fase de grupos de la Copa del Mundo 2014 de Brasil (n = 48 partidos, 38 partidos ajustados) y La Liga 2012-13 (n = 320 partidos ajustados). Las relaciones fueron evaluadas mediante las inferencias en la magnitud de las diferencias y se expresaron como partidos extra ganados o perdidos por cada 10 partidos mediante la variable calculada en 2 desviaciones típicas. Los resultados mostraron que, para los 48 partidos de la fase de grupos de la Copa del Mundo 2014, nueve variables tuvieron un efecto positive en la probabilidad de ganar (tiros, tiros a puerta, tiros de contraataque, tiros dentro del área, posesión de balón, pases en corto, media de secuencia de pases, duelos aéreos y entradas), cuatro tuvieron efectos negativos (tiros bloqueados, centros, regates y tarjetas amarillas), y otras 12 variables tenían efectos triviales o poco claros. Mientras que los 38 partidos ajustados, el efecto de duelos aéreos y tarjetas amarillas fueron triviales y claramente negativos respectivamente. En la La Liga, existió un efecto moderado positive para cada equipo para los tiros a puerta (3.4 victorias extras por cada 10 partidos; 99% IC ±1.0), y un efecto positivo reducido para tiros totales (1.7 victorias extrsa; ±1.0). Los efectos de la mayoría de los eventos se han relacionado con la posesión del balón, la cual obtuvo efectos negativos entre equipos (1.2 derrotas extras; ±1.0) pero un efecto positivo pequeño entra equipos (1.7 victorias extras; ±1.4). La localización del partido mostró un efecto positive reducido dentro de los equipos (1.9 victorias extras; ±0.9). Los resultados obtenidos en los perfiles y el modelado del rendimiento permiten ofrecer una información detallada y avanzada para el entrenamiento, la preparación previa a los partidos, el control de la competición y el análisis post-partido, así como la evaluación e identificación del talento de los jugadores. ABSTRACT Match performance analysis plays an important role in the modern professional football. Although the research in football match analysis is well-developed, there are still some issues and problems remaining in this field, which mainly include the lack of operational definitions of variables, reliability issues, applicability of the findings, the lack of contextual/situational variables, and focusing too much on descriptive and comparative analysis. In order to address these issues, six independent but related studies were conducted in the current thesis. The first study evaluated the inter-operator reliability of football match statistics from OPTA Sportsdata Company which is the data resourse of the thesis. Two groups of experienced operators were required to analyse a Spanish league match independently in the experiment. Results showed that team events and goalkeeper actions coded by independent operators reached a very good agreement (kappa values between 0.86 and 0.94). The inter-operator reliability of match actions and events of individual outfield players was also tested to be at a high level (intra-class correlation coefficients ranged from 0.88 to 1.00, standardised typical error varied from 0.00 to 0.37). These results suggest that the football match statistics collected by well-trained operators from OPTA Sportsdata Company are reliable. The second, third and fourth study aims to enhance the applicability of football match performance analysis and to explore deeply the influences of situational variables. By using a profiling technique, technical and tactical performances of football players and teams can be interpreted, evaluated and compared more easily and straightforwardly, meanwhile, influences and effects from situational variables (match location, strength of team and opposition, and match outcome) on the performances can be properly incorporated. Performance profiles of goalkeepers (n = 46 goalkeepers, 744 full match observations) and outfield players (n = 409 players, 5288 full match observations) from the Spanish First Division Professional Football League (La Liga, season 2012-13), teams (n = 496 matches, 992 observations) from UEFA Champions League (seasons 2009-10 to 2012-13) were set up by presenting the mean, standard deviation, median, lower and upper quartiles of the count values of each performance-related match action and event to represent their typical performances and spreads. Means of goalkeeper from different levels of team in La Liga and teams of different strength in UEFA Champions League when playing under different situational conditions were compared by using one-way ANOVA and independent sample t test (for match location, home and away differences), and were plotted into the same radar charts after unifying all the event counts by standardised score. While differences between the performances of outfield players from Top3 and from Bottom3 teams were compared by magnitude-based inferences. The fifth and sixth study aims to move from the descriptive and comparative football match analysis to a more predictive one. Generalised linear modelling and generalised mixed linear modelling were undertaken to quantify relationships of the performance-related match events, actions and variables with the match outcome in different types of games (close games and all games) in the group stage of 2014 Brazil FIFA World Cup (n = 48 games, 38 close games) and La Liga 2012-13 (n = 320 close games). Relationships were evaluated with magnitude-based inferences and were expressed as extra matches won or lost per 10 matches for an increase of two standard deviations of a variable. Results showed that, for all the 48 games in the group stage of 2014 FIFA World Cup, nine variables had clearly positive effects on the probability of winning (shot, shot on target, shot from counter attack, shot from inside area, ball possession, short pass, average pass streak, aerial advantage, and tackle), four had clearly negative effects (shot blocked, cross, dribble and red card), other 12 variabless had either trivial or unclear effects. While for the 38 close games, the effects of aerial advantage and yellow card turned to trivial and clearly negative, respectively. In the La Liga, there was a moderate positive within-team effect from shots on target (3.4 extra wins per 10 matches; 99% confidence limits ±1.0), and a small positive within-team effect from total shots (1.7 extra wins; ±1.0). Effects of most other match events were related to ball possession, which had a small negative within-team effect (1.2 extra losses; ±1.0) but a small positive between-team effect (1.7 extra wins; ±1.4). Game location showed a small positive within-team effect (1.9 extra wins; ±0.9). Results from the established performance profiles and modelling can provide detailed and straightforward information for training, pre-match preparations, in-match tactical approaches and post-match evaluations, as well as for player identification and development. 摘要 比赛表现分析在现代足球中起着举足轻重的作用。尽管如今对足球比赛表现分析的研究已经相对完善,但仍有很多不足之处。这些不足主要体现在:研究中缺乏对研究变量的清晰定义、数据信效度缺失、研究结果的实用性受限、比赛情境因素缺失以及过于集中在描述性和对比性分析等。针对这些问题,本论文通过六个独立而又相互联系的研究,进一步对足球比赛表现分析进行完善。 第一个研究对本论文的数据源--OPTA Sportsdata公司的足球比赛数据的信效度进行了实验检验。实验中,两组数据收集人员被要求对同一场西班牙足球甲级联赛的比赛进行分析。研究结果显示,两组收集人员记录下的球队比赛事件和守门员比赛行为具有高度的一致性(卡帕系数介于0.86和0.94)。收集人员输出的外场球员的比赛行为和比赛事件也具有很高的组间一致性(ICC相关系数介于0.88和1.00,标准化典型误差介于0.00和0.37)。实验结果证明了OPTA Sportsdata公司收集的足球比赛数据具有足够高的信效度。 第二、三、四个研究旨在提升足球比赛表现分析研究结果的实用性以及深度探讨比赛情境因素对足球比赛表现的影响。通过对足球运动员和运动队的比赛技战术表现进行档案创建,可以对运动员和运动队的比赛表现进行简直接而直观的呈现、评价和对比,同时,情境变量(比赛场地、球队和对手实力、比赛结果)对比赛表现的影响也可以被整合到表现档案中。本部分对2012-13赛季西班牙足球甲级联赛的参赛守门员(n = 46球员人次,744比赛场次)和外场球员(n = 409球员人次, 5288比赛场次)以及2009-10至2012-13赛季欧洲足球冠军联赛的参赛球队(n = 496比赛场次)的比赛技战术表现进行了档案创建。在表现档案中,各项比赛技战术指标的均值、标准差、中位数和大小四分位数被用来展现守门员、外场球员和球队的普遍表现和表现浮动性。方差分析(ANOVA)被用来对西甲不同水平球队的守门员、欧冠中不同水平球队在不同比赛情境下的普遍表现(各项指标的均值)进行对比,独立样本t检验被用来对比主客场比赛普遍表现的差异。数据量级推断(magnitude-based inferences)的方法则被用来对西甲前三名和最后三名球队外场球员的普遍表现进行对比分析。所有来自不同水平球队的运动员和不同水平运动队的各项比赛指标皆被转换成了标准分数,从而能把他们在各种不同比赛情境下的普遍表现(各项比赛指标的均值)投到相同的雷达图中进行直观的对比。 第五和第六个研究目的在于进行预测性足球比赛表现分析,从而跨越之前固有的描述性和对比性分析。广义线性模型和广义混合线性模型被用来对2014年巴西世界杯小组赛(n = 48 比赛场次,38小分差场次)和2012-13赛季西甲联赛(n = 320小分差场次)的比赛中各表现相关比赛事件、行为和变量与比赛结果(胜、平、负)的关系进行建模。模型中的关系通过数据量级推断(magnitude-based inferences)的方法来界定,具体表现为某个变量增加两个标准差对比赛结果的影响(每10场比赛中额外取胜或失利的场数)。研究结果显示,在2014年巴西世界杯小组赛的所有48场比赛中,9个变量(射门、射正、反击中射门、禁区内射门、控球、短传、连续传球平均次数、高空球争抢成功率和抢断)与赢球概率有清晰的正相关关系,4个变量(射门被封堵、传中、过人和红牌)与赢球概率有清晰的负相关关系,其他12个被分析的变量与赢球概率的相关关系微小或不清晰。而在38场小分差比赛中,高空球争抢成功率由正相关变为微小关系,黄牌则由微小关系变为清晰的负相关。在西甲联赛中,每一支球队增加两个标准差的“射正球门”可以给每10场比赛带来3.4场额外胜利(99%置信区间±1.0场),而所有球队作为一个整体,每增加两个标准差的“射正球门”可以给每10场比赛带来1.7场额外胜利(99%置信区间±1.0场)。其他大多数比赛相关事件与比赛结果的相关关系与“控球”相关联。每一支球队增加两个标准差的“控球”将会给每10场比赛带来1.2场额外失利(99%置信区间±1.0场),而所有球队作为一个整体,每增加两个标准差的“控球”可以给每10场比赛带来1.7场额外胜利(99%置信区间±1.4场)。与客场比赛相对,主场能给球队带来1.9 /10场额外胜利(99%置信区间±0.9场)。 比赛表现档案和模型中得出的研究结果可以为俱乐部、足球队、教练组、表现分析师和运动员提供详细而直接的参考信息。这些信息可用于训练指导、赛前备战、赛中技战术调整和赛后技战术表现分析,也可运用于足球运动员选材、培养和发展。
Resumo:
The global economic structure, with its decentralized production and the consequent increase in freight traffic all over the world, creates considerable problems and challenges for the freight transport sector. This situation has led shipping to become the most suitable and cheapest way to transport goods. Thus, ports are configured as nodes with critical importance in the logistics supply chain as a link between two transport systems, sea and land. Increase in activity at seaports is producing three undesirable effects: increasing road congestion, lack of open space in port installations and a significant environmental impact on seaports. These adverse effects can be mitigated by moving part of the activity inland. Implementation of dry ports is a possible solution and would also provide an opportunity to strengthen intermodal solutions as part of an integrated and more sustainable transport chain, acting as a link between road and railway networks. In this sense, implementation of dry ports allows the separation of the links of the transport chain, thus facilitating the shortest possible routes for the lowest capacity and most polluting means of transport. Thus, the decision of where to locate a dry port demands a thorough analysis of the whole logistics supply chain, with the objective of transferring the largest volume of goods possible from road to more energy efficient means of transport, like rail or short-sea shipping, that are less harmful to the environment. However, the decision of where to locate a dry port must also ensure the sustainability of the site. Thus, the main goal of this article is to research the variables influencing the sustainability of dry port location and how this sustainability can be evaluated. With this objective, in this paper we present a methodology for assessing the sustainability of locations by the use of Multi-Criteria Decision Analysis (MCDA) and Bayesian Networks (BNs). MCDA is used as a way to establish a scoring, whilst BNs were chosen to eliminate arbitrariness in setting the weightings using a technique that allows us to prioritize each variable according to the relationships established in the set of variables. In order to determine the relationships between all the variables involved in the decision, giving us the importance of each factor and variable, we built a K2 BN algorithm. To obtain the scores of each variable, we used a complete cartography analysed by ArcGIS. Recognising that setting the most appropriate location to place a dry port is a geographical multidisciplinary problem, with significant economic, social and environmental implications, we consider 41 variables (grouped into 17 factors) which respond to this need. As a case of study, the sustainability of all of the 10 existing dry ports in Spain has been evaluated. In this set of logistics platforms, we found that the most important variables for achieving sustainability are those related to environmental protection, so the sustainability of the locations requires a great respect for the natural environment and the urban environment in which they are framed.
Resumo:
La presente investigación aborda el análisis de las cualidades del entorno urbano que inciden en la experiencia perceptiva de su recorrido. Los intereses que la motivan tienen que ver con la recuperación de la ciudad para el viandante y se sitúa dentro de un marco de intereses más amplio como es el de la ciudad sostenible. En primer lugar, se contextualiza el tema de la experiencia del recorrido a través de disciplinas diferentes como son el cine, la literatura, la arquitectura, el urbanismo, la arquitectura del paisaje, la psicología y la estética ambiental. De este modo, se refuerza el argumento en torno al papel que desempeñan la forma y otros estímulos sensoriales en la experiencia del espacio, ya sea éste urbano, arquitectónico o natural. La experiencia que las personas viven al desplazarse por un entorno viene definida por infinitos factores que van desde la configuración física del entorno recorrido, hasta el aprendizaje cultural de la persona que camina; desde los estímulos sensoriales que se reciben, hasta las condiciones físicas que posibilitan la funcionalidad del recorrido; desde las sutilezas de índole secuencial que se desvelan a medida que avanzamos, hasta las condiciones higrotérmicas en el momento en que se lleva a cabo el desplazamiento. Sin embargo, factores como las motivaciones y atenciones personales del que anda, así como sus recuerdos de experiencias anteriores, deseos, imaginación y estado de ánimo, también definen nuestra experiencia al caminar que, indudablemente, se encuentra ligada a nuestra personalidad y, por lo tanto, no puede ser exacta a la de otra persona. Conscientes de que todos estos factores se funden en un fenómeno global, y que no se dan en la vida por separado, nos hemos esforzado por independizar las cualidades sensoriales y formales, únicamente a efectos de investigación. La importancia del tema a investigar reside en las conexiones que existen entre las cualidades del medio construido y la experiencia cotidiana de su recorrido, en la medida en que dicha vinculación establece relaciones con la calidad de vida, la salud de los ciudadanos y sus sentimientos de identidad. En este sentido, el problema clave encontrado es que si la definición de la experiencia del recorrido urbano es demasiado amplia, se vuelve nebulosa e inoperativa, pero si se precisa demasiado, se pueden excluir variables importantes. Hemos concentrado nuestros esfuerzos en buscar un medio que permita hacer operativo el estudio de aspectos que, por su propia naturaleza, son difíciles de controlar y aplicar en la práctica. En este sentido, el campo de estudio que aúna interés en el tema, producción científica y vocación práctica, es el de la caminabilidad de las ciudades. Por esta razón, hemos situado nuestro trabajo dentro del marco de criterios que este campo establece para valorar un entorno caminable. Sin embargo, hemos detectado que los trabajos sobre caminabilidad tienden a desarrollarse dentro de la disciplina del planeamiento de transporte y, con frecuencia, siguiendo las mismas pautas que la investigación sobre el transporte motorizado. Además, la tendencia a proporcionar datos numéricos producto de la medición controlada de variables, para respaldar iniciativas dentro de los ámbitos de toma de decisiones, lleva consigo un progresivo alejamiento de los aspectos más sutiles y próximos de la experiencia de las personas y la especificidad de los lugares. Por estas razones, hemos estimado necesario profundizar en la experiencia peatonal explorando otras líneas de trabajo como son la habitabilidad de las ciudades, políticas llevadas a cabo para mejorar la calidad del espacio público o las certificaciones de sostenibilidad en el ámbito del urbanismo. También se han estudiado las aproximaciones gráficas a la representación y simulación de recorridos urbanos por su importancia en la comunicación y promoción de caminar. Para detectar las problemáticas implicadas en la definición de una herramienta que permita valorar, de manera operativa, la calidad de la experiencia del recorrido urbano, la metodología de valoración propuesta se ha basado en la combinación de distintos métodos y en la conjugación de tres aproximaciones: valoración por parte del investigador en calidad de experto, valoración realizada por el investigador en el papel de usuario y valoración por el ciudadano. El desarrollo de esta investigación se ha visto condicionado por la intención de aplicar lo estudiado en un caso práctico. El marco de criterios que los trabajos sobre caminabilidad de las ciudades establecen para que se dé un entorno caminable, se puede resumir en que concurran los siguientes factores: mezcla de usos, densidad de población y edificación relativamente altas, destinos públicos accesibles a pie, un alto grado de seguridad con respecto al tráfico y actos delictivos, alta funcionalidad (dimensiones, pendientes, etc.) y atractivo. Una vez definido el modelo de ciudad en que es pertinente realizar una investigación sobre las cualidades del entorno urbano responsables de que éste se perciba como atractivo, se escogió la ciudad de Taipei, entre otras razones, por cumplir con los requisitos restantes. Con respecto al caso práctico, el objetivo es detectar fortalezas y debilidades del área central de la ciudad de Taipei en cuanto a la experiencia perceptiva que proporciona a los viandantes. ABSTRACT This research addresses the analysis of the qualities of the urban environment that affect the perceptual experience of walking. The interests lying behind it are related to the recovery of the city for pedestrians, and is framed within the sustainable city framework of interests. Firstly, the issue of the experience of walking is contextualized through different disciplines such as film, literature, architecture, urban planning, landscape architecture, environmental psychology and aesthetics. This way, the argument about the role that form and other sensory stimuli play in the experience of space, whether it is urban, architectural or natural, is strengthened. The walking experience of people is defined by factors ranging from the physical configuration of the environment to the cultural background of the person who walks, from the sensory stimuli that are perceived to the physical conditions that enable the functionality of the walk, from the subtleties of sequential nature that are revealed as we move around to hygrothermal conditions at the specific time of walking. Nevertheless, factors such as personal motivations and attentions of the walker, as well as memories of past experiences, desires, imagination and mood, also define our walking experience that is undoubtedly linked to our personality and, therefore, it cannot be exactly the same as other people's experience. Being aware that all these factors come together in a total phenomenon and that, in real life, they do not exist separately, we focused on separating sensory and formal qualities only for research purposes. The importance of the research topic lies in the connections between the qualities of the built environment and the quotidian experience of walking through it, to the extent that such link establishes relationships with the quality of life, health and feelings of identity of citizens. In this sense, the key problem encountered is that, if the definition of urban walking experience is too broad, it becomes nebulous and non‐operational, but if it is too precise, important variables can be excluded. We concentrated our efforts on finding a way to operationalize the study of questions that, by their very nature, are difficult to control and apply in practice. In this regard, the field of study that combines interest in the topic, scientific production and practical purpose, is the walkability of cities. For this reason, we placed our work within the framework of the set of criteria that this field establishes for assessing an environment as walkable. However, we found that works on walkability tend to be developed within the discipline of transportation planning and often following the same guidelines that research on motorized transport. Furthermore, the tendency to provide numerical data as a result of the controlled measurement of variables in order to support initiatives in decision making areas, involves a progressive distancing from the proximity and the most subtle aspects of the experience of people, and from the specificity of places. For these reasons, we considered it was necessary to deepen into the study of pedestrians experience, by exploring other lines of work such as the livability of cities, implemented policies to improve the quality of the public space or sustainability certifications in urbanism. Graphic representation and simulation of urban walks are also studied due to their important role in communication and promotion of walking. In order to find the issues involved in defining a tool to assess, in an operational way, the quality of urban walking experience, the proposed assessment methodology is based on the combination of different methods and the synthesis of three approaches: assessment by the researcher as expert, assessment by the researcher playing the role of user and assessment by the citizens. The development of this research has been conditioned by our intention of applying it in a case study. The framework of criteria that works on walkability of cities set to define a walkable environment, can be summarized in the following factors: mix of uses, population and building density rather high, public destinations accessible on foot, high levels of safety in terms of traffic and crime, high functionality (dimensions, slopes, etc.) and attractiveness. After defining the model of city in which it is relevant to conduct an investigation on the qualities that are responsible of the attractiveness of the environment; Taipei City was selected because it meets the remaining requirements, among other reasons. Regarding the case study, the goal is to identify strengths and weaknesses in the central area of Taipei City in terms of the perceptual experience of pedestrians.
Resumo:
Over the last few years, the Data Center market has increased exponentially and this tendency continues today. As a direct consequence of this trend, the industry is pushing the development and implementation of different new technologies that would improve the energy consumption efficiency of data centers. An adaptive dashboard would allow the user to monitor the most important parameters of a data center in real time. For that reason, monitoring companies work with IoT big data filtering tools and cloud computing systems to handle the amounts of data obtained from the sensors placed in a data center.Analyzing the market trends in this field we can affirm that the study of predictive algorithms has become an essential area for competitive IT companies. Complex algorithms are used to forecast risk situations based on historical data and warn the user in case of danger. Considering that several different users will interact with this dashboard from IT experts or maintenance staff to accounting managers, it is vital to personalize it automatically. Following that line of though, the dashboard should only show relevant metrics to the user in different formats like overlapped maps or representative graphs among others. These maps will show all the information needed in a visual and easy-to-evaluate way. To sum up, this dashboard will allow the user to visualize and control a wide range of variables. Monitoring essential factors such as average temperature, gradients or hotspots as well as energy and power consumption and savings by rack or building would allow the client to understand how his equipment is behaving, helping him to optimize the energy consumption and efficiency of the racks. It also would help him to prevent possible damages in the equipment with predictive high-tech algorithms.
Resumo:
The search for a common cause of species richness gradients has spawned more than 100 explanatory hypotheses in just the past two decades. Despite recent conceptual advances, further refinement of the most plausible models has been stifled by the difficulty of compiling high-resolution databases at continental scales. We used a database of the geographic ranges of 2,869 species of birds breeding in South America (nearly a third of the world's living avian species) to explore the influence of climate, quadrat area, ecosystem diversity, and topography on species richness gradients at 10 spatial scales (quadrat area, ≈12,300 to ≈1,225,000 km2). Topography, precipitation, topography × latitude, ecosystem diversity, and cloud cover emerged as the most important predictors of regional variability of species richness in regression models incorporating 16 independent variables, although ranking of variables depended on spatial scale. Direct measures of ambient energy such as mean and maximum temperature were of ancillary importance. Species richness values for 1° × 1° latitude-longitude quadrats in the Andes (peaking at 845 species) were ≈30–250% greater than those recorded at equivalent latitudes in the central Amazon basin. These findings reflect the extraordinary abundance of species associated with humid montane regions at equatorial latitudes and the importance of orography in avian speciation. In a broader context, our data reinforce the hypothesis that terrestrial species richness from the equator to the poles is ultimately governed by a synergism between climate and coarse-scale topographic heterogeneity.
Resumo:
When NMR hydrogen exchange was used previously to monitor the kinetics of RNase A unfolding, some peptide NH protons were found to show EX2 exchange (detected by base catalysis) in addition to the expected EX1 exchange, whose rate is limited by the kinetic unfolding process. In earlier work, two groups showed independently that a restricted two-process model successfully fits published hydrogen exchange rates of native RNase A in the range 0-0.7 M guanidinium chloride. We find that this model predicts properties that are very different from the observed properties of the EX2 exchange reactions of RNase A in conditions where guanidine-induced unfolding takes place. The model predicts that EX2 exchange should be too fast to measure by the technique used, whereas it is readily measurable. Possible explanations for the contradiction are considered here, and we show that removing the restriction from the earlier two-process model is sufficient to resolve the contradiction; instead of specifying that exchange caused by global unfolding occurs by the EX2 mechanism, we allow it to occur by the general mechanism, which includes both the EX1 and EX2 cases. It is logical to remove this restriction because global unfolding of RNase A is known to give rise to EX1 exchange in these unfolding conditions. Resolving the contradiction makes it possible to determine whether populated unfolding intermediates contribute to the EX2 exchange, and this question is considered elsewhere. The results and simulations indicate that moderate or high denaturant concentrations readily give rise to EX1 exchange in native proteins. Earlier studies showed that hydrogen exchange in native proteins typically occurs by the EX2 mechanism but that high temperatures or pH values above 7 may give rise to EX1 exchange. High denaturant concentrations should be added to the list of variables likely to cause EX1 exchange.
Resumo:
This paper deals with stability properties of the feasible set of linear inequality systems having a finite number of variables and an arbitrary number of constraints. Several types of perturbations preserving consistency are considered, affecting respectively, all of the data, the left-hand side data, or the right-hand side coefficients.
Resumo:
Introducción: el consumo de suplementos ergonutricionales está aumentando en los últimos años tanto en España como en el resto del mundo. Los cuestionarios que estiman la ingesta y uso de suplementos presentan limitaciones metodológicas. El objetivo de este estudio es describir las variables utilizadas en los cuestionarios de estimación de consumo de suplementos ergonutricionales para población deportista, descritas en el apartado de metodología de los artículos científicos. Metodología: revisión bibliográfica de variables utilizadas para estimar el uso y consumo de suplementos ergonutricionales. Búsqueda estructurada en la base de datos PubMed y mediante la estrategia bola de nieve. Ecuación de búsqueda: “Questionnaire” AND “Supplementation” AND “Athletes”. Criterios de inclusión: artículos publicados en cualquier país en idioma inglés o en español, que contengan cuestionario en el anexo o las variables que se describen en el apartado de metodología que puedan inferirse a partir de los artículos, que estimen la ingesta y uso de suplementos ergonutricionales y vayan dirigidos a deportistas que entrenan a nivel competitivo (amateurs, élite, etc.). Se realizó un análisis de contenido y un análisis descriptivo. Resultados: de los 122 estudios identificados, 21 cumplieron los criterios de inclusión. Las variables descritas fueron factores sociodemográficos y relacionados con el deporte que se practica, así como la frecuencia de entrenamiento, población deportista a la que iba dirigida, razones/motivos de uso y consumo, fuentes de información, lista de suplementos y frecuencia de uso y consumo. Discusión: existe gran heterogeneidad en cuanto a las variables propuestas por los autores en los cuestionarios de estimación de ingesta y uso de suplementos ergonutricionales. Es necesaria la estandarización de metodologías en el diseño de cuestionarios, así como definir una clasificación de referencia como la propuesta por el Instituto Australiano del Deporte.
Resumo:
This study evaluates the technical efficiency of the learning-teaching process in higher education using a three-stage procedure that offers advances in comparison to previous studies and improves the quality of the results. First, it utilizes a multiple stage Data Envelopment Analysis (DEA) with contextual variables. Second, the levels of super efficiency are calculated in order to prioritize the efficiency units. And finally, through sensitivity analysis, the contribution of each key performance indicator (KPI) is established with respect to the efficiency levels without omission of variables. The analytical data was collected from a survey completed by 633 tourism students during the 2011/12, 2012/13 and 2013/14 academic course years. The results suggest that level of satisfaction with the course, diversity of materials and satisfaction with the teacher were the most important factors affecting teaching performance. Furthermore, the effect of the contextual variables was found to be significant.
Resumo:
A logical reading on Aristotle’s Organon discovers some inconsistencies in the text which have to be solved by reducing them to metaphysical decisions of the author, if they are not just identified as deficiencies in the exposition that should be corrected. The present article tries to display a line of reading paying attention to those so-called inconsistencies, in an attempt to understand them as specific steps in Aristotle’s research. In order to this goal it focuses on the exposition procedure of the Aristotelian figures: the use of variables, whose introduction by Aristotle has been celebrated all over logical tradition. An analysis of the distinctive and internal features in this procedure will allow us to link Aristotle’s logos research and the “being qua being” investigation, and to determine also –though in a negative way– the connection between this reading and the logical-traditional one on Aristotle’s Organon.
Resumo:
In order to increase the use of information and communication technologies (ICT) in the European Union Member States, the European Commission, on the initiative of Commissioner E. Liikannen, launched in December 1999 a bold programme called “eEurope”. Soon after its creation, the eEurope programme was integrated into the so- called Lisbon agenda for Europe to become the “most advanced knowledge based economy” in the world. We try to assess if the programme is successful in achieving its stated objective of promoting a knowledge based economy through the development of an “information society for all”. First, we conclude that eEurope, due to its origins and its procedures, has intrinsic limits both as regards its scope and effectiveness. Second, we show how Member States have adopted different trajectories towards the “knowledge based society”. To identify these heterogeneous paths of growth, we have selected a set of variables that, combined together, represent the institutional arrangements specific to a country or a group of countries. We found sharp differences between two advanced models that we label, respectively, as Scandinavian and Anglo-Saxon. Without asserting the superiority of a model, we propose policy orientations to help Europe overcome those gaps hindering the move towards knowledge economies where information society technologies are widely diffused.
Resumo:
This study gives an overview of the theoretical foundations, empirical procedures and derived results of the literature identifying determinants of land prices. Special attention is given to the effects of different government support policies on land prices. Since almost all empirical studies on the determination of land prices refer either to the net present value method or the hedonic pricing approach as a theoretical basis, a short review of these models is provided. While the two approaches have different theoretical bases, their empirical implementation converges. Empirical studies use a broad range of variables to explain land values and we systematise those into six categories. In order to investigate the influence of different measures of government support on land prices, a meta-regression analysis is carried out. Our results reveal a significantly higher rate of capitalisation for decoupled direct payments and a significantly lower rate of capitalisation for agri-environmental payments, as compared to the rest of government support. Furthermore, the results show that taking theoretically consistent land rents (returns to land) and including non-agricultural variables like urban pressure in the regression implies lower elasticities of capitalisation. In addition, we find a significant influence of the land type, the data type and estimation techniques on the capitalisation rate.
Resumo:
This paper empirically analyses a dataset of more than 7,300 agricultural land sales transactions from 2001 and 2007 to identify the factors influencing agricultural land prices in Bavaria. We use a general spatial model, which combines a spatial lag and a spatial error model, and in addition account for endogeneity introduced by the spatially lagged dependent variable as well as other explanatory variables. Our findings confirm the strong influence of agricultural factors such as land productivity, of variables describing the regional land market structure, and of non-agricultural factors such as urban pressure on agricultural land prices. Moreover, the involvement of public authorities as a seller or buyer increases sales prices in Bavaria. We find a significant capitalisation of government support payments into agricultural land, where a decrease of direct payments by 1% would decrease land prices in 2007 and 2001 by 0.27% and 0.06%, respectively. In addition, we confirm strong spatial relationships in our dataset. Neglecting this leads to biased estimates, especially if aggregated data is used. We find that the price of a specific plot increases by 0.24% when sales prices in surrounding areas increase by 1%.
Resumo:
Senior thesis written for Oceanography 445