21 resultados para General Linear Methods
em Universidad Politécnica de Madrid
Resumo:
- Resumen La hipótesis que anima esta tesis doctoral es que algunas de las características del entorno urbano, en particular las que describen la accesibilidad de su red de espacio público, podrían estar relacionadas con la proporción de viajes a pie o reparto modal, que tiene cada zona o barrio de Madrid. Uno de los puntos de partida de dicha hipótesis que el entorno urbano tiene una mayor influencia sobre los viaje a pie que en sobre otros modos de transporte, por ejemplo que en los viajes de bicicleta o en transporte público; y es que parece razonable suponer que estos últimos van a estar más condicionadas por ejemplo por la disponibilidad de vías ciclistas, en el primer caso, o por la existencia de un servicio fiable y de calidad, en el segundo. Otra de las motivaciones del trabajo es que la investigación en este campo de la accesibilidad del espacio público, en concreto la denominada “Space Syntax”, ha probado en repetidas ocasiones la influencia de la red de espacio público en cómo se distribuye la intensidad del tráfico peatonal por la trama urbana, pero no se han encontrado referencias de la influencia de dicho elemento sobre el reparto modal. De acuerdo con la hipótesis y con otros trabajos anteriores se propone una metodología basada en el análisis empírico y cuantitativo. Su objetivo es comprobar si la red de espacio público, independientemente de otras variables como los usos del suelo, incluso de las variables de ajenas entorno no construido, como las socioeconómicas, está o no relacionada estadísticamente con la proporción de peatones viajes en las zonas urbanas. Las técnicas estadísticas se utilizan para comprobar sistemáticamente la asociación de las variables del entorno urbano, denominadas variables independientes, con el porcentaje de viajes a pie, la variable dependiente. En términos generales, la metodología es similar a la usada en otros trabajos en este campo como los de CERVERÓ y KOCKLEMAN (1997), CERVERÓ y DUNCAN (2003), o para los que se utilizan principalmente en la revisión general de TRB (2005) o, más recientemente, en ZEGRAS (2006) o CHATMAN (2009). Otras opciones metodológicas, como los métodos de preferencias declaradas (ver LOUVIERE, HENSHER y SWAIT, 2000) o el análisis basado en agentes (PENN & TURNER, 2004) fueron descartados, debido a una serie de razones, demasiado extensas para ser descritas aquí. El caso de estudio utilizado es la zona metropolitana de Madrid, abarcándola hasta la M-50, es decir en su mayor parte, con un tamaño aproximado de 31x34 Km y una población de 4.132.820 habitantes (aproximadamente el 80% de la población de la región). Las principales fuentes de datos son la Encuesta Domiciliaria de Movilidad de 2004 (EDM04), del Consorcio Regional de Transportes de Madrid que es la última disponible (muestra: > 35.000 familias,> 95.000 personas), y un modelo espacial del área metropolitana, integrando el modelo para calcular los índices de Space Syntax y un Sistema de Información Geográfica (SIG). La unidad de análisis, en este caso las unidades espaciales, son las zonas de transporte (con una población media de 7.063 personas) y los barrios (con una población media de 26.466 personas). Las variables del entorno urbano son claramente el centro del estudio. Un total de 20 índices (de 21) se seleccionan de entre los más relevantes encontrados en la revisión de la producción científica en este campo siendo que, al mismo tiempo, fueran accesibles. Nueve de ellos se utilizan para describir las características de los usos del suelo, mientras que otros once se usan para describir la red de espacios públicos. Estos últimos incluyen las variables de accesibilidad configuracional, que son, como se desprende de su título, el centro del estudio propuesto. La accesibilidad configuracional es un tipo especial de accesibilidad que se basa en la configuración de la trama urbana, según esta fue definida por HILLIER (1996), el autor de referencia dentro de esta línea de investigación de Space Syntax. Además se incluyen otras variables de la red de espacio público más habituales en los estudios de movilidad, y que aquí se denominan características geométricas de los elementos de la red, tales como su longitud, tipo de intersección, conectividad, etc. Por último se incluye además una variable socioeconómica, es decir ajena al entorno urbano, para evaluar la influencia de los factores externos, pues son varios los que pueden tener un impacto en la decisión de caminar (edad, género, nivel de estudios, ingresos, tasa de motorización, etc.). La asociación entre las variables se han establecido usando análisis de correlación (bivariante) y modelos de análisis multivariante. Las primeras se calculan entre por pares entre cada una de las 21 variables independientes y la dependiente, el porcentaje de viajes a pie. En cuanto a los segundos, se han realizado tres tipos de estudios: modelo multivariante general lineal, modelo multivariante general curvilíneo y análisis discriminante. Todos ellos son capaces de generar modelos de asociación entre diversas variables, pudiéndose de esta manera evaluar con bastante precisión en qué medida cada modelo reproduce el comportamiento de la variable dependiente, y además, el peso o influencia de cada variable en el modelo respecto a las otras. Los resultados fundamentales del estudio se expresan en dos modelos finales alternativos, que demuestran tener una significativa asociación con el porcentaje de viajes a pie (R2 = 0,6789, p <0,0001), al explicar las dos terceras partes de su variabilidad. En ellos, y en general en todo el estudio realizado, se da una influencia constante de tres índices en particular, que quedan como los principales. Dos de ellos, de acuerdo con muchos de los estudios previos, corresponden a la densidad y la mezcla de usos del suelo. Pero lo más novedoso de los resultados obtenidos es que el tercero es una medida de la accesibilidad de la red de espacio público, algo de lo que no había referencias hasta ahora. Pero, ¿cuál es la definición precisa y el peso relativo de cada uno en el modelo, es decir, en la variable independiente? El de mayor peso en la mayor parte de los análisis realizados es el índice de densidad total (n º residentes + n º puestos de trabajo + n º alumnos / Ha). Es decir, una densidad no sólo de población, sino que incluye algunas de las actividades más importantes que pueden darse una zona para generar movilidad a pie. El segundo que mayor peso adquiere, llegando a ser el primero en alguno de los análisis estadísticos efecturados, es el índice de accesibuilidad configuracional denominado integración de radio 5. Se trata de una medida de la accesibilidad de la zona, de su centralidad, a la escala de, más un menor, un distrito o comarca. En cuanto al tercero, obtiene una importancia bastante menor que los anteriores, y es que representa la mezcla de usos. En concreto es una medida del equilibrio entre los comercios especializados de venta al por menor y el número de residentes (n º de tiendas especializadas en alimentación, bebidas y tabaco / n º de habitantes). Por lo tanto, estos resultados confirman buena parte de los de estudios anteriores, especialmente los relativas a los usos del suelo, pero al mismo tiempo, apuntan a que la red de espacio público podría tener una influir mayor de la comprobada hasta ahora en la proporción de peatones sobre el resto de modos de transportes. Las razones de por qué esto puede ser así, se discuten ampliamente en las conclusiones. Finalmente se puede precisar que dicha conclusión principal se refiere a viajes de una sola etapa (no multimodales) que se dan en los barrios y zonas del área metropolitana de Madrid. Por supuesto, esta conclusión tiene en la actualidad, una validez limitada, ya que es el resultado de un solo caso — Abstract The research hypothesis for this Ph.D. Thesis is that some characteristics of the built environment, particularly those describing the accessibility of the public space network, could be associated with the proportion of pedestrians in all trips (modal split), found in the different parts of a city. The underlying idea is that walking trips are more sensitive to built environment than those by other transport modes, such as for example those by bicycle or by public transport, which could be more conditioned by, e.g. infrastructure availability or service frequency and quality. On the other hand, it has to be noted that the previously research on this field, in particular within Space Syntax’s where this study can be referred, have tested similar hypothesis using pedestrian volumes as the dependent variable, but never against modal split. According to such hypothesis, research methodology is based primarily on empirical quantitative analysis, and it is meant to be able to assess whether public space network, no matter other built environment and non-built environment variables, could have a relationship with the proportion of pedestrian trips in urban areas. Statistical techniques are used to check the association of independent variables with the percentage of walking in all trips, the dependent one. Broadly speaking this methodology is similar to that of previous studies in the field such as CERVERO&KOCKLEMAN (1997), CERVERO & DUNCAN (2003), or to those used mainly in the general review of T.R.B. (2005) or, more recently in ZEGRAS (2006) or CHATMAN (2009). Other methodological options such as stated choice methods (see LOUVIERE, HENSHER & SWAIT, 2000) or agent based analysis (PENN & TURNER, 2004), were discarded, due to a number of reasons, too long to be described here. The case study is not the entire Madrid’s metropolitan area, but almost (4.132.820 inhabitants, about 80% of region´s population). Main data sources are the Regional Mobility Home Based Survey 2004 (EDM04), which is the last available (sample: >35.000 families, > 95.000 individuals), and a spatial model of the metropolitan area, developed using Space Syntax and G.I.S. techniques. The analysis unit, in this case spatial units, are both transport zones (mean population = 7.063) and neighborhoods (mean population = 26.466). The variables of the built environment are clearly the core of the study. A total of 20 (out of 21) are selected from among those found in the literature while, at the same time, being accessible. Nine out of them are used to describe land use characteristics while another eleven describe the network of public spaces. Latter ones include configurational accessibility or Space Syntax variables. This is a particular sort of accessibility related with the concept of configuration, by HILLIER (1996), one of the main authors of Space Syntax, But it also include more customary variables used in mobility research to describe the urban design or spatial structure (here public space network), which here are called geometric characteristics of the such as its length, type of intersection, conectivity, density, etc. Finally a single socioeconomic variable was included in order to assess the influence non built environment factors that also may have an impact on walking (age, income, motorization rate, etc.). The association among variables is worked out using bi-variate correlation analysis and multivariate-analysis. Correlations are calculated among the 21 independent variables and the dependent one, the percentage of walking trips. Then, three types of multi-variate studies are run: general linear, curvilinear and discriminant multi-variate analysis. The latter are fully capable of generating complex association models among several variables, assessing quite precisely to what extent each model reproduces the behavior of the dependent variable, and also the weight or influence of each variable in the model. This study’s results show a consistent influence of three particular indexes in the two final alternative models of the multi-variate study (best, R2=0,6789, p<0,0000). Not surprisingly, two of them correspond to density and mix of land uses. But perhaps more interesting is that the third one is a measure of the accessibility of the public space network, a variable less important in the literature up to now. Additional precisions about them and their relative weight could also be of some interest. The density index is not only about population but includes most important activities in an area (nº residents + nº jobs+ nº students/Ha). The configurational index (radius 5 integration) is a measure of the accessibility of the area, i.e. centrality, at the scale of, more a less, a district. Regarding the mix of land uses index, this one is a measure of the balance between retail, in fact local basic retail, and the number of residents (nº of convenience shops / nº of residents). Referring to their weights, configurational index (radius 5 integration) gets the higher standardized coefficient of the final equation. However, in the final equations, there are a higher number of indexes coming from the density or land use mix categories than from public space network enter. Therefore, these findings seem to support part of the field’s knowledge, especially those concerning land uses, but at the same time they seem to bring in the idea that the configuration of the urban grid could have an influence in the proportion of walkers (as a part of total trips on any transport mode) that do single journey trips in the neighborhoods of Madrid, Spain. Of course this conclusion has, at present, a limited validity since it’s the result of a single case. The reasons of why this can be so, are discussed in the last part of the thesis.
Resumo:
In this paper a dynamic analysis of transnational shells is presented. The general linear shell theory is used in conjunction with additional shallow and curved plate approximations. In order to apply some type of extended Levy solution, the shell is assumed to be limited by a rectangular plan form, with two opposite edges simply supported (gable boundary conditions). First, the shells free vibrations are studied in the usual way, obtaining for each Fourier term the natural frequencies as solutions of a transcendental equation. However, solving these equations arises enormous computational difficulties. This paper deals specifically with this problem, trying to reduce its dimension by a discretization procedure. In the shell dynamic characteristics, namely the mass. The shell mass is lumped along a family of coordinate lines. Therefore, the natural frequencies for each harmonic term can be found from the solution of a typical matrix eigenvalues problem and standard numerical techniques can be applied. The shell response to forced vibrations, particularly to earthquake excitation, can be determined by using conventional procedure either in the time or in the frequency domain. Finally, extending the above procedure, any system of translational shells under dynamic loading can be studied. Then, by using matrix methods, a general computer program is written and applied to some illustrative examples. Numerical results has been obtained in two cases: circular cylindrical shell and box girder bridge.
Resumo:
The analysis of the interdependence between time series has become an important field of research in the last years, mainly as a result of advances in the characterization of dynamical systems from the signals they produce, the introduction of concepts such as generalized and phase synchronization and the application of information theory to time series analysis. In neurophysiology, different analytical tools stemming from these concepts have added to the ‘traditional’ set of linear methods, which includes the cross-correlation and the coherency function in the time and frequency domain, respectively, or more elaborated tools such as Granger Causality. This increase in the number of approaches to tackle the existence of functional (FC) or effective connectivity (EC) between two (or among many) neural networks, along with the mathematical complexity of the corresponding time series analysis tools, makes it desirable to arrange them into a unified-easy-to-use software package. The goal is to allow neuroscientists, neurophysiologists and researchers from related fields to easily access and make use of these analysis methods from a single integrated toolbox. Here we present HERMES (http://hermes.ctb.upm.es), a toolbox for the Matlab® environment (The Mathworks, Inc), which is designed to study functional and effective brain connectivity from neurophysiological data such as multivariate EEG and/or MEG records. It includes also visualization tools and statistical methods to address the problem of multiple comparisons. We believe that this toolbox will be very helpful to all the researchers working in the emerging field of brain connectivity analysis.
Resumo:
Existe un creciente interés internacional por el ahorro energético y la sostenibilidad en la edificación con importantes repercusiones en la Arquitectura. La inercia térmica es un parámetro fundamental para poder valorar energéticamente un edificio en condiciones reales. Para ello es necesario cambiar el enfoque tradicional de transmisión de calor en régimen estacionario por otro en régimen dinámico en el que se analizan las ondas térmicas y el flujo de calor oscilante que atraviesan los cerramientos. Los parámetros que definen la inercia térmica son: el espesor, la difusividad y el ciclo térmico. A su vez la difusividad está determinada por la conductividad térmica, la densidad y el calor específico del material. De estos parámetros la conductividad es el más complejo, variable y difícil de medir, especialmente en los cerramientos de tierra debido a su heterogeneidad y complejidad higrotérmica. En general, los métodos de medida de la conductividad o transmitancias en los paramentos presentan inconvenientes a la hora de medir un edificio construido con tierra: dificultades de implementación, el elevado coste o la fiabilidad de los resultados, principalmente. El Método de la Aguja Térmica (MAT) se basa en el principio de la evolución en el tiempo del calor emitido por una fuente lineal al insertarse en el seno de un material. Se ha escogido este método porque resulta práctico, de bajo coste y de fácil aplicación a gran escala pero tiene serios problemas de fiabilidad y exactitud. En esta tesis se desarrolla un método de medida de la conductividad térmica para Piezas de Albañilería de Tierra Cruda en laboratorio basado en el MAT, se mejora su fiabilidad, se analiza su incertidumbre, se compara con otros métodos de referencia y se aplica en adobes, Bloques de Tierra Comprimida y probetas de tierra estabilizada con distintas proporciones de paja. Este método servirá de base a una posterior aplicación in situ. Finalmente se proponen modelos matemáticos para mejorar la exactitud del dispositivo utilizado y para la estimación de la conductividad de cerramientos de tierra en función de su densidad. Con los resultados obtenidos se analizan las posibilidades de amortiguación y retardo de las ondas térmicas y capacidad de almacenaje de energía de los cerramientos en función de su densidad y humedad. There is growing international interest in energy saving and sustainability in buildings with significant impact on Architecture. Thermal inertia is a key parameter to assess energy in buildings in real conditions. This requires changing the traditional approach to heat transfer in steady state by another in dynamic regime which analyzes the thermal waves and oscillating heat flux passing through the external walls. The parameters defining the thermal inertia are: the thickness, the diffusivity and the thermal cycle. In turn, the diffusivity is determined by the thermal conductivity, density and specific heat of the material. Of these parameters, thermal conductivity is the most complex, variable and difficult to measure, especially in earth walls due to their heterogeneity and hygrothermal complexity. In general, the methods of measurement of conductivity and transmittance in walls have drawbacks when measuring a building with earth: implementation difficulties, high cost, or reliability of the results, mainly. The Thermal Needle Procedure (TNP) is based on the principle of evolution in time of heat from a line source when inserted within a material. This method was chosen because it is a practical, low cost and easy to implement on a large scale but has serious problems of reliability and accuracy. This thesis develops a laboratory method for measuring the thermal conductivity of Masonry Units Unfire Earth-based based on TNP, its uncertainty is analyzed, compared to other reference methods and applies in adobes, Compressed Earth Blocks and stabilized soil specimens with different proportions of straw. This method will form the basis of a subsequent application in situ. Finally, mathematical models are proposed to improve the accuracy of the device used, and to estimate the conductivity of earth enclosures depending on its density. With the results obtained earth enclosures are analyzed to estimate their possibilities of delay and buffer of termal waves and energy storage capacity according to their density and moisture.
Resumo:
The analysis of the interdependence between time series has become an important field of research in the last years, mainly as a result of advances in the characterization of dynamical systems from the signals they produce, the introduction of concepts such as generalized and phase synchronization and the application of information theory to time series analysis. In neurophysiology, different analytical tools stemming from these concepts have added to the ?traditional? set of linear methods, which includes the cross-correlation and the coherency function in the time and frequency domain, respectively, or more elaborated tools such as Granger Causality. This increase in the number of approaches to tackle the existence of functional (FC) or effective connectivity (EC) between two (or among many) neural networks, along with the mathematical complexity of the corresponding time series analysis tools, makes it desirable to arrange them into a unified, easy-to-use software package. The goal is to allow neuroscientists, neurophysiologists and researchers from related fields to easily access and make use of these analysis methods from a single integrated toolbox. Here we present HERMES (http://hermes.ctb.upm.es), a toolbox for the Matlab® environment (The Mathworks, Inc), which is designed to study functional and effective brain connectivity from neurophysiological data such as multivariate EEG and/or MEG records. It includes also visualization tools and statistical methods to address the problem of multiple comparisons. We believe that this toolbox will be very helpful to all the researchers working in the emerging field of brain connectivity analysis.
Resumo:
La Ingeniería del Software Empírico (ISE) utiliza como herramientas los estudios empíricos para conseguir evidencias que ayuden a conocer bajo qué circunstancias es mejor usar una tecnología software en lugar de otra. La investigación en la que se enmarca este TFM explora si las intuiciones y/o preferencias de las personas que realizan las pruebas de software, son capaces de predecir la efectividad de tres técnicas de evaluación de código: lectura por abstracciones sucesivas, cobertura de decisión y partición en clases de equivalencia. Para conseguir dicho objetivo, se analizan los datos recogidos en un estudio empírico, realizado por las tutoras de este TFM. En el estudio empírico distintos sujetos aplican las tres técnicas de evaluación de código a tres programas distintos, a los que se les habían introducido una serie de faltas artificialmente. Los sujetos deben reportar los fallos encontrados en los programas, así como, contestar a una serie de preguntas sobre sus intuiciones y preferencias. A la hora de analizar los datos del estudio, se ha comprobado: 1) cuáles son sus intuiciones y preferencias (mediante el test estadístico X2 de Pearson); 2) si los sujetos cambian de opinión después de aplicar las técnicas (para ello se ha utilizado índice de Kappa, el Test de McNemar-Bowker y el Test de Stuart-Maxwell); 3) la consistencia de las distintas preguntas (mediante el índice de Kappa), comparando: intuiciones con intuiciones, preferencias con preferencias e intuiciones con preferencias; 4) Por último, si hay coincidencia entre las intuiciones y preferencias con la efectividad real obtenida (para ello se ha utilizado, el Modelo Lineal General con medidas repetidas). Los resultados muestran que, no hay una intuición clara ni tampoco una preferencia concreta, con respecto a los programas. Además aunque existen cambios de opinión después de aplicar las técnicas, no se encuentran evidencias claras para afirmar que la intuición y preferencias influyen en su efectividad. Finalmente, existen relaciones entre las intuiciones con intuiciones, preferencias con preferencias e intuiciones con preferencias, además esta relación es más notoria después de aplicar las técnicas. ----ABSTRACT----Empirical Software Engineering (ESE) uses empirical studies as a mean to generate evidences to help determine under what circumstances it is convenient to use a given software technology. This Master Thesis is part of a research that explores whether intuitions and/or preferences of testers, can be used to predict the effectiveness of three code evaluation techniques: reading by stepwise abstractions, decision coverage and equivalence partitioning. To achieve this goal, this Master Thesis analyzes the data collected in an empirical study run by the tutors. In the empirical study, different subjects apply three code evaluation techniques to three different programs. A series of faults were artificially introduced to the programs. Subjects are required to report the defects found in the programs, as well as answer a series of questions about their intuitions and preferences. The data analyses test: 1) what are the intuitions and preferences of the subjects (using the Pearson X2 test); 2) whether subjects change their minds after applying the techniques (using the Kappa coefficient, McNemar-Bowker test, and Stuart-Maxwell test); 3) the consistency of the different questions, comparing: intuitions versus intuitions, preferences versus preferences and preferences versus intuitions (using the Kappa coefficient); 4) finally, if intuitions and/or preferences predict the actual effectiveness obtained (using the General Linear Model, repeated measures). The results show that there is not clear intuition or particular preference with respect to the programs. Moreover, although there are changes of mind after applying the techniques, there are not clear evidences to claim that intuition and preferences influence their effectiveness. Finally, there is a relationship between the intuitions versus intuitions, preferences versus preferences and intuitions versus preferences; this relationship is more noticeable after applying the techniques.
Resumo:
El MC en baloncesto es aquel fenómeno relacionado con el juego que presenta unas características particulares determinadas por la idiosincrasia de un equipo y puede afectar a los protagonistas y por ende al devenir del juego. En la presente Tesis se ha estudiado la incidencia del MC en Liga A.C.B. de baloncesto y para su desarrollo en profundidad se ha planteado dos investigaciones una cuantitativa y otra cualitativa cuya metodología se detalla a continuación: La investigación cuantitativa se ha basado en la técnica de estudio del “Performance analysis”, para ello se han estudiado cuatro temporadas de la Liga A.C.B. (del 2007/08 al 2010/11), tal y como refleja en la bibliografía consultada se han tomado como momentos críticos del juego a los últimos cinco minutos de partidos donde la diferencia de puntos fue de seis puntos y todos los Tiempos Extras disputados, de tal manera que se han estudiado 197 momentos críticos. La contextualización del estudio se ha hecho en función de la variables situacionales “game location” (local o visitante), “team quality” (mejores o peores clasificados) y “competition” (fases de LR y Playoff). Para la interpretación de los resultados se han realizado los siguientes análisis descriptivos: 1) Análisis Discriminante, 2) Regresión Lineal Múltiple; y 3) Análisis del Modelo Lineal General Multivariante. La investigación cualitativa se ha basado en la técnica de investigación de la entrevista semiestructurada. Se entrevistaron a 12 entrenadores que militaban en la Liga A.C.B. durante la temporada 2011/12, cuyo objetivo ha sido conocer el punto de vista que tiene el entrenador sobre el concepto del MC y que de esta forma pudiera dar un enfoque más práctico basado en su conocimiento y experiencia acerca de cómo actuar ante el MC en el baloncesto. Los resultados de ambas investigaciones coinciden en señalar la importancia del MC sobre el resultado final del juego. De igual forma, el concepto en sí entraña una gran complejidad por lo que se considera fundamental la visión científica de la observación del juego y la percepción subjetiva que presenta el entrenador ante el fenómeno, para la cual los aspectos psicológicos de sus protagonistas (jugadores y entrenadores) son determinantes. ABSTRACT The Critical Moment (CM) in basketball is a related phenomenon with the game that has particular features determined by the idiosyncrasies of a team and can affect the players and therefore the future of the game. In this Thesis we have studied the impact of CM in the A.C.B. League and from a profound development two investigations have been raised, quantitative and qualitative whose methodology is as follows: The quantitative research is based on the technique of study "Performance analysis", for this we have studied four seasons in the A.C.B. League (2007/08 to 2010/11), and as reflected in the literature the Critical Moments of the games were taken from the last five minutes of games where the point spread was six points and all overtimes disputed, such that 197 critical moments have been studied. The contextualization of the study has been based on the situational variables "game location" (home or away), "team quality" (better or lower classified) and "competition" (LR and Playoff phases). For the interpretation of the results the following descriptive analyzes were performed: 1) Discriminant Analysis, 2) Multiple Linear Regression Analysis; and 3) Analysis of Multivariate General Linear Model. Qualitative research is based on the technique of investigation of a semi-structured interview. 12 coaches who belonged to the A.C.B. League were interviewed in seasons 2011/12, which aimed to determine the point of view that the coach has on the CM concept and thus could give a more practical approach based on their knowledge and experience about how to deal with the CM in basketball. The results of both studies agree on the importance of the CM on the final outcome of the game. Similarly, the concept itself is highly complex so the scientific view of the observation of the game is considered essential as well as the subjective perception the coach presents before the phenomenon, for which the psychological aspects of their characters (players and coaches) are crucial.
Resumo:
Two experiments were conducted to estimate the standardized ileal digestible (SID) Trp:Lys ratio requirement for growth performance of nursery pigs. Experimental diets were formulated to ensure that lysine was the second limiting AA throughout the experiments. In Exp. 1 (6 to 10 kg BW), 255 nursery pigs (PIC 327 × 1050, initially 6.3 ± 0.15 kg, mean ± SD) arranged in pens of 6 or 7 pigs were blocked by pen weight and assigned to experimental diets (7 pens/diet) consisting of SID Trp:Lys ratios of 14.7%, 16.5%, 18.4%, 20.3%, 22.1%, and 24.0% for 14 d with 1.30% SID Lys. In Exp. 2 (11 to 20 kg BW), 1,088 pigs (PIC 337 × 1050, initially 11.2 kg ± 1.35 BW, mean ± SD) arranged in pens of 24 to 27 pigs were blocked by average pig weight and assigned to experimental diets (6 pens/diet) consisting of SID Trp:Lys ratios of 14.5%, 16.5%, 18.0%, 19.5%, 21.0%, 22.5%, and 24.5% for 21 d with 30% dried distillers grains with solubles and 0.97% SID Lys. Each experiment was analyzed using general linear mixed models with heterogeneous residual variances. Competing heteroskedastic models included broken-line linear (BLL), broken-line quadratic (BLQ), and quadratic polynomial (QP). For each response, the best-fitting model was selected using Bayesian information criterion. In Exp. 1 (6 to 10 kg BW), increasing SID Trp:Lys ratio linearly increased (P < 0.05) ADG and G:F. For ADG, the best-fitting model was a QP in which the maximum ADG was estimated at 23.9% (95% confidence interval [CI]: [<14.7%, >24.0%]) SID Trp:Lys ratio. For G:F, the best-fitting model was a BLL in which the maximum G:F was estimated at 20.4% (95% CI: [14.3%, 26.5%]) SID Trp:Lys. In Exp. 2 (11 to 20 kg BW), increasing SID Trp:Lys ratio increased (P < 0.05) ADG and G:F in a quadratic manner. For ADG, the best-fitting model was a QP in which the maximum ADG was estimated at 21.2% (95% CI: [20.5%, 21.9%]) SID Trp:Lys. For G:F, BLL and BLQ models had comparable fit and estimated SID Trp:Lys requirements at 16.6% (95% CI: [16.0%, 17.3%]) and 17.1% (95% CI: [16.6%, 17.7%]), respectively. In conclusion, the estimated SID Trp:Lys requirement in Exp. 1 ranged from 20.4% for maximum G:F to 23.9% for maximum ADG, whereas in Exp. 2 it ranged from 16.6% for maximum G:F to 21.2% for maximum ADG. These results suggest that standard NRC (2012) recommendations may underestimate the SID Trp:Lys requirement for nursery pigs from 11 to 20 kg BW.
Resumo:
Sin duda, el rostro humano ofrece mucha más información de la que pensamos. La cara transmite sin nuestro consentimiento señales no verbales, a partir de las interacciones faciales, que dejan al descubierto nuestro estado afectivo, actividad cognitiva, personalidad y enfermedades. Estudios recientes [OFT14, TODMS15] demuestran que muchas de nuestras decisiones sociales e interpersonales derivan de un previo análisis facial de la cara que nos permite establecer si esa persona es confiable, trabajadora, inteligente, etc. Esta interpretación, propensa a errores, deriva de la capacidad innata de los seres humanas de encontrar estas señales e interpretarlas. Esta capacidad es motivo de estudio, con un especial interés en desarrollar métodos que tengan la habilidad de calcular de manera automática estas señales o atributos asociados a la cara. Así, el interés por la estimación de atributos faciales ha crecido rápidamente en los últimos años por las diversas aplicaciones en que estos métodos pueden ser utilizados: marketing dirigido, sistemas de seguridad, interacción hombre-máquina, etc. Sin embargo, éstos están lejos de ser perfectos y robustos en cualquier dominio de problemas. La principal dificultad encontrada es causada por la alta variabilidad intra-clase debida a los cambios en la condición de la imagen: cambios de iluminación, oclusiones, expresiones faciales, edad, género, etnia, etc.; encontradas frecuentemente en imágenes adquiridas en entornos no controlados. Este de trabajo de investigación estudia técnicas de análisis de imágenes para estimar atributos faciales como el género, la edad y la postura, empleando métodos lineales y explotando las dependencias estadísticas entre estos atributos. Adicionalmente, nuestra propuesta se centrará en la construcción de estimadores que tengan una fuerte relación entre rendimiento y coste computacional. Con respecto a éste último punto, estudiamos un conjunto de estrategias para la clasificación de género y las comparamos con una propuesta basada en un clasificador Bayesiano y una adecuada extracción de características. Analizamos en profundidad el motivo de porqué las técnicas lineales no han logrado resultados competitivos hasta la fecha y mostramos cómo obtener rendimientos similares a las mejores técnicas no-lineales. Se propone un segundo algoritmo para la estimación de edad, basado en un regresor K-NN y una adecuada selección de características tal como se propuso para la clasificación de género. A partir de los experimentos desarrollados, observamos que el rendimiento de los clasificadores se reduce significativamente si los ´estos han sido entrenados y probados sobre diferentes bases de datos. Hemos encontrado que una de las causas es la existencia de dependencias entre atributos faciales que no han sido consideradas en la construcción de los clasificadores. Nuestro resultados demuestran que la variabilidad intra-clase puede ser reducida cuando se consideran las dependencias estadísticas entre los atributos faciales de el género, la edad y la pose; mejorando el rendimiento de nuestros clasificadores de atributos faciales con un coste computacional pequeño. Abstract Surely the human face provides much more information than we think. The face provides without our consent nonverbal cues from facial interactions that reveal our emotional state, cognitive activity, personality and disease. Recent studies [OFT14, TODMS15] show that many of our social and interpersonal decisions derive from a previous facial analysis that allows us to establish whether that person is trustworthy, hardworking, intelligent, etc. This error-prone interpretation derives from the innate ability of human beings to find and interpret these signals. This capability is being studied, with a special interest in developing methods that have the ability to automatically calculate these signs or attributes associated with the face. Thus, the interest in the estimation of facial attributes has grown rapidly in recent years by the various applications in which these methods can be used: targeted marketing, security systems, human-computer interaction, etc. However, these are far from being perfect and robust in any domain of problems. The main difficulty encountered is caused by the high intra-class variability due to changes in the condition of the image: lighting changes, occlusions, facial expressions, age, gender, ethnicity, etc.; often found in images acquired in uncontrolled environments. This research work studies image analysis techniques to estimate facial attributes such as gender, age and pose, using linear methods, and exploiting the statistical dependencies between these attributes. In addition, our proposal will focus on the construction of classifiers that have a good balance between performance and computational cost. We studied a set of strategies for gender classification and we compare them with a proposal based on a Bayesian classifier and a suitable feature extraction based on Linear Discriminant Analysis. We study in depth why linear techniques have failed to provide competitive results to date and show how to obtain similar performances to the best non-linear techniques. A second algorithm is proposed for estimating age, which is based on a K-NN regressor and proper selection of features such as those proposed for the classification of gender. From our experiments we note that performance estimates are significantly reduced if they have been trained and tested on different databases. We have found that one of the causes is the existence of dependencies between facial features that have not been considered in the construction of classifiers. Our results demonstrate that intra-class variability can be reduced when considering the statistical dependencies between facial attributes gender, age and pose, thus improving the performance of our classifiers with a reduced computational cost.
Resumo:
The use of seismic hysteretic dampers for passive control is increasing exponentially in recent years for both new and existing buildings. In order to utilize hysteretic dampers within a structural system, it is of paramount importance to have simplified design procedures based upon knowledge gained from theoretical studies and validated with experimental results. Non-linear Static Procedures (NSPs) are presented as an alternative to the force-based methods more common nowadays. The application of NSPs to conventional structures has been well established; yet there is a lack of experimental information on how NSPs apply to systems with hysteretic dampers. In this research, several shaking table tests were conducted on two single bay and single story 1:2 scale structures with and without hysteretic dampers. The maximum response of the structure with dampers in terms of lateral displacement and base shear obtained from the tests was compared with the prediction provided by three well-known NSPs: (1) the improved version of the Capacity Spectrum Method (CSM) from FEMA 440; (2) the improved version of the Displacement Coefficient Method (DCM) from FEMA 440; and (3) the N2 Method implemented in Eurocode 8. In general, the improved version of the DCM and N2 methods are found to provide acceptable accuracy in prediction, but the CSM tends to underestimate the response.
Resumo:
Question: How do tree species identity, microhabitat and water availability affect inter- and intra-specific interactions between juvenile and adult woody plants? Location: Continental Mediterranean forests in Alto Tajo Natural Park, Guadalajara, Spain. Methods: A total of 2066 juveniles and adults of four co-occurring tree species were mapped in 17 plots. The frequency of juveniles at different microhabitats and water availability levels was analysed using log-linear models. We used nearest-neighbour contingency table analysis of spatial segregation and J-functions to describe the spatial patterns. Results: We found a complex spatial pattern that varied according to species identity and microhabitat. Recruitment was more frequent in gaps for Quercus ilex, while the other three species recruited preferentially under shrubs or trees depending on the water availability level. Juveniles were not spatially associated to conspecific adults, experiencing segregation from them inmany cases. Spatial associations, both positive and negative, were more common at higher water availability levels. Conclusions: Our results do not agree with expectations from the stressgradient hypothesis, suggesting that positive interactions do not increase in importance with increasing aridity in the study ecosystem. Regeneration patterns are species-specific and depend on microhabitat characteristics and dispersal strategies. In general, juveniles do not look for conspecific adult protection. This work contributes to the understanding of species co-existence, proving the importance of considering a multispecies approach at several plots to overcome limitations of simple pair-wise comparisons in a limited number of sites.
Resumo:
The authors are from UPM and are relatively grouped, and all have intervened in different academic or real cases on the subject, at different times as being of different age. With precedent from E. Torroja and A. Páez in Madrid Spain Safety Probabilistic models for concrete about 1957, now in ICOSSAR conferences, author J.M. Antón involved since autumn 1967 for euro-steel construction in CECM produced a math model for independent load superposition reductions, and using it a load coefficient pattern for codes in Rome Feb. 1969, practically adopted for European constructions, giving in JCSS Lisbon Feb. 1974 suggestion of union for concrete-steel-al.. That model uses model for loads like Gumbel type I, for 50 years for one type of load, reduced to 1 year to be added to other independent loads, the sum set in Gumbel theories to 50 years return period, there are parallel models. A complete reliability system was produced, including non linear effects as from buckling, phenomena considered somehow in actual Construction Eurocodes produced from Model Codes. The system was considered by author in CEB in presence of Hydraulic effects from rivers, floods, sea, in reference with actual practice. When redacting a Road Drainage Norm in MOPU Spain an optimization model was realized by authors giving a way to determine the figure of Return Period, 10 to 50 years, for the cases of hydraulic flows to be considered in road drainage. Satisfactory examples were a stream in SE of Spain with Gumbel Type I model and a paper of Ven Te Chow with Mississippi in Keokuk using Gumbel type II, and the model can be modernized with more varied extreme laws. In fact in the MOPU drainage norm the redacting commission acted also as expert to set a table of return periods for elements of road drainage, in fact as a multi-criteria complex decision system. These precedent ideas were used e.g. in wide Codes, indicated in symposia or meetings, but not published in journals in English, and a condensate of contributions of authors is presented. The authors are somehow involved in optimization for hydraulic and agro planning, and give modest hints of intended applications in presence of agro and environment planning as a selection of the criteria and utility functions involved in bayesian, multi-criteria or mixed decision systems. Modest consideration is made of changing in climate, and on the production and commercial systems, and on others as social and financial.
Resumo:
The aim of this thesis is to study the mechanisms of instability that occur in swept wings when the angle of attack increases. For this, a simplified model for the a simplified model for the non-orthogonal swept leading edge boundary layer has been used as well as different numerical techniques in order to solve the linear stability problem that describes the behavior of perturbations superposed upon this base flow. Two different approaches, matrix-free and matrix forming methods, have been validated using direct numerical simulations with spectral resolution. In this way, flow instability in the non-orthogonal swept attachment-line boundary layer is addressed in a linear analysis framework via the solution of the pertinent global (Bi-Global) PDE-based eigenvalue problem. Subsequently, a simple extension of the extended G¨ortler-H¨ammerlin ODEbased polynomial model proposed by Theofilis, Fedorov, Obrist & Dallmann (2003) for orthogonal flow, which includes previous models as particular cases and recovers global instability analysis results, is presented for non-orthogonal flow. Direct numerical simulations have been used to verify the stability results and unravel the limits of validity of the basic flow model analyzed. The effect of the angle of attack, AoA, on the critical conditions of the non-orthogonal problem has been documented; an increase of the angle of attack, from AoA = 0 (orthogonal flow) up to values close to _/2 which make the assumptions under which the basic flow is derived questionable, is found to systematically destabilize the flow. The critical conditions of non-orthogonal flows at 0 _ AoA _ _/2 are shown to be recoverable from those of orthogonal flow, via a simple analytical transformation involving AoA. These results can help to understand the mechanisms of destabilization that occurs in the attachment line of wings at finite angles of attack. Studies taking into account variations of the pressure field in the basic flow or the extension to compressible flows are issues that remain open. El objetivo de esta tesis es estudiar los mecanismos de la inestabilidad que se producen en ciertos dispositivos aerodinámicos cuando se aumenta el ángulo de ataque. Para ello se ha utilizado un modelo simplificado del flujo de base, así como diferentes técnicas numéricas, con el fin de resolver el problema de estabilidad lineal asociado que describe el comportamiento de las perturbaciones. Estos métodos; sin y con formación de matriz, se han validado utilizando simulaciones numéricas directas con resolución espectral. De esta manera, la inestabilidad del flujo de capa límite laminar oblicuo entorno a la línea de estancamiento se aborda en un marco de análisis lineal por medio del método Bi-Global de resolución del problema de valores propios en derivadas parciales. Posteriormente se propone una extensión simple para el flujo no-ortogonal del modelo polinomial de ecuaciones diferenciales ordinarias, G¨ortler-H¨ammerlin extendido, propuesto por Theofilis et al. (2003) para el flujo ortogonal, que incluye los modelos previos como casos particulares y recupera los resultados del analisis global de estabilidad lineal. Se han realizado simulaciones directas con el fin de verificar los resultados del análisis de estabilidad así como para investigar los límites de validez del modelo de flujo base utilizado. En este trabajo se ha documentado el efecto del ángulo de ataque AoA en las condiciones críticas del problema no ortogonal obteniendo que el incremento del ángulo de ataque, de AoA = 0 (flujo ortogonal) hasta valores próximos a _/2, en el cual las hipótesis sobre las que se basa el flujo base dejan de ser válidas, tiende sistemáticamente a desestabilizar el flujo. Las condiciones críticas del caso no ortogonal 0 _ AoA _ _/2 pueden recuperarse a partir del caso ortogonal mediante el uso de una transformación analítica simple que implica el ángulo de ataque AoA. Estos resultados pueden ayudar a comprender los mecanismos de desestabilización que se producen en el borde de ataque de las alas de los aviones a ángulos de ataque finitos. Como tareas pendientes quedaría realizar estudios que tengan en cuenta variaciones del campo de presión en el flujo base así como la extensión de éste al caso de flujos compresibles.
Resumo:
It is well known that the evaluation of the influence matrices in the boundary-element method requires the computation of singular integrals. Quadrature formulae exist which are especially tailored to the specific nature of the singularity, i.e. log(*- x0)9 Ijx- JC0), etc. Clearly the nodes and weights of these formulae vary with the location Xo of the singular point. A drawback of this approach is that a given problem usually includes different types of singularities, and therefore a general-purpose code would have to include many alternative formulae to cater for all possible cases. Recently, several authors1"3 have suggested a type independent alternative technique based on the combination of standard Gaussian rules with non-linear co-ordinate transformations. The transformation approach is particularly appealing in connection with the p.adaptive version, where the location of the collocation points varies at each step of the refinement process. The purpose of this paper is to analyse the technique in eference 3. We show that this technique is asymptotically correct as the number of Gauss points increases. However, the method possesses a 'hidden' source of error that is analysed and can easily be removed.
Resumo:
In this paper, the classic oscillator design methods are reviewed, and their strengths and weaknesses are shown. Provisos for avoiding the misuse of classic methods are also proposed. If the required provisos are satisfied, the solutions provided by the classic methods (oscillator start-up linear approximation) will be correct. The provisos verification needs to use the NDF (Network Determinant Function). The use of the NDF or the most suitable RRT (Return Relation Transponse), which is directly related to the NDF, as a tool to analyze oscillators leads to a new oscillator design method. The RRT is the "true" loop-gain of oscillators. The use of the new method is demonstrated with examples. Finally, a comparison of NDF/RRT results with the HB (Harmonic Balance) simulation and practical implementation measurements prove the universal use of the new methods.