16 resultados para Masculinity in performance

em Universidad Politécnica de Madrid


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Bridges with deck supported on either sliding or elastomeric bearings are very common in mid-seismicity regions. Their main seismic vulnerabilities are related to the pounding of the deck against abutments or between the different deck elements. A simplified model of the longitudinal behavior of those bridges will allow to characterize the reaction forces developed during pounding using the Pacific Earthquake Engineering Research Center framework formula. In order to ensure the general applicability of the results obtained, a large number of system parameter combinations will be considered. The heart of the formula is the identification of suitable intermediate variables. First, the pseudo acceleration spectral value for the fundamental period of the system (Sa(Ts)) will be used as an intensity measure (IM). This IM will result in a very large non-explained variability of the engineering demand parameter. A portion of this variability will be proved to be related to the relative content of high-frequency energy in the input motion. Two vector-valued IMs including a second parameter taking this energy content into account will then be considered. For both of them, a suitable form for the conditional intensity dependence of the response will be obtained. The question of which one to choose will also be analyzed. Finally, additional issues related to the IM will be studied: its applicability to pulse-type records, the validity of scaling records and the sufficiency of the IM.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The aim of the present study was to examine the effect of sodium bicarbonate ingestion on consecutive "all out" sprint tests, analyzing the acid-base status and its influence on performance and perceived effort. Ten elite bicycle motocross (BMX) riders (20.7 ± 1.4 years, training experience 8-12 years) participated in this study which consisted of two trials. Each trial consisted of three consecutive Wingate tests (WTs) separated by 15 min recovery. Ninety minutes prior to exercise subjects ingested either NaHCO(3) (-) (0.3 g kg(-1) body weight) or placebo. Blood samples were collected for the assessment of blood acid-base status: bicarbonate concentration ([HCO(3) (-)]), pH, base excess (BE) and blood lactate concentration ([La(-)]). Performance variables of peak power (PP), mean power (MP), time to peak power and fatigue index were calculated for each sprint. Significant differences (p < 0.05) were observed in acid-base variables [pH before WT1: 7.47 ± 0.05 vs. 7.41 ± 0.03; [HCO(3) (-)] before WT1: 29.08 ± 2.27 vs. 22.85 ± 0.24 mmol L(-1) (bicarbonate vs. placebo conditions, respectively)], but there were not significant differences in performance variables between trials [PP WT1: 1,610 ± 373 vs. 1,599 ± 370 W; PP WT2: 1,548 ± 460 vs. 1,570 ± 428 W; PP WT3: 1,463 ± 361 vs. 1,519 ± 364 W. MP WT1: 809 ± 113 vs. 812 ± 108 W; MP WT2: 799 ± 135 vs. 799 ± 124 W; MP WT3: 762 ± 165 vs. 782 ± 118 W (bicarbonate vs. placebo conditions, respectively)]. Rating of perceived effort (RPE) was not influenced nor ratings of perceived readiness. Sodium bicarbonate ingestion modified significantly the blood acid-base balance, although the induced alkalosis did not improve the Wingate test performance, RPE and perceived readiness across three consecutive WTs in elite BMX cyclists.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Attentional control and Information processing speed are central concepts in cognitive psychology and neuropsychology. Functional neuroimaging and neuropsychological assessment have depicted theoretical models considering attention as a complex and non-unitary process. One of its component processes, Attentional set-shifting ability, is commonly assessed using the Trail Making Test (TMT). Performance in the TMT decreases with increasing age in adults, Mild Cognitive Impairment (MCI) and Alzheimer’s Disease (AD). Besides, speed of information processing (SIP) seems to modulate attentional performance. While neural correlates of attentional control have been widely studied, there are few evidences about the neural substrates of SIP in these groups of patients. Different authors have suggested that it could be a property of cerebral white matter, thus, deterioration of the white matter tracts that connect brain regions related to set-shifting may underlie the age-related, MCI and AD decrease in performance. The aim of this study was to study the anatomical dissociation of attentional and speed mechanisms. Diffusion tensor imaging (DTI) provides a unique insight into the cellular integrity of the brain, offering an in vivo view into the microarchitecture of cerebral white matter. At the same time, the study of ageing, characterized by white matter decline, provides the opportunity to study the anatomical substrates speeded or slowed information processing. We hypothesized that FA values would be inversely correlated with time to completion on Parts A and B of the TMT, but not the derived scores B/A and B-A.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

(Matsukawa and Habeck, 2007) analyse the main instruments for risk mitigation in infrastructure financing with Multilateral Financial Institutions (MFIs). Their review coincided with the global financial crisis of 2007-08, and is highly relevant in current times considering the sovereign debt crisis, the lack of available capital and the increases in bank regulation in Western economies. The current macroeconomic environment has seen a slowdown in the level of finance for infrastructure projects, as they pose a higher credit risk given their requirements for long term investments. The rationale for this work is to look for innovative solutions that are focused on the credit risk mitigation of infrastructure and energy projects whilst optimizing the economic capital allocation for commercial banks. This objective is achieved through risk-sharing with MFIs and looking for capital relief in project finance transactions. This research finds out the answer to the main question: "What is the impact of risk-sharing with MFIs on project finance transactions to increase their efficiency and viability?", and is developed from the perspective of a commercial bank assessing the economic capital used and analysing the relevant variables for it: Probability of Default, Loss Given Default and Recovery Rates, (Altman, 2010). An overview of project finance for the infrastructure and energy sectors in terms of the volume of transactions worldwide is outlined, along with a summary of risk-sharing financing with MFIs. A review of the current regulatory framework beneath risk-sharing in structured finance with MFIs is also analysed. From here, the impact of risk-sharing and the diversification effect in infrastructure and energy projects is assessed, from the perspective of economic capital allocation for a commercial bank. CreditMetrics (J. P. Morgan, 1997) is applied over an existing well diversified portfolio of project finance infrastructure and energy investments, working with the main risk capital measures: economic capital, RAROC, and EVA. The conclusions of this research show that economic capital allocation on a portfolio of project finance along with risk-sharing with MFIs have a huge impact on capital relief whilst increasing performance profitability for commercial banks. There is an outstanding diversification effect due to the portfolio, which is combined with risk mitigation and an improvement in recovery rates through Partial Credit Guarantees issued by MFIs. A stress test scenario analysis is applied to the current assumptions and credit risk model, considering a downgrade in the rating for the commercial bank (lender) and an increase of default in emerging countries, presenting a direct impact on economic capital, through an increase in expected loss and a decrease in performance profitability. Getting capital relief through risk-sharing makes it more viable for commercial banks to finance infrastructure and energy projects, with the beneficial effect of a direct impact of these investments on GDP growth and employment. The main contribution of this work is to promote a strategic economic capital allocation in infrastructure and energy financing through innovative risk-sharing with MFIs and economic pricing to create economic value added for banks, and to allow the financing of more infrastructure and energy projects. This work suggests several topics for further research in relation to issues analysed. (Matsukawa and Habeck, 2007) analizan los principales instrumentos de mitigación de riesgos en las Instituciones Financieras Multilaterales (IFMs) para la financiación de infraestructuras. Su presentación coincidió con el inicio de la crisis financiera en Agosto de 2007, y sus consecuencias persisten en la actualidad, destacando la deuda soberana en economías desarrolladas y los problemas capitalización de los bancos. Este entorno macroeconómico ha ralentizado la financiación de proyectos de infraestructuras. El actual trabajo de investigación tiene su motivación en la búsqueda de soluciones para la financiación de proyectos de infraestructuras y de energía, mitigando los riesgos inherentes, con el objeto de reducir el consumo de capital económico en los bancos financiadores. Este objetivo se alcanza compartiendo el riesgo de la financiación con IFMs, a través de estructuras de risk-sharing. La investigación responde la pregunta: "Cuál es el impacto de risk-sharing con IFMs, en la financiación de proyectos para aumentar su eficiencia y viabilidad?". El trabajo se desarrolla desde el enfoque de un banco comercial, estimando el consumo de capital económico en la financiación de proyectos y analizando las principales variables del riesgo de crédito, Probability of Default, Loss Given Default and Recovery Rates, (Altman, 2010). La investigación presenta las cifras globales de Project Finance en los sectores de infraestructuras y de energía, y analiza el marco regulatorio internacional en relación al consumo de capital económico en la financiación de proyectos en los que participan IFMs. A continuación, el trabajo modeliza una cartera real, bien diversificada, de Project Finance de infraestructuras y de energía, aplicando la metodología CreditMet- rics (J. P. Morgan, 1997). Su objeto es estimar el consumo de capital económico y la rentabilidad de la cartera de proyectos a través del RAROC y EVA. La modelización permite estimar el efecto diversificación y la liberación de capital económico consecuencia del risk-sharing. Los resultados muestran el enorme impacto del efecto diversificación de la cartera, así como de las garantías parciales de las IFMs que mitigan riesgos, mejoran el recovery rate de los proyectos y reducen el consumo de capital económico para el banco comercial, mientras aumentan la rentabilidad, RAROC, y crean valor económico, EVA. En escenarios económicos de inestabilidad, empeoramiento del rating de los bancos, aumentos de default en los proyectos y de correlación en las carteras, hay un impacto directo en el capital económico y en la pérdida de rentabilidad. La liberación de capital económico, como se plantea en la presente investigación, permitirá financiar más proyectos de infraestructuras y de energía, lo que repercutirá en un mayor crecimiento económico y creación de empleo. La principal contribución de este trabajo es promover la gestión activa del capital económico en la financiación de infraestructuras y de proyectos energéticos, a través de estructuras innovadoras de risk-sharing con IFMs y de creación de valor económico en los bancos comerciales, lo que mejoraría su eficiencia y capitalización. La aportación metodológica del trabajo se convierte por su originalidad en una contribución, que sugiere y facilita nuevas líneas de investigación académica en las principales variables del riesgo de crédito que afectan al capital económico en la financiación de proyectos.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

We apply diffusion strategies to propose a cooperative reinforcement learning algorithm, in which agents in a network communicate with their neighbors to improve predictions about their environment. The algorithm is suitable to learn off-policy even in large state spaces. We provide a mean-square-error performance analysis under constant step-sizes. The gain of cooperation in the form of more stability and less bias and variance in the prediction error, is illustrated in the context of a classical model. We show that the improvement in performance is especially significant when the behavior policy of the agents is different from the target policy under evaluation.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The effects of the combined use of long lactation periods (46 days) with alternative cages on the reproductive and growth performance of 104 rabbit does and their litters during five consecutive reproductive cycles were studied. Half of does were housed in conventional polyvalent cages (39 cm×100 cm×30 cm) and the other half in alternative polyvalent cages (39 cm×100 cm×60 cm), with a raised platform. Half of the rabbit does in each type of cage were weaned at 32 and the other half at 46 days after parturition. Longer lactation negatively affected the body weight (P<0.001), fat and energy content (P<0.05) of rabbit does at the end of the lactation period, but this effect decreased with the number of parturitions. Fertility, prolificacy and doe mortality were not affected by lactation length. Late weaning led to higher litter size (by 8.9%) and litter weight (by 11.3%) at the end of growing period (P<0.001) and lower feed conversion ratio per cage during the experimental period (13.5%) than weaning at 32 day (P<0.001). These results were paralleled by lower mortality (12.6 vs. 17.6%; P<0.01) of young rabbits weaned later during the overall experimental period. Differences in performance as a result of different weaning ages were only observed during cycles with worst health status (third and fifth cycles) in which late weaning decreased mortality. Type of cage did not affect doe body weight and body condition, mortality, fertility, prolificacy and litter size during the five reproductive cycles. Nevertheless, at day 21 litter weight and feed conversion ratio between 3 and 21 day were 4.2% higher (P<0.01) and 5.0% lower (P<0.05), respectively, in animals housed in alternative rather than in conventional cages. Alternative cages also led to heavier litters at 59 days (P<0.01). It was concluded that the combined use of longer lactations and cages with higher available surface with a raised platform could be alternatives to improve animal welfare in farmed rabbit.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The effects of the combined use of long lactation periods (46 days) with alternative cages on the reproductive and growth performance of 104 rabbit does and their litters during five consecutive reproductive cycles were studied. Half of does were housed in conven- tional polyvalent cages (39 cm x 100 cm x 30 cm) and the other half in alternative polyvalent cages (39 cm x 100 cm x 60 cm), with a raised platform. Half of the rabbit does in each type of cage were weaned at 32 and the other half at 46 days after parturition. Longer lactation negatively affected the body weight ( P o 0.001), fat and energy content ( P o 0.05) of rabbit does at the end of the lactation period, but this effect decreased with the number of parturitions. Fertility, prolificacy and doe mortality were not affected by lactation length. Late weaning led to higher litter size (by 8.9%) and litter weight (by 11.3%) at the end of growing period ( P o 0.001) and lower feed conversion ratio per cage during the experimental period (13.5%) than weaning at 32 day ( P o 0.001). These results were paralleled by lower mortality (12.6 vs. 17.6%; P o 0.01) of young rabbits weaned later during the overall experimental period. Differences in performance as a result of different weaning ages were only observed during cycles with worst health status (third and fifth cycles) in which late weaning decreased mortality. Type of cage did not affect doe body weight and body condition, mortality, fertility, prolificacy and litter size during the five reproductive cycles. Nevertheless, at day 21 litter weight and feed conversion ratio between 3 and 21 day were 4.2% higher ( P o 0.01) and 5.0% lower ( P o 0.05), respectively, in animals housed in alternative rather than in conventional cages. Alternative cages also led to heavier litters at 59 days ( P o 0.01). It was concluded that the combined use of longer lactations and cages with higher available surface with a raised platform could be alternatives to improve animal welfare in farmed rabbit.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Increased variability in performance has been associated with the emergence of several neurological and psychiatric pathologies. However, whether and how consistency of neuronal activity may also be indicative of an underlying pathology is still poorly understood. Here we propose a novel method for evaluating consistency from non-invasive brain recordings. We evaluate the consistency of the cortical activity recorded with magnetoencephalography in a group of subjects diagnosed with Mild Cognitive Impairment (MCI), a condition sometimes prodromal of dementia, during the execution of a memory task. We use metrics coming from nonlinear dynamics to evaluate the consistency of cortical regions. A representation known as parenclitic networks is constructed, where atypical features are endowed with a network structure, the topological properties of which can be studied at various scales. Pathological conditions correspond to strongly heterogeneous networks, whereas typical or normative conditions are characterized by sparsely connected networks with homogeneous nodes. The analysis of this kind of networks allows identifying the extent to which consistency is affected in the MCI group and the focal points where MCI is especially severe. To the best of our knowledge, these results represent the first attempt at evaluating the consistency of brain functional activity using complex networks theory.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Nowadays computing platforms consist of a very large number of components that require to be supplied with diferent voltage levels and power requirements. Even a very small platform, like a handheld computer, may contain more than twenty diferent loads and voltage regulators. The power delivery designers of these systems are required to provide, in a very short time, the right power architecture that optimizes the performance, meets electrical specifications plus cost and size targets. The appropriate selection of the architecture and converters directly defines the performance of a given solution. Therefore, the designer needs to be able to evaluate a significant number of options in order to know with good certainty whether the selected solutions meet the size, energy eficiency and cost targets. The design dificulties of selecting the right solution arise due to the wide range of power conversion products provided by diferent manufacturers. These products range from discrete components (to build converters) to complete power conversion modules that employ diferent manufacturing technologies. Consequently, in most cases it is not possible to analyze all the alternatives (combinations of power architectures and converters) that can be built. The designer has to select a limited number of converters in order to simplify the analysis. In this thesis, in order to overcome the mentioned dificulties, a new design methodology for power supply systems is proposed. This methodology integrates evolutionary computation techniques in order to make possible analyzing a large number of possibilities. This exhaustive analysis helps the designer to quickly define a set of feasible solutions and select the best trade-off in performance according to each application. The proposed approach consists of two key steps, one for the automatic generation of architectures and other for the optimized selection of components. In this thesis are detailed the implementation of these two steps. The usefulness of the methodology is corroborated by contrasting the results using real problems and experiments designed to test the limits of the algorithms.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

During the last years, there has been much concern about learning management systems' (LMS) effectiveness when compared to traditional learning and about how to assess students' participation during the course. The tracking and monitoring capabilities of most recent LMS have made it possible to analyse every interaction in the system. The issues addressed on this study are: a) Is LMS student's interaction an indicator of academic performance?; b) Are different results in performance expected between distance and in-class LMS-supported education?; c) How can LMS interactions from logs be categorised?; d) May this categorisation detect 'learning witnesses'? To answer these questions, a set of interaction types from Moodle LMS activity record logs has been analysed during two years in online and in-class Master's degrees at the UPM. The results show partial or no evidence of influence between interaction indicators and academic performance, although the proposed categorisation may help detect learning witnesses.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Modern object oriented languages like C# and JAVA enable developers to build complex application in less time. These languages are based on selecting heap allocated pass-by-reference objects for user defined data structures. This simplifies programming by automatically managing memory allocation and deallocation in conjunction with automated garbage collection. This simplification of programming comes at the cost of performance. Using pass-by-reference objects instead of lighter weight pass-by value structs can have memory impact in some cases. These costs can be critical when these application runs on limited resource environments such as mobile devices and cloud computing systems. We explore the problem by using the simple and uniform memory model to improve the performance. In this work we address this problem by providing an automated and sounds static conversion analysis which identifies if a by reference type can be safely converted to a by value type where the conversion may result in performance improvements. This works focus on C# programs. Our approach is based on a combination of syntactic and semantic checks to identify classes that are safe to convert. We evaluate the effectiveness of our work in identifying convertible types and impact of this transformation. The result shows that the transformation of reference type to value type can have substantial performance impact in practice. In our case studies we optimize the performance in Barnes-Hut program which shows total memory allocation decreased by 93% and execution time also reduced by 15%.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

RDF streams are sequences of timestamped RDF statements or graphs, which can be generated by several types of data sources (sensors, social networks, etc.). They may provide data at high volumes and rates, and be consumed by applications that require real-time responses. Hence it is important to publish and interchange them efficiently. In this paper, we exploit a key feature of RDF data streams, which is the regularity of their structure and data values, proposing a compressed, efficient RDF interchange (ERI) format, which can reduce the amount of data transmitted when processing RDF streams. Our experimental evaluation shows that our format produces state-of-the-art streaming compression, remaining efficient in performance.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Los algoritmos basados en registros de desplazamiento con realimentación (en inglés FSR) se han utilizado como generadores de flujos pseudoaleatorios en aplicaciones con recursos limitados como los sistemas de apertura sin llave. Se considera canal primario a aquel que se utiliza para realizar una transmisión de información. La aparición de los ataques de canal auxiliar (en inglés SCA), que explotan información filtrada inintencionadamente a través de canales laterales como el consumo, las emisiones electromagnéticas o el tiempo empleado, supone una grave amenaza para estas aplicaciones, dado que los dispositivos son accesibles por un atacante. El objetivo de esta tesis es proporcionar un conjunto de protecciones que se puedan aplicar de forma automática y que utilicen recursos ya disponibles, evitando un incremento sustancial en los costes y alargando la vida útil de aplicaciones que puedan estar desplegadas. Explotamos el paralelismo existente en algoritmos FSR, ya que sólo hay 1 bit de diferencia entre estados de rondas consecutivas. Realizamos aportaciones en tres niveles: a nivel de sistema, utilizando un coprocesador reconfigurable, a través del compilador y a nivel de bit, aprovechando los recursos disponibles en el procesador. Proponemos un marco de trabajo que nos permite evaluar implementaciones de un algoritmo incluyendo los efectos introducidos por el compilador considerando que el atacante es experto. En el campo de los ataques, hemos propuesto un nuevo ataque diferencial que se adapta mejor a las condiciones de las implementaciones software de FSR, en las que el consumo entre rondas es muy similar. SORU2 es un co-procesador vectorial reconfigurable propuesto para reducir el consumo energético en aplicaciones con paralelismo y basadas en el uso de bucles. Proponemos el uso de SORU2, además, para ejecutar algoritmos basados en FSR de forma segura. Al ser reconfigurable, no supone un sobrecoste en recursos, ya que no está dedicado en exclusiva al algoritmo de cifrado. Proponemos una configuración que ejecuta múltiples algoritmos de cifrado similares de forma simultánea, con distintas implementaciones y claves. A partir de una implementación sin protecciones, que demostramos que es completamente vulnerable ante SCA, obtenemos una implementación segura a los ataques que hemos realizado. A nivel de compilador, proponemos un mecanismo para evaluar los efectos de las secuencias de optimización del compilador sobre una implementación. El número de posibles secuencias de optimizaciones de compilador es extremadamente alto. El marco de trabajo propuesto incluye un algoritmo para la selección de las secuencias de optimización a considerar. Debido a que las optimizaciones del compilador transforman las implementaciones, se pueden generar automáticamente implementaciones diferentes combinamos para incrementar la seguridad ante SCA. Proponemos 2 mecanismos de aplicación de estas contramedidas, que aumentan la seguridad de la implementación original sin poder considerarse seguras. Finalmente hemos propuesto la ejecución paralela a nivel de bit del algoritmo en un procesador. Utilizamos la forma algebraica normal del algoritmo, que automáticamente se paraleliza. La implementación sobre el algoritmo evaluado mejora en rendimiento y evita que se filtre información por una ejecución dependiente de datos. Sin embargo, es más vulnerable ante ataques diferenciales que la implementación original. Proponemos una modificación del algoritmo para obtener una implementación segura, descartando parcialmente ejecuciones del algoritmo, de forma aleatoria. Esta implementación no introduce una sobrecarga en rendimiento comparada con las implementaciones originales. En definitiva, hemos propuesto varios mecanismos originales a distintos niveles para introducir aleatoridad en implementaciones de algoritmos FSR sin incrementar sustancialmente los recursos necesarios. ABSTRACT Feedback Shift Registers (FSR) have been traditionally used to implement pseudorandom sequence generators. These generators are used in Stream ciphers in systems with tight resource constraints, such as Remote Keyless Entry. When communicating electronic devices, the primary channel is the one used to transmit the information. Side-Channel Attack (SCA) use additional information leaking from the actual implementation, including power consumption, electromagnetic emissions or timing information. Side-Channel Attacks (SCA) are a serious threat to FSR-based applications, as an attacker usually has physical access to the devices. The main objective of this Ph.D. thesis is to provide a set of countermeasures that can be applied automatically using the available resources, avoiding a significant cost overhead and extending the useful life of deployed systems. If possible, we propose to take advantage of the inherent parallelism of FSR-based algorithms, as the state of a FSR differs from previous values only in 1-bit. We have contributed in three different levels: architecture (using a reconfigurable co-processor), using compiler optimizations, and at bit level, making the most of the resources available at the processor. We have developed a framework to evaluate implementations of an algorithm including the effects introduced by the compiler. We consider the presence of an expert attacker with great knowledge on the application and the device. Regarding SCA, we have presented a new differential SCA that performs better than traditional SCA on software FSR-based algorithms, where the leaked values are similar between rounds. SORU2 is a reconfigurable vector co-processor. It has been developed to reduce energy consumption in loop-based applications with parallelism. In addition, we propose its use for secure implementations of FSR-based algorithms. The cost overhead is discarded as the co-processor is not exclusively dedicated to the encryption algorithm. We present a co-processor configuration that executes multiple simultaneous encryptions, using different implementations and keys. From a basic implementation, which is proved to be vulnerable to SCA, we obtain an implementation where the SCA applied were unsuccessful. At compiler level, we use the framework to evaluate the effect of sequences of compiler optimization passes on a software implementation. There are many optimization passes available. The optimization sequences are combinations of the available passes. The amount of sequences is extremely high. The framework includes an algorithm for the selection of interesting sequences that require detailed evaluation. As existing compiler optimizations transform the software implementation, using different optimization sequences we can automatically generate different implementations. We propose to randomly switch between the generated implementations to increase the resistance against SCA.We propose two countermeasures. The results show that, although they increase the resistance against SCA, the resulting implementations are not secure. At bit level, we propose to exploit bit level parallelism of FSR-based implementations using pseudo bitslice implementation in a wireless node processor. The bitslice implementation is automatically obtained from the Algebraic Normal Form of the algorithm. The results show a performance improvement, avoiding timing information leakage, but increasing the vulnerability against differential SCA.We provide a secure version of the algorithm by randomly discarding part of the data obtained. The overhead in performance is negligible when compared to the original implementations. To summarize, we have proposed a set of original countermeasures at different levels that introduce randomness in FSR-based algorithms avoiding a heavy overhead on the resources required.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Las futuras misiones para misiles aire-aire operando dentro de la atmósfera requieren la interceptación de blancos a mayores velocidades y más maniobrables, incluyendo los esperados vehículos aéreos de combate no tripulados. La intercepción tiene que lograrse desde cualquier ángulo de lanzamiento. Una de las principales discusiones en la tecnología de misiles en la actualidad es cómo satisfacer estos nuevos requisitos incrementando la capacidad de maniobra del misil y en paralelo, a través de mejoras en los métodos de guiado y control modernos. Esta Tesis aborda estos dos objetivos simultáneamente, al proponer un diseño integrando el guiado y el control de vuelo (autopiloto) y aplicarlo a misiles con control aerodinámico simultáneo en canard y cola. Un primer avance de los resultados obtenidos ha sido publicado recientemente en el Journal of Aerospace Engineering, en Abril de 2015, [Ibarrondo y Sanz-Aranguez, 2015]. El valor del diseño integrado obtenido es que permite al misil cumplir con los requisitos operacionales mencionados empleando únicamente control aerodinámico. El diseño propuesto se compara favorablemente con esquemas más tradicionales, consiguiendo menores distancias de paso al blanco y necesitando de menores esfuerzos de control incluso en presencia de ruidos. En esta Tesis se demostrará cómo la introducción del doble mando, donde tanto el canard como las aletas de cola son móviles, puede mejorar las actuaciones de un misil existente. Comparado con un misil con control en cola, el doble control requiere sólo introducir dos servos adicionales para accionar los canards también en guiñada y cabeceo. La sección de cola será responsable de controlar el misil en balanceo mediante deflexiones diferenciales de los controles. En el caso del doble mando, la complicación añadida es que los vórtices desprendidos de los canards se propagan corriente abajo y pueden incidir sobre las superficies de cola, alterando sus características de control. Como un primer aporte, se ha desarrollado un modelo analítico completo para la aerodinámica no lineal de un misil con doble control, incluyendo la caracterización de este efecto de acoplamiento aerodinámico. Hay dos modos de funcionamiento en picado y guiñada para un misil de doble mando: ”desviación” y ”opuesto”. En modo ”desviación”, los controles actúan en la misma dirección, generando un cambio inmediato en la sustentación y produciendo un movimiento de translación en el misil. La respuesta es rápida, pero en el modo ”desviación” los misiles con doble control pueden tener dificultades para alcanzar grandes ángulos de ataque y altas aceleraciones laterales. Cuando los controles actúan en direcciones opuestas, el misil rota y el ángulo de ataque del fuselaje se incrementa para generar mayores aceleraciones en estado estacionario, aunque el tiempo de respuesta es mayor. Con el modelo aerodinámico completo, es posible obtener una parametrización dependiente de los estados de la dinámica de corto periodo del misil. Debido al efecto de acoplamiento entre los controles, la respuesta en bucle abierto no depende linealmente de los controles. El autopiloto se optimiza para obtener la maniobra requerida por la ley de guiado sin exceder ninguno de los límites aerodinámicos o mecánicos del misil. Una segunda contribución de la tesis es el desarrollo de un autopiloto con múltiples entradas de control y que integra la aerodinámica no lineal, controlando los tres canales de picado, guiñada y cabeceo de forma simultánea. Las ganancias del autopiloto dependen de los estados del misil y se calculan a cada paso de integración mediante la resolución de una ecuación de Riccati de orden 21x21. Las ganancias obtenidas son sub-óptimas, debido a que una solución completa de la ecuación de Hamilton-Jacobi-Bellman no puede obtenerse de manera práctica, y se asumen ciertas simplificaciones. Se incorpora asimismo un mecanismo que permite acelerar la respuesta en caso necesario. Como parte del autopiloto, se define una estrategia para repartir el esfuerzo de control entre el canard y la cola. Esto se consigue mediante un controlador aumentado situado antes del bucle de optimización, que minimiza el esfuerzo total de control para maniobrar. Esta ley de alimentación directa mantiene al misil cerca de sus condiciones de equilibrio, garantizando una respuesta transitoria adecuada. El controlador no lineal elimina la respuesta de fase no-mínima característica de la cola. En esta Tesis se consideran dos diseños para el guiado y control, el control en Doble-Lazo y el control Integrado. En la aproximación de Doble-Lazo, el autopiloto se sitúa dentro de un bucle interior y se diseña independientemente del guiado, que conforma el bucle más exterior del control. Esta estructura asume que existe separación espectral entre los dos, esto es, que los tiempos de respuesta del autopiloto son mucho mayores que los tiempos característicos del guiado. En el estudio se combina el autopiloto desarrollado con una ley de guiado óptimo. Los resultados obtenidos demuestran que se consiguen aumentos muy importantes en las actuaciones frente a misiles con control canard o control en cola, y que la interceptación, cuando se lanza cerca del curso de colisión, se consigue desde cualquier ángulo alrededor del blanco. Para el misil de doble mando, la estrategia óptima resulta en utilizar el modo de control opuesto en la aproximación al blanco y utilizar el modo de desviación justo antes del impacto. Sin embargo la lógica de doble bucle no consigue el impacto cuando hay desviaciones importantes con respecto al curso de colisión. Una de las razones es que parte de la demanda de guiado se pierde, ya que el misil solo es capaz de modificar su aceleración lateral, y no tiene control sobre su aceleración axial, a no ser que incorpore un motor de empuje regulable. La hipótesis de separación mencionada, y que constituye la base del Doble-Bucle, puede no ser aplicable cuando la dinámica del misil es muy alta en las proximidades del blanco. Si se combinan el guiado y el autopiloto en un único bucle, la información de los estados del misil está disponible para el cálculo de la ley de guiado, y puede calcularse la estrategia optima de guiado considerando las capacidades y la actitud del misil. Una tercera contribución de la Tesis es la resolución de este segundo diseño, la integración no lineal del guiado y del autopiloto (IGA) para el misil de doble control. Aproximaciones anteriores en la literatura han planteado este sistema en ejes cuerpo, resultando en un sistema muy inestable debido al bajo amortiguamiento del misil en cabeceo y guiñada. Las simplificaciones que se tomaron también causan que el misil se deslice alrededor del blanco y no consiga la intercepción. En nuestra aproximación el problema se plantea en ejes inerciales y se recurre a la dinámica de los cuaterniones, eliminado estos inconvenientes. No se limita a la dinámica de corto periodo del misil, porque se construye incluyendo de modo explícito la velocidad dentro del bucle de optimización. La formulación resultante en el IGA es independiente de la maniobra del blanco, que sin embargo se ha de incluir en el cálculo del modelo en Doble-bucle. Un típico inconveniente de los sistemas integrados con controlador proporcional, es el problema de las escalas. Los errores de guiado dominan sobre los errores de posición del misil y saturan el controlador, provocando la pérdida del misil. Este problema se ha tratado aquí con un controlador aumentado previo al bucle de optimización, que define un estado de equilibrio local para el sistema integrado, que pasa a actuar como un regulador. Los criterios de actuaciones para el IGA son los mismos que para el sistema de Doble-Bucle. Sin embargo el problema matemático resultante es muy complejo. El problema óptimo para tiempo finito resulta en una ecuación diferencial de Riccati con condiciones terminales, que no puede resolverse. Mediante un cambio de variable y la introducción de una matriz de transición, este problema se transforma en una ecuación diferencial de Lyapunov que puede resolverse mediante métodos numéricos. La solución resultante solo es aplicable en un entorno cercano del blanco. Cuando la distancia entre misil y blanco es mayor, se desarrolla una solución aproximada basada en la solución de una ecuación algebraica de Riccati para cada paso de integración. Los resultados que se han obtenido demuestran, a través de análisis numéricos en distintos escenarios, que la solución integrada es mejor que el sistema de Doble-Bucle. Las trayectorias resultantes son muy distintas. El IGA preserva el guiado del misil y consigue maximizar el uso de la propulsión, consiguiendo la interceptación del blanco en menores tiempos de vuelo. El sistema es capaz de lograr el impacto donde el Doble-Bucle falla, y además requiere un orden menos de magnitud en la cantidad de cálculos necesarios. El efecto de los ruidos radar, datos discretos y errores del radomo se investigan. El IGA es más robusto, resultando menos afectado por perturbaciones que el Doble- Bucle, especialmente porque el núcleo de optimización en el IGA es independiente de la maniobra del blanco. La estimación de la maniobra del blanco es siempre imprecisa y contaminada por ruido, y degrada la precisión de la solución de Doble-Bucle. Finalmente, como una cuarta contribución, se demuestra que el misil con guiado IGA es capaz de realizar una maniobra de defensa contra un blanco que ataque por su cola, sólo con control aerodinámico. Las trayectorias estudiadas consideran una fase pre-programada de alta velocidad de giro, manteniendo siempre el misil dentro de su envuelta de vuelo. Este procedimiento no necesita recurrir a soluciones técnicamente más complejas como el control vectorial del empuje o control por chorro para ejecutar esta maniobra. En todas las demostraciones matemáticas se utiliza el producto de Kronecker como una herramienta practica para manejar las parametrizaciones dependientes de variables, que resultan en matrices de grandes dimensiones. ABSTRACT Future missions for air to air endo-atmospheric missiles require the interception of targets with higher speeds and more maneuverable, including forthcoming unmanned supersonic combat vehicles. The interception will need to be achieved from any angle and off-boresight launch conditions. One of the most significant discussions in missile technology today is how to satisfy these new operational requirements by increasing missile maneuvering capabilities and in parallel, through the development of more advanced guidance and control methods. This Thesis addresses these two objectives by proposing a novel optimal integrated guidance and autopilot design scheme, applicable to more maneuverable missiles with forward and rearward aerodynamic controls. A first insight of these results have been recently published in the Journal of Aerospace Engineering in April 2015, [Ibarrondo and Sanz-Aránguez, 2015]. The value of this integrated solution is that it allows the missile to comply with the aforementioned requirements only by applying aerodynamic control. The proposed design is compared against more traditional guidance and control approaches with positive results, achieving reduced control efforts and lower miss distances with the integrated logic even in the presence of noises. In this Thesis it will be demonstrated how the dual control missile, where canard and tail fins are both movable, can enhance the capabilities of an existing missile airframe. Compared to a tail missile, dual control only requires two additional servos to actuate the canards in pitch and yaw. The tail section will be responsible to maintain the missile stabilized in roll, like in a classic tail missile. The additional complexity is that the vortices shed from the canard propagate downstream where they interact with the tail surfaces, altering the tail expected control characteristics. These aerodynamic phenomena must be properly described, as a preliminary step, with high enough precision for advanced guidance and control studies. As a first contribution we have developed a full analytical model of the nonlinear aerodynamics of a missile with dual control, including the characterization of this cross-control coupling effect. This development has been produced from a theoretical model validated with reliable practical data obtained from wind tunnel experiments available in the scientific literature, complement with computer fluid dynamics and semi-experimental methods. There are two modes of operating a missile with forward and rear controls, ”divert” and ”opposite” modes. In divert mode, controls are deflected in the same direction, generating an increment in direct lift and missile translation. Response is fast, but in this mode, dual control missiles may have difficulties in achieving large angles of attack and high level of lateral accelerations. When controls are deflected in opposite directions (opposite mode) the missile airframe rotates and the body angle of attack is increased to generate greater accelerations in steady-state, although the response time is larger. With the aero-model, a state dependent parametrization of the dual control missile short term dynamics can be obtained. Due to the cross-coupling effect, the open loop dynamics for the dual control missile is not linearly dependent of the fin positions. The short term missile dynamics are blended with the servo system to obtain an extended autopilot model, where the response is linear with the control fins turning rates, that will be the control variables. The flight control loop is optimized to achieve the maneuver required by the guidance law without exceeding any of the missile aerodynamic or mechanical limitations. The specific aero-limitations and relevant performance indicators for the dual control are set as part of the analysis. A second contribution of this Thesis is the development of a step-tracking multi-input autopilot that integrates non-linear aerodynamics. The designed dual control missile autopilot is a full three dimensional autopilot, where roll, pitch and yaw are integrated, calculating command inputs simultaneously. The autopilot control gains are state dependent, and calculated at each integration step solving a matrix Riccati equation of order 21x21. The resulting gains are sub-optimal as a full solution for the Hamilton-Jacobi-Bellman equation cannot be resolved in practical terms and some simplifications are taken. Acceleration mechanisms with an λ-shift is incorporated in the design. As part of the autopilot, a strategy is defined for proper allocation of control effort between canard and tail channels. This is achieved with an augmented feed forward controller that minimizes the total control effort of the missile to maneuver. The feedforward law also maintains the missile near trim conditions, obtaining a well manner response of the missile. The nonlinear controller proves to eliminate the non-minimum phase effect of the tail. Two guidance and control designs have been considered in this Thesis: the Two- Loop and the Integrated approaches. In the Two-Loop approach, the autopilot is placed in an inner loop and designed separately from an outer guidance loop. This structure assumes that spectral separation holds, meaning that the autopilot response times are much higher than the guidance command updates. The developed nonlinear autopilot is linked in the study to an optimal guidance law. Simulations are carried on launching close to collision course against supersonic and highly maneuver targets. Results demonstrate a large boost in performance provided by the dual control versus more traditional canard and tail missiles, where interception with the dual control close to collision course is achieved form 365deg all around the target. It is shown that for the dual control missile the optimal flight strategy results in using opposite control in its approach to target and quick corrections with divert just before impact. However the Two-Loop logic fails to achieve target interception when there are large deviations initially from collision course. One of the reasons is that part of the guidance command is not followed, because the missile is not able to control its axial acceleration without a throttleable engine. Also the separation hypothesis may not be applicable for a high dynamic vehicle like a dual control missile approaching a maneuvering target. If the guidance and autopilot are combined into a single loop, the guidance law will have information of the missile states and could calculate the most optimal approach to the target considering the actual capabilities and attitude of the missile. A third contribution of this Thesis is the resolution of the mentioned second design, the non-linear integrated guidance and autopilot (IGA) problem for the dual control missile. Previous approaches in the literature have posed the problem in body axes, resulting in high unstable behavior due to the low damping of the missile, and have also caused the missile to slide around the target and not actually hitting it. The IGA system is posed here in inertial axes and quaternion dynamics, eliminating these inconveniences. It is not restricted to the missile short term dynamic, and we have explicitly included the missile speed as a state variable. The IGA formulation is also independent of the target maneuver model that is explicitly included in the Two-loop optimal guidance law model. A typical problem of the integrated systems with a proportional control law is the problem of scales. The guidance errors are larger than missile state errors during most of the flight and result in high gains, control saturation and loss of control. It has been addressed here with an integrated feedforward controller that defines a local equilibrium state at each flight point and the controller acts as a regulator to minimize the IGA states excursions versus the defined feedforward state. The performance criteria for the IGA are the same as in the Two-Loop case. However the resulting optimization problem is mathematically very complex. The optimal problem in a finite-time horizon results in an irresoluble state dependent differential Riccati equation with terminal conditions. With a change of variable and the introduction of a transition matrix, the equation is transformed into a time differential Lyapunov equation that can be solved with known numerical methods in real time. This solution results range limited, and applicable when the missile is in a close neighborhood of the target. For larger ranges, an approximate solution is used, obtained from solution of an algebraic matrix Riccati equation at each integration step. The results obtained show, by mean of several comparative numerical tests in diverse homing scenarios, than the integrated approach is a better solution that the Two- Loop scheme. Trajectories obtained are very different in the two cases. The IGA fully preserves the guidance command and it is able to maximize the utilization of the missile propulsion system, achieving interception with lower miss distances and in lower flight times. The IGA can achieve interception against off-boresight targets where the Two- Loop was not able to success. As an additional advantage, the IGA also requires one order of magnitude less calculations than the Two-Loop solution. The effects of radar noises, discrete radar data and radome errors are investigated. IGA solution is robust, and less affected by radar than the Two-Loop, especially because the target maneuvers are not part of the IGA core optimization loop. Estimation of target acceleration is always imprecise and noisy and degrade the performance of the two-Loop solution. The IGA trajectories are such that minimize the impact of radome errors in the guidance loop. Finally, as a fourth contribution, it is demonstrated that the missile with IGA guidance is capable of performing a defense against attacks from its rear hemisphere, as a tail attack, only with aerodynamic control. The studied trajectories have a preprogrammed high rate turn maneuver, maintaining the missile within its controllable envelope. This solution does not recur to more complex features in service today, like vector control of the missile thrust or side thrusters. In all the mathematical treatments and demonstrations, the Kronecker product has been introduced as a practical tool to handle the state dependent parametrizations that have resulted in very high order matrix equations.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

El minuto final de un partido ajustado de baloncesto es un momento crítico que está sujeto a multitud de factores que influyen en su desarrollo. Así, el porcentaje de acierto en los tiros libres durante ese periodo de tiempo va a determinar, en muchas ocasiones, el resultado final del partido. La disminución de rendimiento (drop) en esta faceta de juego en condiciones de presión, puede estar relacionada con múltiples variables propias del contexto deportivo estudiado, como por ejemplo: los segundos restantes de posesión, la situación en el marcador (ir ganando, empatando o perdiendo), la localización del partido (jugar en casa o fuera), la fase de competición (fase regular o eliminatorias) o el nivel del equipo (mejores/peores equipos). Además, las características del jugador que realiza los lanzamientos tienen una gran importancia respecto a su edad y años de experiencia para afrontar los momentos críticos, así como el puesto de juego que ocupa en el equipo. En este sentido, la combinación de factores del contexto y del jugador, permiten interactuar en el rendimiento del lanzador en los momentos finales de partido durante sus lanzamientos de tiro libre. El presente trabajo de tesis doctoral tiene como objetivo encontrar aquellas variables más relacionadas con la disminución de rendimiento del jugador en los tiros libres durante el último minuto de juego, y la última serie de tiros libres en los partidos ajustados de baloncesto. Para alcanzar el objetivo del estudio se analizaron 124 partidos ajustados (diferencias iguales o inferiores a 2 puntos) de todas las competiciones (fase regular, playoff y copa del Rey) de la liga ACB durante las temporadas 2011-2012 a 2014-2015. Para el registro de variables se analizó el porcentaje de acierto en los tiros libres del lanzador en la liga regular, partido completo, último minuto y última serie. De este modo se trató de analizar qué variables del contexto y del jugador permitían explicar el rendimiento en los tiros libres durante el último minuto, y la última serie de tiros libres del partido. Por otro lado, se trató de conocer el grado de asociación entre el descenso del rendimiento (drop) en los momentos finales de partido, y las variables estudiadas del jugador: puesto de juego, edad, y años de experiencia profesional; mientras que las variables situacionales consideradas fueron: fase de competición, localización, clasificación, tiempo restante, y diferencia parcial en el marcador. Para el análisis de los datos se realizaron dos modelos estadísticos: 1º) un modelo de regresión lineal múltiple para conocer el efecto de las variables independientes en el porcentaje de aciertos del lanzador en el último minuto, y en la última serie de tiros libres del partido; y 2º) un análisis de regresión logística binomial para analizar la relación existente entre la probabilidad de tener un drop (disminución del rendimiento) y las características del lanzador, y las variables situacionales. Los resultados del modelo de regresión lineal múltiple mostraron efectos negativos significativos en el porcentaje de acierto en los tiros libres durante el último minuto, cuando los lanzadores son los pívots (-19,45%). Por otro lado, los resultados durante la última serie mostraron el efecto negativo significativo sobre la posición de pívot (- 19,30%) y la diferencia parcial en el marcador (-3,33%, para cada punto de diferencia en el marcador) en el porcentaje de acierto en los tiros libres. Las variables independientes edad, experiencia profesional, clasificación en la liga regular, fase de competición, localización, y tiempo restante, no revelaron efectos significativos en los modelos de regresión lineal. Los resultados de la regresión logística binomial revelaron que las variables experiencia profesional entre 13 y 18 años (OR = 4,63), jugar de alero (OR = 23,01), y jugar de base (OR = 10,68) están relacionadas con una baja probabilidad de disminuir el rendimiento durante el último minuto del partido; mientras que ir ganando, aumenta esta probabilidad (OR = 0,06). Además, los resultados de la última serie mostraron una menor disminución del rendimiento del jugador cuando tiene entre 13 y 18 años de experiencia (OR = 4,28), y juega de alero (OR = 8,06) o base (OR = 6,34). Por el contrario, las variables situacionales relacionadas con esa disminución del rendimiento del jugador son las fases eliminatorias (OR = 0,22) e ir ganando (OR = 0,04). Los resultados principales del estudio mostraron que existe una disminución del rendimiento del jugador en su porcentaje de acierto en los tiros libres durante el último minuto y en la última serie de lanzamientos del partido, y que está relacionada significativamente con la edad, experiencia profesional, puesto de juego del jugador, y diferencia parcial en el marcador. Encontrando relación también con la fase de competición, durante la última serie de tiros libres del partido. Esta información supone una valiosa información para el entrenador, y su aplicación en el ámbito competitivo real. En este sentido, la creación de simulaciones en el apartado de aplicaciones prácticas, permite predecir el porcentaje de acierto en los tiros libres de un jugador durante los momentos de mayor presión del partido, en base a su perfil de rendimiento. Lo que puede servir para realizar una toma de decisiones más idónea, con el objetivo de lograr el mejor resultado. Del mismo modo, orienta el tipo de proceso de entrenamiento que se ha de seguir, en relación a los jugadores más tendentes al drop, con el objetivo de minimizar el efecto de la presión sobre su capacidad para rendir adecuadamente en la ejecución de los tiros libres, y lograr de esta manera un rendimiento más homogéneo en todos los jugadores del equipo en esta faceta del juego, durante el momento crítico del final de partido. ABSTRACT. The final minute of a close game in basketball is a critical moment which is subject to many factors that influence its development. Thus, the success rate in free-throws during that period will determine, in many cases, the outcome of the game. Decrease of performance (drop) in this facet of play under pressure conditions, may be related to studied own multiple sports context variables, such as the remaining seconds of possession, the situation in the score (to be winning, drawing, or losing) the location of the match (playing at home or away), the competition phase (regular season or playoffs) or team level (best/worst teams). In addition, the characteristics of the player are very important related to his age and years of experience to face the critical moments, as well as his playing position into team. In this sense, the combination of factors in context and player, allows interact about performance of shooter in the final moments of the game during his free-throw shooting. The aim of this present doctoral thesis was find the most related variables to player´s drop in free throws in the last minute of the game and the last row of free-throws in closed games of basketball. To achieve the objective of the study, 124 closed games (less or equal than 2 points difference) were analyzed in every copetition in ACB league (regular season, playoff and cup) from 2011-2012 to 2014-2015 seasons. To record the variables, the percentage of success of the shooter in regular season, full game, last minute, and last row were analyzed. This way, it is tried to analyze which player and context variables explain the free-throw performance in last minute and last row of the game. On the other hand, it is tried to determine the degree of association between decrease of performance (drop) of the player in the final moments, and studied player variables: playing position, age, and years of professional experience; while considered situational variables considered were: competition phase, location, classification, remaining time, and score-line. For data analysis were performed two statistical models: 1) A multiple linear regression model to determine the effect of the independent variables in the succsess percentage of shooter at the last minute, and in the last row of free-throws in the game; and 2) A binomial logistic regression analysis to analyze the relationship between the probability of a drop (lower performance) and the characteristics of the shooter and situational variables. The results of multiple linear regression model showed significant negative effects on the free-throw percentage during last minute, when shooters are centers (-19.45%). On the other hand, results in the last series showed the significant negative effect on the center position (-19.30%) and score-line (-3.33% for each point difference in the score) in the free-throw percentage. The independent variables age, professional experience, ranking in the regular season, competition phase, location, and remaining time, revealed no significant effects on linear regression models. The results of the binomial logistic regression showed that the variables professional experience between 13 and 18 years (OR = 4.63), playing forward (OR = 23.01) and playing guard (OR = 10.68) are related to reduce the probability to decrease the performance during the last minute of the game. While wining, increases it (OR = 0.06). Furthermore, the results of the last row showed a reduction in performance degradation when player is between 13 and 18 years of experience (OR = 4.28), and playing forward (OR = 8.06) or guard (OR = 6.34). By contrast, the variables related to the decrease in performance of the player are the knockout phases (OR = 0.22) and wining (OR = 0.04). The main results of the study showed that there is a decrease in performance of the player in the percentage of success in free-throws in the last minute and last row of the game, and it is significantly associated with age, professional experience, and player position. Finding relationship with the competition phase, during last row of free-throws of the game too. This information is a valuable information for the coach, for applying in real competitive environment. In this sense, create simulations in the section of practical applications allows to predict the success rate of free-throw of a player during the most pressing moments of the game, based on their performance profile. What can be used to take more appropriate decisions in order to achieve the best result. Similarly, guides the type of training process must be followed in relation to the most favorable players to drop, in order to minimize the effect of pressure on their ability to perform properly in the execution of the free-throws. And to achieve, in this way, a more consistent performance in all team players in this facet of the game, during the critical moment in the final of the game.