10 resultados para Human factor
em Universidad Politécnica de Madrid
Resumo:
This paper shows the importance of a holistic comprehension of the Earth as a living planet, where man inhabits and is exposed to environmental incidences of different nature. The aim of the paper here summarized is a reflection on all these concepts and scientific considerations related to the important role of men in the handling of natural hazards. Our Planet is an unstable and dynamical system highly sensitive to initial conditions, as proposed by Chaos theory (González-Miranda 2004); it is a complex organic whole, which responds to minimal variations which can affect several natural phenomena such as plate tectonics, solar flares, fluid turbulences, landscape formation, forest fires, growth and migration of populations and biological evolution. This is known as the “butterfly effect” (Lorenz 1972), which means that a small change of the system causes a chain of events leading to large-scale unpredictable consequences. The aim of this work is dwelling on the importance of the knowledge of these natural and catastrophic geological, biological and human systems so much sensible to equilibrium conditions, to prevent, avoid and mend their effects, and to face them in a resilient way
Resumo:
This paper is based on the following postulates taken from a book recently published by this author (Sáez-Vacas, 1990(1)): a) technological innovation in a company is understood to be the process and set of changes that the company undergoes as a result of a specific type of technology; b) the incorporation of technology in the company does not necessarily result in innovation, modernization and progress; c) the very words "modernization" and "progress" are completely bereft of any meaning if isolated from the concept of complexity in its broadest sense, including the human factor. Turning to office technology in specific, the problem of managing office technology for business innovation purposes can be likened to the problem of managing third level complexity, following the guidelines of a three-level complexity model proposed by the author some years ago
Resumo:
Processes of founding and expanding cities in coastal areas have undergone great changes over time driven by environmental conditions. Coastal settlements looked for places above flood levels and away from swamps and other wetlands whenever possible. As populations grew, cities were extending trying to avoid low and wet lands. No city has been able to limit its growth. The risk of flooding can never be eliminated, but only reduced to the extent possible. Flooding of coastal areas is today dramatically attributed to eustasic sea level rise caused by global climate change. This can be inaccurate. Current climate change is generating an average sea level upward trend, but other regional and local factors result in this trend being accentuated in some places or attenuated, and even reversed, in others. Then, the intensity and frequency of coastal flooding around the planet, although not so much as a unique result of this general eustasic elevation, but rather of the superposition of marine and crustal dynamic elements, the former also climate-related, which give rise to a temporary raising in average sea level in the short term. Since the Little Ice Age the planet has been suffering a global warming change leading to sea level rise. The idea of being too obeying to anthropogenic factors may be attributed to Arrhenius (1896), though it is of much later highlight after the sixties of the last century. Never before, the human factor had been able of such an influence on climate. However, other types of changes in sea levels became apparent, resulting from vertical movements of the crust, modifications of sea basins due to continents fracturing, drifting and coming together, or to different types of climate patterns. Coastal zones are then doubly susceptible to floods. Precipitation immediately triggers pluvial flooding. If it continues upland or when snow and glaciers melt eventually fluvial flooding can occur. The urban development presence represents modifying factors. Additional interference is caused by river and waste water drainage systems. Climate also influences sea levels in coastal areas, where tides as well as the structure and dynamic of the geoid and its crust come into play. From the sea, waters can flood and break or push back berms and other coastline borders. The sea level, controlling the mouth of the main channel of the basin's drainage system, is ultimately what governs flood levels. A temporary rise in sea level acts as a dam at the mouth. Even in absence of that global change, so, floods are likely going to increase in many urban coastal areas. Some kind of innovative methodologies and practices should be needed to get more flood resilience cities
Resumo:
This paper introduces the experience of using videoconferencing and recording as a mechanism to support courses which need to be promoted or discontinued within the framework of the European convergence process. Our objective is to make these courses accessible as live streaming during the lessons as well as recorded lectures and associated documents available to the students as soon as the lesson has finished. The technology used has been developed in our university and it is all open source. Although this is a technical project the key is the human factor involved. The people managing the virtual sessions are students of the courses being recorded. However, they lack technical knowledge, so we had to train them in audiovisuals and enhance the usability of the videoconferencing tool and platform. The validation process is being carried out in five real scenarios at our university. During the whole period we are evaluating technical and pedagogical issues of this experience for both students and teachers to guide the future development of the service. Depending on the final results, the service of lectures recording will be available as educational resource for all of the teaching staff of our university.
Resumo:
Any new hospital communication architecture has to support existing services, but at the same time new added features should not affect normal tasks. This article deals with issues regarding old and new systems’ interoperability, as well as the effect the human factor has in a deployed architecture. It also presents valuable information, which is a product of a real scenario. Tracking services are also tested in order to monitor and administer several medical resources.
Resumo:
Las organizaciones son sistemas o unidades sociales, compuestas por personas que interactúan entre sí, para lograr objetivos comunes. Uno de sus objetivos es la productividad. La productividad es un constructo multidimensional en la que influyen aspectos tecnológicos, económicos, organizacionales y humanos. Diversos estudios apoyan la influencia de la motivación de las personas, de las habilidades y destrezas de los individuos, de su talento para desempeñar el trabajo, así como también del ambiente de trabajo presente en la organización, en la productividad. Por esta razón, el objetivo general de la investigación, es analizar la influencia entre los factores humanos y la productividad. Se hará énfasis en la persona como factor productivo clave, para responder a las interrogantes de la investigación, referidas a cuáles son las variables humanas que inciden en la productividad, a la posibilidad de plantear un modelo de productividad que considere el impacto del factor humano y la posibilidad de encontrar un método para la medición de la productividad que contemple la percepción del factor humano. Para resolver estas interrogantes, en esta investigación se busca establecer las relaciones entre las variables humanas y la productividad, vistas desde la perspectiva de tres unidades de análisis diferentes: individuo, grupo y organización, para la formulación de un modelo de productividad humana y el diseño de un instrumento para su medida. Una de las principales fuente de investigación para la elección de las variables humanas, la formulación del modelo, y el método de medición de la productividad, fue la revisión de la literatura disponible sobre la productividad y el factor humano en las organizaciones, lo que facilitó el trazado del marco teórico y conceptual. Otra de las fuentes para la selección fue la opinión de expertos y de especialistas directamente involucrados en el sector eléctrico venezolano, lo cual facilitó la obtención de un modelo, cuyas variables reflejasen la realidad del ámbito en estudio. Para aportar una interpretación explicativa del fenómeno, se planteó el modelo de los Factores Humanos vs Productividad (MFHP), el cual se analizó desde la perspectiva del análisis causal y fue conformado por tres variables latentes exógenas denominadas: factores individuales, factores grupales y factores organizacionales, que estaban relacionadas con una variable latente endógena denominada productividad. El MFHP se formuló mediante la metodología de los modelos de ecuaciones estructurales (SEM). Las relaciones inicialmente propuestas entre las variables latentes fueron corroboradas por los ajustes globales del modelo, se constataron las relaciones entre las variables latentes planteadas y sus indicadores asociados, lo que facilitó el enunciado de 26 hipótesis, de las cuales se comprobaron 24. El modelo fue validado mediante la estrategia de modelos rivales, utilizada para comparar varios modelos SEM, y seleccionar el de mejor ajuste, con sustento teórico. La aceptación del modelo se realizó mediante la evaluación conjunta de los índices de bondad de ajuste globales. Asimismo, para la elaboración del instrumento de medida de la productividad (IMPH), se realizó un análisis factorial exploratorio previo a la aplicación del análisis factorial confirmatorio, aplicando SEM. La revisión de los conceptos de productividad, la incidencia del factor humano, y sus métodos de medición, condujeron al planteamiento de métodos subjetivos que incorporaron la percepción de los principales actores del proceso productivo, tanto para la selección de las variables, como para la formulación de un modelo de productividad y el diseño de un instrumento de medición de la productividad. La contribución metodológica de este trabajo de investigación, ha sido el empleo de los SEM para relacionar variables que tienen que ver con el comportamiento humano en la organización y la productividad, lo cual abre nuevas posibilidades a la investigación en este ámbito. Organizations are social systems or units composed of people who interact with each other to achieve common goals. One objective is productivity, which is a multidimensional construct influenced by technological, economic, organizational and human aspects. Several studies support the influence on productivity of personal motivation, of the skills and abilities of individuals, of their talent for the job, as well as of the work environment present in the organization. Therefore, the overall objective of this research is to analyze the influence between human factors and productivity. The emphasis is on the individual as a productive factor which is key in order to answer the research questions concerning the human variables that affect productivity and to address the ability to propose a productivity model that considers the impact of the human factor and the possibility of finding a method for the measurement of productivity that includes the perception of the human factor. To consider these questions, this research seeks to establish the relationships between human and productivity variables, as seen from the perspective of three different units of analysis: the individual, the group and the organization, in order to formulate a model of human productivity and to design an instrument for its measurement. A major source of research for choosing the human variables, model formulation, and method of measuring productivity, was the review of the available literature on productivity and the human factor in organizations which facilitated the design of the theoretical and conceptual framework. Another source for the selection was the opinion of experts and specialists directly involved in the Venezuelan electricity sector which facilitated obtaining a model whose variables reflect the reality of the area under study. To provide an interpretation explaining the phenomenon, the model of the Human Factors vs. Productivity Model (HFPM) was proposed. This model has been analyzed from the perspective of causal analysis and was composed of three latent exogenous variables denominated: individual, group and organizational factors which are related to a latent variable denominated endogenous productivity. The HFPM was formulated using the methodology of Structural Equation Modeling (SEM). The initially proposed relationships between latent variables were confirmed by the global fits of the model, the relationships between the latent variables and their associated indicators enable the statement of 26 hypotheses, of which 24 were confirmed. The model was validated using the strategy of rival models, used for comparing various SEM models and to select the one that provides the best fit, with theoretical support. The acceptance of the model was performed through the joint evaluation of the adequacy of global fit indices. Additionally, for the development of an instrument to measure productivity, an exploratory factor analysis was performed prior to the application of a confirmatory factor analysis, using SEM. The review of the concepts of productivity, the impact of the human factor, and the measurement methods led to a subjective methods approach that incorporated the perception of the main actors of the production process, both for the selection of variables and for the formulation of a productivity model and the design of an instrument to measure productivity. The methodological contribution of this research has been the use of SEM to relate variables that have to do with human behavior in the organization and with productivity, opening new possibilities for research in this area.
Resumo:
The scope of the present paper is the derivation of a merit function which predicts the visual perception of LED spot lights. The color uniformity level Usl is described by a linear regression function of the spatial color distribution in the far field. Hereby, the function is derived from four basic functions. They describe the color uniformity of spot lights through different features. The result is a reliable prediction for the perceived color uniformity in spot lights. A human factor experiment was performed to evaluate the visual preferences for colors and patterns. A perceived rank order was derived from the subjects’ answers and compared with the four basic functions. The correlation between the perceived rank order and the basic functions was calculated resulting in the definition of the merit function Usl. The application of this function is shown by a comparison of visual evaluations and measurements of LED retrofit spot lamps. The results enable a prediction of color uniformity levels of simulations and measurements concerning the visual perception. The function provides a possibility to evaluate the far field of spot lights without individual subjective judgment. © (2014) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Resumo:
Spotlighting is one illumination field where the application of light emitting diodes (LED) creates many advantages. Commonly, the system for spot lights consists of a LED light engine and collimating secondary optics. Through angular or spatial separated emitted light from the source and imaging optical elements, a non uniform far field appears with colored rings, dots or patterns. Many feasible combinations result in very different spatial color distributions. Several combinations of three multi-chip light sources and secondary optical elements like reflectors and TIR lenses with additional facets or scattering elements were analyzed mainly regarding the color uniformity. They are assessed by the merit function Usl which was derived from human factor experiments and describes the color uniformity based on the visual perception of humans. Furthermore, the optical systems are compared concerning efficiency, peak candela and aspect ratio. Both types of optics differ in the relation between the color uniformity level and other properties. A plain reflector with a slightly color mixing light source performs adequate. The results for the TIR lenses indicate that they need additional elements for good color mixing or blended light source. The most convenient system depends on the requirements of the application.
Resumo:
La iluminación con diodos emisores de luz (LED) está reemplazando cada vez en mayor medida a las fuentes de luz tradicionales. La iluminación LED ofrece ventajas en eficiencia, consumo de energía, diseño, tamaño y calidad de la luz. Durante más de 50 años, los investigadores han estado trabajando en mejoras LED. Su principal relevancia para la iluminación está aumentando rápidamente. Esta tesis se centra en un campo de aplicación importante, como son los focos. Se utilizan para enfocar la luz en áreas definidas, en objetos sobresalientes en condiciones profesionales. Esta iluminación de alto rendimiento requiere una calidad de luz definida, que incluya temperaturas ajustables de color correlacionadas (CCT), de alto índice de reproducción cromática (CRI), altas eficiencias, y colores vivos y brillantes. En el paquete LED varios chips de diferentes colores (rojo, azul, fósforo convertido) se combinan para cumplir con la distribución de energía espectral con alto CRI. Para colimar la luz en los puntos concretos deseados con un ángulo de emisión determinado, se utilizan blancos sintonizables y diversos colores de luz y ópticas secundarias. La combinación de una fuente LED de varios colores con elementos ópticos puede causar falta de homogeneidad cromática en la distribución espacial y angular de la luz, que debe resolverse en el diseño óptico. Sin embargo, no hay necesidad de uniformidad perfecta en el punto de luz debido al umbral en la percepción visual del ojo humano. Por lo tanto, se requiere una descripción matemática del nivel de uniformidad del color con respecto a la percepción visual. Esta tesis está organizada en siete capítulos. Después de un capítulo inicial que presenta la motivación que ha guiado la investigación de esta tesis, en el capítulo 2 se presentan los fundamentos científicos de la uniformidad del color en luces concentradas, como son: el espacio de color aplicado CIELAB, la percepción visual del color, los fundamentos de diseño de focos respecto a los motores de luz y ópticas no formadoras de imágenes, y los últimos avances en la evaluación de la uniformidad del color en el campo de los focos. El capítulo 3 desarrolla diferentes métodos para la descripción matemática de la distribución espacial del color en un área definida, como son la diferencia de color máxima, la desviación media del color, el gradiente de la distribución espacial de color, así como la suavidad radial y axial. Cada función se refiere a los diferentes factores que influyen en la visión, los cuales necesitan un tratamiento distinto que el de los datos que se tendrán en cuenta, además de funciones de ponderación que pre- y post-procesan los datos simulados o medidos para la reducción del ruido, la luminancia de corte, la aplicación de la ponderación de luminancia, la función de sensibilidad de contraste, y la función de distribución acumulativa. En el capítulo 4, se obtiene la función de mérito Usl para la estimación de la uniformidad del color percibida en focos. Se basó en los resultados de dos conjuntos de experimentos con factor humano realizados para evaluar la percepción visual de los sujetos de los patrones de focos típicos. El primer experimento con factor humano dio lugar al orden de importancia percibida de los focos. El orden de rango percibido se utilizó para correlacionar las descripciones matemáticas de las funciones básicas y la función ponderada sobre la distribución espacial del color, que condujo a la función Usl. El segundo experimento con factor humano probó la percepción de los focos bajo condiciones ambientales diversas, con el objetivo de proporcionar una escala absoluta para Usl, para poder así sustituir la opinión subjetiva personal de los individuos por una función de mérito estandarizada. La validación de la función Usl se presenta en relación con el alcance de la aplicación y condiciones, así como las limitaciones y restricciones que se realizan en el capítulo 5. Se compararon los datos medidos y simulados de varios sistemas ópticos. Se discuten los campos de aplicación , así como validaciones y restricciones de la función. El capítulo 6 presenta el diseño del sistema de focos y su optimización. Una evaluación muestra el análisis de sistemas basados en el reflector y la lente TIR. Los sistemas ópticos simulados se comparan en la uniformidad del color Usl, sensibilidad a las sombras coloreadas, eficiencia e intensidad luminosa máxima. Se ha comprobado que no hay un sistema único que obtenga los mejores resultados en todas las categorías, y que una excelente uniformidad de color se pudo alcanzar por la conjunción de dos sistemas diferentes. Finalmente, el capítulo 7 presenta el resumen de esta tesis y la perspectiva para investigar otros aspectos. ABSTRACT Illumination with light-emitting diodes (LED) is more and more replacing traditional light sources. They provide advantages in efficiency, energy consumption, design, size and light quality. For more than 50 years, researchers have been working on LED improvements. Their main relevance for illumination is rapidly increasing. This thesis is focused on one important field of application which are spotlights. They are used to focus light on defined areas, outstanding objects in professional conditions. This high performance illumination required a defined light quality including tunable correlated color temperatures (CCT), high color rendering index (CRI), high efficiencies and bright, vivid colors. Several differently colored chips (red, blue, phosphor converted) in the LED package are combined to meet spectral power distribution with high CRI, tunable white and several light colors and secondary optics are used to collimate the light into the desired narrow spots with defined angle of emission. The combination of multi-color LED source and optical elements may cause chromatic inhomogeneities in spatial and angular light distribution which needs to solved at the optical design. However, there is no need for perfect uniformity in the spot light due to threshold in visual perception of human eye. Therefore, a mathematical description of color uniformity level with regard to visual perception is required. This thesis is organized seven seven chapters. After an initial one presenting the motivation that has guided the research of this thesis, Chapter 2 introduces the scientific basics of color uniformity in spot lights including: the applied color space CIELAB, the visual color perception, the spotlight design fundamentals with regards to light engines and nonimaging optics, and the state of the art for the evaluation of color uniformity in the far field of spotlights. Chapter 3 develops different methods for mathematical description of spatial color distribution in a defined area, which are the maximum color difference, the average color deviation, the gradient of spatial color distribution as well as the radial and axial smoothness. Each function refers to different visual influencing factors, and they need different handling of data be taken into account, along with weighting functions which pre- and post-process the simulated or measured data for noise reduction, luminance cutoff, the implementation of luminance weighting, contrast sensitivity function, and cumulative distribution function. In chapter 4, the merit function Usl for the estimation of the perceived color uniformity in spotlights is derived. It was based on the results of two sets of human factor experiments performed to evaluate the visual perception of typical spotlight patterns by subjects. The first human factor experiment resulted in the perceived rank order of the spotlights. The perceived rank order was used to correlate the mathematical descriptions of basic functions and weighted function concerning the spatial color distribution, which lead to the Usl function. The second human factor experiment tested the perception of spotlights under varied environmental conditions, with to objective to provide an absolute scale for Usl, so the subjective personal opinion of individuals could be replaced by a standardized merit function. The validation of the Usl function is presented concerning the application range and conditions as well as limitations and restrictions in carried out in chapter 5. Measured and simulated data of various optical several systems were compared. Fields of applications are discussed as well as validations and restrictions of the function. Chapter 6 presents spotlight system design and their optimization. An evaluation shows the analysis of reflector-based and TIR lens systems. The simulated optical systems are compared in color uniformity Usl , sensitivity to colored shadows, efficiency, and peak luminous intensity. It has been found that no single system which performed best in all categories, and that excellent color uniformity could be reached by two different system assemblies. Finally, chapter 7 summarizes the conclusions of the present thesis and an outlook for further investigation topics.
Resumo:
El planteamiento tradicional de análisis de la accidentalidad en carretera pasa por la consideración de herramientas paliativas, como son la identificación y gestión de los puntos negros o tramos de concentración de accidentes, o preventivas, como las auditorías e inspecciones de seguridad vial. En esta tesis doctoral se presenta un planteamiento complementario a estas herramientas, desde una perspectiva novedosa: la consideración de los tramos donde no se producen accidentes; son los denominados Tramos Blancos. La tesis persigue demostrar que existen determinados parámetros del diseño de las carreteras y del tráfico que, bajo características generales similares de las vías, tienen influencia en el hecho de que se produzcan o no accidentes, adicionalmente a la exposición al riesgo, como factor principal, y a otros factores. La propia definición de los Tramos Blancos, entendidos como tramos de carreteras de longitud representativa donde no se han producido accidentes con víctimas mortales o heridos graves durante un periodo largo de tiempo, garantiza que esta situación no se produzca como consecuencia de la aleatoriedad de los accidentes, sino que pudiera deberse a una confluencia específica de determinados parámetros de la geometría de la vía y del tráfico total y de vehículos pesados. Para el desarrollo de esta investigación se han considerado la red de autopistas de peaje y las carreteras convencionales de la Red del Estado de España, que supone un total de 17.000 kilómetros, y los datos de accidentes con víctimas mortales y heridos graves en el periodo 2006-2010, ambos incluidos, en estas redes (un total de 10.000 accidentes). La red viaria objeto de análisis supone el 65% de la longitud de la Red de Carreteras del Estado, por la que circula el 33% de su tráfico; en ella se produjeron en el año 2013 el 47% de los accidentes con víctimas y el 60% de las víctimas mortales de la Red de Carreteras del Estado. Durante la investigación se ha desarrollado una base de datos de 250.130 registros y más de 3.5 millones de datos en el caso de las autopistas de peaje de la Red de Carreteras del Estado y de 935.402 registros y más de 14 millones de datos en el caso de la red convencional del Estado analizada. Tanto las autopistas de peaje como las carreteras convencionales han sido clasificadas según sus características de tráfico, de manera que se valoren vías con nivel de exposición al riesgo similar. Para cada tipología de vía, se ha definido como longitud de referencia para que un tramo se considere Tramo Blanco la longitud igual al percentil 95 de las longitudes de tramos sin accidentes con heridos graves o víctimas mortales durante el periodo 2006-2010. En el caso de las autopistas de peaje, en la tipología que ha sido considerada para la definición del modelo, esta longitud de referencia se estableció en 14.5 kilómetros, mientras que en el caso de las carreteras convencionales, se estableció en 7.75 kilómetros. Para cada uno de los tipos de vía considerados se han construido una base de datos en la que se han incluido las variables de existencia o no de Tramo Blanco, así como las variables de tráfico (intensidad media diaria total, intensidad de vehículos pesados y porcentaje de vehículos pesados ), la velocidad media y las variables de geometría (número de carriles, ancho de carril, ancho de arcén derecho e izquierdo, ancho de calzada y plataforma, radio, peralte, pendiente y visibilidad directa e inversa en los casos disponibles); como variables adicionales, se han incluido el número de accidentes con víctimas, los fallecidos y heridos graves, índices de peligrosidad, índices de mortalidad y exposición al riesgo. Los trabajos desarrollados para explicar la presencia de Tramos Blancos en la red de autopistas de peaje han permitido establecer las diferencias entre los valores medios de las variables de tráfico y diseño geométrico en Tramos Blancos respecto a tramos no blancos y comprobar que estas diferencias son significativas. Así mismo, se ha podido calibrar un modelo de regresión logística que explica parcialmente la existencia de Tramos Blancos, para rangos de tráfico inferiores a 10.000 vehículos diarios y para tráficos entre 10.000 y 15.000 vehículos diarios. Para el primer grupo (menos de 10.000 vehículos al día), las variables que han demostrado tener una mayor influencia en la existencia de Tramo Blanco son la velocidad media de circulación, el ancho de carril, el ancho de arcén izquierdo y el porcentaje de vehículos pesados. Para el segundo grupo (entre 10.000 y 15.000 vehículos al día), las variables independientes más influyentes en la existencia de Tramo Blanco han sido la velocidad de circulación, el ancho de calzada y el porcentaje de vehículos pesados. En el caso de las carreteras convencionales, los diferentes análisis realizados no han permitido identificar un modelo que consiga una buena clasificación de los Tramos Blancos. Aun así, se puede afirmar que los valores medios de las variables de intensidad de tráfico, radio, visibilidad, peralte y pendiente presentan diferencias significativas en los Tramos Blancos respecto a los no blancos, que varían en función de la intensidad de tráfico. Los resultados obtenidos deben considerarse como la conclusión de un análisis preliminar, dado que existen otros parámetros, tanto de diseño de la vía como de la circulación, el entorno, el factor humano o el vehículo que podrían tener una influencia en el hecho que se analiza, y no se han considerado por no disponer de esta información. En esta misma línea, el análisis de las circunstancias que rodean al viaje que el usuario de la vía realiza, su tipología y motivación es una fuente de información de interés de la que no se tienen datos y que permitiría mejorar el análisis de accidentalidad en general, y en particular el de esta investigación. Adicionalmente, se reconocen limitaciones en el desarrollo de esta investigación, en las que sería preciso profundizar en el futuro, reconociendo así nuevas líneas de investigación de interés. The traditional approach to road accidents analysis has been based in the use of palliative tools, such as black spot (or road sections) identification and management, or preventive tools, such as road safety audits and inspections. This thesis shows a complementary approach to the existing tools, from a new perspective: the consideration of road sections where no accidents have occurred; these are the so-called White Road Sections. The aim of this thesis is to show that there are certain design parameters and traffic characteristics which, under similar circumstances for roads, have influence in the fact that accidents occur, in addition to the main factor, which is the risk exposure, and others. White Road Sections, defined as road sections of a representative length, where no fatal accidents or accidents involving serious injured have happened during a long period of time, should not be a product of randomness of accidents; on the contrary, they might be the consequence of a confluence of specific parameters of road geometry, traffic volumes and heavy vehicles traffic volumes. For this research, the toll motorway network and single-carriageway network of the Spanish National Road Network have been considered, which is a total of 17.000 kilometers; fatal accidents and those involving serious injured from the period 2006-2010 have been considered (a total number of 10.000 accidents). The road network covered means 65% of the total length of the National Road Network, which allocates 33% of traffic volume; 47% of accidents with victims and 60% of fatalities happened in these road networks during 2013. During the research, a database of 250.130 registers and more than 3.5 million data for toll motorways and 935.042 registers and more than 14 million data for single carriageways of the National Road Network was developed. Both toll motorways and single-carriageways have been classified according to their traffic characteristics, so that the analysis is performed over roads with similar risk exposure. For each road type, a reference length for White Road Section has been defined, as the 95 percentile of all road sections lengths without accidents (with fatalities or serious injured) for 2006-2010. For toll motorways, this reference length concluded to be 14.5 kilometers, while for single-carriageways, it was defined as 7.75 kilometers. A detailed database was developed for each type of road, including the variable “existence of White Road Section”, as well as variables of traffic (average daily traffic volume, heavy vehicles average daily traffic and percentage of heavy vehicles from the total traffic volume), average speed and geometry variables (number of lanes, width of lane, width of shoulders, carriageway width, platform width, radius, superelevation, slope and visibility); additional variables, such as number of accidents with victims, number of fatalities or serious injured, risk and fatality rates and risk exposure, have also been included. Research conducted for the explanation of the presence of White Road Sections in the toll motorway network have shown statistically significant differences in the average values of variables of traffic and geometric design in White Road Sections compared with other road sections. In addition, a binary logistic model for the partial explanation of the presence of White Road Sections was developed, for traffic volumes lower than 10.000 daily vehicles and for those running from 10.000 to 15.000 daily vehicles. For the first group, the most influent variables for the presence of White Road Sections were the average speed, width of lane, width of left shoulder and percentage of heavy vehicles. For the second group, the most influent variables were found to be average speed, carriageway width and percentage of heavy vehicles. For single-carriageways, the different analysis developed did not reach a proper model for the explanation of White Road Sections. However, it can be assumed that the average values of the variables of traffic volume, radius, visibility, superelevation and slope show significant differences in White Road Sections if compared with others, which also vary with traffic volumes. Results obtained should be considered as a conclusion of a preliminary analysis, as there are other parameters, not only design-related, but also regarding traffic, environment, human factor and vehicle which could have an influence in the fact under research, but this information has not been considered in the analysis, as it was not available. In parallel, the analysis of the circumstances around the trip, including its typology and motivation is an interesting source of information, from which data are not available; the availability of this information would be useful for the improvement of accident analysis, in general, and for this research work, in particular. In addition, there are some limitations in the development of the research work; it would be necessary to develop an in-depth analysis in the future, thus assuming new research lines of interest.