52 resultados para Involuntary placement in case of emergency
em Universidad Politécnica de Madrid
Resumo:
Esta Tesis Doctoral tiene como principal objetivo el obtener una cadena de tratamientos seguros de aguas seriados que nos permita asegurar la calidad de las aguas para consumo humano en caso de emergencias, de tal forma que se minimicen los efectos de acciones hostiles, como sabotajes o actos terroristas, desastres naturales, etc y buscar soluciones adecuadas para garantizar en este caso la salud. Las plantas de tratamientos de aguas existentes comercialmente no aseguran dicha calidad y la documentación sobre el tema presenta vacíos de conocimiento, contradicciones entre resultados de investigaciones o insostenibilidad de conclusiones de las mismas. Estas carencias nos permiten determinar los aspectos a tratar durante la investigación. Por ello, este objetivo se concretó en tres acciones: Investigar sobre rendimientos de plantas convencionales en eliminación de microorganismos y productos tóxicos y peligrosos. Introducir mejoras que garanticen el rendimiento de las plantas convencionales. Investigar sobre la conveniencia de complementar las instalaciones existentes buscando seguridad y garantía sanitaria. Y se desarrollaron tres líneas de investigación: LI 1 “Inorgánicos”: Investigación sobre la eliminación de los metales boro, cobre y molibdeno mediante procesos de intercambio iónico y de coagulaciónfloculación- decantación. LI 2 “Compuestos Orgánicos Volátiles”: Investigación sobre la eliminación de los compuestos orgánicos 1,1 dicloroetano, 1,2 dicloroetano, clorobenceno, 1,3 dicloropropeno y hexacloro 1,3 butadieno mediante procesos de carbón activo granular y de oxidación avanzada. LI 3 “Plantas portátiles”: Investigación sobre plantas existentes portátiles para verificar su rendimiento teórico y proponer mejoras. Estas líneas de investigación se desarrollaron tanto en el nivel teórico como en el empírico, bien sea en laboratorio como en campo. A lo largo del documento se demuestra que las principales fuentes de contaminación, salvo la degradación de yacimientos naturales, proceden de la actividad humana (efluentes industriales y agrícolas, aguas residuales y actividades beligerantes) que provocan un amplio espectro de enfermedades por lo que dificultan tanto la definición de la fuente como la anticipada detección de la enfermedad. Las principales conclusiones que se obtuvieron están relacionadas con el rendimiento de eliminación de los parámetros tras la aplicación de los procesos y plantas de tratamiento de aguas anteriormente reseñadas. Sin embargo, el verdadero elemento designador de originalidad de esta Tesis Doctoral, tal como se ha reseñado arriba, radica en la definición de un sistema seriado de procesos de tratamiento de aguas que asegura la calidad en caso de emergencia. Éste se define en el siguiente orden: pretratamiento, oxidación, coagulación-floculación-decantación, filtración por arena, intercambio iónico, carbón activo granular, microfiltración, radiación UV, ósmosis inversa, radiación UV y cloración final. The main objective of this Thesis is to obtain a chain of stepwise safe water treatments that allow us to ensure the quality of water for human consumption in case of emergencies, so that the effects of hostile actions, such as sabotage or terrorism, natural disasters, etc. and seek appropriate solutions in this case to ensure health. The existing commercial water treatment plants do not ensure quality, and the documentation on the subject presents knowledge gaps or contradictions. These gaps allow us to determine the issues to be discussed during the investigation. Therefore, this objective was manifested in three actions: Researching yields in commercial plants and microorganisms, or toxic and dangerous products removal. Improvements to ensure the performance of conventional plants. Inquire about the advisability of implementing existing facilities for safety and health guarantee. And three lines of research are developed: LI 1 “Inorganic elements”: Research removing metals iron, copper and molybdenum by ion exchange processes and coagulation-flocculation-decantation. LI 2 “Volatile Organic Compounds”: Research removing organic compounds 1,1 dichloroethane, 1,2 dichloroethane, chlorobenzene, 1,3-dichloropropene and 1,3-butadiene hexachloro through processes of granular activated carbon and advanced oxidation. LI 3 “Compact Water Treatment Plants”: Research on existing packaged plants to verify theoretical performance and suggest improvements. These lines of research are developed both theoretically and empirically, both in the laboratory and in the field. Throughout the document, it is evident that the main sources of pollution, other than the degradation of natural deposits, come from human activity (industrial and agricultural effluents, sewage and belligerent activities) which cause a broad spectrum of diseases which hamper both the definition of the source and the early detection of the disease. The main conclusions drawn are related to both the removal efficiency parameters after application of processes and treatment plants outlined above water. However, the real designator of originality of this thesis, such as outlined above, lies in the definition of a serial system water treatment processes assuring quality in case of emergency. This is defined in the following order: pretreatment, oxidation, coagulation-flocculation-sedimentation, sand filtration, ion exchange, granular activated carbon, microfiltration, UV radiation, reverse osmosis, UV radiation and final chlorination.
Resumo:
Emergency management is one of the key aspects within the day-to-day operation procedures in a highway. Efficiency in the overall response in case of an incident is paramount in reducing the consequences of any incident. However, the approach of highway operators to the issue of incident management is still usually far from a systematic, standardized way. This paper attempts to address the issue and provide several hints on why this happens, and a proposal on how the situation could be overcome. An introduction to a performance based approach to a general system specification will be described, and then applied to a particular road emergency management task. A real testbed has been implemented to show the validity of the proposed approach. Ad-hoc sensors (one camera and one laser scanner) were efficiently deployed to acquire data, and advanced fusion techniques applied at the processing stage to reach the specific user requirements in terms of functionality, flexibility and accuracy.
Resumo:
In the southern Duero Basin of central Spain, there are vast areas of aeolian sand sheets and dune fields. A comprehensive survey of the sand quarries in this area identified a number of palaeosols in sedimentary sequences. The identification and AMS radiocarbon dating of soil charcoal fragments collected in these palaeosols indicate the persistence of Pinus pinaster in this area throughout most of the Holocene. Although potential natural vegetation models have usually considered the Pinus pinaster forests in this inland area of artificial origin, soil charcoal analysis provides firm evidence of a natural origin. Our data fit perfectly with the pattern of Holocene vegetation development for inland areas of Iberia, which are characterised by stability of pine forests throughout the Holocene. Finally, the growing body of palaeobotanical evidence from Iberia (macrofossils and pollen) is contributing to improve our knowledge of P. pinaster ecology, showing that this species has been present in most Iberian regions during the Holocene, where it has inhabited areas characterised by a very diverse set of climatic and soil conditions.
Resumo:
The Competency-Based Education in the context of training is intended as a comprehensive approach that seeks to link education with the productive sector and increase the potential of individuals, in the face of social, economic, political and cultural transformations that suffers the world and the contemporary society; this is how educational services associated to the rural area takes part of the global revalorization of the role of learning and knowledge. Under the competence approach and taking into account the CONOCER model, we design a Technological Master from the “Colegio de Postgraduados” identifying the competences needed so that the students, professional from different areas of knowledge, managed to develop them, but mainly to achieve the goal of developing the capacities of producers in Mexican rural area.
Resumo:
Starting from the inaugural text of Philibert de L'Orme, stereotomic treatises and manuscripts are subject to the opposing forces of reason and fancy. The Nativity Chapel in Burgos Cathedral provides an outstanding case study on this subject. It was built in 1571-1582 by Martín de Bérriz and Martín de la Haya, using an oval vault resting on trumpet squinches to span a rectangular bay. Bed joints and rib axes are not planar curves, as usual in oval vaults. This warping is not capricious; we shall argue that it is the outcome of a systematic tracing method. As a result of this process, the slope of the bed joints increases slightly in the first courses, but stays fairly constant after the third course; this solution prevents the upper courses from slipping. Thus, in the Nativity Chapel of Burgos Cathedral, the constraints of masonry construction fostered a singular solution verging on capriccio. It is also worthwhile to remark that the warping of the joints is not easily appreciable to the eye and that the tracing process does not seem to start from a previous conception of the resulting form. All this suggests that we should be quite careful when talking about the whimsical character of Late Gothic and Early Renaissance; in some occasions, apparent caprice is the offspring of practical thinking.
Resumo:
The real and potential road influence on soundscape is considered a relevant management aspect that can assess the negative effects of massive visitants on such sensitive species -that have been living for centuries in the area-. As a first approach to the study of human disturbances sound impact, acoustic engineering tools allow us to model noise pollution caused by the main road that crosses the state ?Cabeza de Hierro? (M-604). For these preliminary results we use the French method XPS 31-133, recommend at EU level. Noise emission levels in black vulture nesting area are analyzed to understand the influence of human activities on rural areas and road management on biodiversity conservation. This approach develops a useful tool to make compatible the public enjoyment of forest services such as recreation or landscape scenary, the conservation of biodiversity as well as a suitable social and economic activity level ?timber and firewood harvesting, industry?- at the region.
Resumo:
The first wall armour for the reactor chamber of HiPER will have to face short energy pulses of 5 to 20 MJ mostly in the form of x-rays and charged particles at a repetition rate of 5–10 Hz. Armour material and chamber dimensions have to be chosen to avoid/minimize damage to the chamber, ensuring the proper functioning of the facility during its planned lifetime. The maximum energy fluence that the armour can withstand without risk of failure, is determined by temporal and spatial deposition of the radiation energy inside the material. In this paper, simulations on the thermal effect of the radiation–armour interaction are carried out with an increasing definition of the temporal and spatial deposition of energy to prove their influence on the final results. These calculations will lead us to present the first values of the thermo-mechanical behaviour of the tungsten armour designed for the HiPER project under a shock ignition target of 48 MJ. The results will show that only the crossing of the plasticity limit in the first few micrometres might be a threat after thousands of shots for the survivability of the armour.
Resumo:
This article describes a research project involving students from nine different engineering degrees at the Technical university of Madrid. The purpose of the project was to analyze the use of peer and self assessment and the students? attitudes toward alternative assessment procedures.
Resumo:
European rural development policy is gaining in importance through one of its key instruments, the Protected Geographical Indications (PGI) system, which is designed to improve quality standards. Previous research has shown that PGI-certified beef farms tend to be more extensively managed operations that are better adapted to mountainous areas. This paper describes a comparative study of two production systems, one with PGI certification and one without, focusing on a number of economic variables. The results show a positive association between PGI production and profitability. In efficiency terms, non-certified farms show better pure technical efficiency scores, while PGI-certified holdings score higher on scale efficiency.
Resumo:
The change towards a sustainable economic system represents a big challenge for the present as well as next generations. Such a process requires important long-term changes in technologies, lifestyle, infrastructures and institutions. In this scenario the innovation process is a crucial element for fostering sustainability as well as an egalitarian development in developing countries. For those reasons the concept of Eco-Innovation System is introduced and further considerations are provided for the case of less-developed countries. The paper illustrates that sustainable development is possible by exploiting local potential and traditional knowledge in order to achieve at the same time economic growth, social equality and environmental sustainability. In order to prove such an assumption a specific case study is described: The renewable energy sector in Bolivia. The case study summarizes many important dimensions of the innovation process in developing countries such as technological transfer, diffusion and adaptation, social dimension and development issues.
Resumo:
Public Private Partnerships (PPPs) are mostly implemented to circumvent budgetary constraints, and to encourage efficiency and quality in the provision of public infrastructure in order to reach social welfare. One of the ways of reaching the latter objective is by the introduction of performance based standards tied to bonuses and penalties to reward or punish the performance of the contractor. This paper focuses on the implementation of safety based incentives in PPPs in such a way that the better the safety outcome the greater larger will be the economic reward to the contractor. The main aim of this paper is to identify whether the incentives to improve road safety in PPPs are ultimately effective in improving safety ratios in Spain. To that end, Poisson and negative binomial regression models have been applied using information of motorways of the Spanish network of 2006. The findings indicate that even though road safety is highly influenced by variables that are not much controllable by the contractor such as the Average Annual Daily Traffic and the percentage of heavy vehicles, the implementation of safety incentives in PPPs has a positive influence in the reduction of fatalities, injuries and accidents.
Resumo:
The European Higher Education Area (EHEA) has leaded to a change in the way the subjects are taught. One of the more important aspects of the EHEA is to support the autonomous study of the students. Taking into account this new approach, the virtual laboratory of the subject Mechanisms of the Aeronautical studies at the Technical University of Madrid is being migrated to an on-line scheme. This virtual laboratory consist on two practices: the design of cam-follower mechanisms and the design of trains of gears. Both practices are software applications that, in the current situation, need to be installed on each computer and the students carry out the practice at the computer classroom of the school under the supervision of a teacher. During this year the design of cam-follower mechanisms practice has been moved to a web application using Java and the Google Development Toolkit. In this practice the students has to design and study the running of a cam to perform a specific displacement diagram with a selected follower taking into account that the mechanism must be able to work properly at high speed regime. The practice has maintained its objectives in the new platform but to take advantage of the new methodology and try to avoid the inconveniences that the previous version had shown. Once the new practice has been ready, a pilot study has been carried out to compare both approaches: on-line and in-lab. This paper shows the adaptation of the cam and follower practice to an on-line methodology. Both practices are described and the changes that has been done to the initial one are shown. They are compared and the weak and strong points of each one are analyzed. Finally we explain the pilot study carried out, the students impression and the results obtained.
Resumo:
The European Higher Education Area (EHEA) has leaded to a change in the way the subjects are taught. One of the more important aspects of the EHEA is to support the autonomous study of the students. Taking into account this new approach, the virtual laboratory of the subject Mechanisms of the Aeronautical studies at the Technical University of Madrid is being migrated to an on-line scheme. This virtual laboratory consist on two practices: the design of cam-follower mechanisms and the design of trains of gears. Both practices are software applications that, in the current situation, need to be installed on each computer and the students carry out the practice at the computer classroom of the school under the supervision of a teacher. During this year the design of cam-follower mechanisms practice has been moved to a web application using Java and the Google Development Toolkit. In this practice the students has to design and study the running of a cam to perform a specific displacement diagram with a selected follower taking into account that the mechanism must be able to work properly at high speed regime. The practice has maintained its objectives in the new platform but to take advantage of the new methodology and try to avoid the inconveniences that the previous version had shown. Once the new practice has been ready, a pilot study has been carried out to compare both approaches: on-line and in-lab. This paper shows the adaptation of the cam and follower practice to an on-line methodology. Both practices are described and the changes that has been done to the initial one are shown. They are compared and the weak and strong points of each one are analyzed. Finally we explain the pilot study carried out, the students impression and the results obtained.
Resumo:
La temperatura es una preocupación que juega un papel protagonista en el diseño de circuitos integrados modernos. El importante aumento de las densidades de potencia que conllevan las últimas generaciones tecnológicas ha producido la aparición de gradientes térmicos y puntos calientes durante el funcionamiento normal de los chips. La temperatura tiene un impacto negativo en varios parámetros del circuito integrado como el retardo de las puertas, los gastos de disipación de calor, la fiabilidad, el consumo de energía, etc. Con el fin de luchar contra estos efectos nocivos, la técnicas de gestión dinámica de la temperatura (DTM) adaptan el comportamiento del chip en función en la información que proporciona un sistema de monitorización que mide en tiempo de ejecución la información térmica de la superficie del dado. El campo de la monitorización de la temperatura en el chip ha llamado la atención de la comunidad científica en los últimos años y es el objeto de estudio de esta tesis. Esta tesis aborda la temática de control de la temperatura en el chip desde diferentes perspectivas y niveles, ofreciendo soluciones a algunos de los temas más importantes. Los niveles físico y circuital se cubren con el diseño y la caracterización de dos nuevos sensores de temperatura especialmente diseñados para los propósitos de las técnicas DTM. El primer sensor está basado en un mecanismo que obtiene un pulso de anchura variable dependiente de la relación de las corrientes de fuga con la temperatura. De manera resumida, se carga un nodo del circuito y posteriormente se deja flotando de tal manera que se descarga a través de las corrientes de fugas de un transistor; el tiempo de descarga del nodo es la anchura del pulso. Dado que la anchura del pulso muestra una dependencia exponencial con la temperatura, la conversión a una palabra digital se realiza por medio de un contador logarítmico que realiza tanto la conversión tiempo a digital como la linealización de la salida. La estructura resultante de esta combinación de elementos se implementa en una tecnología de 0,35 _m. El sensor ocupa un área muy reducida, 10.250 nm2, y consume muy poca energía, 1.05-65.5nW a 5 muestras/s, estas cifras superaron todos los trabajos previos en el momento en que se publicó por primera vez y en el momento de la publicación de esta tesis, superan a todas las implementaciones anteriores fabricadas en el mismo nodo tecnológico. En cuanto a la precisión, el sensor ofrece una buena linealidad, incluso sin calibrar; se obtiene un error 3_ de 1,97oC, adecuado para tratar con las aplicaciones de DTM. Como se ha explicado, el sensor es completamente compatible con los procesos de fabricación CMOS, este hecho, junto con sus valores reducidos de área y consumo, lo hacen especialmente adecuado para la integración en un sistema de monitorización de DTM con un conjunto de monitores empotrados distribuidos a través del chip. Las crecientes incertidumbres de proceso asociadas a los últimos nodos tecnológicos comprometen las características de linealidad de nuestra primera propuesta de sensor. Con el objetivo de superar estos problemas, proponemos una nueva técnica para obtener la temperatura. La nueva técnica también está basada en las dependencias térmicas de las corrientes de fuga que se utilizan para descargar un nodo flotante. La novedad es que ahora la medida viene dada por el cociente de dos medidas diferentes, en una de las cuales se altera una característica del transistor de descarga |la tensión de puerta. Este cociente resulta ser muy robusto frente a variaciones de proceso y, además, la linealidad obtenida cumple ampliamente los requisitos impuestos por las políticas DTM |error 3_ de 1,17oC considerando variaciones del proceso y calibrando en dos puntos. La implementación de la parte sensora de esta nueva técnica implica varias consideraciones de diseño, tales como la generación de una referencia de tensión independiente de variaciones de proceso, que se analizan en profundidad en la tesis. Para la conversión tiempo-a-digital, se emplea la misma estructura de digitalización que en el primer sensor. Para la implementación física de la parte de digitalización, se ha construido una biblioteca de células estándar completamente nueva orientada a la reducción de área y consumo. El sensor resultante de la unión de todos los bloques se caracteriza por una energía por muestra ultra baja (48-640 pJ) y un área diminuta de 0,0016 mm2, esta cifra mejora todos los trabajos previos. Para probar esta afirmación, se realiza una comparación exhaustiva con más de 40 propuestas de sensores en la literatura científica. Subiendo el nivel de abstracción al sistema, la tercera contribución se centra en el modelado de un sistema de monitorización que consiste de un conjunto de sensores distribuidos por la superficie del chip. Todos los trabajos anteriores de la literatura tienen como objetivo maximizar la precisión del sistema con el mínimo número de monitores. Como novedad, en nuestra propuesta se introducen nuevos parámetros de calidad aparte del número de sensores, también se considera el consumo de energía, la frecuencia de muestreo, los costes de interconexión y la posibilidad de elegir diferentes tipos de monitores. El modelo se introduce en un algoritmo de recocido simulado que recibe la información térmica de un sistema, sus propiedades físicas, limitaciones de área, potencia e interconexión y una colección de tipos de monitor; el algoritmo proporciona el tipo seleccionado de monitor, el número de monitores, su posición y la velocidad de muestreo _optima. Para probar la validez del algoritmo, se presentan varios casos de estudio para el procesador Alpha 21364 considerando distintas restricciones. En comparación con otros trabajos previos en la literatura, el modelo que aquí se presenta es el más completo. Finalmente, la última contribución se dirige al nivel de red, partiendo de un conjunto de monitores de temperatura de posiciones conocidas, nos concentramos en resolver el problema de la conexión de los sensores de una forma eficiente en área y consumo. Nuestra primera propuesta en este campo es la introducción de un nuevo nivel en la jerarquía de interconexión, el nivel de trillado (o threshing en inglés), entre los monitores y los buses tradicionales de periféricos. En este nuevo nivel se aplica selectividad de datos para reducir la cantidad de información que se envía al controlador central. La idea detrás de este nuevo nivel es que en este tipo de redes la mayoría de los datos es inútil, porque desde el punto de vista del controlador sólo una pequeña cantidad de datos |normalmente sólo los valores extremos| es de interés. Para cubrir el nuevo nivel, proponemos una red de monitorización mono-conexión que se basa en un esquema de señalización en el dominio de tiempo. Este esquema reduce significativamente tanto la actividad de conmutación sobre la conexión como el consumo de energía de la red. Otra ventaja de este esquema es que los datos de los monitores llegan directamente ordenados al controlador. Si este tipo de señalización se aplica a sensores que realizan conversión tiempo-a-digital, se puede obtener compartición de recursos de digitalización tanto en tiempo como en espacio, lo que supone un importante ahorro de área y consumo. Finalmente, se presentan dos prototipos de sistemas de monitorización completos que de manera significativa superan la características de trabajos anteriores en términos de área y, especialmente, consumo de energía. Abstract Temperature is a first class design concern in modern integrated circuits. The important increase in power densities associated to recent technology evolutions has lead to the apparition of thermal gradients and hot spots during run time operation. Temperature impacts several circuit parameters such as speed, cooling budgets, reliability, power consumption, etc. In order to fight against these negative effects, dynamic thermal management (DTM) techniques adapt the behavior of the chip relying on the information of a monitoring system that provides run-time thermal information of the die surface. The field of on-chip temperature monitoring has drawn the attention of the scientific community in the recent years and is the object of study of this thesis. This thesis approaches the matter of on-chip temperature monitoring from different perspectives and levels, providing solutions to some of the most important issues. The physical and circuital levels are covered with the design and characterization of two novel temperature sensors specially tailored for DTM purposes. The first sensor is based upon a mechanism that obtains a pulse with a varying width based on the variations of the leakage currents on the temperature. In a nutshell, a circuit node is charged and subsequently left floating so that it discharges away through the subthreshold currents of a transistor; the time the node takes to discharge is the width of the pulse. Since the width of the pulse displays an exponential dependence on the temperature, the conversion into a digital word is realized by means of a logarithmic counter that performs both the timeto- digital conversion and the linearization of the output. The structure resulting from this combination of elements is implemented in a 0.35_m technology and is characterized by very reduced area, 10250 nm2, and power consumption, 1.05-65.5 nW at 5 samples/s, these figures outperformed all previous works by the time it was first published and still, by the time of the publication of this thesis, they outnumber all previous implementations in the same technology node. Concerning the accuracy, the sensor exhibits good linearity, even without calibration it displays a 3_ error of 1.97oC, appropriate to deal with DTM applications. As explained, the sensor is completely compatible with standard CMOS processes, this fact, along with its tiny area and power overhead, makes it specially suitable for the integration in a DTM monitoring system with a collection of on-chip monitors distributed across the chip. The exacerbated process fluctuations carried along with recent technology nodes jeop-ardize the linearity characteristics of the first sensor. In order to overcome these problems, a new temperature inferring technique is proposed. In this case, we also rely on the thermal dependencies of leakage currents that are used to discharge a floating node, but now, the result comes from the ratio of two different measures, in one of which we alter a characteristic of the discharging transistor |the gate voltage. This ratio proves to be very robust against process variations and displays a more than suficient linearity on the temperature |1.17oC 3_ error considering process variations and performing two-point calibration. The implementation of the sensing part based on this new technique implies several issues, such as the generation of process variations independent voltage reference, that are analyzed in depth in the thesis. In order to perform the time-to-digital conversion, we employ the same digitization structure the former sensor used. A completely new standard cell library targeting low area and power overhead is built from scratch to implement the digitization part. Putting all the pieces together, we achieve a complete sensor system that is characterized by ultra low energy per conversion of 48-640pJ and area of 0.0016mm2, this figure outperforms all previous works. To prove this statement, we perform a thorough comparison with over 40 works from the scientific literature. Moving up to the system level, the third contribution is centered on the modeling of a monitoring system consisting of set of thermal sensors distributed across the chip. All previous works from the literature target maximizing the accuracy of the system with the minimum number of monitors. In contrast, we introduce new metrics of quality apart form just the number of sensors; we consider the power consumption, the sampling frequency, the possibility to consider different types of monitors and the interconnection costs. The model is introduced in a simulated annealing algorithm that receives the thermal information of a system, its physical properties, area, power and interconnection constraints and a collection of monitor types; the algorithm yields the selected type of monitor, the number of monitors, their position and the optimum sampling rate. We test the algorithm with the Alpha 21364 processor under several constraint configurations to prove its validity. When compared to other previous works in the literature, the modeling presented here is the most complete. Finally, the last contribution targets the networking level, given an allocated set of temperature monitors, we focused on solving the problem of connecting them in an efficient way from the area and power perspectives. Our first proposal in this area is the introduction of a new interconnection hierarchy level, the threshing level, in between the monitors and the traditional peripheral buses that applies data selectivity to reduce the amount of information that is sent to the central controller. The idea behind this new level is that in this kind of networks most data are useless because from the controller viewpoint just a small amount of data |normally extreme values| is of interest. To cover the new interconnection level, we propose a single-wire monitoring network based on a time-domain signaling scheme that significantly reduces both the switching activity over the wire and the power consumption of the network. This scheme codes the information in the time domain and allows a straightforward obtention of an ordered list of values from the maximum to the minimum. If the scheme is applied to monitors that employ TDC, digitization resource sharing is achieved, producing an important saving in area and power consumption. Two prototypes of complete monitoring systems are presented, they significantly overcome previous works in terms of area and, specially, power consumption.
Resumo:
- Resumen La hipótesis que anima esta tesis doctoral es que algunas de las características del entorno urbano, en particular las que describen la accesibilidad de su red de espacio público, podrían estar relacionadas con la proporción de viajes a pie o reparto modal, que tiene cada zona o barrio de Madrid. Uno de los puntos de partida de dicha hipótesis que el entorno urbano tiene una mayor influencia sobre los viaje a pie que en sobre otros modos de transporte, por ejemplo que en los viajes de bicicleta o en transporte público; y es que parece razonable suponer que estos últimos van a estar más condicionadas por ejemplo por la disponibilidad de vías ciclistas, en el primer caso, o por la existencia de un servicio fiable y de calidad, en el segundo. Otra de las motivaciones del trabajo es que la investigación en este campo de la accesibilidad del espacio público, en concreto la denominada “Space Syntax”, ha probado en repetidas ocasiones la influencia de la red de espacio público en cómo se distribuye la intensidad del tráfico peatonal por la trama urbana, pero no se han encontrado referencias de la influencia de dicho elemento sobre el reparto modal. De acuerdo con la hipótesis y con otros trabajos anteriores se propone una metodología basada en el análisis empírico y cuantitativo. Su objetivo es comprobar si la red de espacio público, independientemente de otras variables como los usos del suelo, incluso de las variables de ajenas entorno no construido, como las socioeconómicas, está o no relacionada estadísticamente con la proporción de peatones viajes en las zonas urbanas. Las técnicas estadísticas se utilizan para comprobar sistemáticamente la asociación de las variables del entorno urbano, denominadas variables independientes, con el porcentaje de viajes a pie, la variable dependiente. En términos generales, la metodología es similar a la usada en otros trabajos en este campo como los de CERVERÓ y KOCKLEMAN (1997), CERVERÓ y DUNCAN (2003), o para los que se utilizan principalmente en la revisión general de TRB (2005) o, más recientemente, en ZEGRAS (2006) o CHATMAN (2009). Otras opciones metodológicas, como los métodos de preferencias declaradas (ver LOUVIERE, HENSHER y SWAIT, 2000) o el análisis basado en agentes (PENN & TURNER, 2004) fueron descartados, debido a una serie de razones, demasiado extensas para ser descritas aquí. El caso de estudio utilizado es la zona metropolitana de Madrid, abarcándola hasta la M-50, es decir en su mayor parte, con un tamaño aproximado de 31x34 Km y una población de 4.132.820 habitantes (aproximadamente el 80% de la población de la región). Las principales fuentes de datos son la Encuesta Domiciliaria de Movilidad de 2004 (EDM04), del Consorcio Regional de Transportes de Madrid que es la última disponible (muestra: > 35.000 familias,> 95.000 personas), y un modelo espacial del área metropolitana, integrando el modelo para calcular los índices de Space Syntax y un Sistema de Información Geográfica (SIG). La unidad de análisis, en este caso las unidades espaciales, son las zonas de transporte (con una población media de 7.063 personas) y los barrios (con una población media de 26.466 personas). Las variables del entorno urbano son claramente el centro del estudio. Un total de 20 índices (de 21) se seleccionan de entre los más relevantes encontrados en la revisión de la producción científica en este campo siendo que, al mismo tiempo, fueran accesibles. Nueve de ellos se utilizan para describir las características de los usos del suelo, mientras que otros once se usan para describir la red de espacios públicos. Estos últimos incluyen las variables de accesibilidad configuracional, que son, como se desprende de su título, el centro del estudio propuesto. La accesibilidad configuracional es un tipo especial de accesibilidad que se basa en la configuración de la trama urbana, según esta fue definida por HILLIER (1996), el autor de referencia dentro de esta línea de investigación de Space Syntax. Además se incluyen otras variables de la red de espacio público más habituales en los estudios de movilidad, y que aquí se denominan características geométricas de los elementos de la red, tales como su longitud, tipo de intersección, conectividad, etc. Por último se incluye además una variable socioeconómica, es decir ajena al entorno urbano, para evaluar la influencia de los factores externos, pues son varios los que pueden tener un impacto en la decisión de caminar (edad, género, nivel de estudios, ingresos, tasa de motorización, etc.). La asociación entre las variables se han establecido usando análisis de correlación (bivariante) y modelos de análisis multivariante. Las primeras se calculan entre por pares entre cada una de las 21 variables independientes y la dependiente, el porcentaje de viajes a pie. En cuanto a los segundos, se han realizado tres tipos de estudios: modelo multivariante general lineal, modelo multivariante general curvilíneo y análisis discriminante. Todos ellos son capaces de generar modelos de asociación entre diversas variables, pudiéndose de esta manera evaluar con bastante precisión en qué medida cada modelo reproduce el comportamiento de la variable dependiente, y además, el peso o influencia de cada variable en el modelo respecto a las otras. Los resultados fundamentales del estudio se expresan en dos modelos finales alternativos, que demuestran tener una significativa asociación con el porcentaje de viajes a pie (R2 = 0,6789, p <0,0001), al explicar las dos terceras partes de su variabilidad. En ellos, y en general en todo el estudio realizado, se da una influencia constante de tres índices en particular, que quedan como los principales. Dos de ellos, de acuerdo con muchos de los estudios previos, corresponden a la densidad y la mezcla de usos del suelo. Pero lo más novedoso de los resultados obtenidos es que el tercero es una medida de la accesibilidad de la red de espacio público, algo de lo que no había referencias hasta ahora. Pero, ¿cuál es la definición precisa y el peso relativo de cada uno en el modelo, es decir, en la variable independiente? El de mayor peso en la mayor parte de los análisis realizados es el índice de densidad total (n º residentes + n º puestos de trabajo + n º alumnos / Ha). Es decir, una densidad no sólo de población, sino que incluye algunas de las actividades más importantes que pueden darse una zona para generar movilidad a pie. El segundo que mayor peso adquiere, llegando a ser el primero en alguno de los análisis estadísticos efecturados, es el índice de accesibuilidad configuracional denominado integración de radio 5. Se trata de una medida de la accesibilidad de la zona, de su centralidad, a la escala de, más un menor, un distrito o comarca. En cuanto al tercero, obtiene una importancia bastante menor que los anteriores, y es que representa la mezcla de usos. En concreto es una medida del equilibrio entre los comercios especializados de venta al por menor y el número de residentes (n º de tiendas especializadas en alimentación, bebidas y tabaco / n º de habitantes). Por lo tanto, estos resultados confirman buena parte de los de estudios anteriores, especialmente los relativas a los usos del suelo, pero al mismo tiempo, apuntan a que la red de espacio público podría tener una influir mayor de la comprobada hasta ahora en la proporción de peatones sobre el resto de modos de transportes. Las razones de por qué esto puede ser así, se discuten ampliamente en las conclusiones. Finalmente se puede precisar que dicha conclusión principal se refiere a viajes de una sola etapa (no multimodales) que se dan en los barrios y zonas del área metropolitana de Madrid. Por supuesto, esta conclusión tiene en la actualidad, una validez limitada, ya que es el resultado de un solo caso — Abstract The research hypothesis for this Ph.D. Thesis is that some characteristics of the built environment, particularly those describing the accessibility of the public space network, could be associated with the proportion of pedestrians in all trips (modal split), found in the different parts of a city. The underlying idea is that walking trips are more sensitive to built environment than those by other transport modes, such as for example those by bicycle or by public transport, which could be more conditioned by, e.g. infrastructure availability or service frequency and quality. On the other hand, it has to be noted that the previously research on this field, in particular within Space Syntax’s where this study can be referred, have tested similar hypothesis using pedestrian volumes as the dependent variable, but never against modal split. According to such hypothesis, research methodology is based primarily on empirical quantitative analysis, and it is meant to be able to assess whether public space network, no matter other built environment and non-built environment variables, could have a relationship with the proportion of pedestrian trips in urban areas. Statistical techniques are used to check the association of independent variables with the percentage of walking in all trips, the dependent one. Broadly speaking this methodology is similar to that of previous studies in the field such as CERVERO&KOCKLEMAN (1997), CERVERO & DUNCAN (2003), or to those used mainly in the general review of T.R.B. (2005) or, more recently in ZEGRAS (2006) or CHATMAN (2009). Other methodological options such as stated choice methods (see LOUVIERE, HENSHER & SWAIT, 2000) or agent based analysis (PENN & TURNER, 2004), were discarded, due to a number of reasons, too long to be described here. The case study is not the entire Madrid’s metropolitan area, but almost (4.132.820 inhabitants, about 80% of region´s population). Main data sources are the Regional Mobility Home Based Survey 2004 (EDM04), which is the last available (sample: >35.000 families, > 95.000 individuals), and a spatial model of the metropolitan area, developed using Space Syntax and G.I.S. techniques. The analysis unit, in this case spatial units, are both transport zones (mean population = 7.063) and neighborhoods (mean population = 26.466). The variables of the built environment are clearly the core of the study. A total of 20 (out of 21) are selected from among those found in the literature while, at the same time, being accessible. Nine out of them are used to describe land use characteristics while another eleven describe the network of public spaces. Latter ones include configurational accessibility or Space Syntax variables. This is a particular sort of accessibility related with the concept of configuration, by HILLIER (1996), one of the main authors of Space Syntax, But it also include more customary variables used in mobility research to describe the urban design or spatial structure (here public space network), which here are called geometric characteristics of the such as its length, type of intersection, conectivity, density, etc. Finally a single socioeconomic variable was included in order to assess the influence non built environment factors that also may have an impact on walking (age, income, motorization rate, etc.). The association among variables is worked out using bi-variate correlation analysis and multivariate-analysis. Correlations are calculated among the 21 independent variables and the dependent one, the percentage of walking trips. Then, three types of multi-variate studies are run: general linear, curvilinear and discriminant multi-variate analysis. The latter are fully capable of generating complex association models among several variables, assessing quite precisely to what extent each model reproduces the behavior of the dependent variable, and also the weight or influence of each variable in the model. This study’s results show a consistent influence of three particular indexes in the two final alternative models of the multi-variate study (best, R2=0,6789, p<0,0000). Not surprisingly, two of them correspond to density and mix of land uses. But perhaps more interesting is that the third one is a measure of the accessibility of the public space network, a variable less important in the literature up to now. Additional precisions about them and their relative weight could also be of some interest. The density index is not only about population but includes most important activities in an area (nº residents + nº jobs+ nº students/Ha). The configurational index (radius 5 integration) is a measure of the accessibility of the area, i.e. centrality, at the scale of, more a less, a district. Regarding the mix of land uses index, this one is a measure of the balance between retail, in fact local basic retail, and the number of residents (nº of convenience shops / nº of residents). Referring to their weights, configurational index (radius 5 integration) gets the higher standardized coefficient of the final equation. However, in the final equations, there are a higher number of indexes coming from the density or land use mix categories than from public space network enter. Therefore, these findings seem to support part of the field’s knowledge, especially those concerning land uses, but at the same time they seem to bring in the idea that the configuration of the urban grid could have an influence in the proportion of walkers (as a part of total trips on any transport mode) that do single journey trips in the neighborhoods of Madrid, Spain. Of course this conclusion has, at present, a limited validity since it’s the result of a single case. The reasons of why this can be so, are discussed in the last part of the thesis.