13 resultados para extensive cattle system

em Universidad Politécnica de Madrid


Relevância:

90.00% 90.00%

Publicador:

Resumo:

The Bienaventurada mine operates a polymetallic Ag-Pb-Zn (Cu, Au) vein system of the low sulphidation epithermal type. Fluid inclusions, FI, are abundant in quartz, sphalerite and adularia. FI petrography demonstrates typical primary growth zoning which occurs frequently in crystalline quartz, and defines the most common primary FI. These are usually very small, but several types of primary, P, and secondary, S, FI Assemblages (FIAs) comprising FI of measurable size (3 to > 100 μm) can also be identified through careful petrographic work. The fluids are aqueous and undersaturated, and no evidence of CO2 was found; the degree of fill is usually high (~70-80 %) in the L-rich inclusions, but extremely low in V-rich inclusions. The measured microthermometric values are very consistent in the FIAs selected; they are for the most part roughly similar in the P and S assemblages: the median is typically ~258ºC for total homogenization temperatures, Th, and -1.5 ºC for ice melting temperatures, Tm (corresponding to 2.57 wt% NaCl eq). The widespread occurrence of L-rich and V-rich FI in the same FIA and the consistent Th values point to an extensive boiling system along the vein. In these conditions, Th equals T of trapping, and the ores are assumed to have been precipitated from an aqueous low salinity boiling fluid, of likely meteoric origin, at some 250-280º C, under ~500 m hydrostatic head.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The aim of this work was to evaluate different management strategies to optimize rabbit production under chronic heat stress. To achieve it, three trials were conducted. In the first trial, to find the optimal cage density in tropical very dry forest condition, were measured growth performance, mortality rate, injured animals and carcass performance over an initial population of 300 cross-breed rabbits of New Zealand, California, Butterfly, Dutch and Satin, weaned at 30 days (535 ± 8 g, standard error). Treatments evaluated were: 6, 12, 18 and 24 rabbits/m2 (3, 6, 9 and 12 rabbits/cage, respectively, each cage of 0.5 m2). The maximal temperature-humidity index indicated a severe heat stress from weaning to 2.2 kg body weight (experimental time). At the end of experimental period 10, 20, 30 and 30 rabbits from the treatments of 6, 12, 18 and 24 rabbits/m2, respectively, were slaughtered and carcass performance recorded. Average daily gain and feed intake decreased by 0.31 ± 0.070 and 1.20 ± 0.25 g, respectively, per each unit that the density increased at the beginning of the experiment (P = 0.001). It increased the length of the fattening period by 0.91 ± 0.16 d (P = 0.001) per each unit of increment of density. However, rabbit production (kg/m2) increased linear and quadratically with the density (P < 0.008). Animals housed at the highest density compared to the lower one tended to show a higher incidence of ringworm (68.9 vs 39.4%; P = 0.075), injured animals (16.8 vs 3.03%; P = 0.12) and mortality (20.5 vs 9.63%; P = 0.043). The proportion of scapular fat (P = 0.042) increased linearly with increasing levels of density. Increasing density reduced linearly dorsal length (P = 0.001), and reduced linear and quadratically drip loss percentage (P = 0.097 and 0.018, respectively). In the second trial, 46 nulliparous rabbit does (23 clipped and 23 unclipped) with a BW of 3.67 ± 0.05 kg (s.e.) were used to evaluate heat stress and circadian rhythms comparing unclipped and clipped rabbit does, and to study if a more extensive breeding system increase litters performance at weaning without impairing rabbit doe performance,. Rectal temperature, feed and water 4 intake were recorded for 24 h. Rabbit does were mated 7 d after circadian measurements, and randomly assigned to two breeding systems. Control (C): mated at 14 d after parturition + litter weaned at 35 d of age. Extensive (E): mate at 21 after parturition + litter weaned at 42 d of age. The first three cycles were evaluated concerning to rabbit doe and litter performance. Two hundred twenty eight weaned rabbits, were divided into two cage sizes: 0.5 and 0.25 m2 with same density (16 rabbit/m2) and growing performance was recorded. Farm and rectal temperatures were minimal and feed and water intake maximal during the night (P < 0.001). Unclipped rabbit does showed higher rectal temperature (P = 0.045) and lower feed intake respect to clipped does (P = 0.019) which suggest a lower heat stress in the latter. Kits weaned per litter was reduced by 33% (P=0.038) in C group. This reduction was more important in the 2nd and 3rd cycles compared to the first (P ≤ 0.054). Rabbit doe feed efficiency tended to decrease in E respect C group (P = 0.093), whereas it was impaired from the first to the third cycle by 48% (P = 0.014). Growing rabbits from the E group were heavier at weaning (by 38%. P < 0.001), showed a higher feed intake (+7.4%) and lower feed efficiency (-8.4%) throughout the fattening period (P ≤ 0.056) respect to C group. Cage size had minor influence in growing performance. In the third trial, forty five non pregnant and non lactating rabbit does (21 nulliparous and 24 multiparous) were assigned randomly to farm water and to potable water to study if a water quality improvement can affect positively rabbit doe response to heat stress during pregnancy and lactation. A transponder was implanted in each animal to record subcutaneous temperature at 07:30 and 14:30 h. Experimental period extended from pregnancy (with no lactation) to the next lactation (until day 28). Body temperature and milk production were recorded daily, and body condition, feed and water intake weekly. Water quality did not affect any trait (P ≥ 0.15). Pregnant rabbit does were classified as does that weaned (W: 47%), not weaned (NW: 44%) or those pregnant that did not deliver (NB: 9%). Body temperature and feed intake decreased during pregnancy (P ≤ 0.031), but water intake remained constant. In this period body temperature decreased with metabolic weight (P ≤ 0.009). In W and NW does, 5 from mating to birth energy and protein balance impaired (P≤0.011). Body temperature of W does tended to be the lowest (P ≤ 0.090). Pregnancy length and total number of kits born tended to be longer and higher in NW than in W does (P = 0.10 and 0.053, respectively). Kit mortality at birth and from birth to 14 d of lactation was high, being worse for NW than for W does (97 vs. 40%; P<0.001). Body temperature during lactation was maximal at day 12, and milk production increased it (P ≤ 0.025). . In conclusion, in our heat stress conditions densities higher than 18 rabbits/m2 (34 kg/m2) at the end of fattening, are not recommended despite cage size, gestation and lactation productivity impaired not only when lactation is extended and along successive reproductive cycles but also due to a reduced embryo/kit survival and finally water quality improvement did not attenuate negative effect of heat stress. RESUMEN El propósito de éste trabajo fue evaluar diferentes estrategias de manejo para optimizar la producción de conejos bajo estrés térmico. Para lo cual se desarrollaron tres experimentos. En el primer experimento, para encontrar el número óptimo de gazapos por m2 de jaula durante el cebo en condiciones de bosque muy seco tropical, se estudiaron los rendimientos durante el cebo, mortalidad, animales lesionados y rendimiento de la canal sobre una población inicial de 300 conejos mestizos de Nueva Zelanda, California, Mariposa, Holandés y Satin, destetados a los 30 días de edad (535 ± 8g, error estándar). Los tratamientos evaluados fueron: 6, 12, 18 y 24 conejos/m2 (3, 6, 9 y 12 conejos/jaula, respectivamente, en jaulas de 0.5 m2). Durante el período experimental (destete a 2.2 kg de peso vivo), se observaron valores de THI correspondientes con un estrés térmico severo (THI max. De 31 a 35). Al final del período experimental, 10, 20, 30, y 30 conejos de los tratamientos con densidades de 6, 12, 18 y 24 conejos/m2, respectivamente, fueron sacrificados y su canal fue valorada. El promedio de la ganancia diaria y el consumo de alimento disminuyeron en 0.31 ± 0.070 y 1.20 ± 0.25 g, respectivamente, por cada unidad de incremento en la densidad al inicio del experimento (P=0.001). Esto alargó el período de engorde en 0.91 ± 0.16 d (P=0.001) por cada unidad de incremento de la densidad. Sin embargo, la producción de conejos (kg/m2) aumentó lineal y cuadráticamente con la densidad (P<0.008). Los animales alojados en las mayores densidades en comparación con el resto tendieron a mostrar una mayore incidencia de tiña (68.9 vs 39.4%; P=0.075), de cantidad de animales heridos (16.8 vs 3.03%; P=0.12), así como de mortalidad (20.5 vs 9.63%; P=0.043). El aumento en la densidad aumentó linealmente la proporción de grasa escapular (P=0.042) y redujo linealmente la longitud dorsal (P=0.001), y lineal y cuadráticamente el porcentaje de pérdida por goteo (P=0.018). En el segundo experimento, 46 conejas nulliparas (23 rasuradas y 23 no rasuradas) con un peso vivo de 3.67 ± 0.05 kg (e.e.) fueron usadas para evaluar el estrés 8 térmico y los ritmos circadianos comparando conejas rasuradas o no, y estudiar si un sistema de crianza más extensivo mejora el desempeño de la camada al destete sin perjudicar la productividad de la coneja. Durante 24 h se midió la temperatura rectal, consumo de alimento y de agua. Las conejas fueron montadas 7 días después, y distribuidas en dos sistemas de crianza. El control (C): monta a 14 días posparto y destete a 35 d de edad. El extensivo (E): monta a 21 días posparto y destete a 42 d de edad. Se controló la productividad de la coneja y la camada durante los tres primeros ciclos. Doscientos veintiocho gazapos fueron distribuidos en dos tamaños de jaulas (0.5 y 0.25 m2) con la misma densidad (16 conejos/m2) y se controlaron sus rendimientos productivos. Durante la noche se observaron los valores mínimos para la temperatura ambiental y rectal, y los máximos para consumo de alimento y agua (P< 0.001). Las conejas no rasuradas mostraron mayor temperatura rectal (P=0.045) y menores valores de consumo de alimento con respecto a las conejas rasuradas (P=0.019), lo que sugiere un menor estrés térmico en las últimas. El número de gazapos destetados por camada se redujo en 33% (P=0.038) en el grupo C. Este comportamiento se acentuó en el 2do y 3er ciclo en comparación con el primero (P≤0.054). La eficiencia alimenticia de las conejas tendió a disminuir en el grupo E con respecto al grupo C (P=0.093), dicha tendencia se acentúa del primer al tercer ciclo en un 48% (P=0.014). Los gazapos en fase de crecimiento provenientes del grupo E fueron más pesados al momento del destete (en 38% P<0.001), mostrando un mayor consumo de alimento (+7.4%) y menor eficiencia alimenticia (-8.4%) a lo largo del engorde (P≤0.056) con respecto al grupo C. El tamaño de la jaula tuvo una mínima influencia en el comportamiento durante el crecimiento de éstos gazapos. En el tercer experimento, cuarenta y cinco conejas no gestantes ni lactantes (21 nulíparas y 24 multíparas) se les asignó al azar agua dos tipos de agua: común de la granja y agua potable, con el fin de estudiar si una mejora en la calidad del agua puede afectar positivamente la respuesta de la coneja al estrés térmico durante la gestación y la lactancia. Se les implantó un transponder para registrar la temperatura subcutánea a las 7:30 y a las 14:30 h. El período experimental se extendió desde la gestación (sin 9 lactancia) hasta la lactanción consecutiva (hasta los 28 días). La temperatura corporal y la producción de leche se controlaron diariamente, y la condición corporal, consumo de agua y alimento, semanalmente. La calidad del agua no afectó a ninguna variable (P≥0.15). Las conejas preñadas fueron clasificadas como conejas que destetaron (W: 47%), que no destetaron (NW:44%) o aquellas que no parieron (NB: 9%). La temperatura corporal y consumo de alimento disminuyeron durante la gestación (P≤0.031), mientras que el consumo de agua se mantuvo constante. La temperatura corporal descendió con el peso metabólico durante la gestación (P≤0.009). El balance de energía y proteína disminuyó desde la monta al parto para las conejas W y NW (P≤0.011). Durante la gestación la temperatura corporal tendió a ser menor en las conejas W (P≤0.090). La longitud de la gestación y el número total de gazapos nacidos tendieron a ser mayores en conejas NW que en conejas W (P=0.10 y 0.053, respectivamente). La mortalidad de los gazapos al parto y del parto a los 14 días de lactancia fue alta, siendo peor para las conejas NW que para las W (97 vs 40%; P<0.001). Durante la lactancia la temperatura corporal alcanzó su valor máximo para el día 12, y la producción de leche indujo un incremento en la misma (P≤0.025). En conclusión, en nuestras condiciones de estrés térmico y sin importar el tamaño de la jaula, no se recomiendan densidades mayores a 18 conejos/m2 (34 kg/m2) al final del engorde. La productividad de la gestación y la lactancia disminuyen cuando la lactancia es mayor y se suceden varios ciclos reproductivos seguidos. Esto se debe al efecto negativo del estrés térmico sobre la vitalidad y supervivencia del embrión/gazapo. La mejora de la calidad del agua atenuó el efecto negativo del estrés térmico. Las conejas más productoras parece que son aquéllas que consiguen manejar mejor el estrés térmico.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

RESUMEN Los procesos de diseño de zonas o diseño del territorio implican la partición de un espacio geográfico, organizado en un conjunto de unidades de área, en diferentes regiones o zonas según un conjunto especifico de criterios que varían en función del campo de aplicación. En la mayoría de los casos, el objetivo fundamental consiste en crear zonas de tamaño aproximadamente igual respecto a uno o varios atributos de medida -de carácter cuantitativo- (zonas con igual número de habitantes, igual promedio de ventas...). Sin embargo, están apareciendo nuevas aplicaciones, algunas en el contexto de las políticas de desarrollo sostenible, cuya finalidad es la definición de regiones con un tamaño predeterminado, no necesariamente similar. Además, en estos casos las zonas han de formarse en torno a un conjunto específico de posiciones, semillas o generadores. Este tipo de particiones no han sido lo suficientemente investigadas, de manera que no se conocen modelos de solución para la delimitación automática de las zonas. En esta tesis se ha diseñado un nuevo método basado en una versión discreta del diagrama de Voronoi con peso aditivo adaptativo (DVPAA), que permite la partición de un espacio bidimensional en zonas de un tamaño específico, considerando tanto la posición como el peso de cada uno de los generadores. El método consiste en resolver repetidamente un tradicional diagrama de Voronoi con peso aditivo, de forma que los pesos de cada generador se actualizan en cada iteración. En el proceso de cálculo de distancias se usa una métrica basada en el camino más corto, lo que garantiza que la partición obtenida esté formada por un conjunto de zonas conexas. La heurística diseñada se integra en una aplicación prototipo, desarrollada en un entorno SIG (Sistemas de Información Geográfica), que permite el trazado automático de zonas según los criterios anteriormente expuestos. Para analizar la viabilidad del método se ha utilizado como caso de estudio la gestión de los recursos pastorales para la ganadería extensiva en tres municipios de Castilla-La Mancha. Las pruebas realizadas ponen de manifiesto que la heurística diseñada, adaptada a los criterios que se plantean en el contexto de la gestión de sistemas extensivos agropecuarios, es válida para resolver este tipo de problemas de partición. El método propuesto se caracteriza por su eficacia en el tratamiento de un gran número de unidades superficiales en formato vectorial, generando soluciones que convergen con relativa rapidez y verifican los criterios establecidos. En el caso estudiado, aunque la posición prefijada de los generadores reduce considerablemente la complejidad del problema, existen algunas configuraciones espaciales de estos elementos para las que el algoritmo no encuentra una solución satisfactoria, poniéndose de manifiesto una de las limitaciones de este modelo. Tal y como se ha podido comprobar, la localización de los generadores puede tener un considerable impacto en la zonificación resultante, por lo que, de acuerdo con Kalcsics et al. (2005), una selección "inadecuada" difícilmente puede generar regiones válidas que verifiquen los criterios establecidos. ABSTRACT Tenitory or zone design processes entail partitioning a geographic space, organized as a set of basic areal units, into different regions or zones according to a specific set of entena that are dependent on the application context. In most cases the aim is to create zones that have approximately equal sizes with respect to one or several measure attributes (zones with equal numbers of inhabitants, same average sales, etc). However, some of the new applications that have emerged, particularly in the context of sustainable development policies, are aimed at defining zones of a predetermined, though not necessarily similar, size. In addition, the zones should be built around a given set of positions, seeds or generators. This type of partitioning has not been sufñciently researched; therefore there are no known approaches for automated zone delimitation. This thesis proposes a new method based on a discrete versión of the Adaptive Additively Weighted Voronoi Diagram (AAWVD) that makes it possible to partition a 2D space into zones of specific sizes, taking both the position and the weight of each (seed) generator into account. The method consists of repeatedly solving a traditional additively weighted Voronoi diagram, so that the weights of each generator are updated at every iteration. The partition s zones are geographically connected nsing a metric based 011 the shortest path. The proposed heuristic lias been included in an application, developed in a GIS environment that allows the automated zone delimitation according to the mentioned criteria. The management of the extensive farming system of three municipalities of Castilla-La Mancha (Spain) has been used as study case to analyze the viability of the method. The tests carried out have established that the proposed method, adapted to the criteria of this application field, is valid for solving this type of partition problem. The applied algorithm is capable of handling a high number of vector areal units, generating solutions that converge in a reasonable CPU time and comply with the imposed constraints. Although the complexity of this problem is greatly reduced when the generator's positions are fixed, in many cases, these positions impose a spatial confignration that the algorithm proposed is unable to solve, thus revealing one of the limitations of this method. It has been shown that the location of the generators has a considerable impact on the final solution, so that, as Kalcsics et al. (2005) observed, an "inadequate" selection can hardly generate valid zones that comply with the established criteria.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Territory or zone design processes entail partitioning a geographic space, organized as a set of areal units, into different regions or zones according to a specific set of criteria that are dependent on the application context. In most cases, the aim is to create zones of approximately equal sizes (zones with equal numbers of inhabitants, same average sales, etc.). However, some of the new applications that have emerged, particularly in the context of sustainable development policies, are aimed at defining zones of a predetermined, though not necessarily similar, size. In addition, the zones should be built around a given set of seeds. This type of partitioning has not been sufficiently researched; therefore, there are no known approaches for automated zone delimitation. This study proposes a new method based on a discrete version of the adaptive additively weighted Voronoi diagram that makes it possible to partition a two-dimensional space into zones of specific sizes, taking both the position and the weight of each seed into account. The method consists of repeatedly solving a traditional additively weighted Voronoi diagram, so that each seed?s weight is updated at every iteration. The zones are geographically connected using a metric based on the shortest path. Tests conducted on the extensive farming system of three municipalities in Castile-La Mancha (Spain) have established that the proposed heuristic procedure is valid for solving this type of partitioning problem. Nevertheless, these tests confirmed that the given seed position determines the spatial configuration the method must solve and this may have a great impact on the resulting partition.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

There is an increasing tendency of turning the current power grid, essentially unaware of variations in electricity demand and scattered energy sources, into something capable of bringing a degree of intelligence by using tools strongly related to information and communication technologies, thus turning into the so-called Smart Grid. In fact, it could be considered that the Smart Grid is an extensive smart system that spreads throughout any area where power is required, providing a significant optimization in energy generation, storage and consumption. However, the information that must be treated to accomplish these tasks is challenging both in terms of complexity (semantic features, distributed systems, suitable hardware) and quantity (consumption data, generation data, forecasting functionalities, service reporting), since the different energy beneficiaries are prone to be heterogeneous, as the nature of their own activities is. This paper presents a proposal on how to deal with these issues by using a semantic middleware architecture that integrates different components focused on specific tasks, and how it is used to handle information at every level and satisfy end user requests.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A heterogeneous network, mainly based on nodes that use harvested energy to self-energize is presented and its use demonstrated. The network, mostly kinetically powered, has been used for the localization of herds in grazing areas under extreme climate conditions. The network consists of secondary and primary nodes. The former, powered by a kinetic generator, take advantage of animal movements to broadcast a unique identifier. The latter are battery-powered and gather secondarynode transmitted information to provide it, along with position and time data, to a final base station in charge of the animal monitoring. Because a limited human interaction is desirable, the aim of this network is to reduce the battery count of the system.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Mediterranean Dehesas are one of the European natural habitat types of Community interest (43/92/EEC Directive), associated to high diversity levels and producer of important goods and services. In this work, tree contribution and grazing influence over pasture alpha diversity in a Dehesa in Central Spain was studied. We analyzed Richness and Shannon-Wiener (SW) indexes on herbaceous layer under 16 holms oak trees (64 sampling units distributed in two directions and in two distances to the trunk) distributed in four different grazing management zones (depending on species and stocking rate). Floristic composition by species or morphospecies and species abundance were analyzed for each sample unit. Linear mixed models (LMM) and generalized linear mixed models (GLMMs) were used to study relationships between alpha diversity measures and independent factors. Edge crown influence showed the highest values of Richness and SW index. No significant differences were found between orientations under tree crown influence. Grazing management had a significant effect over Richness and SW measures, specially the grazing species (cattle or sheep). We preliminary quantify and analyze the interaction of tree stratum and grazing management over herbaceous diversity in a year of extreme climatic conditions.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

La astronomía de rayos γ estudia las partículas más energéticas que llegan a la Tierra desde el espacio. Estos rayos γ no se generan mediante procesos térmicos en simples estrellas, sino mediante mecanismos de aceleración de partículas en objetos celestes como núcleos de galaxias activos, púlsares, supernovas, o posibles procesos de aniquilación de materia oscura. Los rayos γ procedentes de estos objetos y sus características proporcionan una valiosa información con la que los científicos tratan de comprender los procesos físicos que ocurren en ellos y desarrollar modelos teóricos que describan su funcionamiento con fidelidad. El problema de observar rayos γ es que son absorbidos por las capas altas de la atmósfera y no llegan a la superficie (de lo contrario, la Tierra será inhabitable). De este modo, sólo hay dos formas de observar rayos γ embarcar detectores en satélites, u observar los efectos secundarios que los rayos γ producen en la atmósfera. Cuando un rayo γ llega a la atmósfera, interacciona con las partículas del aire y genera un par electrón - positrón, con mucha energía. Estas partículas secundarias generan a su vez más partículas secundarias cada vez menos energéticas. Estas partículas, mientras aún tienen energía suficiente para viajar más rápido que la velocidad de la luz en el aire, producen una radiación luminosa azulada conocida como radiación Cherenkov durante unos pocos nanosegundos. Desde la superficie de la Tierra, algunos telescopios especiales, conocidos como telescopios Cherenkov o IACTs (Imaging Atmospheric Cherenkov Telescopes), son capaces de detectar la radiación Cherenkov e incluso de tomar imágenes de la forma de la cascada Cherenkov. A partir de estas imágenes es posible conocer las principales características del rayo γ original, y con suficientes rayos se pueden deducir características importantes del objeto que los emitió, a cientos de años luz de distancia. Sin embargo, detectar cascadas Cherenkov procedentes de rayos γ no es nada fácil. Las cascadas generadas por fotones γ de bajas energías emiten pocos fotones, y durante pocos nanosegundos, y las correspondientes a rayos γ de alta energía, si bien producen más electrones y duran más, son más improbables conforme mayor es su energía. Esto produce dos líneas de desarrollo de telescopios Cherenkov: Para observar cascadas de bajas energías son necesarios grandes reflectores que recuperen muchos fotones de los pocos que tienen estas cascadas. Por el contrario, las cascadas de altas energías se pueden detectar con telescopios pequeños, pero conviene cubrir con ellos una superficie grande en el suelo para aumentar el número de eventos detectados. Con el objetivo de mejorar la sensibilidad de los telescopios Cherenkov actuales, en el rango de energía alto (> 10 TeV), medio (100 GeV - 10 TeV) y bajo (10 GeV - 100 GeV), nació el proyecto CTA (Cherenkov Telescope Array). Este proyecto en el que participan más de 27 países, pretende construir un observatorio en cada hemisferio, cada uno de los cuales contará con 4 telescopios grandes (LSTs), unos 30 medianos (MSTs) y hasta 70 pequeños (SSTs). Con un array así, se conseguirán dos objetivos. En primer lugar, al aumentar drásticamente el área de colección respecto a los IACTs actuales, se detectarán más rayos γ en todos los rangos de energía. En segundo lugar, cuando una misma cascada Cherenkov es observada por varios telescopios a la vez, es posible analizarla con mucha más precisión gracias a las técnicas estereoscópicas. La presente tesis recoge varios desarrollos técnicos realizados como aportación a los telescopios medianos y grandes de CTA, concretamente al sistema de trigger. Al ser las cascadas Cherenkov tan breves, los sistemas que digitalizan y leen los datos de cada píxel tienen que funcionar a frecuencias muy altas (≈1 GHz), lo que hace inviable que funcionen de forma continua, ya que la cantidad de datos guardada será inmanejable. En su lugar, las señales analógicas se muestrean, guardando las muestras analógicas en un buffer circular de unos pocos µs. Mientras las señales se mantienen en el buffer, el sistema de trigger hace un análisis rápido de las señales recibidas, y decide si la imagen que hay en el buér corresponde a una cascada Cherenkov y merece ser guardada, o por el contrario puede ignorarse permitiendo que el buffer se sobreescriba. La decisión de si la imagen merece ser guardada o no, se basa en que las cascadas Cherenkov producen detecciones de fotones en píxeles cercanos y en tiempos muy próximos, a diferencia de los fotones de NSB (night sky background), que llegan aleatoriamente. Para detectar cascadas grandes es suficiente con comprobar que más de un cierto número de píxeles en una región hayan detectado más de un cierto número de fotones en una ventana de tiempo de algunos nanosegundos. Sin embargo, para detectar cascadas pequeñas es más conveniente tener en cuenta cuántos fotones han sido detectados en cada píxel (técnica conocida como sumtrigger). El sistema de trigger desarrollado en esta tesis pretende optimizar la sensibilidad a bajas energías, por lo que suma analógicamente las señales recibidas en cada píxel en una región de trigger y compara el resultado con un umbral directamente expresable en fotones detectados (fotoelectrones). El sistema diseñado permite utilizar regiones de trigger de tamaño seleccionable entre 14, 21 o 28 píxeles (2, 3, o 4 clusters de 7 píxeles cada uno), y con un alto grado de solapamiento entre ellas. De este modo, cualquier exceso de luz en una región compacta de 14, 21 o 28 píxeles es detectado y genera un pulso de trigger. En la versión más básica del sistema de trigger, este pulso se distribuye por toda la cámara de forma que todos los clusters sean leídos al mismo tiempo, independientemente de su posición en la cámara, a través de un delicado sistema de distribución. De este modo, el sistema de trigger guarda una imagen completa de la cámara cada vez que se supera el número de fotones establecido como umbral en una región de trigger. Sin embargo, esta forma de operar tiene dos inconvenientes principales. En primer lugar, la cascada casi siempre ocupa sólo una pequeña zona de la cámara, por lo que se guardan muchos píxeles sin información alguna. Cuando se tienen muchos telescopios como será el caso de CTA, la cantidad de información inútil almacenada por este motivo puede ser muy considerable. Por otro lado, cada trigger supone guardar unos pocos nanosegundos alrededor del instante de disparo. Sin embargo, en el caso de cascadas grandes la duración de las mismas puede ser bastante mayor, perdiéndose parte de la información debido al truncamiento temporal. Para resolver ambos problemas se ha propuesto un esquema de trigger y lectura basado en dos umbrales. El umbral alto decide si hay un evento en la cámara y, en caso positivo, sólo las regiones de trigger que superan el nivel bajo son leídas, durante un tiempo más largo. De este modo se evita guardar información de píxeles vacíos y las imágenes fijas de las cascadas se pueden convertir en pequeños \vídeos" que representen el desarrollo temporal de la cascada. Este nuevo esquema recibe el nombre de COLIBRI (Concept for an Optimized Local Image Building and Readout Infrastructure), y se ha descrito detalladamente en el capítulo 5. Un problema importante que afecta a los esquemas de sumtrigger como el que se presenta en esta tesis es que para sumar adecuadamente las señales provenientes de cada píxel, estas deben tardar lo mismo en llegar al sumador. Los fotomultiplicadores utilizados en cada píxel introducen diferentes retardos que deben compensarse para realizar las sumas adecuadamente. El efecto de estos retardos ha sido estudiado, y se ha desarrollado un sistema para compensarlos. Por último, el siguiente nivel de los sistemas de trigger para distinguir efectivamente las cascadas Cherenkov del NSB consiste en buscar triggers simultáneos (o en tiempos muy próximos) en telescopios vecinos. Con esta función, junto con otras de interfaz entre sistemas, se ha desarrollado un sistema denominado Trigger Interface Board (TIB). Este sistema consta de un módulo que irá montado en la cámara de cada LST o MST, y que estará conectado mediante fibras ópticas a los telescopios vecinos. Cuando un telescopio tiene un trigger local, este se envía a todos los vecinos conectados y viceversa, de modo que cada telescopio sabe si sus vecinos han dado trigger. Una vez compensadas las diferencias de retardo debidas a la propagación en las fibras ópticas y de los propios fotones Cherenkov en el aire dependiendo de la dirección de apuntamiento, se buscan coincidencias, y en el caso de que la condición de trigger se cumpla, se lee la cámara en cuestión, de forma sincronizada con el trigger local. Aunque todo el sistema de trigger es fruto de la colaboración entre varios grupos, fundamentalmente IFAE, CIEMAT, ICC-UB y UCM en España, con la ayuda de grupos franceses y japoneses, el núcleo de esta tesis son el Level 1 y la Trigger Interface Board, que son los dos sistemas en los que que el autor ha sido el ingeniero principal. Por este motivo, en la presente tesis se ha incluido abundante información técnica relativa a estos sistemas. Existen actualmente importantes líneas de desarrollo futuras relativas tanto al trigger de la cámara (implementación en ASICs), como al trigger entre telescopios (trigger topológico), que darán lugar a interesantes mejoras sobre los diseños actuales durante los próximos años, y que con suerte serán de provecho para toda la comunidad científica participante en CTA. ABSTRACT -ray astronomy studies the most energetic particles arriving to the Earth from outer space. This -rays are not generated by thermal processes in mere stars, but by means of particle acceleration mechanisms in astronomical objects such as active galactic nuclei, pulsars, supernovas or as a result of dark matter annihilation processes. The γ rays coming from these objects and their characteristics provide with valuable information to the scientist which try to understand the underlying physical fundamentals of these objects, as well as to develop theoretical models able to describe them accurately. The problem when observing rays is that they are absorbed in the highest layers of the atmosphere, so they don't reach the Earth surface (otherwise the planet would be uninhabitable). Therefore, there are only two possible ways to observe γ rays: by using detectors on-board of satellites, or by observing their secondary effects in the atmosphere. When a γ ray reaches the atmosphere, it interacts with the particles in the air generating a highly energetic electron-positron pair. These secondary particles generate in turn more particles, with less energy each time. While these particles are still energetic enough to travel faster than the speed of light in the air, they produce a bluish radiation known as Cherenkov light during a few nanoseconds. From the Earth surface, some special telescopes known as Cherenkov telescopes or IACTs (Imaging Atmospheric Cherenkov Telescopes), are able to detect the Cherenkov light and even to take images of the Cherenkov showers. From these images it is possible to know the main parameters of the original -ray, and with some -rays it is possible to deduce important characteristics of the emitting object, hundreds of light-years away. However, detecting Cherenkov showers generated by γ rays is not a simple task. The showers generated by low energy -rays contain few photons and last few nanoseconds, while the ones corresponding to high energy -rays, having more photons and lasting more time, are much more unlikely. This results in two clearly differentiated development lines for IACTs: In order to detect low energy showers, big reflectors are required to collect as much photons as possible from the few ones that these showers have. On the contrary, small telescopes are able to detect high energy showers, but a large area in the ground should be covered to increase the number of detected events. With the aim to improve the sensitivity of current Cherenkov showers in the high (> 10 TeV), medium (100 GeV - 10 TeV) and low (10 GeV - 100 GeV) energy ranges, the CTA (Cherenkov Telescope Array) project was created. This project, with more than 27 participating countries, intends to build an observatory in each hemisphere, each one equipped with 4 large size telescopes (LSTs), around 30 middle size telescopes (MSTs) and up to 70 small size telescopes (SSTs). With such an array, two targets would be achieved. First, the drastic increment in the collection area with respect to current IACTs will lead to detect more -rays in all the energy ranges. Secondly, when a Cherenkov shower is observed by several telescopes at the same time, it is possible to analyze it much more accurately thanks to the stereoscopic techniques. The present thesis gathers several technical developments for the trigger system of the medium and large size telescopes of CTA. As the Cherenkov showers are so short, the digitization and readout systems corresponding to each pixel must work at very high frequencies (_ 1 GHz). This makes unfeasible to read data continuously, because the amount of data would be unmanageable. Instead, the analog signals are sampled, storing the analog samples in a temporal ring buffer able to store up to a few _s. While the signals remain in the buffer, the trigger system performs a fast analysis of the signals and decides if the image in the buffer corresponds to a Cherenkov shower and deserves to be stored, or on the contrary it can be ignored allowing the buffer to be overwritten. The decision of saving the image or not, is based on the fact that Cherenkov showers produce photon detections in close pixels during near times, in contrast to the random arrival of the NSB phtotons. Checking if more than a certain number of pixels in a trigger region have detected more than a certain number of photons during a certain time window is enough to detect large showers. However, taking also into account how many photons have been detected in each pixel (sumtrigger technique) is more convenient to optimize the sensitivity to low energy showers. The developed trigger system presented in this thesis intends to optimize the sensitivity to low energy showers, so it performs the analog addition of the signals received in each pixel in the trigger region and compares the sum with a threshold which can be directly expressed as a number of detected photons (photoelectrons). The trigger system allows to select trigger regions of 14, 21, or 28 pixels (2, 3 or 4 clusters with 7 pixels each), and with extensive overlapping. In this way, every light increment inside a compact region of 14, 21 or 28 pixels is detected, and a trigger pulse is generated. In the most basic version of the trigger system, this pulse is just distributed throughout the camera in such a way that all the clusters are read at the same time, independently from their position in the camera, by means of a complex distribution system. Thus, the readout saves a complete camera image whenever the number of photoelectrons set as threshold is exceeded in a trigger region. However, this way of operating has two important drawbacks. First, the shower usually covers only a little part of the camera, so many pixels without relevant information are stored. When there are many telescopes as will be the case of CTA, the amount of useless stored information can be very high. On the other hand, with every trigger only some nanoseconds of information around the trigger time are stored. In the case of large showers, the duration of the shower can be quite larger, loosing information due to the temporal cut. With the aim to solve both limitations, a trigger and readout scheme based on two thresholds has been proposed. The high threshold decides if there is a relevant event in the camera, and in the positive case, only the trigger regions exceeding the low threshold are read, during a longer time. In this way, the information from empty pixels is not stored and the fixed images of the showers become to little \`videos" containing the temporal development of the shower. This new scheme is named COLIBRI (Concept for an Optimized Local Image Building and Readout Infrastructure), and it has been described in depth in chapter 5. An important problem affecting sumtrigger schemes like the one presented in this thesis is that in order to add the signals from each pixel properly, they must arrive at the same time. The photomultipliers used in each pixel introduce different delays which must be compensated to perform the additions properly. The effect of these delays has been analyzed, and a delay compensation system has been developed. The next trigger level consists of looking for simultaneous (or very near in time) triggers in neighbour telescopes. These function, together with others relating to interfacing different systems, have been developed in a system named Trigger Interface Board (TIB). This system is comprised of one module which will be placed inside the LSTs and MSTs cameras, and which will be connected to the neighbour telescopes through optical fibers. When a telescope receives a local trigger, it is resent to all the connected neighbours and vice-versa, so every telescope knows if its neighbours have been triggered. Once compensated the delay differences due to propagation in the optical fibers and in the air depending on the pointing direction, the TIB looks for coincidences, and in the case that the trigger condition is accomplished, the camera is read a fixed time after the local trigger arrived. Despite all the trigger system is the result of the cooperation of several groups, specially IFAE, Ciemat, ICC-UB and UCM in Spain, with some help from french and japanese groups, the Level 1 and the Trigger Interface Board constitute the core of this thesis, as they have been the two systems designed by the author of the thesis. For this reason, a large amount of technical information about these systems has been included. There are important future development lines regarding both the camera trigger (implementation in ASICS) and the stereo trigger (topological trigger), which will produce interesting improvements for the current designs during the following years, being useful for all the scientific community participating in CTA.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Dynamic measurements will become a standard for bridge monitoring in the near future. This fact will produce an important cost reduction for maintenance. US Administration has a long term intensive research program in order to diminish the estimated current maintenance cost of US$7 billion per year over 20 years. An optimal intervention maintenance program demands a historical dynamical record, as well as an updated mathematical model of the structure to be monitored. In case that a model of the structure is not actually available it is possible to produce it, however this possibility does not exist for missing measurement records from the past. Current acquisition systems to monitor structures can be made more efficient by introducing the following improvements, under development in the Spanish research Project “Low cost bridge health monitoring by ambient vibration tests using wireless sensors”: (a) a complete wireless system to acquire sensor data, (b) a wireless system that permits the localization and the hardware identification of the whole sensor system. The applied localization system has been object of a recent patent, and (c) automatization of the modal identification process, aimed to diminish human intervention. This system is assembled with cheap components and allows the simultaneous use of a large number of sensors at a low placement cost. The engineer’s intervention is limited to the selection of sensor positions, probably based on a preliminary FE analysis. In case of multiple setups, also the position of a number of fixed reference sensors has to be decided. The wireless localization system will obtain the exact coordinates of all these sensors positions. When the selection of optimal positions is difficult, for example because of the lack of a proper FE model, this can be compensated by using a higher number of measuring (also reference) points. The described low cost acquisition system allows the responsible bridge administration to obtain historical dynamic identification records at reasonable costs that will be used in future maintenance programs. Therefore, due to the importance of the baseline monitoring record of a new bridge, a monitoring test just after its construction should be highly recommended, if not compulsory.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In some countries photovoltaic (PV) technology is at a stage of development at which it can compete with conventional electricity sources. A case in point is Germany where PV market has reached a mature stage. As a manifest of this, the German government has recently reduced the feed-in-tariff, which had been the strongest driver of PV diffusion. This development raises a fundamental question: Why would potential adopters be motivated to adopt PV when feed-in tariff diminishes? The point of departure for the literature on diffusion of PV has been on the effect of subsidies but little attention has paid to adopter motives when the policy support is scaled down. This paper presents an in-depth analysis of the adopter motives for photovoltaic applications. Anchored in an extensive exploratory case study we provide an encompassing explanation of roles of policy, adopters and system suppliers on diffusion of PV.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The new reactor concepts proposed in the Generation IV International Forum (GIF) are conceived to improve the use of natural resources, reduce the amount of high-level radioactive waste and excel in their reliability and safe operation. Among these novel designs sodium fast reactors (SFRs) stand out due to their technological feasibility as demonstrated in several countries during the last decades. As part of the contribution of EURATOM to GIF the CP-ESFR is a collaborative project with the objective, among others, to perform extensive analysis on safety issues involving renewed SFR demonstrator designs. The verification of computational tools able to simulate the plant behaviour under postulated accidental conditions by code-to-code comparison was identified as a key point to ensure reactor safety. In this line, several organizations employed coupled neutronic and thermal-hydraulic system codes able to simulate complex and specific phenomena involving multi-physics studies adapted to this particular fast reactor technology. In the “Introduction” of this paper the framework of this study is discussed, the second section describes the envisaged plant design and the commonly agreed upon modelling guidelines. The third section presents a comparative analysis of the calculations performed by each organisation applying their models and codes to a common agreed transient with the objective to harmonize the models as well as validating the implementation of all relevant physical phenomena in the different system codes.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

From the 60s to the 90s, a great number of events related to the Emergency Core Cooling Systems Strainers have been happened in all kind of reactors all over the world. Thus, the Nuclear Regulatory Commission of the USA emitted some Bulletins to address the concerns about the adequacy of Emergency Core Cooling Systems (ECCS) strainer performance at boiling water reactors (BWR). In Spain the regulatory body (Consejo de Seguridad Nuclear, CSN) adopted the USA regulation and Cofrentes NPP installed new strainers with a considerable bigger size than the old strainers. The nuclear industry conducted significant and extensive research, guidance development, testing, reviews, and hardware and procedure changes during the 90s to resolve the issues related to debris blockage of BWR strainers. In 2001 the NRC and CSN closed the Bulletins. Thereafter, the strainers issues were moved to the PWR reactors. In 2004 the NRC issued a Generic Letter (GL). It requested the resolution of several effects which were not noted in the past. The GL regarded to be resolved by the PWR reactors but the NRC in USA and the CSN in Spain have requested that the BWR reactors investigate differences between the methodologies used by the BWRs and PWRs. The developments and improvements done for Cofrentes NPP are detailed. Studies for this plant show that the head loss due to the considered debris is at most half of the limited head loss for the ECCS strainer and the NPSH (Net Positive Suction Head) required for the ECCS pumps is at least three times lower than the NPSH available.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Intensive farm systems handle large volume of livestock wastes, resulting in adverse environmental effects, such as gaseous losses into the atmosphere in form of ammonia (NH3) and greenhouse gases (GHG), i.e. methane (CH4), carbon dioxide (CO2) and nitrous oxide (N2O). In this study, the manure management continuum of slurry storage with impermeable cover and following cattle slurry band spreading and incorporation to soil was assessed for NH3 and GHG emissions. The experiment was conducted in an outdoor covered storage (flexible bag system) (study I), which collected the slurry produced in 7 dairy cattle farms (2,000 m3 slurry) during 12 days in the northern Spain.