884 resultados para Artificial satellites in telecommunications


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Con el objetivo de consolidar los conocimientos en Sistemas de Telecomunicación, se pretende diseñar una red para la distribución de servicios avanzados, como televisión, internet de banda ancha y telefonía, mediante el uso de la tecnología FTTH. La red será totalmente pasiva, óptica, y de gran ancho de banda, Tipo PON (Passive Optical Network), de manera que el haz de luz del emisor se distribuye hacia múltiples fibras siguiendo diferentes direcciones, o las confina en el sentido opuesto usando técnicas WDM y TDMA. En primer lugar, es preciso realizar un estudio teórico sobre las características de la fibra óptica, junto con las propiedades de los elementos activos y pasivos que interactúan con ella, para poder comprender la tecnología en la que se basa y así ofrecer una solución final acorde a las necesidades que se presenten. Tras un vistazo general a la tecnología de fibra óptica, se procede a estudiar las características y topologías de las redes de acceso basadas en la misma, junto con los requisitos de cara al diseño de la redes y a la gestión de proyectos. Se definirán los objetivos de la red, ya que la complejidad de la red depende de factores como la extensión y cobertura de la misma, o el ancho de banda. Posteriormente se irá diseñando la red en una arquitectura descendente hasta destinar una única fibra óptica para cada hogar dentro de la cobertura de la red. Se detallarán todas las especificaciones necesarias para definir la red, (potencias, modulaciones y tipos de fibra) así como el instrumental y demás medios necesarios para operar con la misma. Aunque el diseño de una red FTTH contempla muchos más aspectos de los estudiados en el presente proyecto, se tomó como objetivo el diseño y planificación correspondientes a una sección de un área de escasa población donde la instalación se tomó como viable, obviando así pequeños detalles y centrándose en el despliegue de fibra óptica. Se pretende así ofrecer una solución de infraestructura doméstica de telecomunicaciones, apta para las futuras necesidades que implican la creciente demanda de servicios como internet de banda ancha o almacenamiento y procesado de aplicaciones en nube. La fibra óptica tiene mucho que ofrecer a las telecomunicaciones, y cada vez es más frecuente encontrar redes de fibra óptica como parte de un servicio de telecomunicaciones, tanto como para particulares como empresas. ABSTRACT. Aiming to strengthen knowledge in Telecommunications Systems, is intended to design a network for advanced services broadcasting, including digital TV, broadband internet and telephony, by using FTTH technology. This network will be entirely passive optic, and high bandwidth, PON type (Passive Optical Network) so that the transmitter beam is broadcasting to multiple optical fibers, branching out in different ways or joining them in the opposite. First, it is necessary to perform a theoretical study on the characteristics of the optical fiber, along with the properties of the active and passive elements that interact with it, to understand the technology that is based and offer a final solution according needs that arise. After an overview of the optical fiber technology, we proceed to study the characteristics and topologies access networks based on that, together with the requirements to face the network design and project management. System objectives will be determined, since the complexity of the network depends on factors such as the size and scope of it, or bandwidth. Later, the network will be designing in downstream architecture to deliver a single optical fiber to each household within the network area. All necessary specifications to define the network (power, modulation and fiber types) will be determined, as instruments and other means to operate it. Although FTTH network design includes many more aspects of those studied in this project, it was taken as objective the network design and planning corresponding to a section of a sparsely populated area where the facility was taken as feasible thus obviating small details and focus on the deployment of optical fiber. This is intended to provide a solution of household telecommunications infrastructure suitable for future needs involving the growing demand for services such as broadband internet or processing storage and cloud applications. The optical fiber has much to offer to telecommunications, and it is increasingly common to find fiber optic networks as part of a telecommunications service, both to individuals and businesses.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Objectives: A recently introduced pragmatic scheme promises to be a useful catalog of interneuron names.We sought to automatically classify digitally reconstructed interneuronal morphologies according tothis scheme. Simultaneously, we sought to discover possible subtypes of these types that might emergeduring automatic classification (clustering). We also investigated which morphometric properties weremost relevant for this classification.Materials and methods: A set of 118 digitally reconstructed interneuronal morphologies classified into thecommon basket (CB), horse-tail (HT), large basket (LB), and Martinotti (MA) interneuron types by 42 of theworld?s leading neuroscientists, quantified by five simple morphometric properties of the axon and fourof the dendrites. We labeled each neuron with the type most commonly assigned to it by the experts. Wethen removed this class information for each type separately, and applied semi-supervised clustering tothose cells (keeping the others? cluster membership fixed), to assess separation from other types and lookfor the formation of new groups (subtypes). We performed this same experiment unlabeling the cells oftwo types at a time, and of half the cells of a single type at a time. The clustering model is a finite mixtureof Gaussians which we adapted for the estimation of local (per-cluster) feature relevance. We performedthe described experiments on three different subsets of the data, formed according to how many expertsagreed on type membership: at least 18 experts (the full data set), at least 21 (73 neurons), and at least26 (47 neurons).Results: Interneurons with more reliable type labels were classified more accurately. We classified HTcells with 100% accuracy, MA cells with 73% accuracy, and CB and LB cells with 56% and 58% accuracy,respectively. We identified three subtypes of the MA type, one subtype of CB and LB types each, andno subtypes of HT (it was a single, homogeneous type). We got maximum (adapted) Silhouette widthand ARI values of 1, 0.83, 0.79, and 0.42, when unlabeling the HT, CB, LB, and MA types, respectively,confirming the quality of the formed cluster solutions. The subtypes identified when unlabeling a singletype also emerged when unlabeling two types at a time, confirming their validity. Axonal morphometricproperties were more relevant that dendritic ones, with the axonal polar histogram length in the [pi, 2pi) angle interval being particularly useful.Conclusions: The applied semi-supervised clustering method can accurately discriminate among CB, HT, LB, and MA interneuron types while discovering potential subtypes, and is therefore useful for neuronal classification. The discovery of potential subtypes suggests that some of these types are more heteroge-neous that previously thought. Finally, axonal variables seem to be more relevant than dendritic ones fordistinguishing among the CB, HT, LB, and MA interneuron types.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

El principal objetivo de la tesis es estudiar el acoplamiento entre los subsistemas de control de actitud y de control térmico de un pequeño satélite, con el fin de buscar la solución a los problemas relacionados con la determinación de los parámetros de diseño. Se considera la evolución de la actitud y de las temperaturas del satélite bajo la influencia de dos estrategias de orientación diferentes: 1) estabilización magnética pasiva de la orientación (PMAS, passive magnetic attitude stabilization), y 2) control de actitud magnético activo (AMAC, active magnetic attitude control). En primer lugar se presenta el modelo matemático del problema, que incluye la dinámica rotacional y el modelo térmico. En el problema térmico se considera un satélite cúbico modelizado por medio de siete nodos (seis externos y uno interno) aplicando la ecuación del balance térmico. Una vez establecido el modelo matemático del problema, se estudia la evolución que corresponde a las dos estrategias mencionadas. La estrategia PMAS se ha seleccionado por su simplicidad, fiabilidad, bajo coste, ahorrando consumo de potencia, masa coste y complejidad, comparado con otras estrategias. Se ha considerado otra estrategia de control que consigue que el satélite gire a una velocidad requerida alrededor de un eje deseado de giro, pudiendo controlar su dirección en un sistema inercial de referencia, ya que frecuentemente el subsistema térmico establece requisitos de giro alrededor de un eje del satélite orientado en una dirección perpendicular a la radiación solar incidente. En relación con el problema térmico, para estudiar la influencia de la velocidad de giro en la evolución de las temperaturas en diversos puntos del satélite, se ha empleado un modelo térmico linealizado, obtenido a partir de la formulación no lineal aplicando un método de perturbaciones. El resultado del estudio muestra que el tiempo de estabilización de la temperatura y la influencia de las cargas periódicas externas disminuye cuando aumenta la velocidad de giro. Los cambios de temperatura se reducen hasta ser muy pequeños para velocidades de rotación altas. En relación con la estrategia PMAC se ha observado que a pesar de su uso extendido entre los micro y nano satélites todavía presenta problemas que resolver. Estos problemas están relacionados con el dimensionamiento de los parámetros del sistema y la predicción del funcionamiento en órbita. Los problemas aparecen debido a la dificultad en la determinación de las características magnéticas de los cuerpos ferromagnéticos (varillas de histéresis) que se utilizan como amortiguadores de oscilaciones en los satélites. Para estudiar este problema se presenta un modelo analítico que permite estimar la eficiencia del amortiguamiento, y que se ha aplicado al estudio del comportamiento en vuelo de varios satélites, y que se ha empleado para comparar los resultados del modelo con los obtenidos en vuelo, observándose que el modelo permite explicar satisfactoriamente el comportamiento registrado. ABSTRACT The main objective of this thesis is to study the coupling between the attitude control and thermal control subsystems of a small satellite, and address the solution to some existing issues concerning the determination of their parameters. Through the thesis the attitude and temperature evolution of the satellite is studied under the influence of two independent attitude stabilization and control strategies: (1) passive magnetic attitude stabilization (PMAS), and (2) active magnetic attitude control (AMAC). In this regard the mathematical model of the problem is explained and presented. The mathematical model includes both the rotational dynamics and the thermal model. The thermal model is derived for a cubic satellite by solving the heat balance equation for 6 external and 1 internal nodes. Once established the mathematical model of the problem, the above mentioned attitude strategies were applied to the system and the temperature evolution of the 7 nodes of the satellite was studied. The PMAS technique has been selected to be studied due to its prevalent use, simplicity, reliability, and cost, as this strategy significantly saves the overall power, weight, cost, and reduces the complexity of the system compared to other attitude control strategies. In addition to that, another control law that provides the satellite with a desired spin rate along a desired axis of the satellite, whose direction can be controlled with respect to the inertial reference frame is considered, as the thermal subsystem of a satellite usually demands a spin requirement around an axis of the satellite which is positioned perpendicular to the direction of the coming solar radiation. Concerning the thermal problem, to study the influence of spin rate on temperature evolution of the satellite a linear approach of the thermal model is used, which is based on perturbation theory applied to the nonlinear differential equations of the thermal model of a spacecraft moving in a closed orbit. The results of this study showed that the temperature stabilization time and the periodic influence of the external thermal loads decreases by increasing the spin rate. However, the changes become insignificant for higher values of spin rate. Concerning the PMAS strategy, it was observed that in spite of its extended application to micro and nano satellites, still there are some issues to be solved regarding this strategy. These issues are related to the sizing of its system parameters and predicting the in-orbit performance. The problems were found to be rooted in the difficulties that exist in determining the magnetic characteristics of the ferromagnetic bodies (hysteresis rods) that are applied as damping devices on-board satellites. To address these issues an analytic model for estimating their damping efficiency is proposed and applied to several existing satellites in order to compare the results with their respective in-flight data. This model can explain the behavior showed by these satellites.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Parte de la investigación biomédica actual se encuentra centrada en el análisis de datos heterogéneos. Estos datos pueden tener distinto origen, estructura, y semántica. Gran cantidad de datos de interés para los investigadores se encuentran en bases de datos públicas, que recogen información de distintas fuentes y la ponen a disposición de la comunidad de forma gratuita. Para homogeneizar estas fuentes de datos públicas con otras de origen privado, existen diversas herramientas y técnicas que permiten automatizar los procesos de homogeneización de datos heterogéneos. El Grupo de Informática Biomédica (GIB) [1] de la Universidad Politécnica de Madrid colabora en el proyecto europeo P-medicine [2], cuya finalidad reside en el desarrollo de una infraestructura que facilite la evolución de los procedimientos médicos actuales hacia la medicina personalizada. Una de las tareas enmarcadas en el proyecto P-medicine que tiene asignado el grupo consiste en elaborar herramientas que ayuden a usuarios en el proceso de integración de datos contenidos en fuentes de información heterogéneas. Algunas de estas fuentes de información son bases de datos públicas de ámbito biomédico contenidas en la plataforma NCBI [3] (National Center for Biotechnology Information). Una de las herramientas que el grupo desarrolla para integrar fuentes de datos es Ontology Annotator. En una de sus fases, la labor del usuario consiste en recuperar información de una base de datos pública y seleccionar de forma manual los resultados relevantes. Para automatizar el proceso de búsqueda y selección de resultados relevantes, por un lado existe un gran interés en conseguir generar consultas que guíen hacia resultados lo más precisos y exactos como sea posible, por otro lado, existe un gran interés en extraer información relevante de elevadas cantidades de documentos, lo cual requiere de sistemas que analicen y ponderen los datos que caracterizan a los mismos. En el campo informático de la inteligencia artificial, dentro de la rama de la recuperación de la información, existen diversos estudios acerca de la expansión de consultas a partir de retroalimentación relevante que podrían ser de gran utilidad para dar solución a la cuestión. Estos estudios se centran en técnicas para reformular o expandir la consulta inicial utilizando como realimentación los resultados que en una primera instancia fueron relevantes para el usuario, de forma que el nuevo conjunto de resultados tenga mayor proximidad con los que el usuario realmente desea. El objetivo de este trabajo de fin de grado consiste en el estudio, implementación y experimentación de métodos que automaticen el proceso de extracción de información trascendente de documentos, utilizándola para expandir o reformular consultas. De esta forma se pretende mejorar la precisión y el ranking de los resultados asociados. Dichos métodos serán integrados en la herramienta Ontology Annotator y enfocados a la fuente de datos de PubMed [4].---ABSTRACT---Part of the current biomedical research is focused on the analysis of heterogeneous data. These data may have different origin, structure and semantics. A big quantity of interesting data is contained in public databases which gather information from different sources and make it open and free to be used by the community. In order to homogenize thise sources of public data with others which origin is private, there are some tools and techniques that allow automating the processes of integration heterogeneous data. The biomedical informatics group of the Universidad Politécnica de Madrid cooperates with the European project P-medicine which main purpose is to create an infrastructure and models to facilitate the transition from current medical practice to personalized medicine. One of the tasks of the project that the group is in charge of consists on the development of tools that will help users in the process of integrating data from diverse sources. Some of the sources are biomedical public data bases from the NCBI platform (National Center for Biotechnology Information). One of the tools in which the group is currently working on for the integration of data sources is called the Ontology Annotator. In this tool there is a phase in which the user has to retrieve information from a public data base and select the relevant data contained in it manually. For automating the process of searching and selecting data on the one hand, there is an interest in automatically generating queries that guide towards the more precise results as possible. On the other hand, there is an interest on retrieve relevant information from large quantities of documents. The solution requires systems that analyze and weigh the data allowing the localization of the relevant items. In the computer science field of the artificial intelligence, in the branch of information retrieval there are diverse studies about the query expansion from relevance feedback that could be used to solve the problem. The main purpose of this studies is to obtain a set of results that is the closer as possible to the information that the user really wants to retrieve. In order to reach this purpose different techniques are used to reformulate or expand the initial query using a feedback the results that where relevant for the user, with this method, the new set of results will have more proximity with the ones that the user really desires. The goal of this final dissertation project consists on the study, implementation and experimentation of methods that automate the process of extraction of relevant information from documents using this information to expand queries. This way, the precision and the ranking of the results associated will be improved. These methods will be integrated in the Ontology Annotator tool and will focus on the PubMed data source.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

online donde el contenido es creado por los propios usuarios mediante el uso de las tecnologías de la web 2.0, que facilitan la edición, la publicación y el intercambio de información, de tal manera que los social media están democratizando las comunicaciones. Según datos publicados por eMarketer, casi el 25% del total de la inversión publicitaria mundial para el 2014 estará dirigida a medios digitales. Los medios de comunicación tradicionales están sumidos desde algún tiempo en una profunda crisis. En los comienzos de Internet, los protagonistas de este sector han creído en la red como un medio de publicación de contenidos en una única dirección, esto es la web 1.0, personas conectándose a páginas web. El típico lector de revista y seminario se ha convertido en un nuevo consumidor con necesidades muy distintas. El lector de un medio tradicional se encuentra con contenido cerrado, sin posibilidad de ser filtrado, que busca dirigir su opinión. Con el método tradicional no es posible ni personalizar el contenido ni la experiencia de uso. De ahí el nacimiento de la web 2.0, personas conectándose a personas (redes sociales). En un futuro cercano, van a triunfar las empresas, sectores y marcas que sepan adaptarse a la llegada de los social media, y que además de desarrollar su línea de negocio habitual, se conviertan en plataformas para la conversación y la construcción de contenidos en colaboración. En este proyecto voy a desarrollar el caso particular de una empresa de telecomunicaciones con una estrategia que use los social media como plataforma de atención al cliente, y para interaccionar también con su almacén en DHL. Los servicios añadidos que dicha empresa de telecomunicaciones ofrece a sus seguidores, clientes externos e internos se pueden agrupar en las siguientes funcionalidades:  Hacer un volcado del sistema de gestión de incidencias de una empresa de telecomunicaciones en las redes sociales, a través de mensajes privados. Dichos mensajes privados se enviarán a las páginas de Facebook y Twitter de los clientes.  Seguimiento por parte del cliente de puestas en marcha de los nuevos servicios a través de mensajes privados en Facebook y Twitter. Mediante este servicio el cliente podrá comprobar a través de las redes sociales el estado de su pedido e intercambiar información de utilidad con la empresa.  Desarrollar a través de las redes sociales una herramienta de marketing para interaccionar con los clientes y posibles seguidores. A través de los canales de los social media se proporciona a los seguidores información de valor sobre novedades del mercado de las telecomunicaciones, soporte técnico, dudas, consejos sobre implementaciones técnicas, equipos de telecomunicaciones,.etc.Es decir, crear una plataforma de conversación entre usuarios, que aporte a los seguidores conocimientos y soluciones útiles para la problemática de sus negocios.  Usar Facebook como base de datos y medio de comunicación entre nuestra empresa de telecomunicaciones y el almacén externalizado en DHL. Supongo que la primera pregunta que debo responder sería, ¿por qué un cliente de una empresa de telecomunicaciones estaría interesado en seguir a su proveedor a través de redes sociales? Facebook tiene en 2014 1,28 billones de usuarios activos. Twitter, la cuarta plataforma social por número de usuarios tras Facebook, alcanzó los 255 millones de usuarios en los primeros meses de 2014. La segunda red social en número de seguidores es Youtube, con un billón de usuarios. El 61,5% de los usuarios de Facebook visita la red todos los días. Los twitteros envían 110 millones tweets por día, tienen fama de muy activos, especialmente por la mañana. En 60 días, se sube más contenido de vídeo a YouTube que el producido por las tres principales cadenas de televisión estadounidense durante 60 años. Como conclusión, las redes sociales están ganando cada día más terreno, es un entorno de comunicación bidireccional en auge cada vez más extendido, por lo cual los usuarios no tienen que dedicar tiempo extra a conocer la herramienta de atención al cliente de su empresa, al usar ésta un canal fácil, conocido, rápido, versátil y gratuito como son las redes sociales. ABSTRACT. Social media are platforms of online communication where content is created by users themselves using technologies of web 2.0, which facilitate the editing, publication and the exchange of information, so that the social media are democratizing communications. According to data published by eMarketer, nearly 25% of the total global ad spend for 2014 will target digital media. The traditional media are trapped for some time in deep crisis. In the early days of the internet, the players of the sector have believed in the network as a means of publishing content in a single direction, this is the web 1.0, connecting people to websites. The typical magazine reader has become a new consumer with very different needs. Traditional medium readers encounter a closed content, unable to be filtered, which seeks handle their opinion. The traditional method doesn’t allow to customize the content and user experience. Hence the birth of Web 2.0, connecting people to people. In the near future, will succeed companies, sectors and brands who can adapt to the advent of social media, and further to develop their line of business as usual, become platforms for conversation and building collaborative content. In this project I will develop the particular case of a telecommunications company with a strategy to use social media as a platform for Customer Service, and also to interact with DHL warehouse. The additional services that the company offers telecommunications followers, external and internal customers can be grouped into the following functionality:  Make a dump of the incident management system of a telecommunications company in social networks through private messages. These private messages are sent Facebook and Twitter clients.  Monitoring by the client launched new services through private messages on Facebook and Twitter. Through this service, customers can check through the social networks of your order status and share useful information with the company.  Develop through social media as a marketing tool to interact with customers and potential followers.Our telco will provide to followers through social media channels valuable information on market developments in telecommunications, technical support, questions, advice on technical implementations, telecommunications equipment,etc. Then we create a platform for discussion between users, to provide followers useful knowledge and solutions to their business problems.  Use Facebook as a database and means of communication between our telco and outsourced telecommunications in DHL warehouse. I guess the first question that must be answered is why a customer of a telecommunications company would be interested in following your provider via social networks. Facebook has 1.28 billion active users in 2014. Twitter, the fourth social platform for many users after Facebook reached 255 million users in the first months of 2014. The second social network on number of followers is YouTube, with a billion users. 61.5% of Facebook users visit the network every day. The twitter users send 110 million tweets per day, have a reputation for very active, especially in the morning. In 60 days, it is produced more video content to YouTube by the three major US television for 60 years. In conclusion, social networks are gaining more ground every day. It is an environment of two-way communication on the rise, so users do not have to spend extra time to learn the tool of Customer Care telco, by using this easy, known, fast, versatile and free channels as are social networks.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Each year more than 250,000 infants in the United States are exposed to artificial lighting in hospital nurseries with little consideration given to environmental lighting cycles. Essential in determining whether environmental lighting cycles need to be considered in hospital nurseries is identifying when the infant’s endogenous circadian clock becomes responsive to light. Using a non-human primate model of the developing human, we examined when the circadian clock, located in the hypothalamic suprachiasmatic nuclei (SCN), becomes responsive to light. Preterm infant baboons of different ages were exposed to light (5,000 lux) at night, and then changes in SCN metabolic activity and gene expression were assessed. After exposure to bright light at night, robust increases in SCN metabolic activity and gene expression were seen at ages that were equivalent to human infants at 24 weeks after conception. These data provide direct evidence that the biological clock of very premature primate infants is responsive to light.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

En el contexto rural de Murcia, la recepción de la modernidad fue un fenómeno episódico y tardío. Hacia 1960 se produjo, sin embargo, un breve e ilusionante periodo cuando algunos maestros de la arquitectura española llegaron para trabajar en el desarrollo turístico de un litoral virgen y lleno de oportunidades. Entre ellos, Antoni Bonet recibió el encargo de urbanizar La Manga, una estrecha franja de tierra de 24km que separa el Mediterráneo del Mar Menor. El Plan retomaba los planteamientos urbanos de Le Corbusier para reelaborar algunas ideas preparatorias: autonomía y jerarquía de los núcleos habitados, eje vertebrador, centro cívico al borde del mar, tratamiento del frente marítimo con islas artificiales. Minimizando su impacto en el ecosistema dunar con una carretera elevada puntualmente para favorecer el aporte de arena, el Plan concentraba la edificación alrededor de torres de 16 plantas, generando unidades compactas de viviendas y equipamientos repetidas cada 2km. Inspirándose en la lectura del territorio de Le Corbusier en Buenos Aires (1929), Bonet respondía a la horizontalidad del plano del mar por contraste topológico, subrayando la presencia vertical de las torres como hitos paisajísticos que, desde la orilla opuesta, marcaban la posición de cada núcleo.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Dissertação apresentada à Escola Superior Agrária do Instituto Politécnico de Castelo Branco para cumprimento dos requisitos necessários à obtenção do grau de Mestre em Fruticultura Integrada.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Globalization is both an integrative and deconstructive process. Globalization integrates states and non-state actors into transnational and global networks (Keohane & Nye, 2000, p. 105). These networks are based on multiple channels of interdependence that include trade, politics, security, environment, and socio-cultural ties (pp. 106-107). Due to advances in telecommunications technology, the expansion of globalization “shrinks” the distance between peoples (p. 105). On the other hand, globalization can also break up the existing political and social order (Mathews, 1997, p. 50). Globalization disperses power and information flows, thus enabling local and transnational identity movements to challenge states (pp. 51-52). This can be exemplified by separatist movements that seek to break away from central authorities.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Der Wiederkäuerklinik oder dem Institut für Genetik der Universität Bern wurden zwi-schen 2012 und 2014 insgesamt 5 Rinder der Rasse Simmental vorgestellt, die je-weils nicht sistierende Blutungen nach Trauma zeigten. Alle betroffenen Tiere waren homozygote Träger für die seit 2007 bekannte RASGRP2 Mutation. Die verfügbaren Eltern wurden als heterozygote Anlageträger genotypisiert, was somit einen rezessi-ven Erbgang bestätigt. Drei erkrankte Tiere sind an den Folgen der unstillbaren Blu-tungen verstorben. Ein Tier konnte stabilisiert werden und wurde einen Monat nach der Entlassung aus der Klinik geschlachtet. Bei einem weiteren Fall wurden wieder-holt andauernde Blutungen sowie mehrmals Hämatome festgestellt und nach der genetischen Analyse wurde das Rind euthanasiert. Die Genotypisierung einer Stich-probe von 145 Stieren, die im Jahr 2013 in der Schweiz in der künstlichen Besamung zum Einsatz kamen, zeigte, dass 10% der getesteten Stiere in der Schweiz Anlage-träger für die assoziierte Mutation sind. Diese Stiere werden mit TP carrier gekenn-zeichnet und sollten zukünftig nicht mehr unkontrolliert eingesetzt werden. Die Zuchtverantwortlichen in der Schweiz nutzen heute den Gentest systematisch zur Selektion von anlagefreien Stieren.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Supersedes U.S. Dept. of Agriculture Technical bulletin 492. Artificial reforestation in the southern pine region, by P.C. Wakeley.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

With the extensive use of pulse modulation methods in telecommunications, much work has been done in the search for a better utilisation of the transmission channel.The present research is an extension of these investigations. A new modulation method, 'Variable Time-Scale Information Processing', (VTSIP), is proposed.The basic principles of this system have been established, and the main advantages and disadvantages investigated. With the proposed system, comparison circuits detect the instants at which the input signal voltage crosses predetermined amplitude levels.The time intervals between these occurrences are measured digitally and the results are temporarily stored, before being transmitted.After reception, an inverse process enables the original signal to be reconstituted.The advantage of this system is that the irregularities in the rate of information contained in the input signal are smoothed out before transmission, allowing the use of a smaller transmission bandwidth. A disadvantage of the system is the time delay necessarily introduced by the storage process.Another disadvantage is a type of distortion caused by the finite store capacity.A simulation of the system has been made using a standard speech signal, to make some assessment of this distortion. It is concluded that the new system should be an improvement on existing pulse transmission systems, allowing the use of a smaller transmission bandwidth, but introducing a time delay.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Substantial altimetry datasets collected by different satellites have only become available during the past five years, but the future will bring a variety of new altimetry missions, both parallel and consecutive in time. The characteristics of each produced dataset vary with the different orbital heights and inclinations of the spacecraft, as well as with the technical properties of the radar instrument. An integral analysis of datasets with different properties offers advantages both in terms of data quantity and data quality. This thesis is concerned with the development of the means for such integral analysis, in particular for dynamic solutions in which precise orbits for the satellites are computed simultaneously. The first half of the thesis discusses the theory and numerical implementation of dynamic multi-satellite altimetry analysis. The most important aspect of this analysis is the application of dual satellite altimetry crossover points as a bi-directional tracking data type in simultaneous orbit solutions. The central problem is that the spatial and temporal distributions of the crossovers are in conflict with the time-organised nature of traditional solution methods. Their application to the adjustment of the orbits of both satellites involved in a dual crossover therefore requires several fundamental changes of the classical least-squares prediction/correction methods. The second part of the thesis applies the developed numerical techniques to the problems of precise orbit computation and gravity field adjustment, using the altimetry datasets of ERS-1 and TOPEX/Poseidon. Although the two datasets can be considered less compatible that those of planned future satellite missions, the obtained results adequately illustrate the merits of a simultaneous solution technique. In particular, the geographically correlated orbit error is partially observable from a dataset consisting of crossover differences between two sufficiently different altimetry datasets, while being unobservable from the analysis of altimetry data of both satellites individually. This error signal, which has a substantial gravity-induced component, can be employed advantageously in simultaneous solutions for the two satellites in which also the harmonic coefficients of the gravity field model are estimated.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Objective: Biomedical events extraction concerns about events describing changes on the state of bio-molecules from literature. Comparing to the protein-protein interactions (PPIs) extraction task which often only involves the extraction of binary relations between two proteins, biomedical events extraction is much harder since it needs to deal with complex events consisting of embedded or hierarchical relations among proteins, events, and their textual triggers. In this paper, we propose an information extraction system based on the hidden vector state (HVS) model, called HVS-BioEvent, for biomedical events extraction, and investigate its capability in extracting complex events. Methods and material: HVS has been previously employed for extracting PPIs. In HVS-BioEvent, we propose an automated way to generate abstract annotations for HVS training and further propose novel machine learning approaches for event trigger words identification, and for biomedical events extraction from the HVS parse results. Results: Our proposed system achieves an F-score of 49.57% on the corpus used in the BioNLP'09 shared task, which is only 2.38% lower than the best performing system by UTurku in the BioNLP'09 shared task. Nevertheless, HVS-BioEvent outperforms UTurku's system on complex events extraction with 36.57% vs. 30.52% being achieved for extracting regulation events, and 40.61% vs. 38.99% for negative regulation events. Conclusions: The results suggest that the HVS model with the hierarchical hidden state structure is indeed more suitable for complex event extraction since it could naturally model embedded structural context in sentences.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Objective: Recently, much research has been proposed using nature inspired algorithms to perform complex machine learning tasks. Ant colony optimization (ACO) is one such algorithm based on swarm intelligence and is derived from a model inspired by the collective foraging behavior of ants. Taking advantage of the ACO in traits such as self-organization and robustness, this paper investigates ant-based algorithms for gene expression data clustering and associative classification. Methods and material: An ant-based clustering (Ant-C) and an ant-based association rule mining (Ant-ARM) algorithms are proposed for gene expression data analysis. The proposed algorithms make use of the natural behavior of ants such as cooperation and adaptation to allow for a flexible robust search for a good candidate solution. Results: Ant-C has been tested on the three datasets selected from the Stanford Genomic Resource Database and achieved relatively high accuracy compared to other classical clustering methods. Ant-ARM has been tested on the acute lymphoblastic leukemia (ALL)/acute myeloid leukemia (AML) dataset and generated about 30 classification rules with high accuracy. Conclusions: Ant-C can generate optimal number of clusters without incorporating any other algorithms such as K-means or agglomerative hierarchical clustering. For associative classification, while a few of the well-known algorithms such as Apriori, FP-growth and Magnum Opus are unable to mine any association rules from the ALL/AML dataset within a reasonable period of time, Ant-ARM is able to extract associative classification rules.