997 resultados para Software geomere Cabri II
Resumo:
Service providers make use of cost-effective wireless solutions to identify, localize, and possibly track users using their carried MDs to support added services, such as geo-advertisement, security, and management. Indoor and outdoor hotspot areas play a significant role for such services. However, GPS does not work in many of these areas. To solve this problem, service providers leverage available indoor radio technologies, such as WiFi, GSM, and LTE, to identify and localize users. We focus our research on passive services provided by third parties, which are responsible for (i) data acquisition and (ii) processing, and network-based services, where (i) and (ii) are done inside the serving network. For better understanding of parameters that affect indoor localization, we investigate several factors that affect indoor signal propagation for both Bluetooth and WiFi technologies. For GSM-based passive services, we developed first a data acquisition module: a GSM receiver that can overhear GSM uplink messages transmitted by MDs while being invisible. A set of optimizations were made for the receiver components to support wideband capturing of the GSM spectrum while operating in real-time. Processing the wide-spectrum of the GSM is possible using a proposed distributed processing approach over an IP network. Then, to overcome the lack of information about tracked devices’ radio settings, we developed two novel localization algorithms that rely on proximity-based solutions to estimate in real environments devices’ locations. Given the challenging indoor environment on radio signals, such as NLOS reception and multipath propagation, we developed an original algorithm to detect and remove contaminated radio signals before being fed to the localization algorithm. To improve the localization algorithm, we extended our work with a hybrid based approach that uses both WiFi and GSM interfaces to localize users. For network-based services, we used a software implementation of a LTE base station to develop our algorithms, which characterize the indoor environment before applying the localization algorithm. Experiments were conducted without any special hardware, any prior knowledge of the indoor layout or any offline calibration of the system.
Resumo:
Service providers make use of cost-effective wireless solutions to identify, localize, and possibly track users using their carried MDs to support added services, such as geo-advertisement, security, and management. Indoor and outdoor hotspot areas play a significant role for such services. However, GPS does not work in many of these areas. To solve this problem, service providers leverage available indoor radio technologies, such as WiFi, GSM, and LTE, to identify and localize users. We focus our research on passive services provided by third parties, which are responsible for (i) data acquisition and (ii) processing, and network-based services, where (i) and (ii) are done inside the serving network. For better understanding of parameters that affect indoor localization, we investigate several factors that affect indoor signal propagation for both Bluetooth and WiFi technologies. For GSM-based passive services, we developed first a data acquisition module: a GSM receiver that can overhear GSM uplink messages transmitted by MDs while being invisible. A set of optimizations were made for the receiver components to support wideband capturing of the GSM spectrum while operating in real-time. Processing the wide-spectrum of the GSM is possible using a proposed distributed processing approach over an IP network. Then, to overcome the lack of information about tracked devices’ radio settings, we developed two novel localization algorithms that rely on proximity-based solutions to estimate in real environments devices’ locations. Given the challenging indoor environment on radio signals, such as NLOS reception and multipath propagation, we developed an original algorithm to detect and remove contaminated radio signals before being fed to the localization algorithm. To improve the localization algorithm, we extended our work with a hybrid based approach that uses both WiFi and GSM interfaces to localize users. For network-based services, we used a software implementation of a LTE base station to develop our algorithms, which characterize the indoor environment before applying the localization algorithm. Experiments were conducted without any special hardware, any prior knowledge of the indoor layout or any offline calibration of the system.
Resumo:
BACKGROUND AND OBJECTIVES Multiple-breath washout (MBW) is an attractive test to assess ventilation inhomogeneity, a marker of peripheral lung disease. Standardization of MBW is hampered as little data exists on possible measurement bias. We aimed to identify potential sources of measurement bias based on MBW software settings. METHODS We used unprocessed data from nitrogen (N2) MBW (Exhalyzer D, Eco Medics AG) applied in 30 children aged 5-18 years: 10 with CF, 10 formerly preterm, and 10 healthy controls. This setup calculates the tracer gas N2 mainly from measured O2 and CO2concentrations. The following software settings for MBW signal processing were changed by at least 5 units or >10% in both directions or completely switched off: (i) environmental conditions, (ii) apparatus dead space, (iii) O2 and CO2 signal correction, and (iv) signal alignment (delay time). Primary outcome was the change in lung clearance index (LCI) compared to LCI calculated with the settings as recommended. A change in LCI exceeding 10% was considered relevant. RESULTS Changes in both environmental and dead space settings resulted in uniform but modest LCI changes and exceeded >10% in only two measurements. Changes in signal alignment and O2 signal correction had the most relevant impact on LCI. Decrease of O2 delay time by 40 ms (7%) lead to a mean LCI increase of 12%, with >10% LCI change in 60% of the children. Increase of O2 delay time by 40 ms resulted in mean LCI decrease of 9% with LCI changing >10% in 43% of the children. CONCLUSIONS Accurate LCI results depend crucially on signal processing settings in MBW software. Especially correct signal delay times are possible sources of incorrect LCI measurements. Algorithms of signal processing and signal alignment should thus be optimized to avoid susceptibility of MBW measurements to this significant measurement bias.
Resumo:
En el presente trabajo se exponen las condiciones de la creación de la única unidad de información para dos Unidades Académicas dentro de la UNLP -la Biblioteca Conjunta de Facultad de Ciencias Agrarias y Forestales y de la Facultad de Ciencias Veterinarias- y cómo esta situación influyó directamente en la conformación del OPAC de la Biblioteca Conjunta en su reconversión al sistema de gestión de bibliotecas Koha-UNLP. Dado que Bienes Patrimoniales requiere que se mantenga el inventario de cadabiblioteca origen, se tuvo que implementar una codificación para asignar losinventarios. Ya que Koha-UNLP genera los inventarios en forma automática, en nuestrocaso decidimos completar manualmente el campo "código de barras", para así poder agregar al número de inventario y la codificación correspondiente a cada Facultad. Dicha codificación es la que dio el CAICYT, DAG (Agrarias/Forestales), DVE (Veterinaria). De esta manera se diferencian los fondos bibliográficos de las distintas facultades, y a su vez el fondo completo lleva la identificación dada por nosotros de Biblioteca Conjunta (DAV) para facilitar la circulación de los libros. Al quedar identificada la Biblioteca Conjunta con la sigla DAV, se pueden realizar movimientos en las colecciones (DAG, DVE) simultáneamente, sin importar a qué Unidad Académica pertenezcan los usuarios, y sin tener que entrar y salir del sistema KOHAUNLP para realizar las transacciones del Área de Circulación.
Resumo:
En el presente trabajo se exponen las condiciones de la creación de la única unidad de información para dos Unidades Académicas dentro de la UNLP -la Biblioteca Conjunta de Facultad de Ciencias Agrarias y Forestales y de la Facultad de Ciencias Veterinarias- y cómo esta situación influyó directamente en la conformación del OPAC de la Biblioteca Conjunta en su reconversión al sistema de gestión de bibliotecas Koha-UNLP. Dado que Bienes Patrimoniales requiere que se mantenga el inventario de cadabiblioteca origen, se tuvo que implementar una codificación para asignar losinventarios. Ya que Koha-UNLP genera los inventarios en forma automática, en nuestrocaso decidimos completar manualmente el campo "código de barras", para así poder agregar al número de inventario y la codificación correspondiente a cada Facultad. Dicha codificación es la que dio el CAICYT, DAG (Agrarias/Forestales), DVE (Veterinaria). De esta manera se diferencian los fondos bibliográficos de las distintas facultades, y a su vez el fondo completo lleva la identificación dada por nosotros de Biblioteca Conjunta (DAV) para facilitar la circulación de los libros. Al quedar identificada la Biblioteca Conjunta con la sigla DAV, se pueden realizar movimientos en las colecciones (DAG, DVE) simultáneamente, sin importar a qué Unidad Académica pertenezcan los usuarios, y sin tener que entrar y salir del sistema KOHAUNLP para realizar las transacciones del Área de Circulación.
Resumo:
En este trabajo se presentan cuatro programas que pueden servir como herramientas auxiliares en la enseñanza y aprendizaje de la Matemática. Mediante su utilización, los docentes, pueden innovar en la forma y el diseño de actividades y situaciones didácticas, editar materiales didácticos de distintos tipos, abordar, plantear y resolver problemas geométricos, algebraicos y referidos al análisis de funciones en los entornos dinámicos que brindan estos programas. En general, GeoGebra, Geonext y Regla y Compás son piezas de software libre y de plataformas múltiples que se utilizan en educación para interactuar dinámicamente con la Matemática, en un ámbito en que se reúnen las Geometría, el Algebra y el Cálculo. Los tres son de uso libre y tienen licencia GNU GPL (General Public License). WIRIS es conjunto de productos (comerciales) dedicados al uso y la enseñanza de la Matemática, no sería nuestra intención relevar software que no sea libre sin embargo se tiene acceso gratuito a las herramientas a través de algunos portales educativos gubernamentales en diversos idiomas (el uso está limitado a la condición de estar conectado a Internet). Con WIRIS se dispone de lo que podríamos llamar "súper calculadora" ya que, mediante un conjunto de solapas, ofrece posibilidad de efectuar diversos tipos de cálculos, representar funciones, representar lugares geométricos, programar una secuencia de comandos, etc.
Resumo:
En este trabajo se presentan cuatro programas que pueden servir como herramientas auxiliares en la enseñanza y aprendizaje de la Matemática. Mediante su utilización, los docentes, pueden innovar en la forma y el diseño de actividades y situaciones didácticas, editar materiales didácticos de distintos tipos, abordar, plantear y resolver problemas geométricos, algebraicos y referidos al análisis de funciones en los entornos dinámicos que brindan estos programas. En general, GeoGebra, Geonext y Regla y Compás son piezas de software libre y de plataformas múltiples que se utilizan en educación para interactuar dinámicamente con la Matemática, en un ámbito en que se reúnen las Geometría, el Algebra y el Cálculo. Los tres son de uso libre y tienen licencia GNU GPL (General Public License). WIRIS es conjunto de productos (comerciales) dedicados al uso y la enseñanza de la Matemática, no sería nuestra intención relevar software que no sea libre sin embargo se tiene acceso gratuito a las herramientas a través de algunos portales educativos gubernamentales en diversos idiomas (el uso está limitado a la condición de estar conectado a Internet). Con WIRIS se dispone de lo que podríamos llamar "súper calculadora" ya que, mediante un conjunto de solapas, ofrece posibilidad de efectuar diversos tipos de cálculos, representar funciones, representar lugares geométricos, programar una secuencia de comandos, etc.
Resumo:
En el presente trabajo se exponen las condiciones de la creación de la única unidad de información para dos Unidades Académicas dentro de la UNLP -la Biblioteca Conjunta de Facultad de Ciencias Agrarias y Forestales y de la Facultad de Ciencias Veterinarias- y cómo esta situación influyó directamente en la conformación del OPAC de la Biblioteca Conjunta en su reconversión al sistema de gestión de bibliotecas Koha-UNLP. Dado que Bienes Patrimoniales requiere que se mantenga el inventario de cadabiblioteca origen, se tuvo que implementar una codificación para asignar losinventarios. Ya que Koha-UNLP genera los inventarios en forma automática, en nuestrocaso decidimos completar manualmente el campo "código de barras", para así poder agregar al número de inventario y la codificación correspondiente a cada Facultad. Dicha codificación es la que dio el CAICYT, DAG (Agrarias/Forestales), DVE (Veterinaria). De esta manera se diferencian los fondos bibliográficos de las distintas facultades, y a su vez el fondo completo lleva la identificación dada por nosotros de Biblioteca Conjunta (DAV) para facilitar la circulación de los libros. Al quedar identificada la Biblioteca Conjunta con la sigla DAV, se pueden realizar movimientos en las colecciones (DAG, DVE) simultáneamente, sin importar a qué Unidad Académica pertenezcan los usuarios, y sin tener que entrar y salir del sistema KOHAUNLP para realizar las transacciones del Área de Circulación.
Resumo:
Software testing is a key aspect of software reliability and quality assurance in a context where software development constantly has to overcome mammoth challenges in a continuously changing environment. One of the characteristics of software testing is that it has a large intellectual capital component and can thus benefit from the use of the experience gained from past projects. Software testing can, then, potentially benefit from solutions provided by the knowledge management discipline. There are in fact a number of proposals concerning effective knowledge management related to several software engineering processes. Objective: We defend the use of a lesson learned system for software testing. The reason is that such a system is an effective knowledge management resource enabling testers and managers to take advantage of the experience locked away in the brains of the testers. To do this, the experience has to be gathered, disseminated and reused. Method: After analyzing the proposals for managing software testing experience, significant weaknesses have been detected in the current systems of this type. The architectural model proposed here for lesson learned systems is designed to try to avoid these weaknesses. This model (i) defines the structure of the software testing lessons learned; (ii) sets up procedures for lesson learned management; and (iii) supports the design of software tools to manage the lessons learned. Results: A different approach, based on the management of the lessons learned that software testing engineers gather from everyday experience, with two basic goals: usefulness and applicability. Conclusion: The architectural model proposed here lays the groundwork to overcome the obstacles to sharing and reusing experience gained in the software testing and test management. As such, it provides guidance for developing software testing lesson learned systems.
Resumo:
La capacidad de transporte es uno de los baremos fundamentales para evaluar la progresión que puede llegar a tener un área económica y social. Es un sector de elevada importancia para la sociedad actual. Englobado en los distintos tipos de transporte, uno de los medios de transporte que se encuentra más en alza en la actualidad, es el ferroviario. Tanto para movilidad de pasajeros como para mercancías, el tren se ha convertido en un medio de transporte muy útil. Se encuentra dentro de las ciudades, entre ciudades con un radio pequeño entre ellas e incluso cada vez más, gracias a la alta velocidad, entre ciudades con gran distancia entre ellas. Esta Tesis pretende ayudar en el diseño de una de las etapas más importantes de los Proyectos de instalación de un sistema ferroviario: el sistema eléctrico de tracción. La fase de diseño de un sistema eléctrico de tracción ferroviaria se enfrenta a muchas dudas que deben ser resueltas con precisión. Del éxito de esta fase dependerá la capacidad de afrontar las demandas de energía de la explotación ferroviaria. También se debe atender a los costes de instalación y de operación, tanto costes directos como indirectos. Con la Metodología que se presenta en esta Tesis se ofrecerá al diseñador la opción de manejar un sistema experto que como soluciones le plantee un conjunto de escenarios de sistemas eléctricos correctos, comprobados por resolución de modelos de ecuaciones. Correctos desde el punto de vista de validez de distintos parámetros eléctrico, como de costes presupuestarios e impacto de costes indirectos. Por tanto, el diseñador al haber hecho uso de esta Metodología, tendría en un espacio de tiempo relativamente corto, un conjunto de soluciones factibles con las que poder elegir cuál convendría más según sus intereses finales. Esta Tesis se ha desarrollado en una vía de investigación integrada dentro del Centro de Investigaciones Ferroviarias CITEF-UPM. Entre otros proyectos y vías de investigación, en CITEF se ha venido trabajando en estudios de validación y dimensionamiento de sistemas eléctricos ferroviarios con diversos y variados clientes y sistemas ferroviarios. A lo largo de los proyectos realizados, el interés siempre ha girado mayoritariamente sobre los siguientes parámetros del sistema eléctrico: - Calcular número y posición de subestaciones de tracción. Potencia de cada subestación. - Tipo de catenaria a lo largo del recorrido. Conductores que componen la catenaria. Características. - Calcular número y posición de autotransformadores para sistemas funcionando en alterna bitensión o 2x25kV. - Posición Zonas Neutras. - Validación según normativa de: o Caídas de tensión en la línea o Tensiones máximas en el retorno de la línea o Sobrecalentamiento de conductores o Sobrecalentamiento de los transformadores de las subestaciones de tracción La idea es que las soluciones aportadas por la Metodología sugieran escenarios donde de estos parámetros estén dentro de los límites que marca la normativa. Tener la posibilidad de tener un repositorio de posibles escenarios donde los parámetros y elementos eléctricos estén calculados como correctos, aporta un avance en tiempos y en pruebas, que mejoraría ostensiblemente el proceso habitual de diseño para los sistemas eléctricos ferroviarios. Los costes directos referidos a elementos como subestaciones de tracción, autotransformadores, zonas neutras, ocupan un gran volumen dentro del presupuesto de un sistema ferroviario. En esta Tesis se ha querido profundizar también en el efecto de los costes indirectos provocados en la instalación y operación de sistemas eléctricos. Aquellos derivados del impacto medioambiental, los costes que se generan al mantener los equipos eléctricos y la instalación de la catenaria, los costes que implican la conexión entre las subestaciones de tracción con la red general o de distribución y por último, los costes de instalación propios de cada elemento compondrían los costes indirectos que, según experiencia, se han pensado relevantes para ejercer un cierto control sobre ellos. La Metodología cubrirá la posibilidad de que los diseños eléctricos propuestos tengan en cuenta variaciones de coste inasumibles o directamente, proponer en igualdad de condiciones de parámetros eléctricos, los más baratos en función de los costes comentados. Analizando los costes directos e indirectos, se ha pensado dividir su impacto entre los que se computan en la instalación y los que suceden posteriormente, durante la operación de la línea ferroviaria. Estos costes normalmente suelen ser contrapuestos, cuánto mejor es uno peor suele ser el otro y viceversa, por lo que hace falta un sistema que trate ambos objetivos por separado. Para conseguir los objetivos comentados, se ha construido la Metodología sobre tres pilares básicos: - Simulador ferroviario Hamlet: Este simulador integra módulos para construir esquemas de vías ferroviarios completos; módulo de simulación mecánica y de la tracción de material rodante; módulo de señalización ferroviaria; módulo de sistema eléctrico. Software realizado en C++ y Matlab. - Análisis y estudio de cómo focalizar los distintos posibles escenarios eléctricos, para que puedan ser examinados rápidamente. Pico de demanda máxima de potencia por el tráfico ferroviario. - Algoritmos de optimización: A partir de un estudio de los posibles algoritmos adaptables a un sistema tan complejo como el que se plantea, se decidió que los algoritmos genéticos serían los elegidos. Se han escogido 3 algoritmos genéticos, permitiendo recabar información acerca del comportamiento y resultados de cada uno de ellos. Los elegidos por motivos de tiempos de respuesta, multiobjetividad, facilidad de adaptación y buena y amplia aplicación en proyectos de ingeniería fueron: NSGA-II, AMGA-II y ɛ-MOEA. - Diseño de funciones y modelo preparado para trabajar con los costes directos e indirectos y las restricciones básicas que los escenarios eléctricos no deberían violar. Estas restricciones vigilan el comportamiento eléctrico y la estabilidad presupuestaria. Las pruebas realizadas utilizando el sistema han tratado o bien de copiar situaciones que se puedan dar en la realidad o directamente sistemas y problemas reales. Esto ha proporcionado además de la posibilidad de validar la Metodología, también se ha posibilitado la comparación entre los algoritmos genéticos, comparar sistemas eléctricos escogidos con los reales y llegar a conclusiones muy satisfactorias. La Metodología sugiere una vía de trabajo muy interesante, tanto por los resultados ya obtenidos como por las oportunidades que puede llegar a crear con la evolución de la misma. Esta Tesis se ha desarrollado con esta idea, por lo que se espera pueda servir como otro factor para trabajar con la validación y diseño de sistemas eléctricos ferroviarios. ABSTRACT Transport capacity is one of the critical points to evaluate the progress than a specific social and economical area is able to reach. This is a sector of high significance for the actual society. Included inside the most common types of transport, one of the means of transport which is elevating its use nowadays is the railway. Such as for passenger transport of weight movements, the train is being consolidated like a very useful mean of transport. Railways are installed in many geography areas. Everyone know train in cities, or connecting cities inside a surrounding area or even more often, taking into account the high-speed, there are railways infrastructure between cities separated with a long distance. This Ph.D work aims to help in the process to design one of the most essential steps in Installation Projects belonging to a railway system: Power Supply System. Design step of the railway power supply, usually confronts to several doubts and uncertainties, which must be solved with high accuracy. Capacity to supply power to the railway traffic depends on the success of this step. On the other hand is very important to manage the direct and indirect costs derived from Installation and Operation. With the Methodology is presented in this Thesis, it will be offered to the designer the possibility to handle an expert system that finally will fill a set of possible solutions. These solutions must be ready to work properly in the railway system, and they were tested using complex equation models. This Thesis has been developed through a research way, integrated inside Citef (Railway Research Centre of Technical University of Madrid). Among other projects and research ways, in Citef has been working in several validation studies and dimensioning of railway power supplies. It is been working by a large range of clients and railways systems. Along the accomplished Projects, the main goal has been rounded mostly about the next list of parameters of the electrical system: - Calculating number and location of traction substations. Power of each substation. - Type of Overhead contact line or catenary through the railway line. The wires which set up the catenary. Main Characteristics. - Calculating number and position of autotransformers for systems working in alternating current bi-voltage of called 2x25 kV. - Location of Neutral Zones. - Validating upon regulation of: o Drop voltages along the line o Maximum return voltages in the line o Overheating/overcurrent of the wires of the catenary o Avoiding overheating in the transformers of the traction substations. Main objective is that the solutions given by the Methodology, could be suggest scenarios where all of these parameters from above, would be between the limits established in the regulation. Having the choice to achieve a repository of possible good scenarios, where the parameters and electrical elements will be assigned like ready to work, that gives a great advance in terms of times and avoiding several tests. All of this would improve evidently the regular railway electrical systems process design. Direct costs referred to elements like traction substations, autotransformers, neutral zones, usually take up a great volume inside the general budget in railway systems. In this Thesis has been thought to bear in mind another kind of costs related to railway systems, also called indirect costs. These could be enveloped by those enmarked during installation and operation of electrical systems. Those derived from environmental impact; costs generated during the maintenance of the electrical elements and catenary; costs involved in the connection between traction substations and general electric grid; finally costs linked with the own installation of the whole electrical elements needed for the correct performance of the railway system. These are integrated inside the set has been collected taking into account own experience and research works. They are relevant to be controlled for our Methodology, just in case for the designers of this type of systems. The Methodology will cover the possibility that the final proposed power supply systems will be hold non-acceptable variations of costs, comparing with initial expected budgets, or directly assuming a threshold of budget for electrical elements in actual scenario, and achieving the cheapest in terms of commented costs from above. Analyzing direct and indirect costs, has been thought to divide their impact between two main categories. First one will be inside the Installation and the other category will comply with the costs often happens during Railway Operation time. These costs normally are opposed, that means when one is better the other turn into worse, in costs meaning. For this reason is necessary treating both objectives separately, in order to evaluate correctly the impact of each one into the final system. The objectives detailed before build the Methodology under three basic pillars: - Railway simulator Hamlet: This software has modules to configure many railway type of lines; mechanical and traction module to simulate the movement of rolling stock; signaling module; power supply module. This software has been developed using C++ and Matlab R13a - Previously has been mandatory to study how would be possible to work properly with a great number of feasible electrical systems. The target comprised the quick examination of these set of scenarios in terms of time. This point is talking about Maximum power demand peaks by railway operation plans. - Optimization algorithms. A railway infrastructure is a very complex system. At the beginning it was necessary to search about techniques and optimization algorithms, which could be adaptable to this complex system. Finally three genetic multiobjective algorithms were the chosen. Final decision was taken attending to reasons such as time complexity, able to multiobjective, easy to integrate in our problem and with a large application in engineering tasks. They are: NSGA-II, AMGA-II and ɛ-MOEA. - Designing objectives functions and equation model ready to work with the direct and indirect costs. The basic restrictions are not able to avoid, like budgetary or electrical, connected hardly with the recommended performance of elements, catenary and safety in a electrical railway systems. The battery of tests launched to the Methodology has been designed to be as real as possible. In fact, due to our work in Citef and with real Projects, has been integrated and configured three real railway lines, in order to evaluate correctly the final results collected by the Methodology. Another topic of our tests has been the comparison between the performances of the three algorithms chosen. Final step has been the comparison again with different possible good solutions, it means power supply system designs, provided by the Methodology, testing the validity of them. Once this work has been finished, the conclusions have been very satisfactory. Therefore this Thesis suggest a very interesting way of research and work, in terms of the results obtained and for the future opportunities can be created with the evolution of this. This Thesis has been developed with this idea in mind, so is expected this work could adhere another factor to work in the difficult task of validation and design of railway power supply systems.
Resumo:
En el presente proyecto se ha procedido a implantar la herramienta de procesado software GNU Radio en la tarjeta EVMK2H, que es un módulo de evaluación fabricado por Texas Instruments que incorpora un System on Chip (SoC) 66AK2H14 de la familia Keystone II, el cual dispone de 4 núcleos ARM y 8 núcleos DSP. Previamente a la instalación de GNU Radio, hubo que configurar la tarjeta, así como instalar el software necesario. De igual manera, se realizó una primera aproximación para comprender el funcionamiento de los sistemas de comunicación entre núcleos de que hace uso la tarjeta, y de los que se hizo uso posteriormente en el proyecto. Tras el portado de GNU Radio se ha comprobado el correcto funcionamiento del mecanismo de comunicación entre núcleos ARM y DSP con un par de aplicaciones de prueba. ABSTRACT. In the present project it was performed the implementation of the software processing toolkit GNU Radio into the EVMK2H board, which is an evaluation module from Texas Instruments that includes a 66AK2H14 System on Chip (SoC) from the Keystone II family, that provides 4 ARM cores and 8 DSP cores. Before installing GNU Radio, it was necessary to configure the board, and as well installing other needed software. Also, a first approach was performed to understand the way the communication system between cores included in the board works, which was used later in the project. After porting GNU Radio, some test applications have been written to test the correct operation of the communication mechanism between ARM and DSP cores.
Resumo:
One saw previously that indications of diversity IT and the one of Shannon permits to characterize globally by only one number one fundamental aspects of the text structure. However a more precise knowledge of this structure requires specific abundance distributions and the use, to represent this one, of a suitable mathematical model. Among the numerous models that would be either susceptible to be proposed, the only one that present a real convenient interest are simplest. One will limit itself to study applied three of it to the language L(MT): the log-linear, the log-normal and Mac Arthur's models very used for the calculation of the diversity of the species of ecosystems, and used, we believe that for the first time, in the calculation of the diversity of a text written in a certain language, in our case L(MT). One will show advantages and inconveniences of each of these model types, methods permitting to adjust them to text data and in short tests that permit to decide if this adjustment is acceptable.
Resumo:
Mode of access: Internet.
Resumo:
O objetivo deste estudo prospectivo foi avaliar os efeitos do aparelho Forsus® nos incisivos centrais superiores e inferiores. A amostra constituiu-se de 22 tomografias computadorizadas de 11 pacientes (sexo masculino e feminino) idade média de 15,8 anos com má oclusão de Classe II que foram tratados com o aparelho Forsus® na clínica do programa de pós-graduação em Odontologia, área de concentração Ortodontia, da Universidade Metodista de São Paulo. As tomografias foram obtidas em dois momentos T1 (final de nivelamento e antes da instalação do Forsus® e T2 (remoção do Forsus®). Para avaliar a distância do ápice até a tábua óssea, as imagens a serem examinadas foram obtidas com o auxílio do viewer do próprio i-CAT® , o iCATVision® e examinadas com o CorelDRAW X5® já para as medidas cefalométricas IMPA e 1.PP as imagens cefalométricas ortogonais foram obtidas em proporção 1:1 com auxílio do software Dolphin 3D® (Dolphin Imaging and Management Solutions, Chatsworth, EUA) e em seguida examinadas com o software Radiocef Studio 2 (Radio Memory, Belo Horizonte, Brasil). Para a obtenção do erro intra-examinador foi feito o teste t de Student pareado para o erro sistemático e a fórmula de DAHLBERG para estimar a ordem de grandeza dos erros casuais e na análise estatística dos resultados utilizou-se: o teste t para a determinação das diferenças entres as fases de observação e o teste de correlação de Pearson para avaliar a correlação entres as alterações. Observou-se: um aumento significativo (p<0,05) tanto no IMPA quanto no 1.PP, aproximação do ápice dos incisivos inferiores da tábua óssea lingual, aproximação do ápice dos incisivos superiores da tábua óssea vestibular, uma correlação negativa muito forte entre o IMPA e a distância do ápice do incisivo até a tábua óssea lingual e uma correlação negativa moderada entre 1.PP e a distância do ápice do incisivo até a tábua óssea vestibular. Sendo assim o aparelho Forsus® no tratamento da Classe II teve como efeito: vestibularização significativa dos incisivos centrais inferiores, uma verticalização significativa dos incisivos centrais superiores, aproximação do ápice dos incisivos inferiores da cortical óssea lingual e aproximação do ápice dos incisivos superiores da cortical óssea vestibular.
Resumo:
Quantitative structure-activity relationship (QSAR) analysis is a cornerstone of modern informatics. Predictive computational models of peptide-major histocompatibility complex (MHC)-binding affinity based on QSAR technology have now become important components of modern computational immunovaccinology. Historically, such approaches have been built around semiqualitative, classification methods, but these are now giving way to quantitative regression methods. We review three methods--a 2D-QSAR additive-partial least squares (PLS) and a 3D-QSAR comparative molecular similarity index analysis (CoMSIA) method--which can identify the sequence dependence of peptide-binding specificity for various class I MHC alleles from the reported binding affinities (IC50) of peptide sets. The third method is an iterative self-consistent (ISC) PLS-based additive method, which is a recently developed extension to the additive method for the affinity prediction of class II peptides. The QSAR methods presented here have established themselves as immunoinformatic techniques complementary to existing methodology, useful in the quantitative prediction of binding affinity: current methods for the in silico identification of T-cell epitopes (which form the basis of many vaccines, diagnostics, and reagents) rely on the accurate computational prediction of peptide-MHC affinity. We have reviewed various human and mouse class I and class II allele models. Studied alleles comprise HLA-A*0101, HLA-A*0201, HLA-A*0202, HLA-A*0203, HLA-A*0206, HLA-A*0301, HLA-A*1101, HLA-A*3101, HLA-A*6801, HLA-A*6802, HLA-B*3501, H2-K(k), H2-K(b), H2-D(b) HLA-DRB1*0101, HLA-DRB1*0401, HLA-DRB1*0701, I-A(b), I-A(d), I-A(k), I-A(S), I-E(d), and I-E(k). In this chapter we show a step-by-step guide into predicting the reliability and the resulting models to represent an advance on existing methods. The peptides used in this study are available from the AntiJen database (http://www.jenner.ac.uk/AntiJen). The PLS method is available commercially in the SYBYL molecular modeling software package. The resulting models, which can be used for accurate T-cell epitope prediction, will be made are freely available online at the URL http://www.jenner.ac.uk/MHCPred.