970 resultados para Large-volume Quartz Latites


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Son numerosos los expertos que predicen que hasta pasado 2050 no se utilizarán masivamente las energías de origen renovable, y que por tanto se mantendrá la emisión de dióxido de carbono de forma incontrolada. Entre tanto, y previendo que este tipo de uso se mantenga hasta un horizonte temporal aún más lejano, la captura, concentración y secuestro o reutilización de dióxido de carbono es y será una de las principales soluciones a implantar para paliar el problema medioambiental causado. Sin embargo, las tecnologías existentes y en desarrollo de captura y concentración de este tipo de gas, presentan dos limitaciones: las grandes cantidades de energía que consumen y los grandes volúmenes de sustancias potencialmente dañinas para el medioambiente que producen durante su funcionamiento. Ambas razones hacen que no sean atractivas para su implantación y uso de forma extensiva. La solución planteada en la presente tesis doctoral se caracteriza por la ausencia de residuos producidos en la operación de captura y concentración del dióxido de carbono, por no utilizar substancias químicas y físicas habituales en las técnicas actuales, por disminuir los consumos energéticos al carecer de sistemas móviles y por evitar la regeneración química y física de los materiales utilizados en la actualidad. Así mismo, plantea grandes retos a futuras innovaciones sobre la idea propuesta que busquen fundamentalmente la disminución de la energía utilizada durante su funcionamiento y la optimización de sus componentes principales. Para conseguir el objetivo antes citado, la presente tesis doctoral, una vez establecido el planteamiento del problema al que se busca solución (capítulo 1), del estudio de las técnicas de separación de gases atmosféricos utilizadas en la actualidad, así como del de los sistemas fundamentales de las instalaciones de captura y concentración del dióxido de carbono (capítulo 2) y tras una definición del marco conceptual y teórico (capítulo 3), aborda el diseño de un prototipo de ionización fotónica de los gases atmosféricos para su posterior separación electrostática, a partir del estudio, adaptación y mejora del funcionamiento de los sistemas de espectrometría de masas. Se diseñarán y desarrollarán los sistemas básicos de fotoionización, mediante el uso de fuentes de fotones coherentes, y los de separación electrostática (capítulo 4), en que se basa el funcionamiento de este sistema de separación de gases atmosféricos y de captura y concentración de dióxido de carbono para construir un prototipo a nivel laboratorio. Posteriormente, en el capítulo 5, serán probados utilizando una matriz experimental que cubra los rangos de funcionamiento previstos y aporte suficientes datos experimentales para corregir y desarrollar el marco teórico real, y con los que se pueda establecer y corregir un modelo físico– matemático de simulación (capítulo 6) aplicable a la unidad en su conjunto. Finalmente, debido a la utilización de unidades de ionización fotónica, sistemas láseres intensos y sistemas eléctricos de gran potencia, es preciso analizar el riesgo biológico a las personas y al medioambiente debido al impacto de la radiación electromagnética producida (capítulo 7), minimizando su impacto y cumpliendo con la legislación vigente. En el capítulo 8 se planteará un diseño escalable a tamaño piloto de la nueva tecnología propuesta y sus principales modos de funcionamiento, así como un análisis de viabilidad económica. Como consecuencia de la tesis doctoral propuesta y del desarrollo de la unidad de separación atmosférica y de captura y concentración de dióxido de carbono, surgen diversas posibilidades de estudio que pueden ser objeto de nuevas tesis doctorales y de futuros desarrollos de ingeniería. El capítulo 9 tratará de incidir en estos aspectos indicando líneas de investigación para futuras tesis y desarrollos industriales. ABSTRACT A large number of experts predict that until at least 2050 renewable energy sources will not be massively used, and for that reason, current Primary Energy sources based on extensive use of fossil fuel will be used maintaining out of control emissions, Carbon Dioxide above all. Meanwhile, under this scenario and considering its extension until at least 2050, Carbon Capture, Concentration, Storage and/or Reuse is and will be one of the main solutions to minimise Greenhouse Gasses environmental effect. But, current Carbon Capture and Storage technology state of development has two main problems: it is a too large energy consuming technology and during normal use it produces a large volume of environmentally dangerous substances. Both reasons are limiting its development and its extensive use. This Ph Degree Thesis document proposes a solution to get the expected effect using a new atmospheric gasses separation system with the following characteristics: absence of wastes produced, it needs no chemical and/or physical substances during its operation, it reduces to minimum the internal energy consumptions due to absence of mobile equipment and it does not need any chemical and/or physical regeneration of substances. This system is beyond the State of the Art of current technology development. Additionally, the proposed solution raises huge challenges for future innovations of the proposed idea finding radical reduction of internal energy consumption during functioning, as well as regarding optimisation of main components, systems and modes of operation. To achieve this target, once established the main problem, main challenge and potential solving solutions (Chapter 1), it is established an initial starting point fixing the Atmospheric Gasses Separation and Carbon Capture and Storage developments (Chapter 2), as well as it will be defined the theoretical and basic model, including existing and potential new governing laws and mathematical formulas to control its system functioning (Chapter 3), this document will deal with the design of an installation of an operating system based on photonic ionization of atmospheric gasses to be separated in a later separation system based on the application of electrostatic fields. It will be developed a basic atmospheric gasses ionization prototype based on intense radioactive sources capable to ionize gasses by coherent photonic radiation, and a basic design of electrostatic separation system (Chapter 4). Both basic designs are the core of the proposed technology that separates Atmospheric Gasses and captures and concentrates Carbon Dioxide. Chapter 5 will includes experimental results obtained from an experimental testing matrix covering expected prototype functioning regimes. With the obtained experimental data, theoretical model will be corrected and improved to act as the real physical and mathematical model capable to simulate real system function (Chapter 6). Finally, it is necessary to assess potential biological risk to public and environment due to the proposed use of units of intense energy photonic ionization, by laser beams or by non–coherent sources and large electromagnetic systems with high energy consumption. It is necessary to know the impact in terms of and electromagnetic radiation taking into account National Legislation (Chapter 7). On Chapter 8, an up scaled pilot plant will be established covering main functioning modes and an economic feasibility assessment. As a consequence of this PhD Thesis, a new field of potential researches and new PhD Thesis are opened, as well as future engineering and industrial developments (Chapter 9).

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The aim of this study was to determine if the soils, waters and plants from the Aliaga dump contained polycyclic aromatic hydrocarbons (PAHs) and their quantification.The results showed that PAHs concentrations in soils are in general higher than the reference levels from the Spanish legislation. Waters and plants contained PAHs but in low concentrations. The possible actions for remediation (photodegradation and bioremediation) seem to be unviable here because of the large volume of materials involved, although its use as an additive for the cement industry and derivatives can be considered. It is proposed that fluorantene in waters, and phenanthrene and benzo[ghi]perilene in soils be considered as pollutants as well as to study the incorporation of PAHs to plants. Key-words: Polycyclic aromatic hydrocarbons, soil, plant and water contamination, fly- ash, power plant. RESUMEN: El objetivo de este estudio fue determinar y cuantificar los hidrocarburos policíclicos aromáticos (PAHs) en los suelos, plantas y aguas de la Escombrera de Aliaga. La concentración de PAHs en las cenizas supera, en general, los valores establecidos en la legislación española.Las aguas y plantas contienen PAHs, aunque en concentraciones bajas. La remoción de los materiales para someterlos a fotodegradación y biorremediación es inviable debido al gran volumen de la escombrera, aunque se plantea su uso como aditivo en la fabricación de productos derivados del cemento. Se propone incluir el fenantreno y benzo[ghi]perileno en la normativa de suelos, así como el naftaleno en la de aguas y la elaboración de una legislación sobre la incorporación de estos compuestos a las plantas.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Intensive farm systems handle large volume of livestock wastes, resulting in adverse environmental effects, such as gaseous losses into the atmosphere in form of ammonia (NH3) and greenhouse gases (GHG), i.e. methane (CH4), carbon dioxide (CO2) and nitrous oxide (N2O). In this study, the manure management continuum of slurry storage with impermeable cover and following cattle slurry band spreading and incorporation to soil was assessed for NH3 and GHG emissions. The experiment was conducted in an outdoor covered storage (flexible bag system) (study I), which collected the slurry produced in 7 dairy cattle farms (2,000 m3 slurry) during 12 days in the northern Spain.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

En la presente Tesis se ha llevado a cabo el contraste y desarrollo de metodologías que permitan mejorar el cálculo de las avenidas de proyecto y extrema empleadas en el cálculo de la seguridad hidrológica de las presas. En primer lugar se ha abordado el tema del cálculo de las leyes de frecuencia de caudales máximos y su extrapolación a altos periodos de retorno. Esta cuestión es de gran relevancia, ya que la adopción de estándares de seguridad hidrológica para las presas cada vez más exigentes, implica la utilización de periodos de retorno de diseño muy elevados cuya estimación conlleva una gran incertidumbre. Es importante, en consecuencia incorporar al cálculo de los caudales de diseño todas la técnicas disponibles para reducir dicha incertidumbre. Asimismo, es importante hacer una buena selección del modelo estadístico (función de distribución y procedimiento de ajuste) de tal forma que se garantice tanto su capacidad para describir el comportamiento de la muestra, como para predecir de manera robusta los cuantiles de alto periodo de retorno. De esta forma, se han realizado estudios a escala nacional con el objetivo de determinar el esquema de regionalización que ofrece mejores resultados para las características hidrológicas de las cuencas españolas, respecto a los caudales máximos anuales, teniendo en cuenta el numero de datos disponibles. La metodología utilizada parte de la identificación de regiones homogéneas, cuyos límites se han determinado teniendo en cuenta las características fisiográficas y climáticas de las cuencas, y la variabilidad de sus estadísticos, comprobando posteriormente su homogeneidad. A continuación, se ha seleccionado el modelo estadístico de caudales máximos anuales con un mejor comportamiento en las distintas zonas de la España peninsular, tanto para describir los datos de la muestra como para extrapolar a los periodos de retorno más altos. El proceso de selección se ha basado, entre otras cosas, en la generación sintética de series de datos mediante simulaciones de Monte Carlo, y el análisis estadístico del conjunto de resultados obtenido a partir del ajuste de funciones de distribución a estas series bajo distintas hipótesis. Posteriormente, se ha abordado el tema de la relación caudal-volumen y la definición de los hidrogramas de diseño en base a la misma, cuestión que puede ser de gran importancia en el caso de presas con grandes volúmenes de embalse. Sin embargo, los procedimientos de cálculo hidrológico aplicados habitualmente no tienen en cuenta la dependencia estadística entre ambas variables. En esta Tesis se ha desarrollado un procedimiento para caracterizar dicha dependencia estadística de una manera sencilla y robusta, representando la función de distribución conjunta del caudal punta y el volumen en base a la función de distribución marginal del caudal punta y la función de distribución condicionada del volumen respecto al caudal. Esta última se determina mediante una función de distribución log-normal, aplicando un procedimiento de ajuste regional. Se propone su aplicación práctica a través de un procedimiento de cálculo probabilístico basado en la generación estocástica de un número elevado de hidrogramas. La aplicación a la seguridad hidrológica de las presas de este procedimiento requiere interpretar correctamente el concepto de periodo de retorno aplicado a variables hidrológicas bivariadas. Para ello, se realiza una propuesta de interpretación de dicho concepto. El periodo de retorno se entiende como el inverso de la probabilidad de superar un determinado nivel de embalse. Al relacionar este periodo de retorno con las variables hidrológicas, el hidrograma de diseño de la presa deja de ser un único hidrograma para convertirse en una familia de hidrogramas que generan un mismo nivel máximo en el embalse, representados mediante una curva en el plano caudal volumen. Esta familia de hidrogramas de diseño depende de la propia presa a diseñar, variando las curvas caudal-volumen en función, por ejemplo, del volumen de embalse o la longitud del aliviadero. El procedimiento propuesto se ilustra mediante su aplicación a dos casos de estudio. Finalmente, se ha abordado el tema del cálculo de las avenidas estacionales, cuestión fundamental a la hora de establecer la explotación de la presa, y que puede serlo también para estudiar la seguridad hidrológica de presas existentes. Sin embargo, el cálculo de estas avenidas es complejo y no está del todo claro hoy en día, y los procedimientos de cálculo habitualmente utilizados pueden presentar ciertos problemas. El cálculo en base al método estadístico de series parciales, o de máximos sobre un umbral, puede ser una alternativa válida que permite resolver esos problemas en aquellos casos en que la generación de las avenidas en las distintas estaciones se deba a un mismo tipo de evento. Se ha realizado un estudio con objeto de verificar si es adecuada en España la hipótesis de homogeneidad estadística de los datos de caudal de avenida correspondientes a distintas estaciones del año. Asimismo, se han analizado los periodos estacionales para los que es más apropiado realizar el estudio, cuestión de gran relevancia para garantizar que los resultados sean correctos, y se ha desarrollado un procedimiento sencillo para determinar el umbral de selección de los datos de tal manera que se garantice su independencia, una de las principales dificultades en la aplicación práctica de la técnica de las series parciales. Por otra parte, la aplicación practica de las leyes de frecuencia estacionales requiere interpretar correctamente el concepto de periodo de retorno para el caso estacional. Se propone un criterio para determinar los periodos de retorno estacionales de forma coherente con el periodo de retorno anual y con una distribución adecuada de la probabilidad entre las distintas estaciones. Por último, se expone un procedimiento para el cálculo de los caudales estacionales, ilustrándolo mediante su aplicación a un caso de estudio. The compare and develop of a methodology in order to improve the extreme flow estimation for dam hydrologic security has been developed. First, the work has been focused on the adjustment of maximum peak flows distribution functions from which to extrapolate values for high return periods. This has become a major issue as the adoption of stricter standards on dam hydrologic security involves estimation of high design return periods which entails great uncertainty. Accordingly, it is important to incorporate all available techniques for the estimation of design peak flows in order to reduce this uncertainty. Selection of the statistical model (distribution function and adjustment method) is also important since its ability to describe the sample and to make solid predictions for high return periods quantiles must be guaranteed. In order to provide practical application of previous methodologies, studies have been developed on a national scale with the aim of determining a regionalization scheme which features best results in terms of annual maximum peak flows for hydrologic characteristics of Spanish basins taking into account the length of available data. Applied methodology starts with the delimitation of regions taking into account basin’s physiographic and climatic characteristics and the variability of their statistical properties, and continues with their homogeneity testing. Then, a statistical model for maximum annual peak flows is selected with the best behaviour for the different regions in peninsular Spain in terms of describing sample data and making solid predictions for high return periods. This selection has been based, among others, on synthetic data series generation using Monte Carlo simulations and statistical analysis of results from distribution functions adjustment following different hypothesis. Secondly, the work has been focused on the analysis of the relationship between peak flow and volume and how to define design flood hydrographs based on this relationship which can be highly important for large volume reservoirs. However, commonly used hydrologic procedures do not take statistical dependence between these variables into account. A simple and sound method for statistical dependence characterization has been developed by the representation of a joint distribution function of maximum peak flow and volume which is based on marginal distribution function of peak flow and conditional distribution function of volume for a given peak flow. The last one is determined by a regional adjustment procedure of a log-normal distribution function. Practical application is proposed by a probabilistic estimation procedure based on stochastic generation of a large number of hydrographs. The use of this procedure for dam hydrologic security requires a proper interpretation of the return period concept applied to bivariate hydrologic data. A standard is proposed in which it is understood as the inverse of the probability of exceeding a determined reservoir level. When relating return period and hydrological variables the only design flood hydrograph changes into a family of hydrographs which generate the same maximum reservoir level and that are represented by a curve in the peak flow-volume two-dimensional space. This family of design flood hydrographs depends on the dam characteristics as for example reservoir volume or spillway length. Two study cases illustrate the application of the developed methodology. Finally, the work has been focused on the calculation of seasonal floods which are essential when determining the reservoir operation and which can be also fundamental in terms of analysing the hydrologic security of existing reservoirs. However, seasonal flood calculation is complex and nowadays it is not totally clear. Calculation procedures commonly used may present certain problems. Statistical partial duration series, or peaks over threshold method, can be an alternative approach for their calculation that allow to solve problems encountered when the same type of event is responsible of floods in different seasons. A study has been developed to verify the hypothesis of statistical homogeneity of peak flows for different seasons in Spain. Appropriate seasonal periods have been analyzed which is highly relevant to guarantee correct results. In addition, a simple procedure has been defined to determine data selection threshold on a way that ensures its independency which is one of the main difficulties in practical application of partial series. Moreover, practical application of seasonal frequency laws requires a correct interpretation of the concept of seasonal return period. A standard is proposed in order to determine seasonal return periods coherently with the annual return period and with an adequate seasonal probability distribution. Finally a methodology is proposed to calculate seasonal peak flows. A study case illustrates the application of the proposed methodology.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

El presente trabajo se ha centrado en la investigación de soluciones para automatizar la tarea del enriquecimiento de fuentes de datos sobre redes de sensores con descripciones lingüísticas, con el fin de facilitar la posterior generación de textos en lenguaje natural. El uso de descripciones en lenguaje natural facilita el acceso a los datos a una mayor diversidad de usuarios y, como consecuencia, permite aprovechar mejor las inversiones en redes de sensores. En el trabajo se ha considerado el uso de bases de datos abiertas para abordar la necesidad de disponer de un gran volumen y diversidad de conocimiento geográfico. Se ha analizado también el enriquecimiento de datos dentro de enfoques metodológicos de curación de datos y métodos de generación de lenguaje natural. Como resultado del trabajo, se ha planteado un método general basado en una estrategia de generación y prueba que incluye una forma de representación y uso del conocimiento heurístico con varias etapas de razonamiento para la construcción de descripciones lingüísticas de enriquecimiento de datos. En la evaluación de la propuesta general se han manejado tres escenarios, dos de ellos para generación de referencias geográficas sobre redes de sensores complejas de dimensión real y otro para la generación de referencias temporales. Los resultados de la evaluación han mostrado la validez práctica de la propuesta general exhibiendo mejoras de rendimiento respecto a otros enfoques. Además, el análisis de los resultados ha permitido identificar y cuantificar el impacto previsible de diversas líneas de mejora en bases de datos abiertas. ABSTRACT This work has focused on the search for solutions to automate the task of enrichment sensor-network-based data sources with textual descriptions, so as to facilitate the generation of natural language texts. Using natural language descriptions facilitates data access to a wider range of users and, therefore, allows better leveraging investments in sensor networks. In this work we have considered the use of open databases to address the need for a large volume and diversity of geographical knowledge. We have also analyzed data enrichment in methodological approaches and data curation methods of natural language generation. As a result, it has raised a general method based on a strategy of generating and testing that includes a representation using heuristic knowledge with several stages of reasoning for the construction of linguistic descriptions of data enrichment. In assessing the overall proposal three scenarios have been addressed, two of them in the environmental domain with complex sensor networks and another real dimension in the time domain. The evaluation results have shown the validity and practicality of our proposal, showing performance improvements over other approaches. Furthermore, the analysis of the results has allowed identifying and quantifying the expected impact of various lines of improvement in open databases.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

La evolución de las redes eléctricas se dirige hacia lo que se conoce como “Smart Grids” o “Redes Eléctricas Inteligentes”. Estas “Smart Grids” se componen de subestaciones eléctricas, que a su vez se componen de unos dispositivos llamados IEDs (Dispositivos Electrónicos Inteligentes – Intelligent Electronic Devices). El diseño de IEDs se encuentra definido en la norma IEC 61850, que especifica además un Lenguaje de Configuración de Subestaciones (Substation Configuration Language SCL) para la definición de la configuración de subestaciones y sus IEDs. Hoy en día, este estándar internacional no sólo se utiliza para diseñar correctamente IEDs y asegurar su interoperabilidad, sino que también se utiliza para el diseño de otros dispositivos de la red eléctrica, como por ejemplo, medidores inteligentes. Sin embargo, aunque existe una tendencia cada vez mayor del uso de este estándar, la comprensión y el manejo del mismo resulta difícil debido al gran volumen de información que lo compone y del nivel de detalle que utiliza, por lo que su uso para el diseño de IEDs se hace tedioso sin la ayuda de un soporte software. Es por ello que, para facilitar la aplicación del estándar IEC 61850 en el diseño de IEDs se han desarrollado herramientas como “Visual SCL”, “SCL Explorer” o “61850 SCLVisual Design Tool”. En concreto, “61850 SCLVisual Design Tool” es una herramienta gráfica para el modelado de subestaciones electricas, generada mediante el uso de los frameworks Eclipse Modeling Framework (EMF) y Epsilon Generative Modeling Technologies (GMT) y desarrollada por el grupo de investigación SYST de la UPM. El objetivo de este proyecto es añadir una nueva funcionalidad a la herramienta “61850 Visual SCL DesignTool”. Esta nueva funcionalidad consiste en la generación automática de un fichero de configuración de subestaciones eléctricas según el estándar IEC 61850 a partir de de una herramienta de diseño gráfico. Este fichero, se denomina SCD (Substation Configuration Description), y se trata de un fichero XML conforme a un esquema XSD (XML Schema Definition) mediante el que se define el lenguaje de configuración de subestaciones SCL del IEC 61850. Para el desarrollo de este proyecto, es necesario el estudio del lenguaje para la configuración de subestaciones SCL, así como del lenguaje gráfico específico de dominio definido por la herramienta “61850 SCLVisual Design Tool”, la estructura de los ficheros SCD, y finalmente, del lenguaje EGL (Epsilon Generation Language) para la transformación y generación automática de código a partir de modelos EMF. ABSTRACT Electrical networks are evolving to “Smart Grids”. Smart Grids are composed of electrical substations that in turn are composed of devices called IEDs (Intelligent Electronic Devices). The design of IEDs is defined by the IEC 61850 standard, which also specifies a Substation Configuration Languaje (SCL) used to define the configuration of substations and their IEDs. Nowadays, this international standard is not only used to design properly IEDs and guarantee their interoperability, but it is also used to design different electrical network devices, such as, smart meters. However, although the use of this standard is growing, its compression as well as its management, is still difficult due to its large volume of information and its level of detail. As a result, designing IEDs becomes a tedious task without a software support. As a consequence of this, in order to make easier the application of the IEC 61850 standard while designing IEDs, some software tools have been developed, such as: “Visual SCL”, “SCL Explorer” or “61850 SCLVisual Design Tool”. In particular, “61850 SCLVisual Design Tool” is a graphical tool used to make electrical substations models, and developed with the Eclipse Modeling Framework (EMF) and Epsilon Generative Modeling Technologies (GMT) by the research group SYST of the UPM. The aim of this project is to add a new functionality to “61850 Visual SCL DesignTool”. This new functionality consists of the automatic code generation of a substation configuration file according to the IEC 61850 standard. This file is called SCD (Substation Configuration Description), and it is a XML file that follows a XSD (XML Schema Definition) that defines the Substation Configuration Language (SCL) of the IEC 61850. In order to develop this project, it is necessary to study the Substation Configuration Language (SCL), the domain-specific graphical languaje defined by the tool “61850 SCLVisual Design Tool”, the structure of a SCD file, and the Epsilon Generation Language (EGL) used for the automatic code generation from EMF models

Relevância:

100.00% 100.00%

Publicador:

Resumo:

We isolated a Dictyostelium cytokinesis mutant with a defect in a novel locus called large volume sphere A (lvsA). lvsA mutants exhibit an unusual phenotype when attempting to undergo cytokinesis in suspension culture. Early in cytokinesis, they initiate furrow formation with concomitant myosin II localization at the cleavage furrow. However, the furrow is later disrupted by a bulge that forms in the middle of the cell. This bulge is bounded by furrows on both sides, which are often enriched in myosin II. The bulge can increase and decrease in size multiple times as the cell attempts to divide. Interestingly, this phenotype is similar to the cytokinesis failure of Dictyostelium clathrin heavy-chain mutants. Furthermore, both cell lines cap ConA receptors but form only a C-shaped loose cap. Unlike clathrin mutants, lvsA mutants are not defective in endocytosis or development. The LvsA protein shares several domains in common with the molecules beige and Chediak–Higashi syndrome proteins that are important for lysosomal membrane traffic. Thus, on the basis of the sequence analysis of the LvsA protein and the phenotype of the lvsA mutants, we postulate that LvsA plays an important role in a membrane-processing pathway that is essential for cytokinesis.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

High throughput genome (HTG) and expressed sequence tag (EST) sequences are currently the most abundant nucleotide sequence classes in the public database. The large volume, high degree of fragmentation and lack of gene structure annotations prevent efficient and effective searches of HTG and EST data for protein sequence homologies by standard search methods. Here, we briefly describe three newly developed resources that should make discovery of interesting genes in these sequence classes easier in the future, especially to biologists not having access to a powerful local bioinformatics environment. trEST and trGEN are regularly regenerated databases of hypothetical protein sequences predicted from EST and HTG sequences, respectively. Hits is a web-based data retrieval and analysis system providing access to precomputed matches between protein sequences (including sequences from trEST and trGEN) and patterns and profiles from Prosite and Pfam. The three resources can be accessed via the Hits home page (http://hits.isb-sib.ch).

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Clinical evidence of hematopoietic restoration with placental/umbilical cord blood (PCB) grafts indicates that PCB can be a useful source of hematopoietic stem cells for routine bone marrow reconstitution. In the unrelated setting, human leukocyte antigen (HLA)-matched donors must be obtained for candidate patients and, hence, large panels of frozen HLA-typed PCB units must be established. The large volume of unprocessed units, consisting mostly of red blood cells, plasma, and cryopreservation medium, poses a serious difficulty in this effort because storage space in liquid nitrogen is limited and costly. We report here that almost all the hematopoietic colony-forming cells present in PCB units can be recovered in a uniform volume of 20 ml by using rouleaux formation induced by hydroxyethyl starch and centrifugation to reduce the bulk of erythrocytes and plasma and, thus, concentrate leukocytes. This method multiples the number of units that can be stored in the same freezer space as much as 10-fold depending on the format of the storage system. We have also investigated the proportion of functional stem/progenitor cells initially present that are actually available to the recipient when thawed cryopreserved PCB units are infused. Progenitor cell viability is measurably decreased when thawed cells, still suspended in hypertonic cryopreservative solutions, are rapidly mixed with large volumes of isotonic solutions or plasma. The osmotic damage inflicted by the severe solute concentration gradient, however, can be averted by a simple 2-fold dilution after thawing, providing almost total recovery of viable hematopoietic progenitor cells.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

We investigate the critical properties of the four-state commutative random permutation glassy Potts model in three and four dimensions by means of Monte Carlo simulations and a finite-size scaling analysis. By using a field programmable gate array, we have been able to thermalize a large number of samples of systems with large volume. This has allowed us to observe a spin-glass ordered phase in d=4 and to study the critical properties of the transition. In d=3, our results are consistent with the presence of a Kosterlitz-Thouless transition, but also with different scenarios: transient effects due to a value of the lower critical dimension slightly below 3 could be very important.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Uma compreensão aprofundada da dinâmica de competição portuária é particularmente importante dado o contexto atual do setor, que orienta à outorga de novos portos e terminais no Brasil, à luz da Nova Lei dos Portos, Lei Nº 12.815 de 2013. A avaliação dos reais impactos decorrentes do aumento de capacidade portuária em cada região será atividade fundamental para que, por um lado, o poder público oriente a alocação efetiva de recursos, sem prejudicar a operação dos complexos existentes; e para que a iniciativa privada, por sua vez, possa compreender os impactos dos possíveis novos empreendimentos sobre as suas operações e delinear estratégias comerciais compatíveis com o novo cenário competitivo. A partir de extensa revisão bibliográfica e da aplicação de técnicas a casos específicos, o presente trabalho detalha a dinâmica competitiva entre terminais de contêineres e avalia criticamente seis métodos utilizados para identificar a existência de competição: correlação de market share, comparação de taxas de ocupação, sobreposição de escalas marítimas, comparação de custos logísticos terrestres, representatividade da região de influência contestável e existência de poder de mercado sobre a região de influência. Dos seis métodos analisados, dois apresentam conclusões fulminantes para a questão, embora sua aplicação demande grande volume de informações; um é assertivo em condições normais de distribuição geográfica de cargas; dois apresentam condições necessárias, porém não suficientes para a identificação de competição; e um deve ser aplicado com ressalvas, uma vez que pode levar a conclusões equivocadas.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

A Internet das Coisas é um novo paradigma de comunicação que estende o mundo virtual (Internet) para o mundo real com a interface e interação entre objetos. Ela possuirá um grande número de dispositivos heteregôneos interconectados, que deverá gerar um grande volume de dados. Um dos importantes desafios para seu desenvolvimento é se guardar e processar esse grande volume de dados em aceitáveis intervalos de tempo. Esta pesquisa endereça esse desafio, com a introdução de serviços de análise e reconhecimento de padrões nas camadas inferiores do modelo de para Internet das Coisas, que procura reduzir o processamento nas camadas superiores. Na pesquisa foram analisados os modelos de referência para Internet das Coisas e plataformas para desenvolvimento de aplicações nesse contexto. A nova arquitetura de implementada estende o LinkSmart Middeware pela introdução de um módulo para reconhecimento de padrões, implementa algoritmos para estimação de valores, detecção de outliers e descoberta de grupos nos dados brutos, oriundos de origens de dados. O novo módulo foi integrado à plataforma para Big Data Hadoop e usa as implementações algorítmicas do framework Mahout. Este trabalho destaca a importância da comunicação cross layer integrada à essa nova arquitetura. Nos experimentos desenvolvidos na pesquisa foram utilizadas bases de dados reais, provenientes do projeto Smart Santander, de modo a validar da nova arquitetura de IoT integrada aos serviços de análise e reconhecimento de padrões e a comunicação cross-layer.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Backbone networks are responsible for long-haul data transport serving many clients with a large volume of data. Since long-haul data transport service must rely on a robust high capacity network the current technology broadly adopted by the industry is Wavelength Division Multiplexing (WDM). WDM networks enable one single ber to operate with multiple high capacity channels, drastically increasing the ber capacity. In WDM networks each channel is associated with an individual wavelength. Therefore a whole wavelength capacity is assigned to a connection, causing waste of bandwidth in case the connection bandwidth requirement is less than the channel total capacity. In the last half decade, Elastic Optical Networks (EON) have been proposed and developed based on the fexible use of the optical spectrum known as the exigrid. EONs are adaptable to clients requirements and may enhance optical networks performance. For these reasons, research community and data transport providers have been demonstrating increasingly high interest in EONs which are likely to replace WDM as the universally adopted technology in backbone networks in the near future. EONs have two characteristics that may limit its ecient resources use. The spectrum fragmentation, inherent to the dynamic EON operation, decrease the network capacity to assign resources to connection requests increasing network blocking probability. The spectrum fragmentation also intensifides the denial of service to higher rate request inducing service unfairness. Due to the fact EONs were just recently developed and proposed, the aforementioned issues were not yet extensively studied and solutions are still being proposed. Furthermore, EONs do not yet provide specific features as differentiated service mechanisms. Differentiated service strategies are important in backbone networks to guarantee client\'s diverse requirements in case of a network failure or the natural congestion and resources contention that may occur at some periods of time in a network. Impelled by the foregoing facts, this thesis objective is three-fold. By means of developing and proposing a mechanism for routing and resources assignment in EONs, we intend to provide differentiated service while decreasing fragmentation level and increasing service fairness. The mechanism proposed and explained in this thesis was tested in a EON simulation environment and performance results indicated that it promotes beneficial performance enhancements when compared to benchmark algorithms.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The waste of plastic beverage bottles creates environmental problems and takes up a large volume of landfill space. The high rate of consumption of plastics in the State of Florida is challenging the disposal capacity of waste authorities. The lack of the reverse vending machines in the State of Florida, including applicable scientific or technical literature represented an opportunity for this research to discuss the applicability of this equipment as a potential solution for the management of the plastic waste in Florida. With this research document, I will propose a recycling system for plastic bottles made with PET based on the implementation of reverse vending machines, stressing the importance of the creation of policies that promote recycling and public participation.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

El presente artículo se centra en la valorización de las cenizas de lodo de depuradora adicionadas en bloques de hormigón prefabricados, por tanto, en una matriz a base de cemento Portland con características particulares (consistencia seca). Se realiza una experiencia piloto mediante la fabricación de bloques en una planta industrial local, utilizando mezclas con porcentajes de sustitución de árido fino del 5, 10 y 15% por cenizas de lodo de depuradora, cuyos resultados se compararán con bloques de referencia (sin adición). Los ensayos aplicados son los preceptivos para su comercialización según normativa europea (marcado CE), lo que permitirá igualmente comprobar las características físicas y mecánicas de los hormigones. Los resultados obtenidos son prometedores: la utilización de este residuo permitiría su comercialización, mejora características de los bloques como la resistencia a compresión, y permitiría reducir un gran volumen de material utilizado habitualmente como materia prima y consumir un residuo destinado a vertedero.