65 resultados para Random linear Network Coding
Resumo:
Topology optimization consists in finding the spatial distribution of a given total volume of material for the resulting structure to have some optimal property, for instance, maximization of structural stiffness or maximization of the fundamental eigenfrequency. In this paper a Genetic Algorithm (GA) employing a representation method based on trees is developed to generate initial feasible individuals that remain feasible upon crossover and mutation and as such do not require any repairing operator to ensure feasibility. Several application examples are studied involving the topology optimization of structures where the objective functions is the maximization of the stiffness and the maximization of the first and the second eigenfrequencies of a plate, all cases having a prescribed material volume constraint.
Resumo:
Os reguladores de tensão LDO são utilizados intensivamente na actual indústria de electrónica, são uma parte essencial de um bloco de gestão de potência para um SoC. O aumento de produtos portáteis alimentados por baterias levou ao crescimento de soluções totalmente integradas, o que degrada o rendimento dos blocos analógicos que o constituem face às perturbações introduzidas na alimentação. Desta forma, surge a necessidade de procurar soluções cada vez mais optimizadas, impondo assim novas soluções, e/ou melhoramentos dos circuitos de gestão de potência, tendo como objectivo final o aumento do desempenho e da autonomia dos dispositivos electrónicos. Normalmente este tipo de reguladores tem a corrente de saída limitada, devido a problemas de estabilidade associados. Numa tentativa de evitar a instabilidade para as correntes de carga definidas e aumentar o PSRR do mesmo, é apresentado um método de implementação que tem como objectivo melhorar estas características, em que se pretende aumentar o rendimento e melhorar a resposta à variação da carga. No entanto, a técnica apresentada utiliza polarização adaptativa do estágio de potência, o que implica um aumento da corrente de consumo. O regulador LDO foi implementado na tecnologia CMOS UMC 0.18μm e ocupa uma área inferior a 0,2mm2. Os resultados da simulação mostram que o mesmo suporta uma transição de corrente 10μA para 100mA, com uma queda de tensão entre a tensão de alimentação e a tensão de saída inferior a 200mV. A estabilidade é assegurada para todas as correntes de carga. O tempo de estabelecimento é inferior a 6μs e as variações da tensão de saída relativamente a seu valor nominal são inferiores a 5mV. A corrente de consumo varia entre os 140μA até 200μA, o que permite atingir as especificações proposta para um PSRR de 40dB@10kHz.
Resumo:
The dynamics of a cylinder rolling on a horizontal plane acted on by an external force applied at an arbitrary angle is studied with emphasis on the directions of the acceleration of the centre-of-mass and the angular acceleration of the body. If rolling occurs without slipping, there is a relationship between the directions of these accelerations. If the linear acceleration points to the right, then the angular acceleration is clockwise. On the other hand, if it points to the left, then the angular acceleration is counterclockwise. In contrast, if rolling and slipping occurs, the direction of the linear acceleration does not determine the direction of the angular acceleration. For example, the linear acceleration may point to the right and the angular acceleration clockwise or counterclockwise depending on the external force orientation and point of application.
Resumo:
Pretende-se, utilizando o Modelo da Elasticidade Linear em freeFEM++, determinar os esforços e deslocamentos de um edifício alto submetido apenas à acção do peso próprio da estrutura e, efectuar estudos comparativos dos resultados obtidos com o SAP2000. O trabalho inicia-se com a introdução da teoria da elasticidade linear, onde são feitas as deduções das Equações de Compatibilidade, Equilíbrio e as Leis Constitutivas, de modo a resolver numericamente o problema de Elasticidade Linear mencionado. O método de elementos finitos será implementado em freeFEM++ com auxílio do GMSH que é uma poderosa ferramenta com capacidade de gerar automaticamente malhas tridimensionais de elementos finitos e com relativa facilidade de pré e pós-processamento.
Resumo:
The advances made in channel-capacity codes, such as turbo codes and low-density parity-check (LDPC) codes, have played a major role in the emerging distributed source coding paradigm. LDPC codes can be easily adapted to new source coding strategies due to their natural representation as bipartite graphs and the use of quasi-optimal decoding algorithms, such as belief propagation. This paper tackles a relevant scenario in distributedvideo coding: lossy source coding when multiple side information (SI) hypotheses are available at the decoder, each one correlated with the source according to different correlation noise channels. Thus, it is proposed to exploit multiple SI hypotheses through an efficient joint decoding technique withmultiple LDPC syndrome decoders that exchange information to obtain coding efficiency improvements. At the decoder side, the multiple SI hypotheses are created with motion compensated frame interpolation and fused together in a novel iterative LDPC based Slepian-Wolf decoding algorithm. With the creation of multiple SI hypotheses and the proposed decoding algorithm, bitrate savings up to 8.0% are obtained for similar decoded quality.
Resumo:
Neste trabalho apresenta-se o desenvolvimento de um programa de elementos finitos tridimensionais denominado AE3D1.0, concebido especificamente para a análise de pavimentos rodoviários, partindo do pressuposto de que todos os materiais incorporados possuem comportamento elástico-linear. Por comparação dos resultados do programa AE3D1.0 com as soluções analíticas da teoria da elasticidade para o semi-espaço homogéneo e multiestratificado, confirma-se que é possível estabelecer uma analogia próxima entre ambas as abordagens. Tirando partindo das potencialidades do método dos elementos finitos, e da capacidade do programa de registar os resultados de cálculo em ficheiros digitais que possibilitam a posterior apreciação visual e tratamento dos dados obtidos, comparam-se pavimentos rígidos expostos a carregamentos de canto e de bordo, e é evidenciado o efeito prejudicial que a erosão da estrutura de apoio subjacente à laje de betão tem na longevidade e integridade estrutural do pavimento. São também aplicadas forças de frenagem a pavimentos rígidos em secções confinadas e não confinadas. Elege-se um modelo de pneu para veículos pesados representativo das características do eixo padrão de 130 kN, e analisa-se o efeito que a correspondente impressão ovalizada e distribuição de pressões verticais não uniforme tem na estrutura de um pavimento semi-rígido. Adapta-se e é aplicada uma malha de elementos finitos ao estudo da avaliação da capacidade de carga de pavimentos através de ensaios com o defletómetro de impacto.
Resumo:
Tuberculosis (TB) is a worldwide infectious disease that has shown over time extremely high mortality levels. The urgent need to develop new antitubercular drugs is due to the increasing rate of appearance of multi-drug resistant strains to the commonly used drugs, and the longer durations of therapy and recovery, particularly in immuno-compromised patients. The major goal of the present study is the exploration of data from different families of compounds through the use of a variety of machine learning techniques so that robust QSAR-based models can be developed to further guide in the quest for new potent anti-TB compounds. Eight QSAR models were built using various types of descriptors (from ADRIANA.Code and Dragon software) with two publicly available structurally diverse data sets, including recent data deposited in PubChem. QSAR methodologies used Random Forests and Associative Neural Networks. Predictions for the external evaluation sets obtained accuracies in the range of 0.76-0.88 (for active/inactive classifications) and Q(2)=0.66-0.89 for regressions. Models developed in this study can be used to estimate the anti-TB activity of drug candidates at early stages of drug development (C) 2011 Elsevier B.V. All rights reserved.
Resumo:
We study a model consisting of particles with dissimilar bonding sites ("patches"), which exhibits self-assembly into chains connected by Y-junctions, and investigate its phase behaviour by both simulations and theory. We show that, as the energy cost epsilon(j) of forming Y-junctions increases, the extent of the liquid-vapour coexistence region at lower temperatures and densities is reduced. The phase diagram thus acquires a characteristic "pinched" shape in which the liquid branch density decreases as the temperature is lowered. To our knowledge, this is the first model in which the predicted topological phase transition between a fluid composed of short chains and a fluid rich in Y-junctions is actually observed. Above a certain threshold for epsilon(j), condensation ceases to exist because the entropy gain of forming Y-junctions can no longer offset their energy cost. We also show that the properties of these phase diagrams can be understood in terms of a temperature-dependent effective valence of the patchy particles. (C) 2011 American Institute of Physics. [doi: 10.1063/1.3605703]
Resumo:
We introduce a microscopic model for particles with dissimilar patches which displays an unconventional "pinched'' phase diagram, similar to the one predicted by Tlusty and Safran in the context of dipolar fluids [Science 290, 1328 (2000)]. The model-based on two types of patch interactions, which account, respectively, for chaining and branching of the self-assembled networks-is studied both numerically via Monte Carlo simulations and theoretically via first-order perturbation theory. The dense phase is rich in junctions, while the less-dense phase is rich in chain ends. The model provides a reference system for a deep understanding of the competition between condensation and self-assembly into equilibrium-polymer chains.
Resumo:
There exist striking analogies in the behaviour of eigenvalues of Hermitian compact operators, singular values of compact operators and invariant factors of homomorphisms of modules over principal ideal domains, namely diagonalization theorems, interlacing inequalities and Courant-Fischer type formulae. Carlson and Sa [D. Carlson and E.M. Sa, Generalized minimax and interlacing inequalities, Linear Multilinear Algebra 15 (1984) pp. 77-103.] introduced an abstract structure, the s-space, where they proved unified versions of these theorems in the finite-dimensional case. We show that this unification can be done using modular lattices with Goldie dimension, which have a natural structure of s-space in the finite-dimensional case, and extend the unification to the countable-dimensional case.
Resumo:
RESUMO: Introdução – A Radioterapia (RT) é uma abordagem terapêutica para tratamento de neoplasia de mama. Contudo, diferentes técnicas de irradiação (TI) podem ser usadas. Objetivos – Comparar 4 TI, considerando a irradiação dos volumes alvo (PTV) e dos órgãos de risco (OAR). Metodologia – Selecionaram-se 7 pacientes com indicação para RT de mama esquerda. Sobre tomografia computorizada foram feitos os contornos do PTV e dos OAR. Foram calculadas 4 planimetrias/paciente para as TI: conformacional externa (EBRT), intensidade modulada com 2 (IMRT2) e 5 campos (IMRT5) e arco dinâmico (DART). Resultados – Histogramas de dose volume foram comparados para todas as TI usando o software de análise estatística, IBM SPSS v20. Com IMRT5 e DART, os OAR recebem mais doses baixas. No entanto, IMRT5 apresenta melhores índices de conformidade e homogeneidade para o PTV. Conclusões – IMRT5 apresenta o melhor índice de conformidade; EBRT e IMRT2 apresentam melhores resultados que DART. Há d.e.s entre as TI, sobretudo em doses mais baixas nos OAR.
Resumo:
A novel high throughput and scalable unified architecture for the computation of the transform operations in video codecs for advanced standards is presented in this paper. This structure can be used as a hardware accelerator in modern embedded systems to efficiently compute all the two-dimensional 4 x 4 and 2 x 2 transforms of the H.264/AVC standard. Moreover, its highly flexible design and hardware efficiency allows it to be easily scaled in terms of performance and hardware cost to meet the specific requirements of any given video coding application. Experimental results obtained using a Xilinx Virtex-5 FPGA demonstrated the superior performance and hardware efficiency levels provided by the proposed structure, which presents a throughput per unit of area relatively higher than other similar recently published designs targeting the H.264/AVC standard. Such results also showed that, when integrated in a multi-core embedded system, this architecture provides speedup factors of about 120x concerning pure software implementations of the transform algorithms, therefore allowing the computation, in real-time, of all the above mentioned transforms for Ultra High Definition Video (UHDV) sequences (4,320 x 7,680 @ 30 fps).
Resumo:
Video coding technologies have played a major role in the explosion of large market digital video applications and services. In this context, the very popular MPEG-x and H-26x video coding standards adopted a predictive coding paradigm, where complex encoders exploit the data redundancy and irrelevancy to 'control' much simpler decoders. This codec paradigm fits well applications and services such as digital television and video storage where the decoder complexity is critical, but does not match well the requirements of emerging applications such as visual sensor networks where the encoder complexity is more critical. The Slepian Wolf and Wyner-Ziv theorems brought the possibility to develop the so-called Wyner-Ziv video codecs, following a different coding paradigm where it is the task of the decoder, and not anymore of the encoder, to (fully or partly) exploit the video redundancy. Theoretically, Wyner-Ziv video coding does not incur in any compression performance penalty regarding the more traditional predictive coding paradigm (at least for certain conditions). In the context of Wyner-Ziv video codecs, the so-called side information, which is a decoder estimate of the original frame to code, plays a critical role in the overall compression performance. For this reason, much research effort has been invested in the past decade to develop increasingly more efficient side information creation methods. This paper has the main objective to review and evaluate the available side information methods after proposing a classification taxonomy to guide this review, allowing to achieve more solid conclusions and better identify the next relevant research challenges. After classifying the side information creation methods into four classes, notably guess, try, hint and learn, the review of the most important techniques in each class and the evaluation of some of them leads to the important conclusion that the side information creation methods provide better rate-distortion (RD) performance depending on the amount of temporal correlation in each video sequence. It became also clear that the best available Wyner-Ziv video coding solutions are almost systematically based on the learn approach. The best solutions are already able to systematically outperform the H.264/AVC Intra, and also the H.264/AVC zero-motion standard solutions for specific types of content. (C) 2013 Elsevier B.V. All rights reserved.
Resumo:
In this contribution, we investigate the low-temperature, low-density behaviour of dipolar hard-sphere (DHS) particles, i.e., hard spheres with dipoles embedded in their centre. We aim at describing the DHS fluid in terms of a network of chains and rings (the fundamental clusters) held together by branching points (defects) of different nature. We first introduce a systematic way of classifying inter-cluster connections according to their topology, and then employ this classification to analyse the geometric and thermodynamic properties of each class of defects, as extracted from state-of-the-art equilibrium Monte Carlo simulations. By computing the average density and energetic cost of each defect class, we find that the relevant contribution to inter-cluster interactions is indeed provided by (rare) three-way junctions and by four-way junctions arising from parallel or anti-parallel locally linear aggregates. All other (numerous) defects are either intra-cluster or associated to low cluster-cluster interaction energies, suggesting that these defects do not play a significant part in the thermodynamic description of the self-assembly processes of dipolar hard spheres. (C) 2013 AIP Publishing LLC.
Resumo:
In the last years it has become increasingly clear that the mammalian transcriptome is highly complex and includes a large number of small non-coding RNAs (sncRNAs) and long noncoding RNAs (lncRNAs). Here we review the biogenesis pathways of the three classes of sncRNAs, namely short interfering RNAs (siRNAs), microRNAs (miRNAs) and PIWI-interacting RNAs (piRNAs). These ncRNAs have been extensively studied and are involved in pathways leading to specific gene silencing and the protection of genomes against virus and transposons, for example. Also, lncRNAs have emerged as pivotal molecules for the transcriptional and post-transcriptional regulation of gene expression which is supported by their tissue-specific expression patterns, subcellular distribution, and developmental regulation. Therefore, we also focus our attention on their role in differentiation and development. SncRNAs and lncRNAs play critical roles in defining DNA methylation patterns, as well as chromatin remodeling thus having a substantial effect in epigenetics. The identification of some overlaps in their biogenesis pathways and functional roles raises the hypothesis that these molecules play concerted functions in vivo, creating complex regulatory networks where cooperation with regulatory proteins is necessary. We also highlighted the implications of biogenesis and gene expression deregulation of sncRNAs and lncRNAs in human diseases like cancer.