989 resultados para linear predictive coding (LPC)
Resumo:
This paper presents a distributed model predictive control (DMPC) for indoor thermal comfort that simultaneously optimizes the consumption of a limited shared energy resource. The control objective of each subsystem is to minimize the heating/cooling energy cost while maintaining the indoor temperature and used power inside bounds. In a distributed coordinated environment, the control uses multiple dynamically decoupled agents (one for each subsystem/house) aiming to achieve satisfaction of coupling constraints. According to the hourly power demand profile, each house assigns a priority level that indicates how much is willing to bid in auction for consume the limited clean resource. This procedure allows the bidding value vary hourly and consequently, the agents order to access to the clean energy also varies. Despite of power constraints, all houses have also thermal comfort constraints that must be fulfilled. The system is simulated with several houses in a distributed environment.
Resumo:
A novel high throughput and scalable unified architecture for the computation of the transform operations in video codecs for advanced standards is presented in this paper. This structure can be used as a hardware accelerator in modern embedded systems to efficiently compute all the two-dimensional 4 x 4 and 2 x 2 transforms of the H.264/AVC standard. Moreover, its highly flexible design and hardware efficiency allows it to be easily scaled in terms of performance and hardware cost to meet the specific requirements of any given video coding application. Experimental results obtained using a Xilinx Virtex-5 FPGA demonstrated the superior performance and hardware efficiency levels provided by the proposed structure, which presents a throughput per unit of area relatively higher than other similar recently published designs targeting the H.264/AVC standard. Such results also showed that, when integrated in a multi-core embedded system, this architecture provides speedup factors of about 120x concerning pure software implementations of the transform algorithms, therefore allowing the computation, in real-time, of all the above mentioned transforms for Ultra High Definition Video (UHDV) sequences (4,320 x 7,680 @ 30 fps).
Resumo:
OBJETIVO: Identificar os determinantes da desnutrição energético-protéica que ocasionam déficits ponderal e de crescimento linear em crianças. MÉTODOS: Estudo transversal envolvendo 1.041 crianças (menores de dois anos de idade) de 10 municípios do Estado da Bahia, de 1999 a 2000. Utilizou-se a técnica de regressão logística e estratégia da abordagem hierárquica para identificar os fatores associados ao estado antropométrico. RESULTADOS: O modelo final para déficit no crescimento linear revelou como determinante básico: a posse de dois ou menos equipamentos domésticos (OR=2,9; IC 95%: 1,74-4,90) e no nível subjacente, a ausência de consulta pré-natal (OR=2,7; IC 95%: 1,47-4,97); entre os determinantes imediatos o baixo peso ao nascer (<2.500 g) (OR=3,6; IC 95%: 1,72-7,70) e relato de hospitalização nos 12 meses anteriores à entrevista (OR=2,4; IC 95%: 1,42-4,10). Fatores determinantes no déficit ponderal nos níveis básico, subjacente e imediato foram, respectivamente: a renda mensal per capita inferior a ¼ do salário-mínimo (OR=3,4; IC 95%: 1,41-8,16), a ausência de pré-natal (OR=2,1; IC 95%: 1,03-4,35), e o baixo peso ao nascer (OR=4,8; IC 95%: 2,00-11,48). CONCLUSÕES: Os déficits ponderal e linear das crianças foram explicados pela intermediação entre as precárias condições materiais de vida e o restrito acesso ao cuidado com a saúde e a carga de morbidade. Intervenções que melhorem as condições de vida e ampliem o acesso às ações do serviço de saúde são estratégias que caminham na busca da eqüidade em saúde e nutrição na infância.
Resumo:
In the last years it has become increasingly clear that the mammalian transcriptome is highly complex and includes a large number of small non-coding RNAs (sncRNAs) and long noncoding RNAs (lncRNAs). Here we review the biogenesis pathways of the three classes of sncRNAs, namely short interfering RNAs (siRNAs), microRNAs (miRNAs) and PIWI-interacting RNAs (piRNAs). These ncRNAs have been extensively studied and are involved in pathways leading to specific gene silencing and the protection of genomes against virus and transposons, for example. Also, lncRNAs have emerged as pivotal molecules for the transcriptional and post-transcriptional regulation of gene expression which is supported by their tissue-specific expression patterns, subcellular distribution, and developmental regulation. Therefore, we also focus our attention on their role in differentiation and development. SncRNAs and lncRNAs play critical roles in defining DNA methylation patterns, as well as chromatin remodeling thus having a substantial effect in epigenetics. The identification of some overlaps in their biogenesis pathways and functional roles raises the hypothesis that these molecules play concerted functions in vivo, creating complex regulatory networks where cooperation with regulatory proteins is necessary. We also highlighted the implications of biogenesis and gene expression deregulation of sncRNAs and lncRNAs in human diseases like cancer.
Resumo:
OBJETIVO: Estratégias metodológicas vêm sendo desenvolvidas para minimizar o efeito do erro de medida da dieta. O objetivo do estudo foi descrever a aplicação de uma estratégia para correção da informação dietética pelo erro de medida. MÉTODOS: Foram obtidos dados de consumo alimentar pela aplicação do Questionário de Freqüência Alimentar a 79 adolescentes do Município de São Paulo em 1999. Os dados dietéticos obtidos foram corrigidos por meio de regressão linear, após o ajuste pela energia usando-se o método dos resíduos. O método de referência utilizado foi o recordatório de 24 horas, aplicado em três momentos distintos. RESULTADOS: Os valores corrigidos aproximaram-se dos valores de referência. O fator de correção lambda foi de 0,89 para energia. Para os macronutrientes, os fatores foram de 0,41; 0,22 e 0,20, para carboidratos, lipídios e proteínas, respectivamente. CONCLUSÕES: As médias e desvios-padrão dos valores corrigidos denotam que houve uma correção do erro de medida. Apesar disso, debate-se o desempenho desses métodos, que são notoriamente imperfeitos quando seus pressupostos teóricos não são atendidos, o que é comum nos estudos da dieta que usam instrumentos de medida baseados no relato dos indivíduos.
Resumo:
Low-density parity-check (LDPC) codes are nowadays one of the hottest topics in coding theory, notably due to their advantages in terms of bit error rate performance and low complexity. In order to exploit the potential of the Wyner-Ziv coding paradigm, practical distributed video coding (DVC) schemes should use powerful error correcting codes with near-capacity performance. In this paper, new ways to design LDPC codes for the DVC paradigm are proposed and studied. The new LDPC solutions rely on merging parity-check nodes, which corresponds to reduce the number of rows in the parity-check matrix. This allows to change gracefully the compression ratio of the source (DCT coefficient bitplane) according to the correlation between the original and the side information. The proposed LDPC codes reach a good performance for a wide range of source correlations and achieve a better RD performance when compared to the popular turbo codes.
Resumo:
A 9.9 kb DNA fragment from the right arm of chromosome VII of Saccharomyces cerevisiae has been sequenced and analysed. The sequence contains four open reading frames (ORFs) longer than 100 amino acids. One gene, PFK1, has already been cloned and sequenced and the other one is the probable yeast gene coding for the beta-subunit of the succinyl-CoA synthetase. The two remaining ORFs share homology with the deduced amino acid sequence (and their physical arrangement is similar to that) of the YHR161c and YHR162w ORFs from chromosome VIII.
Resumo:
As indústrias de componentes e acessórios automóveis são um elo fundamental no processo produtivo da indústria automóvel. Neste leque industrial encontra-se a Preh Portugal, Lda, como empresa fornecedora de componentes electrónicos, mais concretamente, painéis de controlo de climatização. Os painéis fornecidos pela Preh aos seus clientes encontram-se sujeitos a rigorosos testes de qualidade e funcionalidade. Neste sentido o teste funcional das teclas surge, relacionando o curso da tecla em função de uma força actuante. Esta relação está comprometida com uma curva característica padrão para o tipo de tecla. Para além destes compromissos, também é necessário que a tecla feche e abra o seu contacto eléctrico. Esta tese foca-se no desenvolvimento do teste de teclas, apresentando uma alteração ao sistema actual com a introdução de um sistema embebido, no intuito de flexibilizar o sistema de teste e reduzindo custos. O sistema embebido pretende dar capacidade de processamento ao teste e, desta forma, substituir o actual computador como elemento de processamento. A solução implementada consistiu numa mudança estrutural, através da inclusão do sistema embebido entre o computador e o sistema de deslocamento. Passando o foco central do processo de teste a residir no sistema embebido, este tem de estabelecer comunicações com os restantes elementos intervenientes no teste. Estabelece comunicações série RS-232 com o sistema de deslocamento (leitura do curso e força na tecla), Ethernet com o computador (comandos, parâmetros e resultados) e CAN com o painel de controlo de climatização (fecho/abertura do contacto eléctrico). A concretização deste projecto resultou numa nova estrutura e aplicação, a qual é facilmente integrada na linha de produção com as vantagens de ser menos onerosa e mais flexível, conforme o pretendido.
Resumo:
OBJECTIVE: To assess a new impunity index and variables that have been found to predict variation in homicide rates in other geographical levels as predictive of state-level homicide rates in Brazil. METHODS: This was a cross-sectional ecological study. Data from the mortality information system relating to the 27 Brazilian states for the years 1996 to 2005 were analyzed. The outcome variables were taken to be homicide victim rates in 2005, for the entire population and for men aged 20-29 years. Measurements of economic and social development, economic inequality, demographic structure and life expectancy were analyzed as predictors. An "impunity index", calculated as the total number of homicides between 1996 and 2005 divided by the number of individuals in prison in 2007, was constructed. The data were analyzed by means of simple linear regression and negative binomial regression. RESULTS: In 2005, state-level crude total homicide rates ranged from 11 to 51 per 100,000; for young men, they ranged from 39 to 241. The impunity index ranged from 0.4 to 3.5 and was the most important predictor of this variability. From negative binomial regression, it was estimated that the homicide victim rate among young males increased by 50% for every increase of one point in this ratio. CONCLUSIONS: Classic predictive factors were not associated with homicides in this analysis of state-level variation in Brazil. However, the impunity index indicated that the greater the impunity, the higher the homicide rate.
Resumo:
In basaltic dykes the magnetic lineation K1 (maximum magnetic susceptibility axis) is generally taken to indicate the flow direction during solidification of the magma. This assumption was tested in Tertiary basaltic dykes from Greenland displaying independent evidence of subhorizontal flow. The digital processing of microphotographs from thin sections cut in (K1, K2) planes yields the preferred linear orientation of plagioclase, which apparently marks the magma flow lineation. In up to 60% of cases, the angular separation between K1 and the assumed flow direction is greater than 45degrees. This suggests that the uncorroborated use of magnetic lineations in dykes is risky. A simple geometrical method is proposed to infer the flow vector from AMS in dykes based solely on magnetic foliations.
Resumo:
Chpater in Book Proceedings with Peer Review Second Iberian Conference, IbPRIA 2005, Estoril, Portugal, June 7-9, 2005, Proceedings, Part II
Resumo:
We use a simple model of associating fluids which consists of spherical particles having a hard-core repulsion, complemented by three short-ranged attractive sites on the surface (sticky spots). Two of the spots are of type A and one is of type B; the bonding interactions between each pair of spots have strengths epsilon(AA), epsilon(BB), and epsilon(AB). The theory is applied over the whole range of bonding strengths and the results are interpreted in terms of the equilibrium cluster structures of the phases. In addition to our numerical results, we derive asymptotic expansions for the free energy in the limits for which there is no liquid-vapor critical point: linear chains (epsilon(AA)not equal 0, epsilon(AB)=epsilon(BB)=0), hyperbranched polymers (epsilon(AB)not equal 0, epsilon(AA)=epsilon(BB)=0), and dimers (epsilon(BB)not equal 0, epsilon(AA)=epsilon(AB)=0). These expansions also allow us to calculate the structure of the critical fluid by perturbing around the above limits, yielding three different types of condensation: of linear chains (AA clusters connected by a few AB or BB bonds); of hyperbranched polymers (AB clusters connected by AA bonds); or of dimers (BB clusters connected by AA bonds). Interestingly, there is no critical point when epsilon(AA) vanishes despite the fact that AA bonds alone cannot drive condensation.
Resumo:
Screening of topologies developed by hierarchical heuristic procedures can be carried out by comparing their optimal performance. In this work we will be exploiting mono-objective process optimization using two algorithms, simulated annealing and tabu search, and four different objective functions: two of the net present value type, one of them including environmental costs and two of the global potential impact type. The hydrodealkylation of toluene to produce benzene was used as case study, considering five topologies with different complexities mainly obtained by including or not liquid recycling and heat integration. The performance of the algorithms together with the objective functions was observed, analyzed and discussed from various perspectives: average deviation of results for each algorithm, capacity for producing high purity product, screening of topologies, objective functions robustness in screening of topologies, trade-offs between economic and environmental type objective functions and variability of optimum solutions.
Resumo:
Amulti-residue methodology based on a solid phase extraction followed by gas chromatography–tandem mass spectrometry was developed for trace analysis of 32 compounds in water matrices, including estrogens and several pesticides from different chemical families, some of them with endocrine disrupting properties. Matrix standard calibration solutions were prepared by adding known amounts of the analytes to a residue-free sample to compensate matrix-induced chromatographic response enhancement observed for certain pesticides. Validation was done mainly according to the International Conference on Harmonisation recommendations, as well as some European and American validation guidelines with specifications for pesticides analysis and/or GC–MS methodology. As the assumption of homoscedasticity was not met for analytical data, weighted least squares linear regression procedure was applied as a simple and effective way to counteract the greater influence of the greater concentrations on the fitted regression line, improving accuracy at the lower end of the calibration curve. The method was considered validated for 31 compounds after consistent evaluation of the key analytical parameters: specificity, linearity, limit of detection and quantification, range, precision, accuracy, extraction efficiency, stability and robustness.
Resumo:
Ancillary services represent a good business opportunity that must be considered by market players. This paper presents a new methodology for ancillary services market dispatch. The method considers the bids submitted to the market and includes a market clearing mechanism based on deterministic optimization. An Artificial Neural Network is used for day-ahead prediction of Regulation Down, regulation-up, Spin Reserve and Non-Spin Reserve requirements. Two test cases based on California Independent System Operator data concerning dispatch of Regulation Down, Regulation Up, Spin Reserve and Non-Spin Reserve services are included in this paper to illustrate the application of the proposed method: (1) dispatch considering simple bids; (2) dispatch considering complex bids.