971 resultados para Hermitian radial basis function


Relevância:

20.00% 20.00%

Publicador:

Resumo:

Nos últimos vinte e cinco anos o tema da autonomia e da administração e gestão escolar tem ocupado um lugar relevante na agenda política dos sucessivos Governos da República e na preocupação dos diferentes parceiros educativos. Rara tem sido a maioria política que resiste a dar o seu contributo sobre esta matéria, com o objetivo sempre confesso de outorgar maior autonomia às escolas. No enquadramento teórico da nossa investigação começamos por abordar a emergência do conceito de autonomia, nas suas diferentes dimensões e nos seus distintos significados. Não esquecemos também a analise das questões relacionadas com a problemática, cada vez mais atual da regulação múltipla. Analisamos de seguida a evolução da legislação portuguesa, operada a partir da publicação da Lei de Bases do Sistema Educativo com especial destaque às propostas de configuração dos órgãos de Direção e de Gestão das escolas e das competências atribuídas a cada um deles produzido pela CRSE e pelos decretos-leis 43/89, 172/91, 115-A/98 e 75/2008. A investigação empírica teve como objeto de análise dois agrupamentos localizados em concelhos distintos da Área Metropolitana de Lisboa, e procurou determinar se o conselho geral de cada uma dessas unidades orgânicas, assume na totalidade as competências que lhe são conferidas pelo quadro legislativo em vigor, e nessa medida como se articula com os outros órgãos da direção no processo de tomada de decisão. Simultaneamente fizemos o contraponto com a imagem que os intervenientes na gestão intermédia de cada um dos agrupamentos construíram sobre o seu conselho geral e das relações de poder que se estabelecem no interior de cada uma das organizações. Para corresponder aos pressupostos da nossa investigação entrevistaram-se os diretores e os presidentes dos conselhos gerais e facultámos questionários aos docentes que desempenhavam cargos nos dois agrupamentos. Concluímos, em função do que pudemos analisar, que embora o conselho geral veja o seu papel na organização da escola formalmente reconhecido não consegue desempenhar na totalidade as funções que lhe são incumbidas, já que defronta o poder real do diretor e o poder oculto do conselho pedagógico, encontrando dificuldades em libertar-se do reino das sombras.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Nos dias de hoje é necessário criar hábitos de vida mais saudáveis que contribuam para o bem-estar da população. Adoptar medidas e práticas de modo regular e disciplinado, pode diminuir o risco do aparecimento de determinadas doenças, como a obesidade, as doenças cardiovasculares, a hipertensão, a diabetes, alguns tipos de cancro e tantas outras. É também importante salientar que, uma alimentação cuidada dá saúde e aumenta a esperança média de vida. Em Portugal, nos últimos anos, os costumes alimentares da população têm vindo a alterar-se significativamente. As refeições caseiras confeccionadas com produtos frescos dão lugar à designada “cultura do fast food”. Em contrapartida, os consumidores são cada vez mais exigentes, estando em permanente alerta no que se refere ao estado dos alimentos. A rotulagem de um produto, para além da função publicitária, tem vindo a ser objecto de legislação específica de forma a fornecer informação simples e clara, correspondente à composição, qualidade, quantidade, validade ou outras características do produto. Estas informações devem ser acessíveis a qualquer tipo de público, com mais ou menos formação e de qualquer estrato social. A qualidade e segurança dos produtos deve basear-se na garantia de que todos os ingredientes, materiais de embalagem e processos produtivos são adequados à produção de produtos seguros, saudáveis e saborosos. A Silliker Portugal, S.A. é uma empresa independente de prestação de serviços para o sector agro-alimentar, líder mundial na prestação de serviços para a melhoria da qualidade e segurança alimentar. A Silliker dedica-se a ajudar as empresas a encontrar soluções para os desafios actuais do sector, oferecendo uma ampla gama de serviços, onde se inclui o serviço de análises microbiológicas, químicas e sensorial; consultadoria em segurança alimentar e desenvolvimento; auditorias; rotulagem e legislação. A actualização permanente de procedimentos na procura de uma melhoria contínua é um dos objectivos da empresa. Para responder a um dos desafios colocados à Silliker, surgiu este trabalho, que consistiu no desenvolvimento de um novo método para determinação de ácidos gordos e da gordura total em diferentes tipos de alimentos e comparação dos resultados, com os obtidos com o método analítico até então adoptado. Se a gordura é um elemento de grande importância na alimentação, devido às suas propriedades nutricionais e organoléticas, recentemente, os investigadores têm focado a sua atenção nos mais diversos ácidos gordos (saturados, monoinsaturados e polinsaturados), em particular nos ácidos gordos essenciais e nos isómeros do ácido linoleico conjugado (CLA), uma mistura de isómeros posicionais e geométricos do ácido linoleico com actividade biológica importante. A técnica usada nas determinações foi a cromatografia gasosa com ionização de chama, GC-FID, tendo as amostras sido previamente tratadas e extraídas de acordo com o tipo de matriz. A metodologia analítica desenvolvida permitiu a correcta avaliação do perfil em ácidos gordos, tendo-se para isso usado uma mistura de 37 ésteres metílicos, em que o ácido gordo C13:0 foi usado como padrão interno. A identificação baseou-se nos tempos de retenção de cada ácido gordo da mistura e para a quantificação usaram-se os factores de resposta. A validação do método implementado foi baseada nos resultados obtidos no estudo de três matrizes relativas a materiais certificados pela BIPEA (Bureau Interprofessionnel des Etudes Analytiques), para o que foram efectuadas doze réplicas de cada matriz. Para cada réplica efectuada foi calculado o teor de matéria gorda, sendo posteriormente o resultado comparado com o emitido pela entidade certificada. Após análise de cada constituinte foi também possível calcular o teor em ácidos gordos saturados, monoinsaturados e polinsaturados. A determinação do perfil em ácidos gordos dos materiais certificados foi aceitável atendendo aos valores obtidos, os quais se encontravam no intervalo de valores admissíveis indicados nos relatórios. A quantificação da matéria gorda no que se refere à matriz de “Paté à Tartinier” apresentou um z-score de 4,3, o que de acordo com as exigências internas da Silliker, não é válido. Para as outras duas matrizes (“Mélange Nutritif” e “Plat cuisiné à base de viande”) os valores de z-score foram, respectivamente, 0,7 e -1,0, o que permite concluir a validade do método. Para que o método possa vir a ser adoptado como método alternativo é necessário um estudo mais alargado relativamente a amostras com diferentes composições. O método foi aplicado na análise de amostras de fiambre, leite gordo, queijo, ovo com ómega 3, amendoim e óleo de girassol, e os resultados foram comparados com os obtidos pelo método até então adoptado.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We investigate the physical meaning of some of the "texture zeros" which appear in most of the Ansatze on quark masses and mixings. It is shown that starting from arbitrary quark mass matrices and making a suitable weak basis transformation one can obtain some of these sets of zeros which therefore have no physical content. We then analyse the physical implications of a four-texture zero Ansatz which is in agreement with all present experimental data. (C) 2000 Elsevier Science B.V. AU rights reserved.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Aims - To compare reading performance in children with and without visual function anomalies and identify the influence of abnormal visual function and other variables in reading ability. Methods - A cross-sectional study was carried in 110 children of school age (6-11 years) with Abnormal Visual Function (AVF) and 562 children with Normal Visual Function (NVF). An orthoptic assessment (visual acuity, ocular alignment, near point of convergence and accommodation, stereopsis and vergences) and autorefraction was carried out. Oral reading was analyzed (list of 34 words). Number of errors, accuracy (percentage of success) and reading speed (words per minute - wpm) were used as reading indicators. Sociodemographic information from parents (n=670) and teachers (n=34) was obtained. Results - Children with AVF had a higher number of errors (AVF=3.00 errors; NVF=1.00 errors; p<0.001), a lower accuracy (AVF=91.18%; NVF=97.06%; p<0.001) and reading speed (AVF=24.71 wpm; NVF=27.39 wpm; p=0.007). Reading speed in the 3rd school grade was not statistically different between the two groups (AVF=31.41 wpm; NVF=32.54 wpm; p=0.113). Children with uncorrected hyperopia (p=0.003) and astigmatism (p=0.019) had worst reading performance. Children in 2nd, 3rd, or 4th grades presented a lower risk of having reading impairment when compared with the 1st grade. Conclusion - Children with AVF had reading impairment in the first school grade. It seems that reading abilities have a wide variation and this disparity lessens in older children. The slow reading characteristics of the children with AVF are similar to dyslexic children, which suggest the need for an eye evaluation before classifying the children as dyslexic.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Background: Temporal lobe epilepsy (TLE) is a neurological disorder that directly affects cortical areas responsible for auditory processing. The resulting abnormalities can be assessed using event-related potentials (ERP), which have high temporal resolution. However, little is known about TLE in terms of dysfunction of early sensory memory encoding or possible correlations between EEGs, linguistic deficits, and seizures. Mismatch negativity (MMN) is an ERP component – elicited by introducing a deviant stimulus while the subject is attending to a repetitive behavioural task – which reflects pre-attentive sensory memory function and reflects neuronal auditory discrimination and perceptional accuracy. Hypothesis: We propose an MMN protocol for future clinical application and research based on the hypothesis that children with TLE may have abnormal MMN for speech and non-speech stimuli. The MMN can be elicited with a passive auditory oddball paradigm, and the abnormalities might be associated with the location and frequency of epileptic seizures. Significance: The suggested protocol might contribute to a better understanding of the neuropsychophysiological basis of MMN. We suggest that in TLE central sound representation may be decreased for speech and non-speech stimuli. Discussion: MMN arises from a difference to speech and non-speech stimuli across electrode sites. TLE in childhood might be a good model for studying topographic and functional auditory processing and its neurodevelopment, pointing to MMN as a possible clinical tool for prognosis, evaluation, follow-up, and rehabilitation for TLE.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Não existe uma definição única de processo de memória de longo prazo. Esse processo é geralmente definido como uma série que possui um correlograma decaindo lentamente ou um espectro infinito de frequência zero. Também se refere que uma série com tal propriedade é caracterizada pela dependência a longo prazo e por não periódicos ciclos longos, ou que essa característica descreve a estrutura de correlação de uma série de longos desfasamentos ou que é convencionalmente expressa em termos do declínio da lei-potência da função auto-covariância. O interesse crescente da investigação internacional no aprofundamento do tema é justificado pela procura de um melhor entendimento da natureza dinâmica das séries temporais dos preços dos ativos financeiros. Em primeiro lugar, a falta de consistência entre os resultados reclama novos estudos e a utilização de várias metodologias complementares. Em segundo lugar, a confirmação de processos de memória longa tem implicações relevantes ao nível da (1) modelação teórica e econométrica (i.e., dos modelos martingale de preços e das regras técnicas de negociação), (2) dos testes estatísticos aos modelos de equilíbrio e avaliação, (3) das decisões ótimas de consumo / poupança e de portefólio e (4) da medição de eficiência e racionalidade. Em terceiro lugar, ainda permanecem questões científicas empíricas sobre a identificação do modelo geral teórico de mercado mais adequado para modelar a difusão das séries. Em quarto lugar, aos reguladores e gestores de risco importa saber se existem mercados persistentes e, por isso, ineficientes, que, portanto, possam produzir retornos anormais. O objetivo do trabalho de investigação da dissertação é duplo. Por um lado, pretende proporcionar conhecimento adicional para o debate da memória de longo prazo, debruçando-se sobre o comportamento das séries diárias de retornos dos principais índices acionistas da EURONEXT. Por outro lado, pretende contribuir para o aperfeiçoamento do capital asset pricing model CAPM, considerando uma medida de risco alternativa capaz de ultrapassar os constrangimentos da hipótese de mercado eficiente EMH na presença de séries financeiras com processos sem incrementos independentes e identicamente distribuídos (i.i.d.). O estudo empírico indica a possibilidade de utilização alternativa das obrigações do tesouro (OT’s) com maturidade de longo prazo no cálculo dos retornos do mercado, dado que o seu comportamento nos mercados de dívida soberana reflete a confiança dos investidores nas condições financeiras dos Estados e mede a forma como avaliam as respetiva economias com base no desempenho da generalidade dos seus ativos. Embora o modelo de difusão de preços definido pelo movimento Browniano geométrico gBm alegue proporcionar um bom ajustamento das séries temporais financeiras, os seus pressupostos de normalidade, estacionariedade e independência das inovações residuais são adulterados pelos dados empíricos analisados. Por isso, na procura de evidências sobre a propriedade de memória longa nos mercados recorre-se à rescaled-range analysis R/S e à detrended fluctuation analysis DFA, sob abordagem do movimento Browniano fracionário fBm, para estimar o expoente Hurst H em relação às séries de dados completas e para calcular o expoente Hurst “local” H t em janelas móveis. Complementarmente, são realizados testes estatísticos de hipóteses através do rescaled-range tests R/S , do modified rescaled-range test M - R/S e do fractional differencing test GPH. Em termos de uma conclusão única a partir de todos os métodos sobre a natureza da dependência para o mercado acionista em geral, os resultados empíricos são inconclusivos. Isso quer dizer que o grau de memória de longo prazo e, assim, qualquer classificação, depende de cada mercado particular. No entanto, os resultados gerais maioritariamente positivos suportam a presença de memória longa, sob a forma de persistência, nos retornos acionistas da Bélgica, Holanda e Portugal. Isto sugere que estes mercados estão mais sujeitos a maior previsibilidade (“efeito José”), mas também a tendências que podem ser inesperadamente interrompidas por descontinuidades (“efeito Noé”), e, por isso, tendem a ser mais arriscados para negociar. Apesar da evidência de dinâmica fractal ter suporte estatístico fraco, em sintonia com a maior parte dos estudos internacionais, refuta a hipótese de passeio aleatório com incrementos i.i.d., que é a base da EMH na sua forma fraca. Atendendo a isso, propõem-se contributos para aperfeiçoamento do CAPM, através da proposta de uma nova fractal capital market line FCML e de uma nova fractal security market line FSML. A nova proposta sugere que o elemento de risco (para o mercado e para um ativo) seja dado pelo expoente H de Hurst para desfasamentos de longo prazo dos retornos acionistas. O expoente H mede o grau de memória de longo prazo nos índices acionistas, quer quando as séries de retornos seguem um processo i.i.d. não correlacionado, descrito pelo gBm(em que H = 0,5 , confirmando- se a EMH e adequando-se o CAPM), quer quando seguem um processo com dependência estatística, descrito pelo fBm(em que H é diferente de 0,5, rejeitando-se a EMH e desadequando-se o CAPM). A vantagem da FCML e da FSML é que a medida de memória de longo prazo, definida por H, é a referência adequada para traduzir o risco em modelos que possam ser aplicados a séries de dados que sigam processos i.i.d. e processos com dependência não linear. Então, estas formulações contemplam a EMH como um caso particular possível.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Radio Link Quality Estimation (LQE) is a fundamental building block for Wireless Sensor Networks, namely for a reliable deployment, resource management and routing. Existing LQEs (e.g. PRR, ETX, Fourbit, and LQI ) are based on a single link property, thus leading to inaccurate estimation. In this paper, we propose F-LQE, that estimates link quality on the basis of four link quality properties: packet delivery, asymmetry, stability, and channel quality. Each of these properties is defined in linguistic terms, the natural language of Fuzzy Logic. The overall quality of the link is specified as a fuzzy rule whose evaluation returns the membership of the link in the fuzzy subset of good links. Values of the membership function are smoothed using EWMA filter to improve stability. An extensive experimental analysis shows that F-LQE outperforms existing estimators.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This Thesis describes the application of automatic learning methods for a) the classification of organic and metabolic reactions, and b) the mapping of Potential Energy Surfaces(PES). The classification of reactions was approached with two distinct methodologies: a representation of chemical reactions based on NMR data, and a representation of chemical reactions from the reaction equation based on the physico-chemical and topological features of chemical bonds. NMR-based classification of photochemical and enzymatic reactions. Photochemical and metabolic reactions were classified by Kohonen Self-Organizing Maps (Kohonen SOMs) and Random Forests (RFs) taking as input the difference between the 1H NMR spectra of the products and the reactants. The development of such a representation can be applied in automatic analysis of changes in the 1H NMR spectrum of a mixture and their interpretation in terms of the chemical reactions taking place. Examples of possible applications are the monitoring of reaction processes, evaluation of the stability of chemicals, or even the interpretation of metabonomic data. A Kohonen SOM trained with a data set of metabolic reactions catalysed by transferases was able to correctly classify 75% of an independent test set in terms of the EC number subclass. Random Forests improved the correct predictions to 79%. With photochemical reactions classified into 7 groups, an independent test set was classified with 86-93% accuracy. The data set of photochemical reactions was also used to simulate mixtures with two reactions occurring simultaneously. Kohonen SOMs and Feed-Forward Neural Networks (FFNNs) were trained to classify the reactions occurring in a mixture based on the 1H NMR spectra of the products and reactants. Kohonen SOMs allowed the correct assignment of 53-63% of the mixtures (in a test set). Counter-Propagation Neural Networks (CPNNs) gave origin to similar results. The use of supervised learning techniques allowed an improvement in the results. They were improved to 77% of correct assignments when an ensemble of ten FFNNs were used and to 80% when Random Forests were used. This study was performed with NMR data simulated from the molecular structure by the SPINUS program. In the design of one test set, simulated data was combined with experimental data. The results support the proposal of linking databases of chemical reactions to experimental or simulated NMR data for automatic classification of reactions and mixtures of reactions. Genome-scale classification of enzymatic reactions from their reaction equation. The MOLMAP descriptor relies on a Kohonen SOM that defines types of bonds on the basis of their physico-chemical and topological properties. The MOLMAP descriptor of a molecule represents the types of bonds available in that molecule. The MOLMAP descriptor of a reaction is defined as the difference between the MOLMAPs of the products and the reactants, and numerically encodes the pattern of bonds that are broken, changed, and made during a chemical reaction. The automatic perception of chemical similarities between metabolic reactions is required for a variety of applications ranging from the computer validation of classification systems, genome-scale reconstruction (or comparison) of metabolic pathways, to the classification of enzymatic mechanisms. Catalytic functions of proteins are generally described by the EC numbers that are simultaneously employed as identifiers of reactions, enzymes, and enzyme genes, thus linking metabolic and genomic information. Different methods should be available to automatically compare metabolic reactions and for the automatic assignment of EC numbers to reactions still not officially classified. In this study, the genome-scale data set of enzymatic reactions available in the KEGG database was encoded by the MOLMAP descriptors, and was submitted to Kohonen SOMs to compare the resulting map with the official EC number classification, to explore the possibility of predicting EC numbers from the reaction equation, and to assess the internal consistency of the EC classification at the class level. A general agreement with the EC classification was observed, i.e. a relationship between the similarity of MOLMAPs and the similarity of EC numbers. At the same time, MOLMAPs were able to discriminate between EC sub-subclasses. EC numbers could be assigned at the class, subclass, and sub-subclass levels with accuracies up to 92%, 80%, and 70% for independent test sets. The correspondence between chemical similarity of metabolic reactions and their MOLMAP descriptors was applied to the identification of a number of reactions mapped into the same neuron but belonging to different EC classes, which demonstrated the ability of the MOLMAP/SOM approach to verify the internal consistency of classifications in databases of metabolic reactions. RFs were also used to assign the four levels of the EC hierarchy from the reaction equation. EC numbers were correctly assigned in 95%, 90%, 85% and 86% of the cases (for independent test sets) at the class, subclass, sub-subclass and full EC number level,respectively. Experiments for the classification of reactions from the main reactants and products were performed with RFs - EC numbers were assigned at the class, subclass and sub-subclass level with accuracies of 78%, 74% and 63%, respectively. In the course of the experiments with metabolic reactions we suggested that the MOLMAP / SOM concept could be extended to the representation of other levels of metabolic information such as metabolic pathways. Following the MOLMAP idea, the pattern of neurons activated by the reactions of a metabolic pathway is a representation of the reactions involved in that pathway - a descriptor of the metabolic pathway. This reasoning enabled the comparison of different pathways, the automatic classification of pathways, and a classification of organisms based on their biochemical machinery. The three levels of classification (from bonds to metabolic pathways) allowed to map and perceive chemical similarities between metabolic pathways even for pathways of different types of metabolism and pathways that do not share similarities in terms of EC numbers. Mapping of PES by neural networks (NNs). In a first series of experiments, ensembles of Feed-Forward NNs (EnsFFNNs) and Associative Neural Networks (ASNNs) were trained to reproduce PES represented by the Lennard-Jones (LJ) analytical potential function. The accuracy of the method was assessed by comparing the results of molecular dynamics simulations (thermal, structural, and dynamic properties) obtained from the NNs-PES and from the LJ function. The results indicated that for LJ-type potentials, NNs can be trained to generate accurate PES to be used in molecular simulations. EnsFFNNs and ASNNs gave better results than single FFNNs. A remarkable ability of the NNs models to interpolate between distant curves and accurately reproduce potentials to be used in molecular simulations is shown. The purpose of the first study was to systematically analyse the accuracy of different NNs. Our main motivation, however, is reflected in the next study: the mapping of multidimensional PES by NNs to simulate, by Molecular Dynamics or Monte Carlo, the adsorption and self-assembly of solvated organic molecules on noble-metal electrodes. Indeed, for such complex and heterogeneous systems the development of suitable analytical functions that fit quantum mechanical interaction energies is a non-trivial or even impossible task. The data consisted of energy values, from Density Functional Theory (DFT) calculations, at different distances, for several molecular orientations and three electrode adsorption sites. The results indicate that NNs require a data set large enough to cover well the diversity of possible interaction sites, distances, and orientations. NNs trained with such data sets can perform equally well or even better than analytical functions. Therefore, they can be used in molecular simulations, particularly for the ethanol/Au (111) interface which is the case studied in the present Thesis. Once properly trained, the networks are able to produce, as output, any required number of energy points for accurate interpolations.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Penalty and Barrier methods are normally used to solve Nonlinear Optimization Problems constrained problems. The problems appear in areas such as engineering and are often characterised by the fact that involved functions (objective and constraints) are non-smooth and/or their derivatives are not know. This means that optimization methods based on derivatives cannot net used. A Java based API was implemented, including only derivative-free optimizationmethods, to solve both constrained and unconstrained problems, which includes Penalty and Barriers methods. In this work a new penalty function, based on Fuzzy Logic, is presented. This function imposes a progressive penalization to solutions that violate the constraints. This means that the function imposes a low penalization when the violation of the constraints is low and a heavy penalisation when the violation is high. The value of the penalization is not known in beforehand, it is the outcome of a fuzzy inference engine. Numerical results comparing the proposed function with two of the classic penalty/barrier functions are presented. Regarding the presented results one can conclude that the prosed penalty function besides being very robust also exhibits a very good performance.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

OBJECTIVE To understand the job function of caregivers of older adults and contribute to the debate on the consolidation of this professional practice. METHODOLOGICAL PROCEDURES This is a descriptive, qualitative, and exploratory study. Four focal group sessions were performed in 2011 with 11 elderly companions, formal caregivers of older adults in the Programa Acompanhante de Idosos (Program for Caregivers of Older Adults), Sao Paulo, SP, Southeastern Brazil. These sessions, guided by a semi-structured script, were audio-recorded and fully transcribed. Data were analyzed using the Content Analysis technique, Thematic Modality. RESULTS In view of considering the caregivers of older adults as a new category of workers, it was difficult to define their duties. The elderly companions themselves as well as the care receivers, their families, and the professionals that comprised the team were unclear about their duties. The professional practice of these formal caregivers has been built on the basis of constant discussions and negotiations among them and other team members in Programa Acompanhante de Idosos during daily work. This was achieved via a recognition process of their job functions and by setting apart other workers’ exclusive responsibilities. CONCLUSIONS The delimitation of specific job functions for elderly companions is currently one of the greatest challenges faced by these workers to develop and consolidate their professional role as well as improve Programa Acompanhante de Idosos.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

3D laser scanning is becoming a standard technology to generate building models of a facility's as-is condition. Since most constructions are constructed upon planar surfaces, recognition of them paves the way for automation of generating building models. This paper introduces a new logarithmically proportional objective function that can be used in both heuristic and metaheuristic (MH) algorithms to discover planar surfaces in a point cloud without exploiting any prior knowledge about those surfaces. It can also adopt itself to the structural density of a scanned construction. In this paper, a metaheuristic method, genetic algorithm (GA), is used to test this introduced objective function on a synthetic point cloud. The results obtained show the proposed method is capable to find all plane configurations of planar surfaces (with a wide variety of sizes) in the point cloud with a minor distance to the actual configurations. © 2014 IEEE.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper proposes a Genetic Algorithm (GA) for the design of combinational logic circuits. The fitness function evaluation is calculated using Fractional Calculus. This approach extends the classical fitness function by including a fractional-order dynamical evaluation. The experiments reveal superior results when comparing with the classical method.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We study the existence and multiplicity of positive radial solutions of the Dirichlet problem for the Minkowski-curvature equation { -div(del upsilon/root 1-vertical bar del upsilon vertical bar(2)) in B-R, upsilon=0 on partial derivative B-R,B- where B-R is a ball in R-N (N >= 2). According to the behaviour off = f (r, s) near s = 0, we prove the existence of either one, two or three positive solutions. All results are obtained by reduction to an equivalent non-singular one-dimensional problem, to which variational methods can be applied in a standard way.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Refractive indices, n(D), and densities, rho, at 298.15 K were measured for the ternary mixture methanol (MeOH)/propan-1-ol (1-PrOH)/acetonitrile (MeCN) for a total of 22 mole fractions, along with 18 mole fractions of each of the corresponding binary mixtures, methanol/propan-1-ol, propan-1-ol/acetonitrile and methanol/acetonitrile. The variation of excess refractive indices and excess molar volumes with composition was modeled by the Redlich-Kister polynomial function in the case of binary mixtures and by the Cibulka equation for the ternary mixture. A thermodynamic approach to excess refractive indices, recently proposed by other authors, was applied for the first time to ternary liquid mixtures. Structural effects were identified and interpreted both in the binary and ternary systems. A complex relationship between excess refractive indices and excess molar volumes was identified, revealing all four possible sign combinations between these two properties. Structuring of the mixtures was also discussed on the basis of partial molar volumes of the binary and ternary mixtures.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Dynamically reconfigurable SRAM-based field-programmable gate arrays (FPGAs) enable the implementation of reconfigurable computing systems where several applications may be run simultaneously, sharing the available resources according to their own immediate functional requirements. To exclude malfunctioning due to faulty elements, the reliability of all FPGA resources must be guaranteed. Since resource allocation takes place asynchronously, an online structural test scheme is the only way of ensuring reliable system operation. On the other hand, this test scheme should not disturb the operation of the circuit, otherwise availability would be compromised. System performance is also influenced by the efficiency of the management strategies that must be able to dynamically allocate enough resources when requested by each application. As those resources are allocated and later released, many small free resource blocks are created, which are left unused due to performance and routing restrictions. To avoid wasting logic resources, the FPGA logic space must be defragmented regularly. This paper presents a non-intrusive active replication procedure that supports the proposed test methodology and the implementation of defragmentation strategies, assuring both the availability of resources and their perfect working condition, without disturbing system operation.