34 resultados para Vertex Transitive Graph


Relevância:

10.00% 10.00%

Publicador:

Resumo:

Perceber a rede estrutural formada pelos neurónios no cérebro a nível da macro escala é um desafio atual na área das neurociências. Neste estudo analisou-se a conectividade estrutural do cérebro em 22 indivíduos saudáveis e em dois doentes com epilepsia pós-traumática. Avaliaram-se as diferenças entre estes dois grupos. Também se pesquisaram diferenças a nível do género e idade no grupo de indivíduos saudáveis e os que têm valores médios mais elevados nas métricas de caracterização da rede. Para tal, desenvolveu-se um protocolo de análise recorrendo a diversos softwares especializados e usaram-se métricas da Teoria dos Grafos para a caracterização da conectividade estrutural entre 118 regiões encefálicas distintas. Dentro do grupo dos indivíduos saudáveis concluiu-se que os homens, no geral, são os que têm média mais alta para as métricas de caracterização da rede estrutural. Contudo, não se observaram diferenças significativas em relação ao género nas métricas de caracterização global do cérebro. Relativamente à idade, esta correlaciona-se negativamente, no geral, com as métricas de caracterização da rede estrutural. As regiões onde se observaram as diferenças mais importantes entre indivíduos saudáveis e doentes são: o sulco rolândico, o hipocampo, o pré-cuneus, o tálamo e o cerebelo bilateralmente. Estas diferenças são consistentes com as imagens radiológicas dos doentes e com a literatura estudada sobre a epilepsia pós-traumática. Preveem-se desenvolvimentos para o estudo da conectividade estrutural do cérebro humano, uma vez que a sua potencialidade pode ser combinada com outros métodos de modo a caracterizar as alterações dos circuitos cerebrais.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Microarray allow to monitoring simultaneously thousands of genes, where the abundance of the transcripts under a same experimental condition at the same time can be quantified. Among various available array technologies, double channel cDNA microarray experiments have arisen in numerous technical protocols associated to genomic studies, which is the focus of this work. Microarray experiments involve many steps and each one can affect the quality of raw data. Background correction and normalization are preprocessing techniques to clean and correct the raw data when undesirable fluctuations arise from technical factors. Several recent studies showed that there is no preprocessing strategy that outperforms others in all circumstances and thus it seems difficult to provide general recommendations. In this work, it is proposed to use exploratory techniques to visualize the effects of preprocessing methods on statistical analysis of cancer two-channel microarray data sets, where the cancer types (classes) are known. For selecting differential expressed genes the arrow plot was used and the graph of profiles resultant from the correspondence analysis for visualizing the results. It was used 6 background methods and 6 normalization methods, performing 36 pre-processing methods and it was analyzed in a published cDNA microarray database (Liver) available at http://genome-www5.stanford.edu/ which microarrays were already classified by cancer type. All statistical analyses were performed using the R statistical software.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Trabalho de Projeto para obtenção do grau de Mestre em Engenharia Informática e de Computadores

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Relatório do Trabalho Final de Mestrado para obtenção do grau de Mestre em Engenharia de Electrónica e Telecomunicações

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Mestrado em Intervenção Sócio-Organizacional na Saúde - Ramo de especialização: Qualidade e Tecnologias da Saúde

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Although the computational power of mobile devices has been increasing, it is still not enough for some classes of applications. In the present, these applications delegate the computing power burden on servers located on the Internet. This model assumes an always-on Internet connectivity and implies a non-negligible latency. The thesis addresses the challenges and contributions posed to the application of a mobile collaborative computing environment concept to wireless networks. The goal is to define a reference architecture for high performance mobile applications. Current work is focused on efficient data dissemination on a highly transitive environment, suitable to many mobile applications and also to the reputation and incentive system available on this mobile collaborative computing environment. For this we are improving our already published reputation/incentive algorithm with knowledge from the usage pattern from the eduroam wireless network in the Lisbon area.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This letter presents a new parallel method for hyperspectral unmixing composed by the efficient combination of two popular methods: vertex component analysis (VCA) and sparse unmixing by variable splitting and augmented Lagrangian (SUNSAL). First, VCA extracts the endmember signatures, and then, SUNSAL is used to estimate the abundance fractions. Both techniques are highly parallelizable, which significantly reduces the computing time. A design for the commodity graphics processing units of the two methods is presented and evaluated. Experimental results obtained for simulated and real hyperspectral data sets reveal speedups up to 100 times, which grants real-time response required by many remotely sensed hyperspectral applications.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Trabalho de projeto realizado para obtenção do grau de Mestre em Engenharia Informática e de Computadores

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We start by presenting the current status of a complex flavour conserving two-Higgs doublet model. We will focus on some very interesting scenarios where unexpectedly the light Higgs couplings to leptons and to b-quarks can have a large pseudoscalar component with a vanishing scalar component. Predictions for the allowed parameter space at end of the next run with a total collected luminosity of 300 fb(-1) and 3000 fb(-1) are also discussed. These scenarios are not excluded by present data and most probably will survive the next LHC run. However, a measurement of the mixing angle phi(tau), between the scalar and pseudoscalar component of the 125 GeV Higgs, in the decay h -> tau(+)tau(-) will be able to probe many of these scenarios, even with low luminosity. Similarly, a measurement of phi(t) in the vertex (t) over bar th could help to constrain the low tan beta region in the Type I model.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper proposes an FPGA-based architecture for onboard hyperspectral unmixing. This method based on the Vertex Component Analysis (VCA) has several advantages, namely it is unsupervised, fully automatic, and it works without dimensionality reduction (DR) pre-processing step. The architecture has been designed for a low cost Xilinx Zynq board with a Zynq-7020 SoC FPGA based on the Artix-7 FPGA programmable logic and tested using real hyperspectral datasets. Experimental results indicate that the proposed implementation can achieve real-time processing, while maintaining the methods accuracy, which indicate the potential of the proposed platform to implement high-performance, low cost embedded systems.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

One of the most challenging task underlying many hyperspectral imagery applications is the linear unmixing. The key to linear unmixing is to find the set of reference substances, also called endmembers, that are representative of a given scene. This paper presents the vertex component analysis (VCA) a new method to unmix linear mixtures of hyperspectral sources. The algorithm is unsupervised and exploits a simple geometric fact: endmembers are vertices of a simplex. The algorithm complexity, measured in floating points operations, is O (n), where n is the sample size. The effectiveness of the proposed scheme is illustrated using simulated data.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Os sensores hiperespectrais que estão a ser desenvolvidos para aplicações em detecção remota, produzem uma elevada quantidade de dados. Tal quantidade de dados obriga a que as ferramentas de análise e processamento sejam eficientes e tenham baixa complexidade computacional. Uma tarefa importante na detecção remota é a determinação das substâncias presentes numa imagem hiperespectral e quais as suas concentrações. Neste contexto, Vertex component analysis (VCA), é um método não-supervisionado recentemente proposto que é eficiente e tem a complexidade computacional mais baixa de todos os métodos conhecidos. Este método baseia-se no facto de os vértices do simplex corresponderem às assinaturas dos elementos presentes nos dados. O VCA projecta os dados em direcções ortogonais ao subespaço gerado pelas assinaturas das substâncias já encontradas, correspondendo o extremo desta projecção à assinatura da nova substância encontrada. Nesta comunicação apresentam-se várias optimizações ao VCA nomeadamente: 1) a introdução de um método de inferência do sub-espaço de sinal que permite para além de reduzir a dimensionalidade dos dados, também permite estimar o número de substâncias presentes. 2) projeção dos dados é executada em várias direcções para garantir maior robustez em situações de baixa relação sinal-ruído. As potencialidades desta técnica são ilustradas num conjunto de experiências com dados simulados e reais, estes últimos adquiridos pela plataforma AVIRIS.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The reaction between 2-aminobenzenesulfonic acid and 2-hydroxy-3-methoxybenzaldehyde produces the acyclic Schiff base 2-[(2-hydroxy-3-methoxyphenyl) methylideneamino] benzenesulfonic acid (H2L center dot 3H(2)O) (1). In situ reactions of this compound with Cu(II) salts and, eventually, in the presence of pyridine (py) or 2,2'-bipyridine (2,2'-bipy) lead to the formation of the mononuclear complexes [CuL(H2O)(2)] (2) and [CuL(2,2'-bipy)]center dot DMF center dot H2O (3) and the diphenoxo-bridged dicopper compounds [CuL(py)](2) (4) and [CuL(EtOH)](2)center dot 2H(2)O (5). In 2-5 the L-2-ligand acts as a tridentate chelating species by means of one of the O-sulfonate atoms, the O-phenoxo and the N-atoms. The remaining coordination sites are then occupied by H2O (in 2), 2,2'-bipyridine (in 3), pyridine (in 4) or EtOH (in 5). Hydrogen bond interactions resulted in R-2(2) (14) and in R-4(4)(12) graph sets leading to dimeric species (in 2 and 3, respectively), 1D chain associations (in 2 and 5) or a 2D network (1). Complexes 2-5 are applied as selective catalysts for the homogeneous peroxidative (with tert-butylhydroperoxide, TBHP) oxidation of primary and secondary alcohols, under solvent-and additive-free conditions and under low power microwave (MW) irradiation. A quantitative yield of acetophenone was obtained by oxidation of 1-phenylethanol with compound 4 [TOFs up to 7.6 x 10(3) h(-1)] after 20 min of MW irradiation, whereas the oxidation of benzyl alcohol to benzaldehyde is less effective (TOF 992 h(-1)). The selectivity of 4 to oxidize the alcohol relative to the ene function is demonstrated when using cinnamyl alcohol as substrate.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Hyperspectral instruments have been incorporated in satellite missions, providing large amounts of data of high spectral resolution of the Earth surface. This data can be used in remote sensing applications that often require a real-time or near-real-time response. To avoid delays between hyperspectral image acquisition and its interpretation, the last usually done on a ground station, onboard systems have emerged to process data, reducing the volume of information to transfer from the satellite to the ground station. For this purpose, compact reconfigurable hardware modules, such as field-programmable gate arrays (FPGAs), are widely used. This paper proposes an FPGA-based architecture for hyperspectral unmixing. This method based on the vertex component analysis (VCA) and it works without a dimensionality reduction preprocessing step. The architecture has been designed for a low-cost Xilinx Zynq board with a Zynq-7020 system-on-chip FPGA-based on the Artix-7 FPGA programmable logic and tested using real hyperspectral data. Experimental results indicate that the proposed implementation can achieve real-time processing, while maintaining the methods accuracy, which indicate the potential of the proposed platform to implement high-performance, low-cost embedded systems, opening perspectives for onboard hyperspectral image processing.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Many Hyperspectral imagery applications require a response in real time or near-real time. To meet this requirement this paper proposes a parallel unmixing method developed for graphics processing units (GPU). This method is based on the vertex component analysis (VCA), which is a geometrical based method highly parallelizable. VCA is a very fast and accurate method that extracts endmember signatures from large hyperspectral datasets without the use of any a priori knowledge about the constituent spectra. Experimental results obtained for simulated and real hyperspectral datasets reveal considerable acceleration factors, up to 24 times.