81 resultados para Weighted adjacency matrix
em CentAUR: Central Archive University of Reading - UK
Resumo:
For many networks in nature, science and technology, it is possible to order the nodes so that most links are short-range, connecting near-neighbours, and relatively few long-range links, or shortcuts, are present. Given a network as a set of observed links (interactions), the task of finding an ordering of the nodes that reveals such a range-dependent structure is closely related to some sparse matrix reordering problems arising in scientific computation. The spectral, or Fiedler vector, approach for sparse matrix reordering has successfully been applied to biological data sets, revealing useful structures and subpatterns. In this work we argue that a periodic analogue of the standard reordering task is also highly relevant. Here, rather than encouraging nonzeros only to lie close to the diagonal of a suitably ordered adjacency matrix, we also allow them to inhabit the off-diagonal corners. Indeed, for the classic small-world model of Watts & Strogatz (1998, Collective dynamics of ‘small-world’ networks. Nature, 393, 440–442) this type of periodic structure is inherent. We therefore devise and test a new spectral algorithm for periodic reordering. By generalizing the range-dependent random graph class of Grindrod (2002, Range-dependent random graphs and their application to modeling large small-world proteome datasets. Phys. Rev. E, 66, 066702-1–066702-7) to the periodic case, we can also construct a computable likelihood ratio that suggests whether a given network is inherently linear or periodic. Tests on synthetic data show that the new algorithm can detect periodic structure, even in the presence of noise. Further experiments on real biological data sets then show that some networks are better regarded as periodic than linear. Hence, we find both qualitative (reordered networks plots) and quantitative (likelihood ratios) evidence of periodicity in biological networks.
Resumo:
We explore the influence of the choice of attenuation factor on Katz centrality indices for evolving communication networks. For given snapshots of a network observed over a period of time, recently developed communicability indices aim to identify best broadcasters and listeners in the network. In this article, we looked into the sensitivity of communicability indices on the attenuation factor constraint, in relation to spectral radius (the largest eigenvalue) of the network at any point in time and its computation in the case of large networks. We proposed relaxed communicability measures where the spectral radius bound on attenuation factor is relaxed and the adjacency matrix is normalised in order to maintain the convergence of the measure. Using a vitality based measure of both standard and relaxed communicability indices we looked at the ways of establishing the most important individuals for broadcasting and receiving of messages related to community bridging roles. We illustrated our findings with two examples of real-life networks, MIT reality mining data set of daily communications between 106 individuals during one year and UK Twitter mentions network, direct messages on Twitter between 12.4k individuals during one week.
Resumo:
In this article, we investigate how the choice of the attenuation factor in an extended version of Katz centrality influences the centrality of the nodes in evolving communication networks. For given snapshots of a network, observed over a period of time, recently developed communicability indices aim to identify the best broadcasters and listeners (receivers) in the network. Here we explore the attenuation factor constraint, in relation to the spectral radius (the largest eigenvalue) of the network at any point in time and its computation in the case of large networks. We compare three different communicability measures: standard, exponential, and relaxed (where the spectral radius bound on the attenuation factor is relaxed and the adjacency matrix is normalised, in order to maintain the convergence of the measure). Furthermore, using a vitality-based measure of both standard and relaxed communicability indices, we look at the ways of establishing the most important individuals for broadcasting and receiving of messages related to community bridging roles. We compare those measures with the scores produced by an iterative version of the PageRank algorithm and illustrate our findings with two examples of real-life evolving networks: the MIT reality mining data set, consisting of daily communications between 106 individuals over the period of one year, a UK Twitter mentions network, constructed from the direct \emph{tweets} between 12.4k individuals during one week, and a subset the Enron email data set.
Resumo:
We characterize the essential spectra of Toeplitz operators Ta on weighted Bergman spaces with matrix-valued symbols; in particular we deal with two classes of symbols, the Douglas algebra C+H∞ and the Zhu class Q := L∞ ∩VMO∂ . In addition, for symbols in C+H∞ , we derive a formula for the index of Ta in terms of its symbol a in the scalar-valued case, while in the matrix-valued case we indicate that the standard reduction to the scalar-valued case fails to work analogously to the Hardy space case. Mathematics subject classification (2010): 47B35,
Resumo:
A discrete-time algorithm is presented which is based on a predictive control scheme in the form of dynamic matrix control. A set of control inputs are calculated and made available at each time instant, the actual input applied being a weighted summation of the inputs within the set. The algorithm is directly applicable in a self-tuning format and is therefore suitable for slowly time-varying systems in a noisy environment.
Resumo:
Results are presented from a matrix of coupled model integrations, using atmosphere resolutions of 135 and 90 km, and ocean resolutions of 1° and 1/3°, to study the impact of resolution on simulated climate. The mean state of the tropical Pacific is found to be improved in the models with a higher ocean resolution. Such an improved mean state arises from the development of tropical instability waves, which are poorly resolved at low resolution; these waves reduce the equatorial cold tongue bias. The improved ocean state also allows for a better simulation of the atmospheric Walker circulation. Several sensitivity studies have been performed to further understand the processes involved in the different component models. Significantly decreasing the horizontal momentum dissipation in the coupled model with the lower-resolution ocean has benefits for the mean tropical Pacific climate, but decreases model stability. Increasing the momentum dissipation in the coupled model with the higher-resolution ocean degrades the simulation toward that of the lower-resolution ocean. These results suggest that enhanced ocean model resolution can have important benefits for the climatology of both the atmosphere and ocean components of the coupled model, and that some of these benefits may be achievable at lower ocean resolution, if the model formulation allows.
Resumo:
A generic Nutrient Export Risk Matrix (NERM) approach is presented. This provides advice to farmers and policy makers on good practice for reducing nutrient loss and is intended to persuade them to implement such measures. Combined with a range of nutrient transport modelling tools and field experiments, NERMs can play an important role in reducing nutrient export from agricultural land. The Phosphorus Export Risk Matrix (PERM) is presented as an example NERM. The PERM integrates hydrological understanding of runoff with a number of agronomic and policy factors into a clear problem-solving framework. This allows farmers and policy makers to visualise strategies for reducing phosphorus loss through proactive land management. The risk Of Pollution is assessed by a series of informed questions relating to farming intensity and practice. This information is combined with the concept of runoff management to point towards simple, practical remedial strategies which do not compromise farmers' ability to obtain sound economic returns from their crop and livestock.
Resumo:
The influence matrix is used in ordinary least-squares applications for monitoring statistical multiple-regression analyses. Concepts related to the influence matrix provide diagnostics on the influence of individual data on the analysis - the analysis change that would occur by leaving one observation out, and the effective information content (degrees of freedom for signal) in any sub-set of the analysed data. In this paper, the corresponding concepts have been derived in the context of linear statistical data assimilation in numerical weather prediction. An approximate method to compute the diagonal elements of the influence matrix (the self-sensitivities) has been developed for a large-dimension variational data assimilation system (the four-dimensional variational system of the European Centre for Medium-Range Weather Forecasts). Results show that, in the boreal spring 2003 operational system, 15% of the global influence is due to the assimilated observations in any one analysis, and the complementary 85% is the influence of the prior (background) information, a short-range forecast containing information from earlier assimilated observations. About 25% of the observational information is currently provided by surface-based observing systems, and 75% by satellite systems. Low-influence data points usually occur in data-rich areas, while high-influence data points are in data-sparse areas or in dynamically active regions. Background-error correlations also play an important role: high correlation diminishes the observation influence and amplifies the importance of the surrounding real and pseudo observations (prior information in observation space). Incorrect specifications of background and observation-error covariance matrices can be identified, interpreted and better understood by the use of influence-matrix diagnostics for the variety of observation types and observed variables used in the data assimilation system. Copyright © 2004 Royal Meteorological Society
Resumo:
With its highly fluctuating ion production matrix-assisted laser desorption/ionization (MALDI) poses many practical challenges for its application in mass spectrometry. Instrument tuning and quantitative ion abundance measurements using ion signal alone depend on a stable ion beam. Liquid MALDI matrices have been shown to be a promising alternative to the commonly used solid matrices. Their application in areas where a stable ion current is essential has been discussed but only limited data have been provided to demonstrate their practical use and advantages in the formation of stable MALDI ion beams. In this article we present experimental data showing high MALDI ion beam stability over more than two orders of magnitude at high analytical sensitivity (low femtomole amount prepared) for quantitative peptide abundance measurements and instrument tuning in a MALDI Q-TOF mass spectrometer. Samples were deposited on an inexpensive conductive hydrophobic surface and shrunk to droplets <10 nL in size. By using a sample droplet <10 nL it was possible to acquire data from a single irradiated spot for roughly 10,000 shots with little variation in ion signal intensity at a laser repetition rate of 5-20 Hz.