838 resultados para Networks analysis
Resumo:
The plethora, and mass take up, of digital communication tech- nologies has resulted in a wealth of interest in social network data collection and analysis in recent years. Within many such networks the interactions are transient: thus those networks evolve over time. In this paper we introduce a class of models for such networks using evolving graphs with memory dependent edges, which may appear and disappear according to their recent history. We consider time discrete and time continuous variants of the model. We consider the long term asymptotic behaviour as a function of parameters controlling the memory dependence. In particular we show that such networks may continue evolving forever, or else may quench and become static (containing immortal and/or extinct edges). This depends on the ex- istence or otherwise of certain infinite products and series involving age dependent model parameters. To test these ideas we show how model parameters may be calibrated based on limited samples of time dependent data, and we apply these concepts to three real networks: summary data on mobile phone use from a developing region; online social-business network data from China; and disaggregated mobile phone communications data from a reality mining experiment in the US. In each case we show that there is evidence for memory dependent dynamics, such as that embodied within the class of models proposed here.
Resumo:
The themes of awareness and influence within the innovation diffusion process are addressed. The innovation diffusion process is typically represented as stages, yet awareness and influence are somewhat under-represented in the literature. Awareness and influence are situated within the contextual setting of individual actors but also within the broader institutional forces. Understanding how actors become aware of an innovation and then how their opinion is influenced is important for creating a more innovation-active UK construction sector. Social network analysis is proposed as one technique for mapping how awareness and influence occur and what they look like as a network. Empirical data are gathered using two modes of enquiry. This is done through a pilot study consisting of chartered professionals and then through a case study organization as it attempted to diffuse an innovation. The analysis demonstrates significant variations across actors’ awareness and influence networks. It is argued that social network analysis can complement other research methods in order to present a richer picture of how actors become aware of innovations and where they draw their influences regarding adopting innovations. In summarizing the findings, a framework for understanding awareness and influence associated with innovation within the UK construction sector is presented. Finally, with the UK construction sector continually being encouraged to be innovative, understanding and managing an actor’s awareness and influence network will be beneficial. The overarching conclusion thus describes the need not only to build research capacity in this area but also to push the boundaries related to the research methods employed.
Resumo:
The Stochastic Diffusion Search algorithm -an integral part of Stochastic Search Networks is investigated. Stochastic Diffusion Search is an alternative solution for invariant pattern recognition and focus of attention. It has been shown that the algorithm can be modelled as an ergodic, finite state Markov Chain under some non-restrictive assumptions. Sub-linear time complexity for some settings of parameters has been formulated and proved. Some properties of the algorithm are then characterised and numerical examples illustrating some features of the algorithm are presented.
Resumo:
Deep Brain Stimulation has been used in the study of and for treating Parkinson’s Disease (PD) tremor symptoms since the 1980s. In the research reported here we have carried out a comparative analysis to classify tremor onset based on intraoperative microelectrode recordings of a PD patient’s brain Local Field Potential (LFP) signals. In particular, we compared the performance of a Support Vector Machine (SVM) with two well known artificial neural network classifiers, namely a Multiple Layer Perceptron (MLP) and a Radial Basis Function Network (RBN). The results show that in this study, using specifically PD data, the SVM provided an overall better classification rate achieving an accuracy of 81% recognition.
Resumo:
The paper analyses the emergence of group-specific attitudes and beliefs about tax compliance when individuals interact in a social network. It develops a model in which taxpayers possess a range of individual characteristics – including attitude to risk, potential for success in self-employment, and the weight attached to the social custom for honesty – and make an occupational choice based on these characteristics. Occupations differ in the possibility for evading tax. The social network determines which taxpayers are linked, and information about auditing and compliance is transmitted at meetings between linked taxpayers. Using agent-based simulations, the analysis demonstrates how attitudes and beliefs endogenously emerge that differ across sub-groups of the population. Compliance behaviour is different across occupational groups, and this is reinforced by the development of group-specific attitudes and beliefs. Taxpayers self-select into occupations according to the degree of risk aversion, the subjective probability of audit is sustained above the objective probability, and the weight attached to the social custom differs across occupations. These factors combine to lead to compliance levels that differ across occupations.
Resumo:
The K-Means algorithm for cluster analysis is one of the most influential and popular data mining methods. Its straightforward parallel formulation is well suited for distributed memory systems with reliable interconnection networks, such as massively parallel processors and clusters of workstations. However, in large-scale geographically distributed systems the straightforward parallel algorithm can be rendered useless by a single communication failure or high latency in communication paths. The lack of scalable and fault tolerant global communication and synchronisation methods in large-scale systems has hindered the adoption of the K-Means algorithm for applications in large networked systems such as wireless sensor networks, peer-to-peer systems and mobile ad hoc networks. This work proposes a fully distributed K-Means algorithm (EpidemicK-Means) which does not require global communication and is intrinsically fault tolerant. The proposed distributed K-Means algorithm provides a clustering solution which can approximate the solution of an ideal centralised algorithm over the aggregated data as closely as desired. A comparative performance analysis is carried out against the state of the art sampling methods and shows that the proposed method overcomes the limitations of the sampling-based approaches for skewed clusters distributions. The experimental analysis confirms that the proposed algorithm is very accurate and fault tolerant under unreliable network conditions (message loss and node failures) and is suitable for asynchronous networks of very large and extreme scale.
Resumo:
Working memory (WM) is not a unitary construct. There are distinct processes involved in encoding information, maintaining it on-line, and using it to guide responses. The anatomical configurations of these processes are more accurately analyzed as functionally connected networks than collections of individual regions. In the current study we analyzed event-related functional magnetic resonance imaging (fMRI) data from a Sternberg Item Recognition Paradigm WM task using a multivariate analysis method that allowed the linking of functional networks to temporally-separated WM epochs. The length of the delay epochs was varied to optimize isolation of the hemodynamic response (HDR) for each task epoch. All extracted functional networks displayed statistically significant sensitivity to delay length. Novel information extracted from these networks that was not apparent in the univariate analysis of these data included involvement of the hippocampus in encoding/probe, and decreases in BOLD signal in the superior temporal gyrus (STG), along with default-mode regions, during encoding/delay. The bilateral hippocampal activity during encoding/delay fits with theoretical models of WM in which memoranda held across the short term are activated long-term memory representations. The BOLD signal decreases in the STG were unexpected, and may reflect repetition suppression effects invoked by internal repetition of letter stimuli. Thus, analysis methods focusing on how network dynamics relate to experimental conditions allowed extraction of novel information not apparent in univariate analyses, and are particularly recommended for WM experiments for which task epochs cannot be randomized.
Resumo:
The UK has a target for an 80% reduction in CO2 emissions by 2050 from a 1990 base. Domestic energy use accounts for around 30% of total emissions. This paper presents a comprehensive review of existing models and modelling techniques and indicates how they might be improved by considering individual buying behaviour. Macro (top-down) and micro (bottom-up) models have been reviewed and analysed. It is found that bottom-up models can project technology diffusion due to their higher resolution. The weakness of existing bottom-up models at capturing individual green technology buying behaviour has been identified. Consequently, Markov chains, neural networks and agent-based modelling are proposed as possible methods to incorporate buying behaviour within a domestic energy forecast model. Among the three methods, agent-based models are found to be the most promising, although a successful agent approach requires large amounts of input data. A prototype agent-based model has been developed and tested, which demonstrates the feasibility of an agent approach. This model shows that an agent-based approach is promising as a means to predict the effectiveness of various policy measures.
Resumo:
In the recent years, the area of data mining has been experiencing considerable demand for technologies that extract knowledge from large and complex data sources. There has been substantial commercial interest as well as active research in the area that aim to develop new and improved approaches for extracting information, relationships, and patterns from large datasets. Artificial neural networks (NNs) are popular biologically-inspired intelligent methodologies, whose classification, prediction, and pattern recognition capabilities have been utilized successfully in many areas, including science, engineering, medicine, business, banking, telecommunication, and many other fields. This paper highlights from a data mining perspective the implementation of NN, using supervised and unsupervised learning, for pattern recognition, classification, prediction, and cluster analysis, and focuses the discussion on their usage in bioinformatics and financial data analysis tasks. © 2012 Wiley Periodicals, Inc.
Resumo:
The article features a conversation between Rob Cross and Martin Kilduff about organizational network analysis in research and practice. It demonstrates the value of using social network perspectives in HRM. Drawing on the discussion about managing personal networks; managing the networks of others; the impact of social networking sites on perceptions of relationships; and ethical issues in organizational network analysis, we propose specific suggestions to bring social network perspectives closer to HRM researchers and practitioners and rebalance our attention to people and to their relationships.
Resumo:
An incidence matrix analysis is used to model a three-dimensional network consisting of resistive and capacitive elements distributed across several interconnected layers. A systematic methodology for deriving a descriptor representation of the network with random allocation of the resistors and capacitors is proposed. Using a transformation of the descriptor representation into standard state-space form, amplitude and phase admittance responses of three-dimensional random RC networks are obtained. Such networks display an emergent behavior with a characteristic Jonscher-like response over a wide range of frequencies. A model approximation study of these networks is performed to infer the admittance response using integral and fractional order models. It was found that a fractional order model with only seven parameters can accurately describe the responses of networks composed of more than 70 nodes and 200 branches with 100 resistors and 100 capacitors. The proposed analysis can be used to model charge migration in amorphous materials, which may be associated to specific macroscopic or microscopic scale fractal geometrical structures in composites displaying a viscoelastic electromechanical response, as well as to model the collective responses of processes governed by random events described using statistical mechanics.
Resumo:
In this paper, we develop a method, termed the Interaction Distribution (ID) method, for analysis of quantitative ecological network data. In many cases, quantitative network data sets are under-sampled, i.e. many interactions are poorly sampled or remain unobserved. Hence, the output of statistical analyses may fail to differentiate between patterns that are statistical artefacts and those which are real characteristics of ecological networks. The ID method can support assessment and inference of under-sampled ecological network data. In the current paper, we illustrate and discuss the ID method based on the properties of plant-animal pollination data sets of flower visitation frequencies. However, the ID method may be applied to other types of ecological networks. The method can supplement existing network analyses based on two definitions of the underlying probabilities for each combination of pollinator and plant species: (1), pi,j: the probability for a visit made by the i’th pollinator species to take place on the j’th plant species; (2), qi,j: the probability for a visit received by the j’th plant species to be made by the i’th pollinator. The method applies the Dirichlet distribution to estimate these two probabilities, based on a given empirical data set. The estimated mean values for pi,j and qi,j reflect the relative differences between recorded numbers of visits for different pollinator and plant species, and the estimated uncertainty of pi,j and qi,j decreases with higher numbers of recorded visits.
Resumo:
This study aims to characterise the rainfall exceptionality and the meteorological context of the 20 February 2010 flash-floods in Madeira (Portugal). Daily and hourly precipitation records from the available rain-gauge station networks are evaluated in order to reconstitute the temporal evolution of the rainstorm, as its geographic incidence, contributing to understand the flash-flood dynamics and the type and spatial distribution of the associated impacts. The exceptionality of the rainstorm is further confirmed by the return period associated with the daily precipitation registered at the two long-term record stations, with 146.9 mm observed in the city of Funchal and 333.8 mm on the mountain top, corresponding to an estimated return period of approximately 290 yr and 90 yr, respectively. Furthermore, the synoptic associated situation responsible for the flash-floods is analysed using different sources of information, e.g., weather charts, reanalysis data, Meteosat images and radiosounding data, with the focus on two main issues: (1) the dynamical conditions that promoted such anomalous humidity availability over the Madeira region on 20 February 2010 and (2) the uplift mechanism that induced deep convection activity.
Resumo:
Mobile-to-mobile (M-to-M) communications are expected to play a crucial role in future wireless systems and networks. In this paper, we consider M-to-M multiple-input multiple-output (MIMO) maximal ratio combining system and assess its performance in spatially correlated channels. The analysis assumes double-correlated Rayleigh-and-Lognormal fading channels and is performed in terms of average symbol error probability, outage probability, and ergodic capacity. To obtain the receive and transmit spatial correlation functions needed for the performance analysis, we used a three-dimensional (3D) M-to-M MIMO channel model, which takes into account the effects of fast fading and shadowing. The expressions for the considered metrics are derived as a function of the average signal-to-noise ratio per receive antenna in closed-form and are further approximated using the recursive adaptive Simpson quadrature method. Numerical results are provided to show the effects of system parameters, such as distance between antenna elements, maximum elevation angle of scatterers, orientation angle of antenna array in the x–y plane, angle between the x–y plane and the antenna array orientation, and degree of scattering in the x–y plane, on the system performance. Copyright © 2011 John Wiley & Sons, Ltd.
Resumo:
Body Sensor Networks (BSNs) have been recently introduced for the remote monitoring of human activities in a broad range of application domains, such as health care, emergency management, fitness and behaviour surveillance. BSNs can be deployed in a community of people and can generate large amounts of contextual data that require a scalable approach for storage, processing and analysis. Cloud computing can provide a flexible storage and processing infrastructure to perform both online and offline analysis of data streams generated in BSNs. This paper proposes BodyCloud, a SaaS approach for community BSNs that supports the development and deployment of Cloud-assisted BSN applications. BodyCloud is a multi-tier application-level architecture that integrates a Cloud computing platform and BSN data streams middleware. BodyCloud provides programming abstractions that allow the rapid development of community BSN applications. This work describes the general architecture of the proposed approach and presents a case study for the real-time monitoring and analysis of cardiac data streams of many individuals.