7 resultados para B-2-10-SF
em Greenwich Academic Literature Archive - UK
Resumo:
Network analysis is distinguished from traditional social science by the dyadic nature of the standard data set. Whereas in traditional social science we study monadic attributes of individuals, in network analysis we study dyadic attributes of pairs of individuals. These dyadic attributes (e.g. social relations) may be represented in matrix form by a square 1-mode matrix. In contrast, the data in traditional social science are represented as 2-mode matrices. However, network analysis is not completely divorced from traditional social science, and often has occasion to collect and analyze 2-mode matrices. Furthermore, some of the methods developed in network analysis have uses in analysing non-network data. This paper presents and discusses ways of applying and interpreting traditional network analytic techniques to 2-mode data, as well as developing new techniques. Three areas are covered in detail: displaying 2-mode data as networks, detecting clusters and measuring centrality.
Resumo:
Procedures are described for solving the equations governing a multi-physics process. Finite volume techniques are used to discretise, using the same unstructured mesh, the equations of fluid flow, heat transfer with solidification, and solid deformation. These discretised equations are then solved in an integrated manner. The computational mechanics environment, PHYSICA, which facilitates the building of multi-physics models, is described. Comparisons between model predictions and experimental data are presented for the casting of metal components.
Resumo:
A common but informal notion in social network analysis and other fields is the concept of a core/periphery structure. The intuitive conception entails a dense, cohesive core and a sparse, unconnected periphery. This paper seeks to formalize the intuitive notion of a core/periphery structure and suggests algorithms for detecting this structure, along with statistical tests for testing a priori hypotheses. Different models are presented for different kinds of graphs (directed and undirected, valued and nonvalued). In addition, the close relation of the continuous models developed to certain centrality measures is discussed.
Resumo:
The paper considers the job shop scheduling problem to minimize the makespan. It is assumed that each job consists of at most two operations, one of which is to be processed on one of m⩾2 machines, while the other operation must be performed on a single bottleneck machine, the same for all jobs. For this strongly NP-hard problem we present two heuristics with improved worst-case performance. One of them guarantees a worst-case performance ratio of 3/2. The other algorithm creates a schedule with the makespan that exceeds the largest machine workload by at most the length of the largest operation.
Resumo:
It is known that for the open shop scheduling problem to minimize the makespan there exists no polynomial-time heuristic algorithm that guarantees a worst-case performance ratio better than 5/4, unless P6≠NP. However, this result holds only if the instance of the problem contains jobs consisting of at least three operations. This paper considers the open shop scheduling problem, provided that each job consists of at most two operations, one of which is to be processed on one of the m⩾2 machines, while the other operation must be performed on the bottleneck machine, the same for all jobs. For this NP-hard problem we present a heuristic algorithm and show that its worst-case performance ratio is 5/4.
Resumo:
Melting of metallic samples in a cold crucible causes inclusions to concentrate on the surface owing to the action of the electromagnetic force in the skin layer. This process is dynamic, involving the melting stage, then quasi-stationary particle separation, and finally the solidification in the cold crucible. The proposed modeling technique is based on the pseudospectral solution method for coupled turbulent fluid flow, thermal and electromagnetic fields within the time varying fluid volume contained by the free surface, and partially the solid crucible wall. The model uses two methods for particle tracking: (1) a direct Lagrangian particle path computation and (2) a drifting concentration model. Lagrangian tracking is implemented for arbitrary unsteady flow. A specific numerical time integration scheme is implemented using implicit advancement that permits relatively large time-steps in the Lagrangian model. The drifting concentration model is based on a local equilibrium drift velocity assumption. Both methods are compared and demonstrated to give qualitatively similar results for stationary flow situations. The particular results presented are obtained for iron alloys. Small size particles of the order of 1 μm are shown to be less prone to separation by electromagnetic field action. In contrast, larger particles, 10 to 100 μm, are easily “trapped” by the electromagnetic field and stay on the sample surface at predetermined locations depending on their size and properties. The model allows optimization for melting power, geometry, and solidification rate.
Resumo:
A communication system model for mutual information performance analysis of multiple-symbol differential M-phase shift keying over time-correlated, time-varying flat-fading communication channels is developed. This model is a finite-state Markov (FSM) equivalent channel representing the cascade of the differential encoder, FSM channel model and differential decoder. A state-space approach is used to model channel phase time correlations. The equivalent model falls in a class that facilitates the use of the forward backward algorithm, enabling the important information theoretic results to be evaluated. Using such a model, one is able to calculate mutual information for differential detection over time-varying fading channels with an essentially finite time set of correlations, including the Clarke fading channel. Using the equivalent channel, it is proved and corroborated by simulations that multiple-symbol differential detection preserves the channel information capacity when the observation interval approaches infinity.