962 resultados para Vertex Transitive Graph
Resumo:
Many extensions of the Standard Model posit the existence of heavy particles with long lifetimes. In this Letter, results are presented of a search for events containing one or more such particles, which decay at a significant distance from their production point, using a final state containing charged hadrons and an associated muon. This analysis uses a data sample of proton-proton collisions at root s = 7 TeV corresponding to an integrated luminosity of 4.4 fb(-1) collected in 2011 by the ATLAS detector operating at the Large Hadron Collider. Results are interpreted in the context of R-parity violating supersymmetric scenarios. No events in the signal region are observed and limits are set on the production cross section for pair production of supersymmetric particles, multiplied by the square of the branching fraction for a neutralino to decay to charged hadrons and a muon, based on the scenario where both of the produced supersymmetric particles give rise to neutralinos that decay in this way. However, since the search strategy is based on triggering on and reconstructing the decay products of individual long-lived particles, irrespective of the rest of the event, these limits can easily be reinterpreted in scenarios with different numbers of long-lived particles per event. The limits are presented as a function of neutralino lifetime, and for a range of squark and neutralino masses.
Resumo:
This chapter presents fuzzy cognitive maps (FCM) as a vehicle for Web knowledge aggregation, representation, and reasoning. The corresponding Web KnowARR framework incorporates findings from fuzzy logic. To this end, a first emphasis is particularly on the Web KnowARR framework along with a stakeholder management use case to illustrate the framework’s usefulness as a second focal point. This management form is to help projects to acceptance and assertiveness where claims for company decisions are actively involved in the management process. Stakeholder maps visually (re-) present these claims. On one hand, they resort to non-public content and on the other they resort to content that is available to the public (mostly on the Web). The Semantic Web offers opportunities not only to present public content descriptively but also to show relationships. The proposed framework can serve as the basis for the public content of stakeholder maps.
Resumo:
addplot adds twoway plot objects to an existing twoway graph. This is useful if you want to add additional objects such as titles or extra data points to a twoway graph after it has been created. Most of what addplot can do, can also be done by rerunning the original graph command including additional options or plot statements. addplot, however, might be useful if you have to modify a graph for which you cannot rerun the original command, for example, because you only have the graph file but not the data that were used to create the graph. Furthermore, addplot can do certain things that would be difficult to achieve in a single graph command (e.g. customizing individual subgraphs within a by-graph). addplot also provides a substitute for some of the functionality of the graph editor.
Resumo:
In Stata, graphs are usually generated by one call to the graph command. Sometimes, however, it would be convenient to be able to add objects to a graph after the graph has been created. In this article, I provide a command called addplot that offers such functionality for twoway graphs, capitalizing on an undocumented feature of Stata's graphics system.
Resumo:
This paper addresses the issue of fully automatic segmentation of a hip CT image with the goal to preserve the joint structure for clinical applications in hip disease diagnosis and treatment. For this purpose, we propose a Multi-Atlas Segmentation Constrained Graph (MASCG) method. The MASCG method uses multi-atlas based mesh fusion results to initialize a bone sheetness based multi-label graph cut for an accurate hip CT segmentation which has the inherent advantage of automatic separation of the pelvic region from the bilateral proximal femoral regions. We then introduce a graph cut constrained graph search algorithm to further improve the segmentation accuracy around the bilateral hip joint regions. Taking manual segmentation as the ground truth, we evaluated the present approach on 30 hip CT images (60 hips) with a 15-fold cross validation. When the present approach was compared to manual segmentation, an average surface distance error of 0.30 mm, 0.29 mm, and 0.30 mm was found for the pelvis, the left proximal femur, and the right proximal femur, respectively. A further look at the bilateral hip joint regions demonstrated an average surface distance error of 0.16 mm, 0.21 mm and 0.20 mm for the acetabulum, the left femoral head, and the right femoral head, respectively.
Resumo:
-pshare- computes and graphs percentile shares from individual level data. Percentile shares are often used in inequality research to study the distribution of income or wealth. They are defined as differences between Lorenz ordinates of the outcome variable. Technically, the observations are sorted in increasing order of the outcome variable and the specified percentiles are computed from the running sum of the outcomes. Percentile shares are then computed as differences between percentiles, divided by total outcome. pshare requires moremata to be installed on the system; see ssc describe moremata.
Resumo:
A series of long-chain (C37, C38, C39), primarily di and tri-unsaturated methyl and ethyl ketones, first identified in sediments from Walvis Ridge off West Africa and from Black Sea (de Leeuw et al., 1979), has been found in marine sediments throughout the world (Brassell et al., 1986 doi:10.1038/320129a0). The marine coccolithophorid Emiliania huxleyi and members of the class Prymnesiophyceae are now the recognized sources of these compounds (Volkman et al., 1979; Marlowe, et al., 1984). Experiments with laboratory cultures of algae showed the degree of unsaturation in the ketone seris biosynthesized depends on growth temperature (Brassell et al., 1986; Marlowe, 1984), a physiological respons observed for classical membrane lipids (vanDeenen et al., 1972). Brassell and co-workers (Brassell et al., 198; Brassell et al., 1986b) thus proposed that systematic fluctuations in the unsaturation of these alkenones noted down-core in sediments from the Kane Gap region of the north-east tropical Atlantic Ocean and correlated with glacial-interglacial cycles provide an organic geochemical measure of past sea-surface water temperatures. Using laboratory cultures of E. huxleyi, we have calibrated changes in the unsaturation pattern of the long-chain ketone series versus growth temperature. The calibration curve is linear and accurtely predicts unsuturation patterns observed in natural particulate materials collected from oceanic waters of known temperature. We present evidence supporting the proposed paleotemperature hypothesis (Brassell et al., 1986, Brassel et al., 1986b) and suggesting absolute 'sea-surface temperatures' for a given oceanic location can be estimated from an analysis of long-chain ketone compositions preserved in glacial and interglacial horizons of deep-sea sediment cores.
Resumo:
We propose a weakly supervised method to arrange images of a given category based on the relative pose between the camera and the object in the scene. Relative poses are points on a sphere centered at the object in a given canonical pose, which we call object viewpoints. Our method builds a graph on this sphere by assigning images with similar viewpoint to the same node and by connecting nodes if they are related by a small rotation. The key idea is to exploit a large unlabeled dataset to validate the likelihood of dominant 3D planes of the object geometry. A number of 3D plane hypotheses are evaluated by applying small 3D rotations to each hypothesis and by measuring how well the deformed images match other images in the dataset. Correct hypotheses will result in deformed images that correspond to plausible views of the object, and thus will likely match well other images in the same category. The identified 3D planes are then used to compute affinities between images related by a change of viewpoint. We then use the affinities to build a view graph via a greedy method and the maximum spanning tree.
Resumo:
This paper describes a new exact algorithm PASS for the vertex coloring problem based on the well known DSATUR algorithm. At each step DSATUR maximizes saturation degree to select a new candidate vertex to color, breaking ties by maximum degree w.r.t. uncolored vertices. Later Sewell introduced a new tiebreaking strategy, which evaluated available colors for each vertex explicitly. PASS differs from Sewell in that it restricts its application to a particular set of vertices. Overall performance is improved when the new strategy is applied selectively instead of at every step. The paper also reports systematic experiments over 1500 random graphs and a subset of the DIMACS color benchmark.
Resumo:
A number of thrombectomy devices using a variety of methods have now been developed to facilitate clot removal. We present research involving one such experimental device recently developed in the UK, called a ‘GP’ Thrombus Aspiration Device (GPTAD). This device has the potential to bring about the extraction of a thrombus. Although the device is at a relatively early stage of development, the results look encouraging. In this work, we present an analysis and modeling of the GPTAD by means of the bond graph technique; it seems to be a highly effective method of simulating the device under a variety of conditions. Such modeling is useful in optimizing the GPTAD and predicting the result of clot extraction. The aim of this simulation model is to obtain the minimum pressure necessary to extract the clot and to verify that both the pressure and the time required to complete the clot extraction are realistic for use in clinical situations, and are consistent with any experimentally obtained data. We therefore consider aspects of rheology and mechanics in our modeling.
Resumo:
Tree-reweighted belief propagation is a message passing method that has certain advantages compared to traditional belief propagation (BP). However, it fails to outperform BP in a consistent manner, does not lend itself well to distributed implementation, and has not been applied to distributions with higher-order interactions. We propose a method called uniformly-reweighted belief propagation that mitigates these drawbacks. After having shown in previous works that this method can substantially outperform BP in distributed inference with pairwise interaction models, in this paper we extend it to higher-order interactions and apply it to LDPC decoding, leading performance gains over BP.
Resumo:
We present a novel framework for encoding latency analysis of arbitrary multiview video coding prediction structures. This framework avoids the need to consider an specific encoder architecture for encoding latency analysis by assuming an unlimited processing capacity on the multiview encoder. Under this assumption, only the influence of the prediction structure and the processing times have to be considered, and the encoding latency is solved systematically by means of a graph model. The results obtained with this model are valid for a multiview encoder with sufficient processing capacity and serve as a lower bound otherwise. Furthermore, with the objective of low latency encoder design with low penalty on rate-distortion performance, the graph model allows us to identify the prediction relationships that add higher encoding latency to the encoder. Experimental results for JMVM prediction structures illustrate how low latency prediction structures with a low rate-distortion penalty can be derived in a systematic manner using the new model.
Resumo:
We show a procedure for constructing a probabilistic atlas based on affine moment descriptors. It uses a normalization procedure over the labeled atlas. The proposed linear registration is defined by closed-form expressions involving only geometric moments. This procedure applies both to atlas construction as atlas-based segmentation. We model the likelihood term for each voxel and each label using parametric or nonparametric distributions and the prior term is determined by applying the vote-rule. The probabilistic atlas is built with the variability of our linear registration. We have two segmentation strategy: a) it applies the proposed affine registration to bring the target image into the coordinate frame of the atlas or b) the probabilistic atlas is non-rigidly aligning with the target image, where the probabilistic atlas is previously aligned to the target image with our affine registration. Finally, we adopt a graph cut - Bayesian framework for implementing the atlas-based segmentation.