68 resultados para Vertex Folkman Graph
Resumo:
Stemmatology, or the reconstruction of the transmission history of texts, is a field that stands particularly to gain from digital methods. Many scholars already take stemmatic approaches that rely heavily on computational analysis of the collated text (e.g. Robinson and O’Hara 1996; Salemans 2000; Heikkilä 2005; Windram et al. 2008 among many others). Although there is great value in computationally assisted stemmatology, providing as it does a reproducible result and allowing access to the relevant methodological process in related fields such as evolutionary biology, computational stemmatics is not without its critics. The current state-of-the-art effectively forces scholars to choose between a preconceived judgment of the significance of textual differences (the Lachmannian or neo-Lachmannian approach, and the weighted phylogenetic approach) or to make no judgment at all (the unweighted phylogenetic approach). Some basis for judgment of the significance of variation is sorely needed for medieval text criticism in particular. By this, we mean that there is a need for a statistical empirical profile of the text-genealogical significance of the different sorts of variation in different sorts of medieval texts. The rules that apply to copies of Greek and Latin classics may not apply to copies of medieval Dutch story collections; the practices of copying authoritative texts such as the Bible will most likely have been different from the practices of copying the Lives of local saints and other commonly adapted texts. It is nevertheless imperative that we have a consistent, flexible, and analytically tractable model for capturing these phenomena of transmission. In this article, we present a computational model that captures most of the phenomena of text variation, and a method for analysis of one or more stemma hypotheses against the variation model. We apply this method to three ‘artificial traditions’ (i.e. texts copied under laboratory conditions by scholars to study the properties of text variation) and four genuine medieval traditions whose transmission history is known or deduced in varying degrees. Although our findings are necessarily limited by the small number of texts at our disposal, we demonstrate here some of the wide variety of calculations that can be made using our model. Certain of our results call sharply into question the utility of excluding ‘trivial’ variation such as orthographic and spelling changes from stemmatic analysis.
Resumo:
A dedicated reconstruction algorithm to find decay vertices in the ATLAS muon spectrometer is presented. The algorithm searches the region just upstream of or inside the muon spectrometer volume for multi-particle vertices that originate from the decay of particles with long decay paths. The performance of the algorithm is evaluated using both a sample of simulated Higgs boson events, in which the Higgs boson decays to long-lived neutral particles that in turn decay to b final states, and pp collision data at √s = 7 TeV collected with the ATLAS detector at the LHC during 2011.
Resumo:
A search for the direct production of charginos and neutralinos in final states with three electrons or muons and missing transverse momentum is presented. The analysis is based on 4.7 fb(-1) of root s = 7 TeV proton-proton collision data delivered by the Large Hadron Collider and recorded with the ATLAS detector. Observations are consistent with Standard Model expectations in three signal regions that are either depleted or enriched in Z-boson decays. Upper limits at 95% confidence level are set in R-parity conserving phenomenological minimal supersymmetric models and in simplified models, significantly extending previous results.
Resumo:
Many extensions of the Standard Model posit the existence of heavy particles with long lifetimes. In this Letter, results are presented of a search for events containing one or more such particles, which decay at a significant distance from their production point, using a final state containing charged hadrons and an associated muon. This analysis uses a data sample of proton-proton collisions at root s = 7 TeV corresponding to an integrated luminosity of 4.4 fb(-1) collected in 2011 by the ATLAS detector operating at the Large Hadron Collider. Results are interpreted in the context of R-parity violating supersymmetric scenarios. No events in the signal region are observed and limits are set on the production cross section for pair production of supersymmetric particles, multiplied by the square of the branching fraction for a neutralino to decay to charged hadrons and a muon, based on the scenario where both of the produced supersymmetric particles give rise to neutralinos that decay in this way. However, since the search strategy is based on triggering on and reconstructing the decay products of individual long-lived particles, irrespective of the rest of the event, these limits can easily be reinterpreted in scenarios with different numbers of long-lived particles per event. The limits are presented as a function of neutralino lifetime, and for a range of squark and neutralino masses.
Resumo:
This chapter presents fuzzy cognitive maps (FCM) as a vehicle for Web knowledge aggregation, representation, and reasoning. The corresponding Web KnowARR framework incorporates findings from fuzzy logic. To this end, a first emphasis is particularly on the Web KnowARR framework along with a stakeholder management use case to illustrate the framework’s usefulness as a second focal point. This management form is to help projects to acceptance and assertiveness where claims for company decisions are actively involved in the management process. Stakeholder maps visually (re-) present these claims. On one hand, they resort to non-public content and on the other they resort to content that is available to the public (mostly on the Web). The Semantic Web offers opportunities not only to present public content descriptively but also to show relationships. The proposed framework can serve as the basis for the public content of stakeholder maps.
Resumo:
addplot adds twoway plot objects to an existing twoway graph. This is useful if you want to add additional objects such as titles or extra data points to a twoway graph after it has been created. Most of what addplot can do, can also be done by rerunning the original graph command including additional options or plot statements. addplot, however, might be useful if you have to modify a graph for which you cannot rerun the original command, for example, because you only have the graph file but not the data that were used to create the graph. Furthermore, addplot can do certain things that would be difficult to achieve in a single graph command (e.g. customizing individual subgraphs within a by-graph). addplot also provides a substitute for some of the functionality of the graph editor.