921 resultados para topological complexity
Resumo:
Immigrants from the West Indies and other nations challenge the simple United States dichotomy of blacks versus whites. Many apparently black Caribbean immigrants proclaim that they did not know they were “black” until they arrived in the U.S. They seek to maintain their national identity and resist identity and solidarity with Black Americans. In response, many Black Americans respond that the immigrants are simply being naive, that U.S. society demands simple racial identity. Regardless of one's self-identity and personal history, in the U.S., if you look black, you are black, was their thinking. ^ This study examines the contemporary struggle of identity and solidarity among and between Black Americans and Jamaicans living in South Florida (Broward and Miami-Dade counties). Even though the primary focus of this study is to examine the relationship between Black Americans and Jamaicans, other West Indian nationals will be addressed more generally. The primary research problem of this study is to determine why the existence of common ancestry and physical traits are insufficient for an assumption of ethnic solidarity between Black Americans and Jamaicans. ^ In examining this problem, I felt that depth rather than breadth would provide insight into the current state of polarization between Black Americans and Jamaicans. To this end, a qualitative study was designed. A non-random snowball sample consisting of forty-seven informants was selected for this study. Realizing that such a technique presents problems with generalizations beyond the sample, this approach was, nonetheless, the most suitable for the current research problem. One of the initial challenges of this research was the use of the label “black” in discussing Caribbean immigrants. Unlike America, where distinctions based on skin color were at the bedrock of America's formation, this was not the case in the Caribbean. In the Caribbean skin color was an important marker as an indicator of class, rather than of race. Therefore, I refrained from using the label, “black Jamaicans,” but rather used Jamaicans throughout. ^
Resumo:
The Unified Modeling Language (UML) has quickly become the industry standard for object-oriented software development. It is being widely used in organizations and institutions around the world. However, UML is often found to be too complex for novice systems analysts. Although prior research has identified difficulties novice analysts encounter in learning UML, no viable solution has been proposed to address these difficulties. Sequence-diagram modeling, in particular, has largely been overlooked. The sequence diagram models the behavioral aspects of an object-oriented software system in terms of interactions among its building blocks, i.e. objects and classes. It is one of the most commonly-used UML diagrams in practice. However, there has been little research on sequence-diagram modeling. The current literature scarcely provides effective guidelines for developing a sequence diagram. Such guidelines will be greatly beneficial to novice analysts who, unlike experienced systems analysts, do not possess relevant prior experience to easily learn how to develop a sequence diagram. There is the need for an effective sequence-diagram modeling technique for novices. This dissertation reports a research study that identified novice difficulties in modeling a sequence diagram and proposed a technique called CHOP (CHunking, Ordering, Patterning), which was designed to reduce the cognitive load by addressing the cognitive complexity of sequence-diagram modeling. The CHOP technique was evaluated in a controlled experiment against a technique recommended in a well-known textbook, which was found to be representative of approaches provided in many textbooks as well as practitioner literatures. The results indicated that novice analysts were able to perform better using the CHOP technique. This outcome seems have been enabled by pattern-based heuristics provided by the technique. Meanwhile, novice analysts rated the CHOP technique more useful although not significantly easier to use than the control technique. The study established that the CHOP technique is an effective sequence-diagram modeling technique for novice analysts.
Resumo:
While most studies take a dyadic view when examining the environmental difference between the home country of a multinational enterprise (MNE) and a particular foreign country, they ignore that an MNE is managing a network of subsidiaries embedded in diverse environments. Additionally, neither the impacts of global environments on top executives nor the effects of top executives’ capabilities to handle institutional complexity are fully explored. Thus, using a three-essay format, this dissertation tried to fill these gaps by addressing the effects of institutional complexity and top management characteristics on top executive compensation and firm performance. ^ Essay 1 investigated the impact of an MNE’s institutional complexity, or the diversity of national institutions facing an MNE’s network of subsidiaries, on the top management team (TMT) compensation. This essay proposed that greater political and cultural complexity leads to not only greater TMT total compensation but also to a greater portion of TMT compensation linked with long-term performance. The arguments are supported in this essay by using an unbalanced panel dataset including 296 U.S. firms with 1,340 observations. ^ Essay 2 explored TMT social capital and its moderating role on value creation and appropriation by the chief executive officer (CEO). Using a sample with 548 U.S. firms and 2,010 observations, it found that greater TMT social capital does facilitate the effects of CEO intellectual capital and social capital on firm growth. Finally, essay 3 examined the performance implications for the fit between managerial information-processing capabilities and institutional complexity. It proposed that institutional complexity is associated with the needs of information-processing. On the other hand, smaller TMT turnover and larger TMT size reflect larger managerial information-processing capabilities. Consequently, superior performance is achieved by the match among institutional complexity, TMT turnover, and TMT size. All hypotheses in essay 3 are supported in a sample of 301 U.S. firms and 1,404 observations. ^ To conclude, this dissertation advances and extends our knowledge on the roles of institutional environments and top executives on firm performance and top executive compensation.^
Resumo:
ackground Following incomplete spinal cord injury (iSCI), descending drive is impaired, possibly leading to a decrease in the complexity of gait. To test the hypothesis that iSCI impairs gait coordination and decreases locomotor complexity, we collected 3D joint angle kinematics and muscle parameters of rats with a sham or an incomplete spinal cord injury. Methods 12 adult, female, Long-Evans rats, 6 sham and 6 mild-moderate T8 iSCI, were tested 4 weeks following injury. The Basso Beattie Bresnahan locomotor score was used to verify injury severity. Animals had reflective markers placed on the bony prominences of their limb joints and were filmed in 3D while walking on a treadmill. Joint angles and segment motion were analyzed quantitatively, and complexity of joint angle trajectory and overall gait were calculated using permutation entropy and principal component analysis, respectively. Following treadmill testing, the animals were euthanized and hindlimb muscles removed. Excised muscles were tested for mass, density, fiber length, pennation angle, and relaxed sarcomere length. Results Muscle parameters were similar between groups with no evidence of muscle atrophy. The animals showed overextension of the ankle, which was compensated for by a decreased range of motion at the knee. Left-right coordination was altered, leading to left and right knee movements that are entirely out of phase, with one joint moving while the other is stationary. Movement patterns remained symmetric. Permutation entropy measures indicated changes in complexity on a joint specific basis, with the largest changes at the ankle. No significant difference was seen using principal component analysis. Rats were able to achieve stable weight bearing locomotion at reasonable speeds on the treadmill despite these deficiencies. Conclusions Decrease in supraspinal control following iSCI causes a loss of complexity of ankle kinematics. This loss can be entirely due to loss of supraspinal control in the absence of muscle atrophy and may be quantified using permutation entropy. Joint-specific differences in kinematic complexity may be attributed to different sources of motor control. This work indicates the importance of the ankle for rehabilitation interventions following spinal cord injury.
Resumo:
The Unified Modeling Language (UML) has quickly become the industry standard for object-oriented software development. It is being widely used in organizations and institutions around the world. However, UML is often found to be too complex for novice systems analysts. Although prior research has identified difficulties novice analysts encounter in learning UML, no viable solution has been proposed to address these difficulties. Sequence-diagram modeling, in particular, has largely been overlooked. The sequence diagram models the behavioral aspects of an object-oriented software system in terms of interactions among its building blocks, i.e. objects and classes. It is one of the most commonly-used UML diagrams in practice. However, there has been little research on sequence-diagram modeling. The current literature scarcely provides effective guidelines for developing a sequence diagram. Such guidelines will be greatly beneficial to novice analysts who, unlike experienced systems analysts, do not possess relevant prior experience to easily learn how to develop a sequence diagram. There is the need for an effective sequence-diagram modeling technique for novices. This dissertation reports a research study that identified novice difficulties in modeling a sequence diagram and proposed a technique called CHOP (CHunking, Ordering, Patterning), which was designed to reduce the cognitive load by addressing the cognitive complexity of sequence-diagram modeling. The CHOP technique was evaluated in a controlled experiment against a technique recommended in a well-known textbook, which was found to be representative of approaches provided in many textbooks as well as practitioner literatures. The results indicated that novice analysts were able to perform better using the CHOP technique. This outcome seems have been enabled by pattern-based heuristics provided by the technique. Meanwhile, novice analysts rated the CHOP technique more useful although not significantly easier to use than the control technique. The study established that the CHOP technique is an effective sequence-diagram modeling technique for novice analysts.
Resumo:
We study the helical edge states of a two-dimensional topological insulator without axial spin symmetry due to the Rashba spin-orbit interaction. Lack of axial spin symmetry can lead to so-called generic helical edge states, which have energy-dependent spin orientation. This opens the possibility of inelastic backscattering and thereby nonquantized transport. Here we find analytically the new dispersion relations and the energy dependent spin orientation of the generic helical edge states in the presence of Rashba spin-orbit coupling within the Bernevig-Hughes-Zhang model, for both a single isolated edge and for a finite width ribbon. In the single-edge case, we analytically quantify the energy dependence of the spin orientation, which turns out to be weak for a realistic HgTe quantum well. Nevertheless, finite size effects combined with Rashba spin-orbit coupling result in two avoided crossings in the energy dispersions, where the spin orientation variation of the edge states is very significantly increased for realistic parameters. Finally, our analytical results are found to compare well to a numerical tight-binding regularization of the model.
Resumo:
Topological quantum error correction codes are currently among the most promising candidates for efficiently dealing with the decoherence effects inherently present in quantum devices. Numerically, their theoretical error threshold can be calculated by mapping the underlying quantum problem to a related classical statistical-mechanical spin system with quenched disorder. Here, we present results for the general fault-tolerant regime, where we consider both qubit and measurement errors. However, unlike in previous studies, here we vary the strength of the different error sources independently. Our results highlight peculiar differences between toric and color codes. This study complements previous results published in New J. Phys. 13, 083006 (2011).
Resumo:
Acknowledgements We acknowledge, with thanks the contributions, of the following people who co-designed Boot Camp: Angus JM Watson (Highland Surgical Research Unit, NHSH & UoS), Morag E Hogg (NHSH Raigmore Hospital) and Ailsa Armstrong (NHSH). We also thank Angus JM Watson and Morag E Hogg for helping with the preparation of the funding application which supported this work. Funding Our thanks to the Clinical Skills Managed Educational Network (CSMEN) of Scotland for funding this research.
Resumo:
10 pages, 5 figures, conference or other essential info Acknowledgments LK and JCS were supported by Blue Brain Project. P.D. and R.L. were supported in part by the Blue Brain Project and by the start-up grant of KH. Partial support for P.D. has been provided by the Advanced Grant of the European Research Council GUDHI (Geometric Understanding in Higher Dimensions). MS was supported by the SNF NCCR ”Synapsy”.
Resumo:
The presence of high phase noise in addition to additive white Gaussian noise in coherent optical systems affects the performance of forward error correction (FEC) schemes. In this paper, we propose a simple scheme for such systems, using block interleavers and binary Bose–Chaudhuri–Hocquenghem (BCH) codes. The block interleavers are specifically optimized for differential quadrature phase shift keying modulation. We propose a method for selecting BCH codes that, together with the interleavers, achieve a target post-FEC bit error rate (BER). This combination of interleavers and BCH codes has very low implementation complexity. In addition, our approach is straightforward, requiring only short pre-FEC simulations to parameterize a model, based on which we select codes analytically. We aim to correct a pre-FEC BER of around (Formula presented.). We evaluate the accuracy of our approach using numerical simulations. For a target post-FEC BER of (Formula presented.), codes selected using our method result in BERs around 3(Formula presented.) target and achieve the target with around 0.2 dB extra signal-to-noise ratio.
Resumo:
We experimentally demonstrate 7-dB reduction of nonlinearity penalty in 40-Gb/s CO-OFDM at 2000-km using support vector machine regression-based equalization. Simulation in WDM-CO-OFDM shows up to 12-dB enhancement in Q-factor compared to linear equalization.
Investigating optical complexity of the phase transition in the intensity of a fibre laser radiation
Resumo:
Fibre lasers have been shown to manifest a laminar-to-turbulent transition when increasing its pump power. In order to study the dynamical complexity of this transition we use advanced statistical tools of time-series analysis. We apply ordinal analysis and the horizontal visibility graph to the experimentally measured laser output intensity. This reveal the presence of temporal correlations during the transition from the laminar to the turbulent lasing regimes. Both methods allow us to unveil coherent structures with well defined time-scales and strong correlations both, in the timing of the laser pulses and in their peak intensities.
Resumo:
Pre-treatment HCV quasispecies complexity and diversity may predict response to interferon based anti-viral therapy. The objective of this study was to retrospectively (1) examine temporal changes in quasispecies prior to the start of therapy and (2) investigate extensively quasispecies evolution in a group of 10 chronically infected patients with genotype 3a, treated with pegylated alpha 2a-Interferon and ribavirin. The degree of sequence heterogeneity within the hypervariable region 1 was assessed by analyzing 20-30 individual clones in serial serum samples. Genetic parameters, including amino acid Shannon entropy, Hamming distance and genetic distance were calculated for each sample. Treatment outcome was divided into (1) sustained virological responders (SVR) and (2) treatment failure (TF).Our results indicate, (1) quasispecies complexity and diversity are lower in the SVR group, (2) quasispecies vary temporally and (3) genetic heterogeneity at baseline can be used to predict treatment outcome. We discuss the results from the perspective of replicative homeostasis. We discuss the results from the perspective of replicative homeostasis.
Resumo:
Highlights of Data Expedition: • Students explored daily observations of local climate data spanning the past 35 years. • Topological Data Analysis, or TDA for short, provides cutting-edge tools for studying the geometry of data in arbitrarily high dimensions. • Using TDA tools, students discovered intrinsic dynamical features of the data and learned how to quantify periodic phenomenon in a time-series. • Since nature invariably produces noisy data which rarely has exact periodicity, students also considered the theoretical basis of almost-periodicity and even invented and tested new mathematical definitions of almost-periodic functions. Summary The dataset we used for this data expedition comes from the Global Historical Climatology Network. “GHCN (Global Historical Climatology Network)-Daily is an integrated database of daily climate summaries from land surface stations across the globe.” Source: https://www.ncdc.noaa.gov/oa/climate/ghcn-daily/ We focused on the daily maximum and minimum temperatures from January 1, 1980 to April 1, 2015 collected from RDU International Airport. Through a guided series of exercises designed to be performed in Matlab, students explore these time-series, initially by direct visualization and basic statistical techniques. Then students are guided through a special sliding-window construction which transforms a time-series into a high-dimensional geometric curve. These high-dimensional curves can be visualized by projecting down to lower dimensions as in the figure below (Figure 1), however, our focus here was to use persistent homology to directly study the high-dimensional embedding. The shape of these curves has meaningful information but how one describes the “shape” of data depends on which scale the data is being considered. However, choosing the appropriate scale is rarely an obvious choice. Persistent homology overcomes this obstacle by allowing us to quantitatively study geometric features of the data across multiple-scales. Through this data expedition, students are introduced to numerically computing persistent homology using the rips collapse algorithm and interpreting the results. In the specific context of sliding-window constructions, 1-dimensional persistent homology can reveal the nature of periodic structure in the original data. I created a special technique to study how these high-dimensional sliding-window curves form loops in order to quantify the periodicity. Students are guided through this construction and learn how to visualize and interpret this information. Climate data is extremely complex (as anyone who has suffered from a bad weather prediction can attest) and numerous variables play a role in determining our daily weather and temperatures. This complexity coupled with imperfections of measuring devices results in very noisy data. This causes the annual seasonal periodicity to be far from exact. To this end, I have students explore existing theoretical notions of almost-periodicity and test it on the data. They find that some existing definitions are also inadequate in this context. Hence I challenged them to invent new mathematics by proposing and testing their own definition. These students rose to the challenge and suggested a number of creative definitions. While autocorrelation and spectral methods based on Fourier analysis are often used to explore periodicity, the construction here provides an alternative paradigm to quantify periodic structure in almost-periodic signals using tools from topological data analysis.
Resumo:
This research has explored the relationship between system test complexity and tacit knowledge. It is proposed as part of this thesis, that the process of system testing (comprising of test planning, test development, test execution, test fault analysis, test measurement, and case management), is directly affected by both complexity associated with the system under test, and also by other sources of complexity, independent of the system under test, but related to the wider process of system testing. While a certain amount of knowledge related to the system under test is inherent, tacit in nature, and therefore difficult to make explicit, it has been found that a significant amount of knowledge relating to these other sources of complexity, can indeed be made explicit. While the importance of explicit knowledge has been reinforced by this research, there has been a lack of evidence to suggest that the availability of tacit knowledge to a test team is of any less importance to the process of system testing, when operating in a traditional software development environment. The sentiment was commonly expressed by participants, that even though a considerable amount of explicit knowledge relating to the system is freely available, that a good deal of knowledge relating to the system under test, which is demanded for effective system testing, is actually tacit in nature (approximately 60% of participants operating in a traditional development environment, and 60% of participants operating in an agile development environment, expressed similar sentiments). To cater for the availability of tacit knowledge relating to the system under test, and indeed, both explicit and tacit knowledge required by system testing in general, an appropriate knowledge management structure needs to be in place. This would appear to be required, irrespective of the employed development methodology.