749 resultados para Reed
Resumo:
We present measurements of the charge balance function, from the charged particles, for diverse pseudorapidity and transverse momentum ranges in Au + Au collisions at root S-NN = 200 GeV using the STAR detector at RHIC. We observe that the balance function is boost-invariant within the pseudorapidity coverage vertical bar-1.3, 1.3 vertical bar. The balance function properly scaled by the width of the observed pseudorapidity window does not depend on the position or size of the pseudorapidity window. This scaling property also holds for particles in different transverse momentum ranges. In addition, we find that the width of the balance function decreases monotonically with increasing transverse momentum for all centrality classes. (c) 2010 Elsevier B.V. All rights reserved.
Resumo:
We present the multiplicity and pseudorapidity distributions of photons produced in Au + Au and Cu + Cu collisions at root(NN)-N-s = 62.4 and 200 GeV. The photons are measured in the region -3.7 < eta < -2.3 using the photon Multiplicity detector in the STAR experiment at RHIC. The number of photons produced per average number of participating nucleon pairs increases with the beam energy and is independent of (lie collision centrality. For collisions with similar average numbers of participating nucleons the photon multiplicities are observed to be similar for An + Au and Cu + Cu collisions at a given beam energy. The ratios of the number of charged particles to photons in the measured pseudorapidity range are found to be 1.4 +/- 0.1 and 1.2 +/- 0.1 for root(NN)-N-s = 62.4 and 200 GeV, respectively. The energy dependence of this ratio could reflect varying contributions from baryons to charged particles, while mesons are the dominant contributors to photon production in the given kinematic region. The photon pseudorapidity distributions normalized by average number of participating nucleon pairs, when plotted as a function of eta-Y-beam, are found to follow a longitudinal scaling independent of centrality and colliding ion species at both beam energies. (C) 2009 Elsevier B.V. All rights reserved.
Resumo:
The results of the examination showed that some wetland plants' leaves and stems above the surface of water have little ability to supply water body with oxygen through roots of themselves while they are photosynthesizing. These plants are calamus(Acorus calamus), cattail(Typha angustifolia), wild rice stem(Zizania caduciflora), Cyoerus alternifokius, and water hyacinth(Eichhornia crassipes). It means that there is no relationship between these plants' photosynthesis and the breath of root cells. But duckweed(Lemna minor) has a small to raise DO 0.44mg·L -1 in average, while it is photosynthesizing during the examination. Reed(Phragmitas communis) may have a little the to provide oxygen for water body through root of itself while it is photosynthesizing. It raised DO 0.30mg·L -1 in average during the examination.
Resumo:
The nutrient cycling in the reed field has been studied via field investigation and laboratory analysis. The results indicate that N absorption in higher reed productvity field is 0.6044% (high level); The P absorption is 0.1100% (intermediate); The K absorption is 0.0153% (low level).the cycling coefficiencies for N,P and K are 0.16、 0.15 and 0.12,respectively. They are all at low levels.
Resumo:
This research is concerned with the development of tactual displays to supplement the information available through lipreading. Because voicing carries a high informational load in speech and is not well transmitted through lipreading, the efforts are focused on providing tactual displays of voicing to supplement the information available on the lips of the talker. This research includes exploration of 1) signal-processing schemes to extract information about voicing from the acoustic speech signal, 2) methods of displaying this information through a multi-finger tactual display, and 3) perceptual evaluations of voicing reception through the tactual display alone (T), lipreading alone (L), and the combined condition (L+T). Signal processing for the extraction of voicing information used amplitude-envelope signals derived from filtered bands of speech (i.e., envelopes derived from a lowpass-filtered band at 350 Hz and from a highpass-filtered band at 3000 Hz). Acoustic measurements made on the envelope signals of a set of 16 initial consonants represented through multiple tokens of C1VC2 syllables indicate that the onset-timing difference between the low- and high-frequency envelopes (EOA: envelope-onset asynchrony) provides a reliable and robust cue for distinguishing voiced from voiceless consonants. This acoustic cue was presented through a two-finger tactual display such that the envelope of the high-frequency band was used to modulate a 250-Hz carrier signal delivered to the index finger (250-I) and the envelope of the low-frequency band was used to modulate a 50-Hz carrier delivered to the thumb (50T). The temporal-onset order threshold for these two signals, measured with roving signal amplitude and duration, averaged 34 msec, sufficiently small for use of the EOA cue. Perceptual evaluations of the tactual display of EOA with speech signal indicated: 1) that the cue was highly effective for discrimination of pairs of voicing contrasts; 2) that the identification of 16 consonants was improved by roughly 15 percentage points with the addition of the tactual cue over L alone; and 3) that no improvements in L+T over L were observed for reception of words in sentences, indicating the need for further training on this task
Resumo:
O objetivo deste trabalho foi desenvolver e testar um protocolo de PCR específico para a amplificação de espécies de bactérias fixadoras de nitrogênio do gênero Paenibacillus, utilizando-se a estratégia de clonagem e sequenciamento do gene nifH. .
Resumo:
Mavron, Vassili; McDonough, T.P.; Key, J.D., (2006) 'Information sets and partial permutation decoding for codes from finite geometries', Finite Fields and their applications 12(2) pp.232-247 RAE2008
Resumo:
Ioan Fazey, John A. Fazey, Joern Fischer, Kate Sherren, John Warren, Reed F. Noss, Stephen R. Dovers (2007) Adaptive capacity and learning to learn as leverage for social?ecological resilience. Frontiers in Ecology and the Environment 5(7),375-380. RAE2008
Resumo:
http://moa.umdl.umich.edu/cgi/sgml/moa-idx?notisid=AAU8319 View book via University of Michigan
Resumo:
This collection primarily contains correspondence from Wright’s years as president of ASOR. Material dates as far back as 1957, and proceed into the early 1970’s. Some of Wright’s more notable correspondents include William F. Albright, A. Henry Detweiler, Paul W. Lapp, William Reed, and Dean Seiler. Subject-specific correspondence includes records of expenditures, budget planning, corporate memberships, and the Jerusalem School.
Resumo:
Communities of faith have appeared online since the inception of computer - mediated communication (CMC)and are now ubiquitous. Yet the character and legitimacy of Internet communities as ecclesial bodies is often disputed by traditional churches; and the Internet's ability to host the church as church for online Christians remains a question. This dissertation carries out a practical theological conversation between three main sources: the phenomenon of the church online; ecclesiology (especially that characteristic of Reformed communities); and communication theory. After establishing the need for this study in Chapter 1, Chapter 2 investigates the online presence of Christians and trends in their Internet use, including its history and current expressions. Chapter 3 sets out an historical overview of the Reformed Tradition, focusing on the work of John Calvin and Karl Barth, as well as more contemporary theologians. With a theological context in which to consider online churches in place, Chapter 4 introduces four theological themes prominent in both ecclesiology and CMC studies: authority; community; mediation; and embodiment. These themes constitute the primary lens through which the dissertation conducts a critical-confessional interface between communication theory and ecclesiology in the examination of CMC. Chapter 5 continues the contextualization of online churches with consideration of communication theories that impact CMC, focusing on three major communication theories: Narrative Theory; Interpretive Theory; and Speech Act Theory. Chapter 6 contains the critical conversation between ecclesiology and communication theory by correlating the aforementioned communication theories with Narrative Theology, Communities of Practice, and Theo-Drama, and applying these to the four theological themes noted above. In addition, new or anticipated developments in CMC investigated in relationship to traditional ecclesiologies and the prospect of cyber-ecclesiology. Chapter 7 offers an evaluative tool consisting of a three-step hermeneutical process that examines: 1) the history, tradition, and ecclesiology of the particular community being evaluated; 2) communication theories and the process of religious-social shaping of technology; and 3) CMC criteria for establishing the presence of a stable, interactive, and relational community. As this hermeneutical process unfolds, it holds the church at the center of the process, seeking a contextual yet faithful understanding of the church.
Resumo:
An increasing number of applications, such as distributed interactive simulation, live auctions, distributed games and collaborative systems, require the network to provide a reliable multicast service. This service enables one sender to reliably transmit data to multiple receivers. Reliability is traditionally achieved by having receivers send negative acknowledgments (NACKs) to request from the sender the retransmission of lost (or missing) data packets. However, this Automatic Repeat reQuest (ARQ) approach results in the well-known NACK implosion problem at the sender. Many reliable multicast protocols have been recently proposed to reduce NACK implosion. But, the message overhead due to NACK requests remains significant. Another approach, based on Forward Error Correction (FEC), requires the sender to encode additional redundant information so that a receiver can independently recover from losses. However, due to the lack of feedback from receivers, it is impossible for the sender to determine how much redundancy is needed. In this paper, we propose a new reliable multicast protocol, called ARM for Adaptive Reliable Multicast. Our protocol integrates ARQ and FEC techniques. The objectives of ARM are (1) reduce the message overhead due to NACK requests, (2) reduce the amount of data transmission, and (3) reduce the time it takes for all receivers to receive the data intact (without loss). During data transmission, the sender periodically informs the receivers of the number of packets that are yet to be transmitted. Based on this information, each receiver predicts whether this amount is enough to recover its losses. Only if it is not enough, that the receiver requests the sender to encode additional redundant packets. Using ns simulations, we show the superiority of our hybrid ARQ-FEC protocol over the well-known Scalable Reliable Multicast (SRM) protocol.
Resumo:
SomeCast is a novel paradigm for the reliable multicast of real-time data to a large set of receivers over the Internet. SomeCast is receiver-initiated and thus scalable in the number of receivers, the diverse characteristics of paths between senders and receivers (e.g. maximum bandwidth and round-trip-time), and the dynamic conditions of such paths (e.g. congestion-induced delays and losses). SomeCast enables receivers to dynamically adjust the rate at which they receive multicast information to enable the satisfaction of real-time QoS constraints (e.g. rate, deadlines, or jitter). This is done by enabling a receiver to join SOME number of concurrent multiCAST sessions, whereby each session delivers a portion of an encoding of the real-time data. By adjusting the number of such sessions dynamically, client-specific QoS constraints can be met independently. The SomeCast paradigm can be thought of as a generalization of the AnyCast (e.g. Dynamic Server Selection) and ManyCast (e.g. Digital Fountain) paradigms, which have been proposed in the literature to address issues of scalability of UniCast and MultiCast environments, respectively. In this paper we overview the SomeCast paradigm, describe an instance of a SomeCast protocol, and present simulation results that quantify the significant advantages gained from adopting such a protocol for the reliable multicast of data to a diverse set of receivers subject to real-time QoS constraints.
Resumo:
There is much common ground between the areas of coding theory and systems theory. Fitzpatrick has shown that a Göbner basis approach leads to efficient algorithms in the decoding of Reed-Solomon codes and in scalar interpolation and partial realization. This thesis simultaneously generalizes and simplifies that approach and presents applications to discrete-time modeling, multivariable interpolation and list decoding. Gröbner basis theory has come into its own in the context of software and algorithm development. By generalizing the concept of polynomial degree, term orders are provided for multivariable polynomial rings and free modules over polynomial rings. The orders are not, in general, unique and this adds, in no small way, to the power and flexibility of the technique. As well as being generating sets for ideals or modules, Gröbner bases always contain a element which is minimal with respect tot the corresponding term order. Central to this thesis is a general algorithm, valid for any term order, that produces a Gröbner basis for the solution module (or ideal) of elements satisfying a sequence of generalized congruences. These congruences, based on shifts and homomorphisms, are applicable to a wide variety of problems, including key equations and interpolations. At the core of the algorithm is an incremental step. Iterating this step lends a recursive/iterative character to the algorithm. As a consequence, not all of the input to the algorithm need be available from the start and different "paths" can be taken to reach the final solution. The existence of a suitable chain of modules satisfying the criteria of the incremental step is a prerequisite for applying the algorithm.
Resumo:
Error correcting codes are combinatorial objects, designed to enable reliable transmission of digital data over noisy channels. They are ubiquitously used in communication, data storage etc. Error correction allows reconstruction of the original data from received word. The classical decoding algorithms are constrained to output just one codeword. However, in the late 50’s researchers proposed a relaxed error correction model for potentially large error rates known as list decoding. The research presented in this thesis focuses on reducing the computational effort and enhancing the efficiency of decoding algorithms for several codes from algorithmic as well as architectural standpoint. The codes in consideration are linear block codes closely related to Reed Solomon (RS) codes. A high speed low complexity algorithm and architecture are presented for encoding and decoding RS codes based on evaluation. The implementation results show that the hardware resources and the total execution time are significantly reduced as compared to the classical decoder. The evaluation based encoding and decoding schemes are modified and extended for shortened RS codes and software implementation shows substantial reduction in memory footprint at the expense of latency. Hermitian codes can be seen as concatenated RS codes and are much longer than RS codes over the same aphabet. A fast, novel and efficient VLSI architecture for Hermitian codes is proposed based on interpolation decoding. The proposed architecture is proven to have better than Kötter’s decoder for high rate codes. The thesis work also explores a method of constructing optimal codes by computing the subfield subcodes of Generalized Toric (GT) codes that is a natural extension of RS codes over several dimensions. The polynomial generators or evaluation polynomials for subfield-subcodes of GT codes are identified based on which dimension and bound for the minimum distance are computed. The algebraic structure for the polynomials evaluating to subfield is used to simplify the list decoding algorithm for BCH codes. Finally, an efficient and novel approach is proposed for exploiting powerful codes having complex decoding but simple encoding scheme (comparable to RS codes) for multihop wireless sensor network (WSN) applications.