986 resultados para Binary codes


Relevância:

20.00% 20.00%

Publicador:

Resumo:

Aims: X-ray emission is an important diagnostics to study magnetic activity in very low mass stars that are presumably fully convective and have an effectively neutral photosphere. Methods: We investigate an XMM-Newton observation of SCR 1845-6357, a nearby, ultracool M 8.5 / T 5.5 dwarf binary. The binary is unresolved in the XMM detectors, but the X-ray emission is very likely from the M 8.5 dwarf. We compare its flaring emission to those of similar very low mass stars and additionally present an XMM observation of the M 8 dwarf VB 10. Results: We detect quasi-quiescent X-ray emission from SCR 1845-6357 at soft X-ray energies in the 0.2-2.0 keV band, as well as a strong flare with a count rate increase of a factor of 30 and a duration of only 10 min. The quasi-quiescent X-ray luminosity of log LX = 26.2 erg/s and the corresponding activity level of log LX/Lbol = -3.8 point to a fairly active star. Coronal temperatures of up to 5 MK and frequent minor variability support this picture. During the flare, which is accompanied by a significant brightening in the near-UV, plasma temperatures of 25-30 MK are observed and an X-ray luminosity of LX = 8 × 1027 erg/s is reached. Conclusions: The source SCR 1845-6357 is a nearby, very low mass star that emits X-rays at detectable levels in quasi-quiescence, implying the existence of a corona. The high activity level, coronal temperatures and the observed large flare point to a rather active star, despite its estimated age of a few Gyr.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A comparison of collision strengths and effective collision strengths has been undertaken for the Cr II ion based on the model of Wasson et al [2010 A & A. 524 A35]. Calculations have been completed using the Breit-Pauli, RMATRX II and DARC suites of codes.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A complete review of the published data on the mixing enthalpies of mixtures containing ionic liquids, measured directly using calorimetric techniques, is presented in this paper. The field of ionic liquids is very active and a number of research groups in the world are dealing with different applications of these fluids in the fields of chemistry, chemical engineering, energy, gas storage and separation or materials science. In all these fields, the knowledge of the energetics of mixing is capital both to understand the interactions between these fluids and the different substrates and also to establish the energy and environmental cost of possible applications. Due to the relative novelty of the field, the published data is sometimes controversial and recent reviews are fragmentary and do not represent a set of reliable data. This fact can be attributed to different reasons: (i) difficulties in controlling the purity and stability of the ionic liquid samples; (ii) availability of accurate experimental techniques, appropriate for the measurement of viscous, charged, complex fluids; and (iii) choice of an appropriate clear thermodynamic formalism to be used by an interdisciplinary scientific community. In this paper, we address all these points and propose a critical review of the published data, advise on the most appropriate apparatus and experimental procedure to measure this type of physical-chemical data in ionic liquids as well as the way to treat the information obtained by an appropriate thermodynamic formalism.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

O tema principal desta tese é o problema de cancelamento de interferência para sistemas multi-utilizador, com antenas distribuídas. Como tal, ao iniciar, uma visão geral das principais propriedades de um sistema de antenas distribuídas é apresentada. Esta descrição inclui o estudo analítico do impacto da ligação, dos utilizadores do sistema, a mais antenas distribuídas. Durante essa análise é demonstrado que a propriedade mais importante do sistema para obtenção do ganho máximo, através da ligação de mais antenas de transmissão, é a simetria espacial e que os utilizadores nas fronteiras das células são os mais bene ciados. Tais resultados são comprovados através de simulação. O problema de cancelamento de interferência multi-utilizador é considerado tanto para o caso unidimensional (i.e. sem codi cação) como para o multidimensional (i.e. com codi cação). Para o caso unidimensional um algoritmo de pré-codi cação não-linear é proposto e avaliado, tendo como objectivo a minimização da taxa de erro de bit. Tanto o caso de portadora única como o de multipla-portadora são abordados, bem como o cenário de antenas colocadas e distribuidas. É demonstrado que o esquema proposto pode ser visto como uma extensão do bem conhecido esquema de zeros forçados, cuja desempenho é provado ser um limite inferior para o esquema generalizado. O algoritmo é avaliado, para diferentes cenários, através de simulação, a qual indica desempenho perto do óptimo, com baixa complexidade. Para o caso multi-dimensional um esquema para efectuar "dirty paper coding" binário, tendo como base códigos de dupla camada é proposto. No desenvolvimento deste esquema, a compressão com perdas de informação, é considerada como um subproblema. Resultados de simulação indicam transmissão dedigna proxima do limite de Shannon.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this thesis we consider two-dimensional (2D) convolutional codes. As happens in the one-dimensional (1D) case one of the major issues is obtaining minimal state-space realizations for these codes. It turns out that the problem of minimal realization of codes is not equivalent to the minimal realization of encoders. This is due to the fact that the same code may admit different encoders with different McMillan degrees. Here we focus on the study of minimality of the realizations of 2D convolutional codes by means of separable Roesser models. Such models can be regarded as a series connection between two 1D systems. As a first step we provide an algorithm to obtain a minimal realization of a 1D convolutional code starting from a minimal realization of an encoder of the code. Then, we restrict our study to two particular classes of 2D convolutional codes. The first class to be considered is the one of codes which admit encoders of type n 1. For these codes, minimal encoders (i.e., encoders for which a minimal realization is also minimal as a code realization) are characterized enabling the construction of minimal code realizations starting from such encoders. The second class of codes to be considered is the one constituted by what we have called composition codes. For a subclass of these codes, we propose a method to obtain minimal realizations by means of separable Roesser models.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Contemporary studies of spatial and social cognition frequently use human figures as stimuli. The interpretation of such studies may be complicated by spatial compatibility effects that emerge when researchers employ spatial responses, and participants spontaneously code spatial relationships about an observed body. Yet, the nature of these spatial codes – whether they are location- or object-based, and coded from the perspective of the observer or the figure – has not been determined. Here, we investigated this issue by exploring spatial compatibility effects arising for objects held by a visually presented whole-bodied schematic human figure. In three experiments, participants responded to the colour of the object held in the figure’s left or right hand, using left or right key presses. Left-right compatibility effects were found relative to the participant’s egocentric perspective, rather than the figure’s. These effects occurred even when the figure was rotated by 90 degrees to the left or to the right, and the coloured objects were aligned with the participant’s midline. These findings are consistent with spontaneous spatial coding from the participant’s perspective and relative to the normal upright orientation of the body. This evidence for object-based spatial coding implies that the domain general cognitive mechanisms that result in spatial compatibility effects may contribute to certain spatial perspective-taking and social cognition phenomena.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

An experimental study aimed at assessing the influence of redundancy and neutrality on the performance of an (1+1)-ES evolution strategy modeled using Markov chains and applied to NK fitness landscapes is presented. For the study, two families of redundant binary representations, one non-neutral family which is based on linear transformations and that allows the phenotypic neighborhoods to be designed in a simple and effective way, and the neutral family based on the mathematical formulation of error control codes are used. The results indicate whether redundancy or neutrality affects more strongly the behavior of the algorithm used.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

La compression des données est la technique informatique qui vise à réduire la taille de l’information pour minimiser l’espace de stockage nécessaire et accélérer la transmission des données dans les réseaux à bande passante limitée. Plusieurs techniques de compression telles que LZ77 et ses variantes souffrent d’un problème que nous appelons la redondance causée par la multiplicité d’encodages. La multiplicité d’encodages (ME) signifie que les données sources peuvent être encodées de différentes manières. Dans son cas le plus simple, ME se produit lorsqu’une technique de compression a la possibilité, au cours du processus d’encodage, de coder un symbole de différentes manières. La technique de compression par recyclage de bits a été introduite par D. Dubé et V. Beaudoin pour minimiser la redondance causée par ME. Des variantes de recyclage de bits ont été appliquées à LZ77 et les résultats expérimentaux obtenus conduisent à une meilleure compression (une réduction d’environ 9% de la taille des fichiers qui ont été compressés par Gzip en exploitant ME). Dubé et Beaudoin ont souligné que leur technique pourrait ne pas minimiser parfaitement la redondance causée par ME, car elle est construite sur la base du codage de Huffman qui n’a pas la capacité de traiter des mots de code (codewords) de longueurs fractionnaires, c’est-à-dire qu’elle permet de générer des mots de code de longueurs intégrales. En outre, le recyclage de bits s’appuie sur le codage de Huffman (HuBR) qui impose des contraintes supplémentaires pour éviter certaines situations qui diminuent sa performance. Contrairement aux codes de Huffman, le codage arithmétique (AC) peut manipuler des mots de code de longueurs fractionnaires. De plus, durant ces dernières décennies, les codes arithmétiques ont attiré plusieurs chercheurs vu qu’ils sont plus puissants et plus souples que les codes de Huffman. Par conséquent, ce travail vise à adapter le recyclage des bits pour les codes arithmétiques afin d’améliorer l’efficacité du codage et sa flexibilité. Nous avons abordé ce problème à travers nos quatre contributions (publiées). Ces contributions sont présentées dans cette thèse et peuvent être résumées comme suit. Premièrement, nous proposons une nouvelle technique utilisée pour adapter le recyclage de bits qui s’appuie sur les codes de Huffman (HuBR) au codage arithmétique. Cette technique est nommée recyclage de bits basé sur les codes arithmétiques (ACBR). Elle décrit le cadriciel et les principes de l’adaptation du HuBR à l’ACBR. Nous présentons aussi l’analyse théorique nécessaire pour estimer la redondance qui peut être réduite à l’aide de HuBR et ACBR pour les applications qui souffrent de ME. Cette analyse démontre que ACBR réalise un recyclage parfait dans tous les cas, tandis que HuBR ne réalise de telles performances que dans des cas très spécifiques. Deuxièmement, le problème de la technique ACBR précitée, c’est qu’elle requiert des calculs à précision arbitraire. Cela nécessite des ressources illimitées (ou infinies). Afin de bénéficier de cette dernière, nous proposons une nouvelle version à précision finie. Ladite technique devienne ainsi efficace et applicable sur les ordinateurs avec les registres classiques de taille fixe et peut être facilement interfacée avec les applications qui souffrent de ME. Troisièmement, nous proposons l’utilisation de HuBR et ACBR comme un moyen pour réduire la redondance afin d’obtenir un code binaire variable à fixe. Nous avons prouvé théoriquement et expérimentalement que les deux techniques permettent d’obtenir une amélioration significative (moins de redondance). À cet égard, ACBR surpasse HuBR et fournit une classe plus étendue des sources binaires qui pouvant bénéficier d’un dictionnaire pluriellement analysable. En outre, nous montrons qu’ACBR est plus souple que HuBR dans la pratique. Quatrièmement, nous utilisons HuBR pour réduire la redondance des codes équilibrés générés par l’algorithme de Knuth. Afin de comparer les performances de HuBR et ACBR, les résultats théoriques correspondants de HuBR et d’ACBR sont présentés. Les résultats montrent que les deux techniques réalisent presque la même réduction de redondance sur les codes équilibrés générés par l’algorithme de Knuth.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The UMTS turbo encoder is composed of parallel concatenation of two Recursive Systematic Convolutional (RSC) encoders which start and end at a known state. This trellis termination directly affects the performance of turbo codes. This paper presents performance analysis of multi-point trellis termination of turbo codes which is to terminate RSC encoders at more than one point of the current frame while keeping the interleaver length the same. For long interleaver lengths, this approach provides dividing a data frame into sub-frames which can be treated as independent blocks. A novel decoding architecture using multi-point trellis termination and collision-free interleavers is presented. Collision-free interleavers are used to solve memory collision problems encountered by parallel decoding of turbo codes. The proposed parallel decoding architecture reduces the decoding delay caused by the iterative nature and forward-backward metric computations of turbo decoding algorithms. Our simulations verified that this turbo encoding and decoding scheme shows Bit Error Rate (BER) performance very close to that of the UMTS turbo coding while providing almost %50 time saving for the 2-point termination and %80 time saving for the 5-point termination.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Turbo codes experience a significant decoding delay because of the iterative nature of the decoding algorithms, the high number of metric computations and the complexity added by the (de)interleaver. The extrinsic information is exchanged sequentially between two Soft-Input Soft-Output (SISO) decoders. Instead of this sequential process, a received frame can be divided into smaller windows to be processed in parallel. In this paper, a novel parallel processing methodology is proposed based on the previous parallel decoding techniques. A novel Contention-Free (CF) interleaver is proposed as part of the decoding architecture which allows using extrinsic Log-Likelihood Ratios (LLRs) immediately as a-priori LLRs to start the second half of the iterative turbo decoding. The simulation case studies performed in this paper show that our parallel decoding method can provide %80 time saving compared to the standard decoding and %30 time saving compared to the previous parallel decoding methods at the expense of 0.3 dB Bit Error Rate (BER) performance degradation.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Paper presented at Geo-Spatial Crossroad GI_Forum, Salzburg, Austria.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In population surveys of the exposure to medical X-rays both the frequency of examinations and the effective dose per examination are required. The use of the Swiss medical tariffication system (TARMED) for establishing the frequency of X-ray medical examinations was explored. The method was tested for radiography examinations performed in 2008 at the Lausanne University Hospital. The annual numbers of radiographies determined from the "TARMED" database are in good agreement with the figures extracted from the local RIS (Radiology Information System). The "TARMED" is a reliable and fast method for establishing the frequency of radiography examination, if we respect the context in which the "TARMED" code is used. In addition, this billing context provides most valuable information on the average number of radiographs per examination as well as the age and sex distributions. Radiographies represent the major part of X-ray examinations and are performed by about 4,000 practices and hospitals in Switzerland. Therefore this method has the potential to drastically simplify the organisation of nationwide surveys. There are still some difficulties to overcome if the method is to be used to assess the frequency of computed tomography or fluoroscopy examinations; procedures that deliver most of the radiation dose to the population. This is due to the poor specificity of "TARMED" codes concerning these modalities. However, the use of CT and fluoroscopy installations is easier to monitor using conventional survey methods since there are fewer centres. Ways to overcome the "TARMED" limitations for these two modalities are still being explored.