120 resultados para categorization IT PFC computational neuroscience model HMAX
Resumo:
Context. The turbulent pumping effect corresponds to the transport of magnetic flux due to the presence of density and turbulence gradients in convectively unstable layers. In the induction equation it appears as an advective term and for this reason it is expected to be important in the solar and stellar dynamo processes. Aims. We explore the effects of turbulent pumping in a flux-dominated Babcock-Leighton solar dynamo model with a solar-like rotation law. Methods. As a first step, only vertical pumping has been considered through the inclusion of a radial diamagnetic term in the induction equation. In the second step, a latitudinal pumping term was included and then, a near-surface shear was included. Results. The results reveal the importance of the pumping mechanism in solving current limitations in mean field dynamo modeling, such as the storage of the magnetic flux and the latitudinal distribution of the sunspots. If a meridional flow is assumed to be present only in the upper part of the convective zone, it is the full turbulent pumping that regulates both the period of the solar cycle and the latitudinal distribution of the sunspot activity. In models that consider shear near the surface, a second shell of toroidal field is generated above r = 0.95 R(circle dot) at all latitudes. If the full pumping is also included, the polar toroidal fields are efficiently advected inwards, and the toroidal magnetic activity survives only at the observed latitudes near the equator. With regard to the parity of the magnetic field, only models that combine turbulent pumping with near-surface shear always converge to the dipolar parity. Conclusions. This result suggests that, under the Babcock-Leighton approach, the equartorward motion of the observed magnetic activity is governed by the latitudinal pumping of the toroidal magnetic field rather than by a large scale coherent meridional flow. Our results support the idea that the parity problem is related to the quadrupolar imprint of the meridional flow on the poloidal component of the magnetic field and the turbulent pumping positively contributes to wash out this imprint.
Resumo:
Context. About 2/3 of the Be stars present the so-called V/R variations, a phenomenon characterized by the quasi-cyclic variation in the ratio between the violet and red emission peaks of the HI emission lines. These variations are generally explained by global oscillations in the circumstellar disk forming a one-armed spiral density pattern that precesses around the star with a period of a few years. Aims. This paper presents self-consistent models of polarimetric, photometric, spectrophotometric, and interferometric observations of the classical Be star zeta Tauri. The primary goal is to conduct a critical quantitative test of the global oscillation scenario. Methods. Detailed three-dimensional, NLTE radiative transfer calculations were carried out using the radiative transfer code HDUST. The most up-to-date research on Be stars was used as input for the code in order to include a physically realistic description for the central star and the circumstellar disk. The model adopts a rotationally deformed, gravity darkened central star, surrounded by a disk whose unperturbed state is given by a steady-state viscous decretion disk model. It is further assumed that this disk is in vertical hydrostatic equilibrium. Results. By adopting a viscous decretion disk model for zeta Tauri and a rigorous solution of the radiative transfer, a very good fit of the time-average properties of the disk was obtained. This provides strong theoretical evidence that the viscous decretion disk model is the mechanism responsible for disk formation. The global oscillation model successfully fitted spatially resolved VLTI/AMBER observations and the temporal V/R variations in the H alpha and Br gamma lines. This result convincingly demonstrates that the oscillation pattern in the disk is a one-armed spiral. Possible model shortcomings, as well as suggestions for future improvements, are also discussed.
Resumo:
Background: The dust mite Blomia tropicalis is an important source of aeroallergens in tropical areas. Although a mouse model for B. tropicalis extract (BtE)-induced asthma has been described, no study comparing different mouse strains in this asthma model has been reported. The relevance and reproducibility of experimental animal models of allergy depends on the genetic background of the animal, the molecular composition of the allergen and the experimental protocol. Objectives: This work had two objectives. The first was to study the anti-B. tropicalis allergic responses in different mouse strains using a short-term model of respiratory allergy to BtE. This study included the comparison of the allergic responses elicited by BtE with those elicited by ovalbumin in mice of the strain that responded better to BtE sensitization. The second objective was to investigate whether the best responder mouse strain could be used in an experimental model of allergy employing relatively low BtE doses. Methods: Groups of mice of four different syngeneic strains were sensitized subcutaneously with 100 mu g of BtE on days 0 and 7 and challenged four times intranasally, at days 8, 10, 12, and 14, with 10 mu g of BtE. A/J mice, that were the best responders to BtE sensitization, were used to compare the B. tropicalis-specific asthma experimental model with the conventional experimental model of ovalbumin (OVA)-specific asthma. A/J mice were also sensitized with a lower dose of BtE. Results: Mice of all strains had lung inflammatory-cell infiltration and increased levels of anti-BtE IgE antibodies, but these responses were significantly more intense in A/J mice than in CBA/J, BALB/c or C57BL/6J mice. Immunization of A/J mice with BtE induced a more intense airway eosinophil influx, higher levels of total IgE, similar airway hyperreactivity to methacholine but less intense mucous production, and lower levels of specific IgE, IgG1 and IgG2 antibodies than sensitization with OVA. Finally, immunization with a relatively low BtE dose (10 mu g per subcutaneous injection per mouse) was able to sensitize A/J mice, which were the best responders to high-dose BtE immunization, for the development of allergy-associated immune and lung inflammatory responses. Conclusions: The described short-term model of BtE-induced allergic lung disease is reproducible in different syngeneic mouse strains, and mice of the A/J strain was the most responsive to it. In addition, it was shown that OVA and BtE induce quantitatively different immune responses in A/J mice and that the experimental model can be set up with low amounts of BtE.
Resumo:
In Natural Language Processing (NLP) symbolic systems, several linguistic phenomena, for instance, the thematic role relationships between sentence constituents, such as AGENT, PATIENT, and LOCATION, can be accounted for by the employment of a rule-based grammar. Another approach to NLP concerns the use of the connectionist model, which has the benefits of learning, generalization and fault tolerance, among others. A third option merges the two previous approaches into a hybrid one: a symbolic thematic theory is used to supply the connectionist network with initial knowledge. Inspired on neuroscience, it is proposed a symbolic-connectionist hybrid system called BIO theta PRED (BIOlogically plausible thematic (theta) symbolic-connectionist PREDictor), designed to reveal the thematic grid assigned to a sentence. Its connectionist architecture comprises, as input, a featural representation of the words (based on the verb/noun WordNet classification and on the classical semantic microfeature representation), and, as output, the thematic grid assigned to the sentence. BIO theta PRED is designed to ""predict"" thematic (semantic) roles assigned to words in a sentence context, employing biologically inspired training algorithm and architecture, and adopting a psycholinguistic view of thematic theory.
Resumo:
We use the Kharzeev-Levin-Nardi (KLN) model of the low x gluon distributions to fit recent HERA data on F(L) and F(2)(c)(F(2)(b)). Having checked that this model gives a good description of the data, we use it to predict F(L) and F(2)(c) to be measured in a future electron-ion collider. The results are similar to those obtained with the de Florian-Sassot and Eskola-Paukkunen-Salgado nuclear gluon distributions. The conclusion of this exercise is that the KLN model, simple as it is, may still be used as an auxiliary tool to make estimates for both heavy-ion and electron-ion collisions.
Resumo:
In the last decade the Sznajd model has been successfully employed in modeling some properties and scale features of both proportional and majority elections. We propose a version of the Sznajd model with a generalized bounded confidence rule-a rule that limits the convincing capability of agents and that is essential to allow coexistence of opinions in the stationary state. With an appropriate choice of parameters it can be reduced to previous models. We solved this model both in a mean-field approach (for an arbitrary number of opinions) and numerically in a Barabaacutesi-Albert network (for three and four opinions), studying the transient and the possible stationary states. We built the phase portrait for the special cases of three and four opinions, defining the attractors and their basins of attraction. Through this analysis, we were able to understand and explain discrepancies between mean-field and simulation results obtained in previous works for the usual Sznajd model with bounded confidence and three opinions. Both the dynamical system approach and our generalized bounded confidence rule are quite general and we think it can be useful to the understanding of other similar models.
Resumo:
The Sznajd model is a sociophysics model that mimics the propagation of opinions in a closed society, where the interactions favor groups of agreeing people. It is based in the Ising and Potts ferromagnetic models and, although the original model used only linear chains, it has since been adapted to general networks. This model has a very rich transient, which has been used to model several aspects of elections, but its stationary states are always consensus states. In order to model more complex behaviors, we have, in a recent work, introduced the idea of biases and prejudices to the Sznajd model by generalizing the bounded confidence rule, which is common to many continuous opinion models, to what we called confidence rules. In that work we have found that the mean field version of this model (corresponding to a complete network) allows for stationary states where noninteracting opinions survive, but never for the coexistence of interacting opinions. In the present work, we provide networks that allow for the coexistence of interacting opinions for certain confidence rules. Moreover, we show that the model does not become inactive; that is, the opinions keep changing, even in the stationary regime. This is an important result in the context of understanding how a rule that breeds local conformity is still able to sustain global diversity while avoiding a frozen stationary state. We also provide results that give some insights on how this behavior approaches the mean field behavior as the networks are changed.
Resumo:
Using the superfield formalism, we study the dynamical breaking of gauge symmetry and super-conformal invariance in the N = 1 three-dimensional supersymmetric Chern-Simons model, coupled to a complex scalar superfield with a quartic self-coupling. This is an analogue of the conformally invariant Coleman-Weinberg model in four spacetime dimensions. We show that a mass for the gauge and matter superfields are dynamically generated after two-loop corrections to the effective superpotential. We also discuss the N = 2 extension of our work, showing that the Coleman-Weinberg mechanism in such model is not feasible, because it is incompatible with perturbation theory.
Resumo:
We adopt the Dirac model for graphene and calculate the Casimir interaction energy between a plane suspended graphene sample and a parallel plane perfect conductor. This is done in two ways. First, we use the quantum-field-theory approach and evaluate the leading-order diagram in a theory with 2+1-dimensional fermions interacting with 3+1-dimensional photons. Next, we consider an effective theory for the electromagnetic field with matching conditions induced by quantum quasiparticles in graphene. The first approach turns out to be the leading order in the coupling constant of the second one. The Casimir interaction for this system appears to be rather weak. It exhibits a strong dependence on the mass of the quasiparticles in graphene.
Resumo:
We investigate the quantum integrability of the Landau-Lifshitz (LL) model and solve the long-standing problem of finding the local quantum Hamiltonian for the arbitrary n-particle sector. The particular difficulty of the LL model quantization, which arises due to the ill-defined operator product, is dealt with by simultaneously regularizing the operator product and constructing the self-adjoint extensions of a very particular structure. The diagonalizibility difficulties of the Hamiltonian of the LL model, due to the highly singular nature of the quantum-mechanical Hamiltonian, are also resolved in our method for the arbitrary n-particle sector. We explicitly demonstrate the consistency of our construction with the quantum inverse scattering method due to Sklyanin [Lett. Math. Phys. 15, 357 (1988)] and give a prescription to systematically construct the general solution, which explains and generalizes the puzzling results of Sklyanin for the particular two-particle sector case. Moreover, we demonstrate the S-matrix factorization and show that it is a consequence of the discontinuity conditions on the functions involved in the construction of the self-adjoint extensions.
Resumo:
We consider scalar perturbations in the time dependent Horava-Witten model in order to probe its stability. We show that during the nonsingular epoque the model evolves without instabilities until it encounters the curvature singularity where a big crunch is supposed to occur. We compute the frequencies of the scalar field oscillation during the stable period and show how the oscillations can be used to prove the presence of such a singularity.
Resumo:
Background: The inference of gene regulatory networks (GRNs) from large-scale expression profiles is one of the most challenging problems of Systems Biology nowadays. Many techniques and models have been proposed for this task. However, it is not generally possible to recover the original topology with great accuracy, mainly due to the short time series data in face of the high complexity of the networks and the intrinsic noise of the expression measurements. In order to improve the accuracy of GRNs inference methods based on entropy (mutual information), a new criterion function is here proposed. Results: In this paper we introduce the use of generalized entropy proposed by Tsallis, for the inference of GRNs from time series expression profiles. The inference process is based on a feature selection approach and the conditional entropy is applied as criterion function. In order to assess the proposed methodology, the algorithm is applied to recover the network topology from temporal expressions generated by an artificial gene network (AGN) model as well as from the DREAM challenge. The adopted AGN is based on theoretical models of complex networks and its gene transference function is obtained from random drawing on the set of possible Boolean functions, thus creating its dynamics. On the other hand, DREAM time series data presents variation of network size and its topologies are based on real networks. The dynamics are generated by continuous differential equations with noise and perturbation. By adopting both data sources, it is possible to estimate the average quality of the inference with respect to different network topologies, transfer functions and network sizes. Conclusions: A remarkable improvement of accuracy was observed in the experimental results by reducing the number of false connections in the inferred topology by the non-Shannon entropy. The obtained best free parameter of the Tsallis entropy was on average in the range 2.5 <= q <= 3.5 (hence, subextensive entropy), which opens new perspectives for GRNs inference methods based on information theory and for investigation of the nonextensivity of such networks. The inference algorithm and criterion function proposed here were implemented and included in the DimReduction software, which is freely available at http://sourceforge.net/projects/dimreduction and http://code.google.com/p/dimreduction/.
Resumo:
We study a general stochastic rumour model in which an ignorant individual has a certain probability of becoming a stifler immediately upon hearing the rumour. We refer to this special kind of stifler as an uninterested individual. Our model also includes distinct rates for meetings between two spreaders in which both become stiflers or only one does, so that particular cases are the classical Daley-Kendall and Maki-Thompson models. We prove a Law of Large Numbers and a Central Limit Theorem for the proportions of those who ultimately remain ignorant and those who have heard the rumour but become uninterested in it.
Resumo:
Currently there is a trend for the expansion of the area cropped with sugarcane (Saccharum officinarum L.), driven by an increase in the world demand for biofuels, due to economical, environmental, and geopolitical issues. Although sugarcane is traditionally harvested by burning dried leaves and tops, the unburned, mechanized harvest has been progressively adopted. The use of process based models is useful in understanding the effects of plant litter in soil C dynamics. The objective of this work was to use the CENTURY model in evaluating the effect of sugarcane residue management in the temporal dynamics of soil C. The approach taken in this work was to parameterize the CENTURY model for the sugarcane crop, to simulate the temporal dynamics of soil C, validating the model through field experiment data, and finally to make predictions in the long term regarding soil C. The main focus of this work was the comparison of soil C stocks between the burned and unburned litter management systems, but the effect of mineral fertilizer and organic residue applications were also evaluated. The simulations were performed with data from experiments with different durations, from 1 to 60 yr, in Goiana and Timbauba, Pernambuco, and Pradopolis, Sao Paulo, all in Brazil; and Mount Edgecombe, Kwazulu-Natal, South Africa. It was possible to simulate the temporal dynamics of soil C (R(2) = 0.89). The predictions made with the model revealed that there is, in the long term, a trend for higher soil C stocks with the unburned management. This increase is conditioned by factors such as climate, soil texture, time of adoption of the unburned system, and N fertilizer management.
Resumo:
In this work we study an agent based model to investigate the role of asymmetric information degrees for market evolution. This model is quite simple and may be treated analytically since the consumers evaluate the quality of a certain good taking into account only the quality of the last good purchased plus her perceptive capacity beta. As a consequence, the system evolves according to a stationary Markov chain. The value of a good offered by the firms increases along with quality according to an exponent alpha, which is a measure of the technology. It incorporates all the technological capacity of the production systems such as education, scientific development and techniques that change the productivity rates. The technological level plays an important role to explain how the asymmetry of information may affect the market evolution in this model. We observe that, for high technological levels, the market can detect adverse selection. The model allows us to compute the maximum asymmetric information degree before the market collapses. Below this critical point the market evolves during a limited period of time and then dies out completely. When beta is closer to 1 (symmetric information), the market becomes more profitable for high quality goods, although high and low quality markets coexist. The maximum asymmetric information level is a consequence of an ergodicity breakdown in the process of quality evaluation. (C) 2011 Elsevier B.V. All rights reserved.