883 resultados para Linear transect
Resumo:
A systolic array to implement lattice-reduction-aided lineardetection is proposed for a MIMO receiver. The lattice reductionalgorithm and the ensuing linear detections are operated in the same array, which can be hardware-efficient. All-swap lattice reduction algorithm (ASLR) is considered for the systolic design.ASLR is a variant of the LLL algorithm, which processes all lattice basis vectors within one iteration. Lattice-reduction-aided linear detection based on ASLR and LLL algorithms have very similarbit-error-rate performance, while ASLR is more time efficient inthe systolic array, especially for systems with a large number ofantennas.
Resumo:
An important statistical development of the last 30 years has been the advance in regression analysis provided by generalized linear models (GLMs) and generalized additive models (GAMs). Here we introduce a series of papers prepared within the framework of an international workshop entitled: Advances in GLMs/GAMs modeling: from species distribution to environmental management, held in Riederalp, Switzerland, 6-11 August 2001.We first discuss some general uses of statistical models in ecology, as well as provide a short review of several key examples of the use of GLMs and GAMs in ecological modeling efforts. We next present an overview of GLMs and GAMs, and discuss some of their related statistics used for predictor selection, model diagnostics, and evaluation. Included is a discussion of several new approaches applicable to GLMs and GAMs, such as ridge regression, an alternative to stepwise selection of predictors, and methods for the identification of interactions by a combined use of regression trees and several other approaches. We close with an overview of the papers and how we feel they advance our understanding of their application to ecological modeling.
Resumo:
The detailed geological mapping and structural study of a complete transect across the northwestern Himalaya allow to describe the tectonic evolution of the north Indian continental margin during the Tethys ocean opening and the Himalayan Orogeny. The Late Paleozoic Tethys rifting is associated with several tectonomagmatic events. In Upper Lahul and SE Zanskar, this extensional phase is recorded by Lower Carboniferous synsedimentary transtensional faults, a Lower Permian stratigraphic unconformity, a Lower Permian granitic intrusion and middle Permian basaltic extrusions (Panjal Traps). In eastern Ladakh, a Permian listric normal fault is also related to this phase. The scarcity of synsedimentary faults and the gradual increase of the Permian syn-rift sediment thickness towards the NE suggest a flexural type margin. The collision of India and Asia is characterized by a succession of contrasting orogenic phases. South of the Suture Zone, the initiation of the SW vergent Nyimaling-Tsarap Nappe corresponds to an early phase of continental underthrusting. To the S, in Lahul, an opposite underthrusting within the Indian plate is recorded by the NE vergent Tandi Syncline. This structure is associated with the newly defined Shikar Beh Nappe, now partly eroded, which is responsible for the high grade (amphibolite facies) regional metamorphism of South Lahul. The main thrusting of the Nyimaling-Tsarap Nappe followed the formation of the Shikar Beh Nappe. The Nyimaling-Tsarap Nappe developed by ductile shear of the upper part of the subducted Indian continental margin and is responsible for the progressive regional metamorphism of SE Zanskar, reaching amphibolite facies below the frontal part of the nappe, near Sarchu. In Upper Lahul, the frontal parts of the Nyimaling-Tsarap and Shikar Beh nappes are separated by a zone of low grade metamorphic rocks (pumpellyite-actinolite facies to lower greenschist facies). At high structural level, the Nyimaling-Tsarap Nappe is characterized by imbricate structures, which grade into a large ductile shear zone with depth. The related crustal shortening is about 87 km. The root zone and the frontal part of this nappe have been subsequently affected by two zones of dextral transpression and underthrusting: the Nyimaling Shear Zone and the Sarchu Shear Zone. These shear zones are interpreted as consequences of the counterclockwise rotation of the continental underthrusting direction of India relative to Asia, which occurred some 45 and 36 Ma ago, according to plate tectonic models. Later, a phase of NE vergent `'backfolding'' developed on these two zones of dextral transpression, creating isoclinal folds in SE Zanskar and more open folds in the Nyimaling Dome and in the Indus Molasse sediments. During a late stage of the Himalayan Orogeny, the frontal part of the Nyimaling-Tsarap Nappe underwent an extension of about 15 km. This phase is represented by two types of structures, responsible for the tectonic unroofing of the amphibolite facies rocks of the Sarchu area: the Sarchu high angle Normal Fault, cutting a first set of low angle normal faults, which have been created by reactivation of older thrust planes related to the Nyimaling-Tsarap Nappe.
Resumo:
A Investigação Operacional vem demonstrando ser uma valiosa ferramenta de gestão nos dias de hoje em que se vive num mercado cada vez mais competitivo. Através da Programação Linear pode-se reproduzir matematicamente um problema de maximização dos resultados ou minimização dos custos de produção com o propósito de auxiliar os gestores na tomada de decisão. A Programação Linear é um método matemático em que a função objectivo e as restrições assumem características lineares, com diversas aplicações no controlo de gestão, envolvendo normalmente problemas de utilização dos recursos disponíveis sujeitos a limitações impostas pelo processo produtivo ou pelo mercado. O objectivo geral deste trabalho é o de propor um modelo de Programação Linear para a programação ou produção e alocação de recursos necessários. Optimizar uma quantidade física designada função objectivo, tendo em conta um conjunto de condicionalismos endógenas às actividades em gestão. O objectivo crucial é dispor um modelo de apoio à gestão contribuindo assim para afectação eficiente de recursos escassos à disposição da unidade económica. Com o trabalho desenvolvido ficou patente a importância da abordagem quantitativa como recurso imprescindível de apoio ao processo de decisão. The operational research has proven to be a valuable management tool today we live in an increasingly competitive market. Through Linear Programming can be mathematically reproduce a problem of maximizing performance or minimizing production costs in order to assist managers in decision making. The Linear Programming is a mathematical method in which the objective function and constraints are linear features, with several applications in the control of management, usually involving problems of resource use are available subject to limitations imposed by the production process or the market. The overall objective of this work is to propose a Linear Programming model for scheduling or production and allocation of necessary resources. Optimizing a physical quantity called the objective function, given a set of endogenous constraints on management thus contributing to efficient allocation of scarce resources available to the economic unit. With the work has demonstrated the importance of the quantitative approach as essential resource to support the decision process.
Resumo:
The mathematical representation of Brunswik s lens model has been usedextensively to study human judgment and provides a unique opportunity to conduct ameta-analysis of studies that covers roughly five decades. Specifically, we analyzestatistics of the lens model equation (Tucker, 1964) associated with 259 different taskenvironments obtained from 78 papers. In short, we find on average fairly high levelsof judgmental achievement and note that people can achieve similar levels of cognitiveperformance in both noisy and predictable environments. Although overall performancevaries little between laboratory and field studies, both differ in terms of components ofperformance and types of environments (numbers of cues and redundancy). An analysisof learning studies reveals that the most effective form of feedback is information aboutthe task. We also analyze empirically when bootstrapping is more likely to occur. Weconclude by indicating shortcomings of the kinds of studies conducted to date, limitationsin the lens model methodology, and possibilities for future research.
Resumo:
We consider the application of normal theory methods to the estimation and testing of a general type of multivariate regressionmodels with errors--in--variables, in the case where various data setsare merged into a single analysis and the observable variables deviatepossibly from normality. The various samples to be merged can differ on the set of observable variables available. We show that there is a convenient way to parameterize the model so that, despite the possiblenon--normality of the data, normal--theory methods yield correct inferencesfor the parameters of interest and for the goodness--of--fit test. Thetheory described encompasses both the functional and structural modelcases, and can be implemented using standard software for structuralequations models, such as LISREL, EQS, LISCOMP, among others. An illustration with Monte Carlo data is presented.
Resumo:
The network revenue management (RM) problem arises in airline, hotel, media,and other industries where the sale products use multiple resources. It can be formulatedas a stochastic dynamic program but the dynamic program is computationallyintractable because of an exponentially large state space, and a number of heuristicshave been proposed to approximate it. Notable amongst these -both for their revenueperformance, as well as their theoretically sound basis- are approximate dynamic programmingmethods that approximate the value function by basis functions (both affinefunctions as well as piecewise-linear functions have been proposed for network RM)and decomposition methods that relax the constraints of the dynamic program to solvesimpler dynamic programs (such as the Lagrangian relaxation methods). In this paperwe show that these two seemingly distinct approaches coincide for the network RMdynamic program, i.e., the piecewise-linear approximation method and the Lagrangianrelaxation method are one and the same.
Resumo:
The choice network revenue management model incorporates customer purchase behavioras a function of the offered products, and is the appropriate model for airline and hotel networkrevenue management, dynamic sales of bundles, and dynamic assortment optimization.The optimization problem is a stochastic dynamic program and is intractable. A certainty-equivalencerelaxation of the dynamic program, called the choice deterministic linear program(CDLP) is usually used to generate dyamic controls. Recently, a compact linear programmingformulation of this linear program was given for the multi-segment multinomial-logit (MNL)model of customer choice with non-overlapping consideration sets. Our objective is to obtaina tighter bound than this formulation while retaining the appealing properties of a compactlinear programming representation. To this end, it is natural to consider the affine relaxationof the dynamic program. We first show that the affine relaxation is NP-complete even for asingle-segment MNL model. Nevertheless, by analyzing the affine relaxation we derive a newcompact linear program that approximates the dynamic programming value function betterthan CDLP, provably between the CDLP value and the affine relaxation, and often comingclose to the latter in our numerical experiments. When the segment consideration sets overlap,we show that some strong equalities called product cuts developed for the CDLP remain validfor our new formulation. Finally we perform extensive numerical comparisons on the variousbounds to evaluate their performance.
Resumo:
Standard methods for the analysis of linear latent variable models oftenrely on the assumption that the vector of observed variables is normallydistributed. This normality assumption (NA) plays a crucial role inassessingoptimality of estimates, in computing standard errors, and in designinganasymptotic chi-square goodness-of-fit test. The asymptotic validity of NAinferences when the data deviates from normality has been calledasymptoticrobustness. In the present paper we extend previous work on asymptoticrobustnessto a general context of multi-sample analysis of linear latent variablemodels,with a latent component of the model allowed to be fixed across(hypothetical)sample replications, and with the asymptotic covariance matrix of thesamplemoments not necessarily finite. We will show that, under certainconditions,the matrix $\Gamma$ of asymptotic variances of the analyzed samplemomentscan be substituted by a matrix $\Omega$ that is a function only of thecross-product moments of the observed variables. The main advantage of thisis thatinferences based on $\Omega$ are readily available in standard softwareforcovariance structure analysis, and do not require to compute samplefourth-order moments. An illustration with simulated data in the context ofregressionwith errors in variables will be presented.
Resumo:
We introduce several exact nonparametric tests for finite sample multivariatelinear regressions, and compare their powers. This fills an important gap inthe literature where the only known nonparametric tests are either asymptotic,or assume one covariate only.
Resumo:
A new algorithm called the parameterized expectations approach(PEA) for solving dynamic stochastic models under rational expectationsis developed and its advantages and disadvantages are discussed. Thisalgorithm can, in principle, approximate the true equilibrium arbitrarilywell. Also, this algorithm works from the Euler equations, so that theequilibrium does not have to be cast in the form of a planner's problem.Monte--Carlo integration and the absence of grids on the state variables,cause the computation costs not to go up exponentially when the numberof state variables or the exogenous shocks in the economy increase. \\As an application we analyze an asset pricing model with endogenousproduction. We analyze its implications for time dependence of volatilityof stock returns and the term structure of interest rates. We argue thatthis model can generate hump--shaped term structures.
Resumo:
Electron microscopy was used to monitor the fate of reconstituted nucleosome cores during in vitro transcription of long linear and supercoiled multinucleosomic templates by the prokaryotic T7 RNA polymerase and the eukaryotic RNA polymerase II. Transcription by T7 RNA polymerase disrupted the nucleosomal configuration in the transcribed region, while nucleosomes were preserved upstream of the transcription initiation site and in front of the polymerase. Nucleosome disruption was independent of the topology of the template, linear or supercoiled, and of the presence or absence of nucleosome positioning sequences in the transcribed region. In contrast, the nucleosomal configuration was preserved during transcription from the vitellogenin B1 promoter with RNA polymerase II in a rat liver total nuclear extract. However, the persistence of nucleosomes on the template was not RNA polymerase II-specific, but was dependent on another activity present in the nuclear extract. This was demonstrated by addition of the extract to the T7 RNA polymerase transcription reaction, which resulted in retention of the nucleosomal configuration. This nuclear activity, also found in HeLa cell nuclei, is heat sensitive and could not be substituted by nucleoplasmin, chromatin assembly factor (CAF-I) or a combination thereof. Altogether, these results identify a novel nuclear activity, called herein transcription-dependent chromatin stabilizing activity I or TCSA-I, which may be involved in a nucleosome transfer mechanism during transcription.