949 resultados para variable line-space gratings
Resumo:
We analyze high temporal and spatial resolution time-series of spectralscans of the Hα line obtained with the CRisp Imaging SpectroPolarimeter instrument mounted on the Swedish Solar Telescope.The data reveal highly dynamic, dark, short-lived structures known asRapid Redshifted and Blueshifted Excursions (RREs, RBEs) that areon-disk absorption features observed in the red and blue wings ofspectral lines formed in the chromosphere. We study the dynamics of RREsand RBEs by tracking their evolution in space and time, measuring thespeed of the apparent motion, line of sight (LOS) Doppler velocity, andtransverse velocity of individual structures. A statistical study oftheir measured properties shows that RREs and RBEs have similaroccurrence rates, lifetimes, lengths, and widths. They also displaynon-periodic, nonlinear transverse motions perpendicular to their axesat speeds of 4-31 km s-1. Furthermore, both typesof structures either appear as high speed jets and blobs that aredirected outwardly from a magnetic bright point with speeds of50-150 km s-1, or emerge within a few seconds. Astudy of the different velocity components suggests that the transversemotions along the LOS of the chromospheric flux tubes are responsiblefor the formation and appearance of these redshifted/blueshiftedstructures. The short lifetime and fast disappearance of the RREs/RBEssuggests that, similar to type II spicules, they are rapidly heated totransition region or even coronal temperatures. We speculate that theKelvin-Helmholtz instability triggered by observed transversemotions of these structures may be a viable mechanism for their heating.
Resumo:
We investigate line formation processes in Type IIb supernovae (SNe) from 100 to 500 days post-explosion using spectral synthesis calculations. The modelling identifies the nuclear burning layers and physical mechanisms that produce the major emission lines, and the diagnostic potential of these. We compare the model calculations with data on the three best observed Type IIb SNe to-date - SN 1993J, SN 2008ax, and SN 2011dh. Oxygen nucleosynthesis depends sensitively on the main-sequence mass of the star and modelling of the [O I] lambda lambda 6300, 6364 lines constrains the progenitors of these three SNe to the M-ZAMS = 12-16 M-circle dot range (ejected oxygen masses 0.3-0.9 M-circle dot), with SN 2011dh towards the lower end and SN 1993J towards the upper end of the range. The high ejecta masses from M-ZAMS greater than or similar to 17 M-circle dot progenitors give rise to brighter nebular phase emission lines than observed. Nucleosynthesis analysis thus supports a scenario of low-to-moderate mass progenitors for Type IIb SNe, and by implication an origin in binary systems. We demonstrate how oxygen and magnesium recombination lines may be combined to diagnose the magnesium mass in the SN ejecta. For SN 2011dh, a magnesium mass of 0.02-0.14 M-circle dot is derived, which gives a Mg/O production ratio consistent with the solar value. Nitrogen left in the He envelope from CNO burning gives strong [N II] lambda lambda 6548, 6583 emission lines that dominate over Ha emission in our models. The hydrogen envelopes of Type IIb SNe are too small and dilute to produce any noticeable H alpha emission or absorption after similar to 150 days, and nebular phase emission seen around 6550 angstrom is in many cases likely caused by [N II] lambda lambda 6548, 6583. Finally, the influence of radiative transport on the emergent line profiles is investigated. Significant line blocking in the metal core remains for several hundred days, which affects the emergent spectrum. These radiative transfer effects lead to early-time blueshifts of the emission line peaks, which gradually disappear as the optical depths decrease with time. The modelled evolution of this effect matches the observed evolution in SN 2011dh.
A comparison of theoretical Mg VI emission line strengths with active-region observations from SERTS
Resumo:
R-matrix calculations of electron impact excitation rates in N-like Mg VI are used to derive theoretical electron-density-sensitive emission line ratios involving 2s22p3 - 2s2p4 transitions in the 269-403 Å wavelength range. A comparison of these with observations of a solar active region, obtained during the 1989 flight of the Solar EUV Rocket Telescope and Spectrograph (SERTS), reveals good agreement between theory and observation for the 2s22p3 4S - 2s2p 4 4p transitions at 399.28, 400.67, and 403.30 Å, and the 2s22p3 2p - 2s2p4 2D lines at 387.77 and 387.97 Å. However, intensities for the other lines attributed to Mg VI in this spectrum by various authors do not match the present theoretical predictions. We argue that these discrepancies are not due to errors in the adopted atomic data, as previously suggested, but rather to observational uncertainties or mis-identifications. Some of the features previously identified as Mg VI lines in the SERTS spectrum, such as 291.36 and 293.15 Å, are judged to be noise, while others (including 349.16 Å) appear to be blended.
Resumo:
O mercado imobiliário tem um papel importante nas economias modernas, tanto a nível macro como a nível micro. Ao nível macro, a construção de habitação representa um sector importante e influente na economia, com efeitos multiplicadores significativos sobre a produção e o emprego. Ao nível micro, uma residência representa o activo mais valioso da maioria dos indivíduos e uma parcela muito relevante da riqueza das famílias. Para estas, o custo e a qualidade das suas habitações influencia directa e indirectamente a sua qualidade de vida. A habitação é por isso mesmo um tema, que avaliado nas suas múltiplas dimensões, se caracteriza por ser bastante complexo, mas também ao mesmo tempo desafiante. De modo a delimitar o objecto de análise do trabalho de investigação, esta tese realça os aspectos de localização e distribuição espacial das habitações urbanas. Será desenvolvido um quadro conceptual e respectiva metodologia para a compreender a estrutura espacial da habitação urbana realçando os três aspectos fundamentais da análise espacial: heterogenidade espacial, dependência espacial e escala espacial. A metodologia, aplicada à área urbana de Aveiro e Ílhavo é baseada numa análise hedónica factorial de preços e na noção não geométrica do espaço. Primeiro, é fixada uma escala territorial e são definidos submercados habitacional. Posteriormente, quer a heterogeneidade quer a dependência espaciais são estudados utilizando métodos econométricos, sem considerar qualquer padrão fixo e conhecido de interações espaciais. Em vez disso, são desenvolvidos novos métodos,tendo como base o modelo hedónico factorial, para inferir sobre os potenciais drivers de difusão espacial no valor de uma habitação. Este modelo, foi aplicado a duas diferentes escalas espaciais, para compreender as preferências dos indivíduos em Aveiro ao escolher os seus locais de residencia, e como estas afectam os preços da habitação. O trabalho empírico, utilizando duas bases de dados de habitação distintas, aplicadas ao mercado de habitação de Aveiro mostram: i) em linha com a literatura, a dificuldade de definir submercados e compreender as inter-relações entre esses mercados; ii) a utilidade de uma abordagem híbrida, combinando análise factorial com regressão; iii) a importância fundamental que o efeito escala espacial desempenha no estudo da heterogeneidade e dos spillovers e, finalmente, iv) uma metodologia inovadora para analisar spillovers sem assumir aprioristicamente uma estrutura espacial específica de difusão espacial. Esta metodologia considera a matriz de pesos espaciais (W) desconhecida e estimatima as interações espaciais dentro e entre submercados habitação.
Resumo:
This article aims to interrogate law's ambivalent relationship with urban space. It deals with the paradoxical relation between law and the city, visibility and invisibility, materiality and abstraction, and polis and metropolis. It builds on previous work on the lawscape, namely the priority of invitation by law or the city to be conditioned by the other, and expands this line of thought towards a more tangible understanding of visibility and its mutual constitution with invisibility. We believe that spatialisation is a relevant avenue for law's (re)conceptualisation because it moves away from a description of humanism based on the universality of subjectivity, and paves the way for a particularised and material description of law's multiplicity that specifically addresses law's social positioning. This inevitably leads to a dematerialisation of space and the reinstatement of circularity between concreteness and abstraction. Inspired by some of the themes addressed by the contributors in this issue, we begin constructing a vocabulary of lawscaping, where law and urban space are brought together in an epistemological embrace that targets and eventually questions the solipsistic way in which the two of them have been conceptualised so far.
Resumo:
Este trabalho surgiu do âmbito da Tese de Dissertação do Mestrado em Energias Sustentáveis do Instituto Superior de Engenharia do Porto, tendo o acompanhamento dos orientadores da empresa Laboratório Ecotermolab do Instituto de Soldadura e Qualidade e do Instituto Superior de Engenharia do Porto, de forma a garantir a linha traçada indo de acordo aos objectivos propostos. A presente tese abordou o estudo do impacto da influência do ar novo na climatização de edifícios, tendo como base de apoio à análise a simulação dinâmica do edifício em condições reais num programa adequado, acreditado pela norma ASHRAE 140-2004. Este trabalho pretendeu evidenciar qual o impacto da influência do ar novo na climatização de um edifício com a conjugação de vários factores, tais como, ocupação, actividades e padrões de utilização (horários), iluminação e equipamentos, estudando ainda a possibilidade do sistema funcionar em regime de “Free-Cooling”. O princípio partiu fundamentalmente por determinar até que ponto se pode climatizar recorrendo único e exclusivamente à introdução de ar novo em regime de “Free-Cooling”, através de um sistema tudo-ar de Volume de Ar Variável - VAV, sem o apoio de qualquer outro sistema de climatização auxiliar localizado no espaço, respeitando os caudais mínimos impostos pelo RSECE (Decreto-Lei 79/2006). Numa primeira fase foram identificados todos os dados relativos à determinação das cargas térmicas do edifício, tendo em conta todos os factores e contributos alusivos ao valor da carga térmica, tais como a transmissão de calor e seus constituintes, a iluminação, a ventilação, o uso de equipamentos e os níveis de ocupação. Consequentemente foram elaboradas diversas simulações dinâmicas com o recurso ao programa EnergyPlus integrado no DesignBuilder, conjugando variáveis desde as envolventes à própria arquitectura, perfis de utilização ocupacional, equipamentos e taxas de renovação de ar nos diferentes espaços do edifício em estudo. Obtiveram-se vários modelos de forma a promover um estudo comparativo e aprofundado que permitisse determinar o impacto do ar novo na climatização do edifício, perspectivando a capacidade funcional do sistema funcionar em regime de “Free-Cooling”. Deste modo, a análise e comparação dos dados obtidos permitiram chegar às seguintes conclusões: Tendo em consideração que para necessidades de arrefecimento bastante elevadas, o “Free-Cooling” diurno revelou-se pouco eficaz ou quase nulo, para o tipo de clima verificado em Portugal, pois o diferencial de temperatura existente entre o exterior e o interior não é suficiente de modo a tornar possível a remoção das cargas de forma a baixar a temperatura interior para o intervalo de conforto. Em relação ao “Free-Cooling” em horário nocturno ou pós-laboral, este revelou-se bem mais eficiente. Obtiveram-se prestações muito interessantes sobretudo durante as estações de aquecimento e meia-estação, tendo em consideração o facto de existir necessidades de arrefecimento mesmo durante a estação de aquecimento. Referente à ventilação nocturna, isto é, em períodos de madrugada e fecho do edifício, concluiu-se que tal contribui para um abaixamento do calor acumulado durante o dia nos materiais construtivos do edifício e que é libertado ou restituído posteriormente para os espaços em períodos mais tardios. De entre as seguintes variáveis, aumento de caudal de ar novo insuflado e o diferencial de temperatura existente entre o ar exterior e interior, ficou demonstrado que este último teria maior peso contributivo na remoção do calor. Por fim, é ponto assente que de um modo geral, um sistema de climatização será sempre indispensável devido a cargas internas elevadas, requisitos interiores de temperatura e humidade, sendo no entanto aconselhado o “Free- Cooling” como um opção viável a incorporar na solução de climatização, de forma a promover o arrefecimento natural, a redução do consumo energético e a introdução activa de ar novo.
Resumo:
With the current complexity of communication protocols, implementing its layers totally in the kernel of the operating system is too cumbersome, and it does not allow use of the capabilities only available in user space processes. However, building protocols as user space processes must not impair the responsiveness of the communication. Therefore, in this paper we present a layer of a communication protocol, which, due to its complexity, was implemented in a user space process. Lower layers of the protocol are, for responsiveness issues, implemented in the kernel. This protocol was developed to support large-scale power-line communication (PLC) with timing requirements.
Resumo:
Latent variable models in finance originate both from asset pricing theory and time series analysis. These two strands of literature appeal to two different concepts of latent structures, which are both useful to reduce the dimension of a statistical model specified for a multivariate time series of asset prices. In the CAPM or APT beta pricing models, the dimension reduction is cross-sectional in nature, while in time-series state-space models, dimension is reduced longitudinally by assuming conditional independence between consecutive returns, given a small number of state variables. In this paper, we use the concept of Stochastic Discount Factor (SDF) or pricing kernel as a unifying principle to integrate these two concepts of latent variables. Beta pricing relations amount to characterize the factors as a basis of a vectorial space for the SDF. The coefficients of the SDF with respect to the factors are specified as deterministic functions of some state variables which summarize their dynamics. In beta pricing models, it is often said that only the factorial risk is compensated since the remaining idiosyncratic risk is diversifiable. Implicitly, this argument can be interpreted as a conditional cross-sectional factor structure, that is, a conditional independence between contemporaneous returns of a large number of assets, given a small number of factors, like in standard Factor Analysis. We provide this unifying analysis in the context of conditional equilibrium beta pricing as well as asset pricing with stochastic volatility, stochastic interest rates and other state variables. We address the general issue of econometric specifications of dynamic asset pricing models, which cover the modern literature on conditionally heteroskedastic factor models as well as equilibrium-based asset pricing models with an intertemporal specification of preferences and market fundamentals. We interpret various instantaneous causality relationships between state variables and market fundamentals as leverage effects and discuss their central role relative to the validity of standard CAPM-like stock pricing and preference-free option pricing.
Resumo:
Ma thèse est composée de trois chapitres reliés à l'estimation des modèles espace-état et volatilité stochastique. Dans le première article, nous développons une procédure de lissage de l'état, avec efficacité computationnelle, dans un modèle espace-état linéaire et gaussien. Nous montrons comment exploiter la structure particulière des modèles espace-état pour tirer les états latents efficacement. Nous analysons l'efficacité computationnelle des méthodes basées sur le filtre de Kalman, l'algorithme facteur de Cholesky et notre nouvelle méthode utilisant le compte d'opérations et d'expériences de calcul. Nous montrons que pour de nombreux cas importants, notre méthode est plus efficace. Les gains sont particulièrement grands pour les cas où la dimension des variables observées est grande ou dans les cas où il faut faire des tirages répétés des états pour les mêmes valeurs de paramètres. Comme application, on considère un modèle multivarié de Poisson avec le temps des intensités variables, lequel est utilisé pour analyser le compte de données des transactions sur les marchés financières. Dans le deuxième chapitre, nous proposons une nouvelle technique pour analyser des modèles multivariés à volatilité stochastique. La méthode proposée est basée sur le tirage efficace de la volatilité de son densité conditionnelle sachant les paramètres et les données. Notre méthodologie s'applique aux modèles avec plusieurs types de dépendance dans la coupe transversale. Nous pouvons modeler des matrices de corrélation conditionnelles variant dans le temps en incorporant des facteurs dans l'équation de rendements, où les facteurs sont des processus de volatilité stochastique indépendants. Nous pouvons incorporer des copules pour permettre la dépendance conditionnelle des rendements sachant la volatilité, permettant avoir différent lois marginaux de Student avec des degrés de liberté spécifiques pour capturer l'hétérogénéité des rendements. On tire la volatilité comme un bloc dans la dimension du temps et un à la fois dans la dimension de la coupe transversale. Nous appliquons la méthode introduite par McCausland (2012) pour obtenir une bonne approximation de la distribution conditionnelle à posteriori de la volatilité d'un rendement sachant les volatilités d'autres rendements, les paramètres et les corrélations dynamiques. Le modèle est évalué en utilisant des données réelles pour dix taux de change. Nous rapportons des résultats pour des modèles univariés de volatilité stochastique et deux modèles multivariés. Dans le troisième chapitre, nous évaluons l'information contribuée par des variations de volatilite réalisée à l'évaluation et prévision de la volatilité quand des prix sont mesurés avec et sans erreur. Nous utilisons de modèles de volatilité stochastique. Nous considérons le point de vue d'un investisseur pour qui la volatilité est une variable latent inconnu et la volatilité réalisée est une quantité d'échantillon qui contient des informations sur lui. Nous employons des méthodes bayésiennes de Monte Carlo par chaîne de Markov pour estimer les modèles, qui permettent la formulation, non seulement des densités a posteriori de la volatilité, mais aussi les densités prédictives de la volatilité future. Nous comparons les prévisions de volatilité et les taux de succès des prévisions qui emploient et n'emploient pas l'information contenue dans la volatilité réalisée. Cette approche se distingue de celles existantes dans la littérature empirique en ce sens que ces dernières se limitent le plus souvent à documenter la capacité de la volatilité réalisée à se prévoir à elle-même. Nous présentons des applications empiriques en utilisant les rendements journaliers des indices et de taux de change. Les différents modèles concurrents sont appliqués à la seconde moitié de 2008, une période marquante dans la récente crise financière.
Resumo:
Après des décennies de développement, l'ablation laser est devenue une technique importante pour un grand nombre d'applications telles que le dépôt de couches minces, la synthèse de nanoparticules, le micro-usinage, l’analyse chimique, etc. Des études expérimentales ainsi que théoriques ont été menées pour comprendre les mécanismes physiques fondamentaux mis en jeu pendant l'ablation et pour déterminer l’effet de la longueur d'onde, de la durée d'impulsion, de la nature de gaz ambiant et du matériau de la cible. La présente thèse décrit et examine l'importance relative des mécanismes physiques qui influencent les caractéristiques des plasmas d’aluminium induits par laser. Le cadre général de cette recherche forme une étude approfondie de l'interaction entre la dynamique de la plume-plasma et l’atmosphère gazeuse dans laquelle elle se développe. Ceci a été réalisé par imagerie résolue temporellement et spatialement de la plume du plasma en termes d'intensité spectrale, de densité électronique et de température d'excitation dans différentes atmosphères de gaz inertes tel que l’Ar et l’He et réactifs tel que le N2 et ce à des pressions s’étendant de 10‾7 Torr (vide) jusqu’à 760 Torr (pression atmosphérique). Nos résultats montrent que l'intensité d'émission de plasma dépend généralement de la nature de gaz et qu’elle est fortement affectée par sa pression. En outre, pour un délai temporel donné par rapport à l'impulsion laser, la densité électronique ainsi que la température augmentent avec la pression de gaz, ce qui peut être attribué au confinement inertiel du plasma. De plus, on observe que la densité électronique est maximale à proximité de la surface de la cible où le laser est focalisé et qu’elle diminue en s’éloignant (axialement et radialement) de cette position. Malgré la variation axiale importante de la température le long du plasma, on trouve que sa variation radiale est négligeable. La densité électronique et la température ont été trouvées maximales lorsque le gaz est de l’argon et minimales pour l’hélium, tandis que les valeurs sont intermédiaires dans le cas de l’azote. Ceci tient surtout aux propriétés physiques et chimiques du gaz telles que la masse des espèces, leur énergie d'excitation et d'ionisation, la conductivité thermique et la réactivité chimique. L'expansion de la plume du plasma a été étudiée par imagerie résolue spatio-temporellement. Les résultats montrent que la nature de gaz n’affecte pas la dynamique de la plume pour des pressions inférieures à 20 Torr et pour un délai temporel inférieur à 200 ns. Cependant, pour des pressions supérieures à 20 Torr, l'effet de la nature du gaz devient important et la plume la plus courte est obtenue lorsque la masse des espèces du gaz est élevée et lorsque sa conductivité thermique est relativement faible. Ces résultats sont confirmés par la mesure de temps de vol de l’ion Al+ émettant à 281,6 nm. D’autre part, on trouve que la vitesse de propagation des ions d’aluminium est bien définie juste après l’ablation et près de la surface de la cible. Toutefois, pour un délai temporel important, les ions, en traversant la plume, se thermalisent grâce aux collisions avec les espèces du plasma et du gaz.
Resumo:
Study on variable stars is an important topic of modern astrophysics. After the invention of powerful telescopes and high resolving powered CCD’s, the variable star data is accumulating in the order of peta-bytes. The huge amount of data need lot of automated methods as well as human experts. This thesis is devoted to the data analysis on variable star’s astronomical time series data and hence belong to the inter-disciplinary topic, Astrostatistics. For an observer on earth, stars that have a change in apparent brightness over time are called variable stars. The variation in brightness may be regular (periodic), quasi periodic (semi-periodic) or irregular manner (aperiodic) and are caused by various reasons. In some cases, the variation is due to some internal thermo-nuclear processes, which are generally known as intrinsic vari- ables and in some other cases, it is due to some external processes, like eclipse or rotation, which are known as extrinsic variables. Intrinsic variables can be further grouped into pulsating variables, eruptive variables and flare stars. Extrinsic variables are grouped into eclipsing binary stars and chromospheri- cal stars. Pulsating variables can again classified into Cepheid, RR Lyrae, RV Tauri, Delta Scuti, Mira etc. The eruptive or cataclysmic variables are novae, supernovae, etc., which rarely occurs and are not periodic phenomena. Most of the other variations are periodic in nature. Variable stars can be observed through many ways such as photometry, spectrophotometry and spectroscopy. The sequence of photometric observa- xiv tions on variable stars produces time series data, which contains time, magni- tude and error. The plot between variable star’s apparent magnitude and time are known as light curve. If the time series data is folded on a period, the plot between apparent magnitude and phase is known as phased light curve. The unique shape of phased light curve is a characteristic of each type of variable star. One way to identify the type of variable star and to classify them is by visually looking at the phased light curve by an expert. For last several years, automated algorithms are used to classify a group of variable stars, with the help of computers. Research on variable stars can be divided into different stages like observa- tion, data reduction, data analysis, modeling and classification. The modeling on variable stars helps to determine the short-term and long-term behaviour and to construct theoretical models (for eg:- Wilson-Devinney model for eclips- ing binaries) and to derive stellar properties like mass, radius, luminosity, tem- perature, internal and external structure, chemical composition and evolution. The classification requires the determination of the basic parameters like pe- riod, amplitude and phase and also some other derived parameters. Out of these, period is the most important parameter since the wrong periods can lead to sparse light curves and misleading information. Time series analysis is a method of applying mathematical and statistical tests to data, to quantify the variation, understand the nature of time-varying phenomena, to gain physical understanding of the system and to predict future behavior of the system. Astronomical time series usually suffer from unevenly spaced time instants, varying error conditions and possibility of big gaps. This is due to daily varying daylight and the weather conditions for ground based observations and observations from space may suffer from the impact of cosmic ray particles. Many large scale astronomical surveys such as MACHO, OGLE, EROS, xv ROTSE, PLANET, Hipparcos, MISAO, NSVS, ASAS, Pan-STARRS, Ke- pler,ESA, Gaia, LSST, CRTS provide variable star’s time series data, even though their primary intention is not variable star observation. Center for Astrostatistics, Pennsylvania State University is established to help the astro- nomical community with the aid of statistical tools for harvesting and analysing archival data. Most of these surveys releases the data to the public for further analysis. There exist many period search algorithms through astronomical time se- ries analysis, which can be classified into parametric (assume some underlying distribution for data) and non-parametric (do not assume any statistical model like Gaussian etc.,) methods. Many of the parametric methods are based on variations of discrete Fourier transforms like Generalised Lomb-Scargle peri- odogram (GLSP) by Zechmeister(2009), Significant Spectrum (SigSpec) by Reegen(2007) etc. Non-parametric methods include Phase Dispersion Minimi- sation (PDM) by Stellingwerf(1978) and Cubic spline method by Akerlof(1994) etc. Even though most of the methods can be brought under automation, any of the method stated above could not fully recover the true periods. The wrong detection of period can be due to several reasons such as power leakage to other frequencies which is due to finite total interval, finite sampling interval and finite amount of data. Another problem is aliasing, which is due to the influence of regular sampling. Also spurious periods appear due to long gaps and power flow to harmonic frequencies is an inherent problem of Fourier methods. Hence obtaining the exact period of variable star from it’s time series data is still a difficult problem, in case of huge databases, when subjected to automation. As Matthew Templeton, AAVSO, states “Variable star data analysis is not always straightforward; large-scale, automated analysis design is non-trivial”. Derekas et al. 2007, Deb et.al. 2010 states “The processing of xvi huge amount of data in these databases is quite challenging, even when looking at seemingly small issues such as period determination and classification”. It will be beneficial for the variable star astronomical community, if basic parameters, such as period, amplitude and phase are obtained more accurately, when huge time series databases are subjected to automation. In the present thesis work, the theories of four popular period search methods are studied, the strength and weakness of these methods are evaluated by applying it on two survey databases and finally a modified form of cubic spline method is intro- duced to confirm the exact period of variable star. For the classification of new variable stars discovered and entering them in the “General Catalogue of Vari- able Stars” or other databases like “Variable Star Index“, the characteristics of the variability has to be quantified in term of variable star parameters.
Resumo:
Globalization is widely regarded as the rise of the borderless world. However in practice, true globalization points rather to a “spatial logic” by which globalization is manifested locally in the shape of insular space. Globalization in this sense is not merely about the creation of physical fragmentation of space but also the creation of social disintegration. This study tries to proof that global processes also create various forms of insular space leading also to specific social implications. In order to examine the problem this study looks at two cases: China’s Pearl River Delta (PRD) and Jakarta in Indonesia. The PRD case reveals three forms of insular space namely the modular, concealed and the hierarchical. The modular points to the form of enclosed factories where workers are vulnerable for human-right violations due to the absent of public control. The concealed refers to the production of insular space by subtle discrimination against certain social groups in urban space. And the hierarchical points to a production of insular space that is formed by an imbalanced population flow. The Jakarta case attempts to show more types of insularity in relation to the complexity of a mega-city which is shaped by a culture of exclusion. Those are dormant and hollow insularity. The dormant refers to the genesis of insular– radical – community from a culture of resistance. The last type, the hollow, points to the process of making a “pseudo community” where sense of community is not really developed as well as weak social relationship with its surrounding. Although global process creates various expressions of territorial insularization, however, this study finds that the “line of flight” is always present, where the border of insularity is crossed. The PRD’s produces vernacular modernization done by peasants which is less likely to be controlled by the politics of insularization. In Jakarta, the culture of insularization causes urban informalities that have no space, neither spatially nor socially; hence their state of ephemerality continues as a tactic of place-making. This study argues that these crossings possess the potential for reconciling venue to defuse the power of insularity.
Resumo:
This paper investigates the linear degeneracies of projective structure estimation from point and line features across three views. We show that the rank of the linear system of equations for recovering the trilinear tensor of three views reduces to 23 (instead of 26) in the case when the scene is a Linear Line Complex (set of lines in space intersecting at a common line) and is 21 when the scene is planar. The LLC situation is only linearly degenerate, and we show that one can obtain a unique solution when the admissibility constraints of the tensor are accounted for. The line configuration described by an LLC, rather than being some obscure case, is in fact quite typical. It includes, as a particular example, the case of a camera moving down a hallway in an office environment or down an urban street. Furthermore, an LLC situation may occur as an artifact such as in direct estimation from spatio-temporal derivatives of image brightness. Therefore, an investigation into degeneracies and their remedy is important also in practice.
Resumo:
Compositional data analysis motivated the introduction of a complete Euclidean structure in the simplex of D parts. This was based on the early work of J. Aitchison (1986) and completed recently when Aitchinson distance in the simplex was associated with an inner product and orthonormal bases were identified (Aitchison and others, 2002; Egozcue and others, 2003). A partition of the support of a random variable generates a composition by assigning the probability of each interval to a part of the composition. One can imagine that the partition can be refined and the probability density would represent a kind of continuous composition of probabilities in a simplex of infinitely many parts. This intuitive idea would lead to a Hilbert-space of probability densities by generalizing the Aitchison geometry for compositions in the simplex into the set probability densities
Resumo:
It is well known that regression analyses involving compositional data need special attention because the data are not of full rank. For a regression analysis where both the dependent and independent variable are components we propose a transformation of the components emphasizing their role as dependent and independent variables. A simple linear regression can be performed on the transformed components. The regression line can be depicted in a ternary diagram facilitating the interpretation of the analysis in terms of components. An exemple with time-budgets illustrates the method and the graphical features