40 resultados para Spatial Database Systems
Resumo:
Although extensively studied within the lidar community, the multiple scattering phenomenon has always been considered a rare curiosity by radar meteorologists. Up to few years ago its appearance has only been associated with two- or three-body-scattering features (e.g. hail flares and mirror images) involving highly reflective surfaces. Recent atmospheric research aimed at better understanding of the water cycle and the role played by clouds and precipitation in affecting the Earth's climate has driven the deployment of high frequency radars in space. Examples are the TRMM 13.5 GHz, the CloudSat 94 GHz, the upcoming EarthCARE 94 GHz, and the GPM dual 13-35 GHz radars. These systems are able to detect the vertical distribution of hydrometeors and thus provide crucial feedbacks for radiation and climate studies. The shift towards higher frequencies increases the sensitivity to hydrometeors, improves the spatial resolution and reduces the size and weight of the radar systems. On the other hand, higher frequency radars are affected by stronger extinction, especially in the presence of large precipitating particles (e.g. raindrops or hail particles), which may eventually drive the signal below the minimum detection threshold. In such circumstances the interpretation of the radar equation via the single scattering approximation may be problematic. Errors will be large when the radiation emitted from the radar after interacting more than once with the medium still contributes substantially to the received power. This is the case if the transport mean-free-path becomes comparable with the instrument footprint (determined by the antenna beam-width and the platform altitude). This situation resembles to what has already been experienced in lidar observations, but with a predominance of wide- versus small-angle scattering events. At millimeter wavelengths, hydrometeors diffuse radiation rather isotropically compared to the visible or near infrared region where scattering is predominantly in the forward direction. A complete understanding of radiation transport modeling and data analysis methods under wide-angle multiple scattering conditions is mandatory for a correct interpretation of echoes observed by space-borne millimeter radars. This paper reviews the status of research in this field. Different numerical techniques currently implemented to account for higher order scattering are reviewed and their weaknesses and strengths highlighted. Examples of simulated radar backscattering profiles are provided with particular emphasis given to situations in which the multiple scattering contributions become comparable or overwhelm the single scattering signal. We show evidences of multiple scattering effects from air-borne and from CloudSat observations, i.e. unique signatures which cannot be explained by single scattering theory. Ideas how to identify and tackle the multiple scattering effects are discussed. Finally perspectives and suggestions for future work are outlined. This work represents a reference-guide for studies focused at modeling the radiation transport and at interpreting data from high frequency space-borne radar systems that probe highly opaque scattering media such as thick ice clouds or precipitating clouds.
Resumo:
Information modelling is a topic that has been researched a great deal, but still many questions around it have not been solved. An information model is essential in the design of a database which is the core of an information system. Currently most of databases only deal with information that represents facts, or asserted information. The ability of capturing semantic aspect has to be improved, and yet other types, such as temporal and intentional information, should be considered. Semantic Analysis, a method of information modelling, has offered a way to handle various aspects of information. It employs the domain knowledge and communication acts as sources of information modelling. It lends itself to a uniform structure whereby semantic, temporal and intentional information can be captured, which builds a sound foundation for building a semantic temporal database.
Resumo:
The main objective is to develop methods that automatically generate kinematic models for the movements of biological and robotic systems. Two methods for the identification of the kinematics are presented. The first method requires the elimination of the displacement variables that cannot be measured while the second method attempts to estimate the changes in these variables. The methods were tested using a planar two-revolute-joint linkage. Results show that the model parameters obtained agree with the actual parameters to within 5%. Moreover, the methods were applied to model head and neck movements in the sagittal plane. The results indicate that these movements are well modeled by a two-revolute-joint system. A spatial three-revolute-joint model was also discussed and tested.
Resumo:
Control systems theory can be a discipline difficult to learn without some laboratory help. With the help of focused laboratories this discipline turns to be very interesting to the students involved. The main problem is that laboratories aren't always available to students, and sometimes, when they are available, aren't big enough to a growing student population. Thus, with computer networks growing so fast, why don't create remote control labs that can be used by a large number of students? Why don't create remote control labs using Internetⓒ Copyright ?2001 IFAC Keywords: Remote Control, Computer Networks, Database, Educational Aids, Laboratory Education, Communication Control Applications.
Resumo:
This paper investigates the impact of policies to promote the adoption of LEED-certified buildings across CBSA in the United States. Drawing upon a unique database that combines data from a large number of sources and using a number of regression procedures, the determinants of the proportion LEED-certified space for more than 170 CBSA in the US is modeled. LEED-certified space still accounts for a relatively small proportion of commercial stock in all markets. The average proportion is less than 1%. There is no conclusive evidence of a positive impact of policy intervention on the levels of LEED-certified space. However, after accounting for bias introduced by non-random assignment of policies, we find preliminary evidence of a positive impact of city-level green building incentives. There is a significant positive association between market size and indicators of economic vitality on proportions of LEED-certified space.
Resumo:
Developing high-quality scientific research will be most effective if research communities with diverse skills and interests are able to share information and knowledge, are aware of the major challenges across disciplines, and can exploit economies of scale to provide robust answers and better inform policy. We evaluate opportunities and challenges facing the development of a more interactive research environment by developing an interdisciplinary synthesis of research on a single geographic region. We focus on the Amazon as it is of enormous regional and global environmental importance and faces a highly uncertain future. To take stock of existing knowledge and provide a framework for analysis we present a set of mini-reviews from fourteen different areas of research, encompassing taxonomy, biodiversity, biogeography, vegetation dynamics, landscape ecology, earth-atmosphere interactions, ecosystem processes, fire, deforestation dynamics, hydrology, hunting, conservation planning, livelihoods, and payments for ecosystem services. Each review highlights the current state of knowledge and identifies research priorities, including major challenges and opportunities. We show that while substantial progress is being made across many areas of scientific research, our understanding of specific issues is often dependent on knowledge from other disciplines. Accelerating the acquisition of reliable and contextualized knowledge about the fate of complex pristine and modified ecosystems is partly dependent on our ability to exploit economies of scale in shared resources and technical expertise, recognise and make explicit interconnections and feedbacks among sub-disciplines, increase the temporal and spatial scale of existing studies, and improve the dissemination of scientific findings to policy makers and society at large. Enhancing interaction among research efforts is vital if we are to make the most of limited funds and overcome the challenges posed by addressing large-scale interdisciplinary questions. Bringing together a diverse scientific community with a single geographic focus can help increase awareness of research questions both within and among disciplines, and reveal the opportunities that may exist for advancing acquisition of reliable knowledge. This approach could be useful for a variety of globally important scientific questions.
Resumo:
A new database of weather and circulation type catalogs is presented comprising 17 automated classification methods and five subjective classifications. It was compiled within COST Action 733 "Harmonisation and Applications of Weather Type Classifications for European regions" in order to evaluate different methods for weather and circulation type classification. This paper gives a technical description of the included methods using a new conceptual categorization for classification methods reflecting the strategy for the definition of types. Methods using predefined types include manual and threshold based classifications while methods producing types derived from the input data include those based on eigenvector techniques, leader algorithms and optimization algorithms. In order to allow direct comparisons between the methods, the circulation input data and the methods' configuration were harmonized for producing a subset of standard catalogs of the automated methods. The harmonization includes the data source, the climatic parameters used, the classification period as well as the spatial domain and the number of types. Frequency based characteristics of the resulting catalogs are presented, including variation of class sizes, persistence, seasonal and inter-annual variability as well as trends of the annual frequency time series. The methodological concept of the classifications is partly reflected by these properties of the resulting catalogs. It is shown that the types of subjective classifications compared to automated methods show higher persistence, inter-annual variation and long-term trends. Among the automated classifications optimization methods show a tendency for longer persistence and higher seasonal variation. However, it is also concluded that the distance metric used and the data preprocessing play at least an equally important role for the properties of the resulting classification compared to the algorithm used for type definition and assignment.
Resumo:
This article examines the relevance of the concept of autopoiesis for public participation in spatial planning. The concept of autopoiesis within social science, as advanced by Niklas Luhmann (1995) argues for a systemic, consistent and sophisticated theory of society based on a systems view, as opposed to more familiar action-based theories (for instance as in the case of the Habermasian tradition). By examining the relevance of the concept for public participation in spatial planning, this article highlights specific aspects of public participation that draw attention to dimensions of planning that are not currently explicitly highlighted within mainstream discourses in planning
Resumo:
Using the formalism of the Ruelle response theory, we study how the invariant measure of an Axiom A dynamical system changes as a result of adding noise, and describe how the stochastic perturbation can be used to explore the properties of the underlying deterministic dynamics. We first find the expression for the change in the expectation value of a general observable when a white noise forcing is introduced in the system, both in the additive and in the multiplicative case. We also show that the difference between the expectation value of the power spectrum of an observable in the stochastically perturbed case and of the same observable in the unperturbed case is equal to the variance of the noise times the square of the modulus of the linear susceptibility describing the frequency-dependent response of the system to perturbations with the same spatial patterns as the considered stochastic forcing. This provides a conceptual bridge between the change in the fluctuation properties of the system due to the presence of noise and the response of the unperturbed system to deterministic forcings. Using Kramers-Kronig theory, it is then possible to derive the real and imaginary part of the susceptibility and thus deduce the Green function of the system for any desired observable. We then extend our results to rather general patterns of random forcing, from the case of several white noise forcings, to noise terms with memory, up to the case of a space-time random field. Explicit formulas are provided for each relevant case analysed. As a general result, we find, using an argument of positive-definiteness, that the power spectrum of the stochastically perturbed system is larger at all frequencies than the power spectrum of the unperturbed system. We provide an example of application of our results by considering the spatially extended chaotic Lorenz 96 model. These results clarify the property of stochastic stability of SRB measures in Axiom A flows, provide tools for analysing stochastic parameterisations and related closure ansatz to be implemented in modelling studies, and introduce new ways to study the response of a system to external perturbations. Taking into account the chaotic hypothesis, we expect that our results have practical relevance for a more general class of system than those belonging to Axiom A.
Resumo:
Tropical-extratropical cloud band systems over southern Africa, known as tropical temperate troughs (TTTs), are known to contribute substantially to South African summer rainfall. This study performs a comprehensive assessment of the seasonal cycle and rainfall contribution of TTTs by using a novel object-based strategy that explicitly tracks these systems for their full life cycle. The methodology incorporates a simple assignment of station rainfall data to each event, thereby creating a database containing detailed rainfall characteristics for each TTT. This is used to explore the importance of TTTs for rain days and climatological rainfall totals in October–March. Average contributions range from 30 to 60 % with substantial spatial heterogeneity observed. TTT rainfall contributions over the Highveld and eastern escarpment are lower than expected. A short analysis of TTT rainfall variability indicates TTTs provide substantial, but not dominant, intraseasonal and interannual variability in station rainfall totals. TTTs are however responsible for a high proportion of heavy rainfall days. Of 52 extreme rainfall events in the 1979–1999 period, 30 are associated with these tropical-extratropical interactions. Cut-off lows were included in the evolution of 6 of these TTTs. The study concludes with an analysis of the question: does the Madden-Julian Oscillation influence the intensity of TTT rainfall over South Africa? Results suggest a weak but significant suppression (enhancement) of intensity during phase 1(6).
Resumo:
Platelets in the circulation are triggered by vascular damage to activate, aggregate and form a thrombus that prevents excessive blood loss. Platelet activation is stringently regulated by intracellular signalling cascades, which when activated inappropriately lead to myocardial infarction and stroke. Strategies to address platelet dysfunction have included proteomics approaches which have lead to the discovery of a number of novel regulatory proteins of potential therapeutic value. Global analysis of platelet proteomes may enhance the outcome of these studies by arranging this information in a contextual manner that recapitulates established signalling complexes and predicts novel regulatory processes. Platelet signalling networks have already begun to be exploited with interrogation of protein datasets using in silico methodologies that locate functionally feasible protein clusters for subsequent biochemical validation. Characterization of these biological systems through analysis of spatial and temporal organization of component proteins is developing alongside advances in the proteomics field. This focused review highlights advances in platelet proteomics data mining approaches that complement the emerging systems biology field. We have also highlighted nucleated cell types as key examples that can inform platelet research. Therapeutic translation of these modern approaches to understanding platelet regulatory mechanisms will enable the development of novel anti-thrombotic strategies.
Resumo:
In order to best utilize the limited resource of medical resources, and to reduce the cost and improve the quality of medical treatment, we propose to build an interoperable regional healthcare systems among several levels of medical treatment organizations. In this paper, our approaches are as follows:(1) the ontology based approach is introduced as the methodology and technological solution for information integration; (2) the integration framework of data sharing among different organizations are proposed(3)the virtual database to realize data integration of hospital information system is established. Our methods realize the effective management and integration of the medical workflow and the mass information in the interoperable regional healthcare system. Furthermore, this research provides the interoperable regional healthcare system with characteristic of modularization, expansibility and the stability of the system is enhanced by hierarchy structure.
Resumo:
Mobile-to-mobile (M-to-M) communications are expected to play a crucial role in future wireless systems and networks. In this paper, we consider M-to-M multiple-input multiple-output (MIMO) maximal ratio combining system and assess its performance in spatially correlated channels. The analysis assumes double-correlated Rayleigh-and-Lognormal fading channels and is performed in terms of average symbol error probability, outage probability, and ergodic capacity. To obtain the receive and transmit spatial correlation functions needed for the performance analysis, we used a three-dimensional (3D) M-to-M MIMO channel model, which takes into account the effects of fast fading and shadowing. The expressions for the considered metrics are derived as a function of the average signal-to-noise ratio per receive antenna in closed-form and are further approximated using the recursive adaptive Simpson quadrature method. Numerical results are provided to show the effects of system parameters, such as distance between antenna elements, maximum elevation angle of scatterers, orientation angle of antenna array in the x–y plane, angle between the x–y plane and the antenna array orientation, and degree of scattering in the x–y plane, on the system performance. Copyright © 2011 John Wiley & Sons, Ltd.
Cross-layer design for MIMO systems over spatially correlated and keyhole Nakagami-m fading channels
Resumo:
Cross-layer design is a generic designation for a set of efficient adaptive transmission schemes, across multiple layers of the protocol stack, that are aimed at enhancing the spectral efficiency and increasing the transmission reliability of wireless communication systems. In this paper, one such cross-layer design scheme that combines physical layer adaptive modulation and coding (AMC) with link layer truncated automatic repeat request (T-ARQ) is proposed for multiple-input multiple-output (MIMO) systems employing orthogonal space--time block coding (OSTBC). The performance of the proposed cross-layer design is evaluated in terms of achievable average spectral efficiency (ASE), average packet loss rate (PLR) and outage probability, for which analytical expressions are derived, considering transmission over two types of MIMO fading channels, namely, spatially correlated Nakagami-m fading channels and keyhole Nakagami-m fading channels. Furthermore, the effects of the maximum number of ARQ retransmissions, numbers of transmit and receive antennas, Nakagami fading parameter and spatial correlation parameters, are studied and discussed based on numerical results and comparisons. Copyright © 2009 John Wiley & Sons, Ltd.
Resumo:
We evaluate the effects of spatial resolution on the ability of a regional climate model to reproduce observed extreme precipitation for a region in the Southwestern United States. A total of 73 National Climate Data Center observational sites spread throughout Arizona and New Mexico are compared with regional climate simulations at the spatial resolutions of 50 km and 10 km for a 31 year period from 1980 to 2010. We analyze mean, 3-hourly and 24-hourly extreme precipitation events using WRF regional model simulations driven by NCEP-2 reanalysis. The mean climatological spatial structure of precipitation in the Southwest is well represented by the 10 km resolution but missing in the coarse (50 km resolution) simulation. However, the fine grid has a larger positive bias in mean summer precipitation than the coarse-resolution grid. The large overestimation in the simulation is in part due to scale-dependent deficiencies in the Kain-Fritsch convective parameterization scheme that generate excessive precipitation and induce a slow eastward propagation of the moist convective summer systems in the high-resolution simulation. Despite this overestimation in the mean, the 10 km simulation captures individual extreme summer precipitation events better than the 50 km simulation. In winter, however, the two simulations appear to perform equally in simulating extremes.