950 resultados para Spatial Database Systems


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Thesis (Master's)--University of Washington, 2016-08

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this paper the problem of the evolution of an object-oriented database in the context of orthogonal persistent programming systems is addressed. We have observed two characteristics in that type of systems that offer particular conditions to implement the evolution in a semi-transparent fashion. That transparency can further be enhanced with the obliviousness provided by the Aspect-Oriented Programming techniques. Was conceived a meta-model and developed a prototype to test the feasibility of our approach. The system allows programs, written to a schema, access semi-transparently to data in other versions of the schema.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Abstract : Images acquired from unmanned aerial vehicles (UAVs) can provide data with unprecedented spatial and temporal resolution for three-dimensional (3D) modeling. Solutions developed for this purpose are mainly operating based on photogrammetry concepts, namely UAV-Photogrammetry Systems (UAV-PS). Such systems are used in applications where both geospatial and visual information of the environment is required. These applications include, but are not limited to, natural resource management such as precision agriculture, military and police-related services such as traffic-law enforcement, precision engineering such as infrastructure inspection, and health services such as epidemic emergency management. UAV-photogrammetry systems can be differentiated based on their spatial characteristics in terms of accuracy and resolution. That is some applications, such as precision engineering, require high-resolution and high-accuracy information of the environment (e.g. 3D modeling with less than one centimeter accuracy and resolution). In other applications, lower levels of accuracy might be sufficient, (e.g. wildlife management needing few decimeters of resolution). However, even in those applications, the specific characteristics of UAV-PSs should be well considered in the steps of both system development and application in order to yield satisfying results. In this regard, this thesis presents a comprehensive review of the applications of unmanned aerial imagery, where the objective was to determine the challenges that remote-sensing applications of UAV systems currently face. This review also allowed recognizing the specific characteristics and requirements of UAV-PSs, which are mostly ignored or not thoroughly assessed in recent studies. Accordingly, the focus of the first part of this thesis is on exploring the methodological and experimental aspects of implementing a UAV-PS. The developed system was extensively evaluated for precise modeling of an open-pit gravel mine and performing volumetric-change measurements. This application was selected for two main reasons. Firstly, this case study provided a challenging environment for 3D modeling, in terms of scale changes, terrain relief variations as well as structure and texture diversities. Secondly, open-pit-mine monitoring demands high levels of accuracy, which justifies our efforts to improve the developed UAV-PS to its maximum capacities. The hardware of the system consisted of an electric-powered helicopter, a high-resolution digital camera, and an inertial navigation system. The software of the system included the in-house programs specifically designed for camera calibration, platform calibration, system integration, onboard data acquisition, flight planning and ground control point (GCP) detection. The detailed features of the system are discussed in the thesis, and solutions are proposed in order to enhance the system and its photogrammetric outputs. The accuracy of the results was evaluated under various mapping conditions, including direct georeferencing and indirect georeferencing with different numbers, distributions and types of ground control points. Additionally, the effects of imaging configuration and network stability on modeling accuracy were assessed. The second part of this thesis concentrates on improving the techniques of sparse and dense reconstruction. The proposed solutions are alternatives to traditional aerial photogrammetry techniques, properly adapted to specific characteristics of unmanned, low-altitude imagery. Firstly, a method was developed for robust sparse matching and epipolar-geometry estimation. The main achievement of this method was its capacity to handle a very high percentage of outliers (errors among corresponding points) with remarkable computational efficiency (compared to the state-of-the-art techniques). Secondly, a block bundle adjustment (BBA) strategy was proposed based on the integration of intrinsic camera calibration parameters as pseudo-observations to Gauss-Helmert model. The principal advantage of this strategy was controlling the adverse effect of unstable imaging networks and noisy image observations on the accuracy of self-calibration. The sparse implementation of this strategy was also performed, which allowed its application to data sets containing a lot of tie points. Finally, the concepts of intrinsic curves were revisited for dense stereo matching. The proposed technique could achieve a high level of accuracy and efficiency by searching only through a small fraction of the whole disparity search space as well as internally handling occlusions and matching ambiguities. These photogrammetric solutions were extensively tested using synthetic data, close-range images and the images acquired from the gravel-pit mine. Achieving absolute 3D mapping accuracy of 11±7 mm illustrated the success of this system for high-precision modeling of the environment.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In today’s big data world, data is being produced in massive volumes, at great velocity and from a variety of different sources such as mobile devices, sensors, a plethora of small devices hooked to the internet (Internet of Things), social networks, communication networks and many others. Interactive querying and large-scale analytics are being increasingly used to derive value out of this big data. A large portion of this data is being stored and processed in the Cloud due the several advantages provided by the Cloud such as scalability, elasticity, availability, low cost of ownership and the overall economies of scale. There is thus, a growing need for large-scale cloud-based data management systems that can support real-time ingest, storage and processing of large volumes of heterogeneous data. However, in the pay-as-you-go Cloud environment, the cost of analytics can grow linearly with the time and resources required. Reducing the cost of data analytics in the Cloud thus remains a primary challenge. In my dissertation research, I have focused on building efficient and cost-effective cloud-based data management systems for different application domains that are predominant in cloud computing environments. In the first part of my dissertation, I address the problem of reducing the cost of transactional workloads on relational databases to support database-as-a-service in the Cloud. The primary challenges in supporting such workloads include choosing how to partition the data across a large number of machines, minimizing the number of distributed transactions, providing high data availability, and tolerating failures gracefully. I have designed, built and evaluated SWORD, an end-to-end scalable online transaction processing system, that utilizes workload-aware data placement and replication to minimize the number of distributed transactions that incorporates a suite of novel techniques to significantly reduce the overheads incurred both during the initial placement of data, and during query execution at runtime. In the second part of my dissertation, I focus on sampling-based progressive analytics as a means to reduce the cost of data analytics in the relational domain. Sampling has been traditionally used by data scientists to get progressive answers to complex analytical tasks over large volumes of data. Typically, this involves manually extracting samples of increasing data size (progressive samples) for exploratory querying. This provides the data scientists with user control, repeatable semantics, and result provenance. However, such solutions result in tedious workflows that preclude the reuse of work across samples. On the other hand, existing approximate query processing systems report early results, but do not offer the above benefits for complex ad-hoc queries. I propose a new progressive data-parallel computation framework, NOW!, that provides support for progressive analytics over big data. In particular, NOW! enables progressive relational (SQL) query support in the Cloud using unique progress semantics that allow efficient and deterministic query processing over samples providing meaningful early results and provenance to data scientists. NOW! enables the provision of early results using significantly fewer resources thereby enabling a substantial reduction in the cost incurred during such analytics. Finally, I propose NSCALE, a system for efficient and cost-effective complex analytics on large-scale graph-structured data in the Cloud. The system is based on the key observation that a wide range of complex analysis tasks over graph data require processing and reasoning about a large number of multi-hop neighborhoods or subgraphs in the graph; examples include ego network analysis, motif counting in biological networks, finding social circles in social networks, personalized recommendations, link prediction, etc. These tasks are not well served by existing vertex-centric graph processing frameworks whose computation and execution models limit the user program to directly access the state of a single vertex, resulting in high execution overheads. Further, the lack of support for extracting the relevant portions of the graph that are of interest to an analysis task and loading it onto distributed memory leads to poor scalability. NSCALE allows users to write programs at the level of neighborhoods or subgraphs rather than at the level of vertices, and to declaratively specify the subgraphs of interest. It enables the efficient distributed execution of these neighborhood-centric complex analysis tasks over largescale graphs, while minimizing resource consumption and communication cost, thereby substantially reducing the overall cost of graph data analytics in the Cloud. The results of our extensive experimental evaluation of these prototypes with several real-world data sets and applications validate the effectiveness of our techniques which provide orders-of-magnitude reductions in the overheads of distributed data querying and analysis in the Cloud.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A natural way to generalize tensor network variational classes to quantum field systems is via a continuous tensor contraction. This approach is first illustrated for the class of quantum field states known as continuous matrix-product states (cMPS). As a simple example of the path-integral representation we show that the state of a dynamically evolving quantum field admits a natural representation as a cMPS. A completeness argument is also provided that shows that all states in Fock space admit a cMPS representation when the number of variational parameters tends to infinity. Beyond this, we obtain a well-behaved field limit of projected entangled-pair states (PEPS) in two dimensions that provide an abstract class of quantum field states with natural symmetries. We demonstrate how symmetries of the physical field state are encoded within the dynamics of an auxiliary field system of one dimension less. In particular, the imposition of Euclidean symmetries on the physical system requires that the auxiliary system involved in the class' definition must be Lorentz-invariant. The physical field states automatically inherit entropy area laws from the PEPS class, and are fully described by the dissipative dynamics of a lower dimensional virtual field system. Our results lie at the intersection many-body physics, quantum field theory and quantum information theory, and facilitate future exchanges of ideas and insights between these disciplines.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We consider an LTE network where a secondary user acts as a relay, transmitting data to the primary user using a decode-and-forward mechanism, transparent to the base-station (eNodeB). Clearly, the relay can decode symbols more reliably if the employed precoder matrix indicators (PMIs) are known. However, for closed loop spatial multiplexing (CLSM) transmit mode, this information is not always embedded in the downlink signal, leading to a need for effective methods to determine the PMI. In this thesis, we consider 2x2 MIMO and 4x4 MIMO downlink channels corresponding to CLSM and formulate two techniques to estimate the PMI at the relay using a hypothesis testing framework. We evaluate their performance via simulations for various ITU channel models over a range of SNR and for different channel quality indicators (CQIs). We compare them to the case when the true PMI is known at the relay and show that the performance of the proposed schemes are within 2 dB at 10% block error rate (BLER) in almost all scenarios. Furthermore, the techniques add minimal computational overhead over existent receiver structure. Finally, we also identify scenarios when using the proposed precoder detection algorithms in conjunction with the cooperative decode-and-forward relaying mechanism benefits the PUE and improves the BLER performance for the PUE. Therefore, we conclude from this that the proposed algorithms as well as the cooperative relaying mechanism at the CMR can be gainfully employed in a variety of real-life scenarios in LTE networks.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Nitrous oxide (N2O) emissions from soil are often measured using the manual static chamber method. Manual gas sampling is labour intensive, so a minimal sampling frequency that maintains the accuracy of measurements would be desirable. However, the high temporal (diurnal, daily and seasonal) variabilities of N2O emissions can compromise the accuracy of measurements if not addressed adequately when formulating a sampling schedule. Assessments of sampling strategies to date have focussed on relatively low emission systems with high episodicity, where a small number of the highest emission peaks can be critically important in the measurement of whole season cumulative emissions. Using year-long, automated sub-daily N2O measurements from three fertilised sugarcane fields, we undertook an evaluation of the optimum gas sampling strategies in high emission systems with relatively long emission episodes. The results indicated that sampling in the morning between 09:00–12:00, when soil temperature was generally close to the daily average, best approximated the daily mean N2O emission within 4–7% of the ‘actual’ daily emissions measured by automated sampling. Weekly sampling with biweekly sampling for one week after >20 mm of rainfall was the recommended sampling regime. It resulted in no extreme (>20%) deviations from the ‘actuals’, had a high probability of estimating the annual cumulative emissions within 10% precision, with practicable sampling numbers in comparison to other sampling regimes. This provides robust and useful guidance for manual gas sampling in sugarcane cropping systems, although further adjustments by the operators in terms of expected measurement accuracy and resource availability are encouraged. By implementing these sampling strategies together, labour inputs and errors in measured cumulative N2O emissions can be minimised. Further research is needed to quantify the spatial variability of N2O emissions within sugarcane cropping and to develop techniques for effectively addressing both spatial and temporal variabilities simultaneously.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

By considering the spatial character of sensor-based interactive systems, this paper investigates how discussions of seams and seamlessness in ubiquitous computing neglect the complex spatial character that is constructed as a side-effect of deploying sensor technology within a space. Through a study of a torch (`flashlight') based interface, we develop a framework for analysing this spatial character generated by sensor technology. This framework is then used to analyse and compare a range of other systems in which sensor technology is used, in order to develop a design spectrum that contrasts the revealing and hiding of a system's structure to users. Finally, we discuss the implications for interfaces situated in public spaces and consider the benefits of hiding structure from users.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Fault tolerance allows a system to remain operational to some degree when some of its components fail. One of the most common fault tolerance mechanisms consists on logging the system state periodically, and recovering the system to a consistent state in the event of a failure. This paper describes a general fault tolerance logging-based mechanism, which can be layered over deterministic systems. Our proposal describes how a logging mechanism can recover the underlying system to a consistent state, even if an action or set of actions were interrupted mid-way, due to a server crash. We also propose different methods of storing the logging information, and describe how to deploy a fault tolerant master-slave cluster for information replication. We adapt our model to a previously proposed framework, which provided common relational features, like transactions with atomic, consistent, isolated and durable properties, to NoSQL database management systems.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In database applications, access control security layers are mostly developed from tools provided by vendors of database management systems and deployed in the same servers containing the data to be protected. This solution conveys several drawbacks. Among them we emphasize: 1) if policies are complex, their enforcement can lead to performance decay of database servers; 2) when modifications in the established policies implies modifications in the business logic (usually deployed at the client-side), there is no other possibility than modify the business logic in advance and, finally, 3) malicious users can issue CRUD expressions systematically against the DBMS expecting to identify any security gap. In order to overcome these drawbacks, in this paper we propose an access control stack characterized by: most of the mechanisms are deployed at the client-side; whenever security policies evolve, the security mechanisms are automatically updated at runtime and, finally, client-side applications do not handle CRUD expressions directly. We also present an implementation of the proposed stack to prove its feasibility. This paper presents a new approach to enforce access control in database applications, this way expecting to contribute positively to the state of the art in the field.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In database applications, access control security layers are mostly developed from tools provided by vendors of database management systems and deployed in the same servers containing the data to be protected. This solution conveys several drawbacks. Among them we emphasize: (1) if policies are complex, their enforcement can lead to performance decay of database servers; (2) when modifications in the established policies implies modifications in the business logic (usually deployed at the client-side), there is no other possibility than modify the business logic in advance and, finally, 3) malicious users can issue CRUD expressions systematically against the DBMS expecting to identify any security gap. In order to overcome these drawbacks, in this paper we propose an access control stack characterized by: most of the mechanisms are deployed at the client-side; whenever security policies evolve, the security mechanisms are automatically updated at runtime and, finally, client-side applications do not handle CRUD expressions directly. We also present an implementation of the proposed stack to prove its feasibility. This paper presents a new approach to enforce access control in database applications, this way expecting to contribute positively to the state of the art in the field.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

L’utilisation de nanovecteurs pour la livraison contrôlée de principes actifs est un concept commun de nous jours. Les systèmes de livraison actuels présentent encore cependant des limites au niveau du taux de relargage des principes actifs ainsi que de la stabilité des transporteurs. Les systèmes composés à la fois de nanovecteurs (liposomes, microgels et nanogels) et d’hydrogels peuvent cependant permettre de résoudre ces problèmes. Dans cette étude, nous avons développé un système de livraison contrôlé se basant sur l’incorporation d’un nanovecteur dans une matrice hydrogel dans le but de combler les lacunes des systèmes se basant sur un vecteur uniquement. Une telle combinaison pourrait permettre un contrôle accru du relargage par stabilisation réciproque. Plus spécifiquement, nous avons développé un hydrogel structuré intégrant des liposomes, microgels et nanogels séparément chargés en principes actifs modèles potentiellement relargués de manière contrôlé. Ce contrôle a été obtenu par la modification de différents paramètres tels que la température ainsi que la composition et la concentration en nanovecteurs. Nous avons comparé la capacité de chargement et la cinétique de relargage de la sulforhodamine B et de la rhodamine 6G en utilisant des liposomes de DOPC et DPPC à différents ratios, des nanogels de chitosan/acide hyaluronique et des microgels de N-isopropylacrylamide (NIPAM) à différents ratios d’acide méthacrylique, incorporés dans un hydrogel modèle d’acrylamide. Les liposomes présentaient des capacités de chargement modérés avec un relargage prolongé sur plus de dix jours alors que les nanogels présentaient des capacités de chargement plus élevées mais une cinétique de relargage plus rapide avec un épuisement de la cargaison en deux jours. Comparativement, les microgels relarguaient complétement leur contenu en un jour. Malgré une cinétique de relargage plus rapide, les microgels ont démontré la possibilité de contrôler finement le chargement en principe actif. Ce contrôle peut être atteint par la modification des propriétés structurelles ou en changeant le milieu d’incubation, comme l’a montré la corrélation avec les isothermes de Langmuir. Chaque système développé a démontré un potentiel contrôle du taux de relargage, ce qui en fait des candidats pour des investigations futures.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

When considering deployment of wave energy converters at a given site, it is of prime importance from both a technical and an economical point of view to accurately assess the total yearly energy that can be extracted by the given device. Especially, to be considered is the assessment of the efficiency of the device over the widest span of the sea-states spectral bandwidth. Hence, the aim of this study is to assess the biases and errors introduced on extracted power classically computed using spectral data derived from analytical functions such as a JONSWAP spectrum, compared to the power derived using actual wave spectra obtained from a spectral hindcast database.