282 resultados para decimation filter
Resumo:
The ninth release of the Toolbox, represents over fifteen years of development and a substantial level of maturity. This version captures a large number of changes and extensions generated over the last two years which support my new book “Robotics, Vision & Control”. The Toolbox has always provided many functions that are useful for the study and simulation of classical arm-type robotics, for example such things as kinematics, dynamics, and trajectory generation. The Toolbox is based on a very general method of representing the kinematics and dynamics of serial-link manipulators. These parameters are encapsulated in MATLAB ® objects - robot objects can be created by the user for any serial-link manipulator and a number of examples are provided for well know robots such as the Puma 560 and the Stanford arm amongst others. The Toolbox also provides functions for manipulating and converting between datatypes such as vectors, homogeneous transformations and unit-quaternions which are necessary to represent 3-dimensional position and orientation. This ninth release of the Toolbox has been significantly extended to support mobile robots. For ground robots the Toolbox includes standard path planning algorithms (bug, distance transform, D*, PRM), kinodynamic planning (RRT), localization (EKF, particle filter), map building (EKF) and simultaneous localization and mapping (EKF), and a Simulink model a of non-holonomic vehicle. The Toolbox also including a detailed Simulink model for a quadcopter flying robot.
Resumo:
EMR (Electronic Medical Record) is an emerging technology that is highly-blended between non-IT and IT area. One methodology is to link the non-IT and IT area is to construct databases. Nowadays, it supports before and after-treatment for patients and should satisfy all stakeholders such as practitioners, nurses, researchers, administrators and financial departments and so on. In accordance with the database maintenance, DAS (Data as Service) model is one solution for outsourcing. However, there are some scalability and strategy issues when we need to plan to use DAS model properly. We constructed three kinds of databases such as plan-text, MS built-in encryption which is in-house model and custom AES (Advanced Encryption Standard) - DAS model scaling from 5K to 2560K records. To perform custom AES-DAS better, we also devised Bucket Index using Bloom Filter. The simulation showed the response times arithmetically increased in the beginning but after a certain threshold, exponentially increased in the end. In conclusion, if the database model is close to in-house model, then vendor technology is a good way to perform and get query response times in a consistent manner. If the model is DAS model, it is easy to outsource the database, however, some techniques like Bucket Index enhances its utilization. To get faster query response times, designing database such as consideration of the field type is also important. This study suggests cloud computing would be a next DAS model to satisfy the scalability and the security issues.
Resumo:
Electronic Health Record (EHR) retrieval processes are complex demanding Information Technology (IT) resources exponentially in particular memory usage. Database-as-a-service (DAS) model approach is proposed to meet the scalability factor of EHR retrieval processes. A simulation study using ranged of EHR records with DAS model was presented. The bucket-indexing model incorporated partitioning fields and bloom filters in a Singleton design pattern were used to implement custom database encryption system. It effectively provided faster responses in the range query compared to different types of queries used such as aggregation queries among the DAS, built-in encryption and the plain-text DBMS. The study also presented with constraints around the approach should consider for other practical applications.
Resumo:
In the recent past, there are some social issues when personal sensitive data in medical database were exposed. The personal sensitive data should be protected and access must be accounted for. Protecting the sensitive information is possible by encrypting such information. The challenge is querying the encrypted information when making the decision. Encrypted query is practically somewhat tedious task. So we present the more effective method using bucket index and bloom filter technology. We find that our proposed method shows low memory and fast efficiency comparatively. Simulation approaches on data encryption techniques to improve health care decision making processes are presented in this paper as a case scenario.
Resumo:
In the medical and healthcare arena, patients‟ data is not just their own personal history but also a valuable large dataset for finding solutions for diseases. While electronic medical records are becoming popular and are used in healthcare work places like hospitals, as well as insurance companies, and by major stakeholders such as physicians and their patients, the accessibility of such information should be dealt with in a way that preserves privacy and security. Thus, finding the best way to keep the data secure has become an important issue in the area of database security. Sensitive medical data should be encrypted in databases. There are many encryption/ decryption techniques and algorithms with regard to preserving privacy and security. Currently their performance is an important factor while the medical data is being managed in databases. Another important factor is that the stakeholders should decide more cost-effective ways to reduce the total cost of ownership. As an alternative, DAS (Data as Service) is a popular outsourcing model to satisfy the cost-effectiveness but it takes a consideration that the encryption/ decryption modules needs to be handled by trustworthy stakeholders. This research project is focusing on the query response times in a DAS model (AES-DAS) and analyses the comparison between the outsourcing model and the in-house model which incorporates Microsoft built-in encryption scheme in a SQL Server. This research project includes building a prototype of medical database schemas. There are 2 types of simulations to carry out the project. The first stage includes 6 databases in order to carry out simulations to measure the performance between plain-text, Microsoft built-in encryption and AES-DAS (Data as Service). Particularly, the AES-DAS incorporates implementations of symmetric key encryption such as AES (Advanced Encryption Standard) and a Bucket indexing processor using Bloom filter. The results are categorised such as character type, numeric type, range queries, range queries using Bucket Index and aggregate queries. The second stage takes the scalability test from 5K to 2560K records. The main result of these simulations is that particularly as an outsourcing model, AES-DAS using the Bucket index shows around 3.32 times faster than a normal AES-DAS under the 70 partitions and 10K record-sized databases. Retrieving Numeric typed data takes shorter time than Character typed data in AES-DAS. The aggregation query response time in AES-DAS is not as consistent as that in MS built-in encryption scheme. The scalability test shows that the DBMS reaches in a certain threshold; the query response time becomes rapidly slower. However, there is more to investigate in order to bring about other outcomes and to construct a secured EMR (Electronic Medical Record) more efficiently from these simulations.
Resumo:
Disposal of mud and ash, particularly in wet weather conditions, is a significant expense for mills. This paper reports on part of a process to pelletise mud and ash, aimed at making mud and ash more attractive to growers across entire mill districts. The full process and the re-constituting and centrifuging rotary vacuum filter mud part of the process were described in two papers to the 2011 conference. The component described in this paper involves aspects of mixing mud and ash with subsequent drying using boiler exit gas. The mud material needs to mix easily with boiler ash and the mixture has to feed easily into and be pneumatically conveyed by a flue gas dryer. The performance of a pilot flue gas dryer for drying mud and ash was evaluated. The mud and ash mixture was found to dry much faster than final bagasse, provided the mud and ash material was broken up into individual particles. A previously developed computer model of bagasse drying was updated to take into account the smaller particle size of the mud and ash mixture. This upgraded model predicted the performance of the pilot flue gas dryer well.
Resumo:
Sorghum (Sorghum bicolor (L.) Moench) is the world’s fifth major cereal crop and holds importance as a construction material, food and fodder source. More recently, the potential of this plant as a biofuel source has been noted. Despite its agronomic importance, the use of sorghum production is being constrained by both biotic and abiotic factors. These challenges could be addressed by the use of genetic engineering strategies to complement conventional breeding techniques. However, sorghum is one of the most recalcitrant crops for genetic modification with the lack of an efficient tissue culture system being amongst the chief reasons. Therefore, the aim of this study was to develop an efficient tissue culture system for establishing regenerable embryogenic cell lines, micropropagation and acclimatisation for Sorghum bicolor and use this to optimise parameters for genetic transformation via Agrobacterium-mediated transformation and microprojectile bombardment. Using five different sorghum cultivars, SA281, 296B, SC49, Wray and Rio, numerous parameters were investigated in an attempt to establish an efficient and reproducible tissue culture and transformation system. Using immature embryos (IEs) as explants, regenerable embryogenic cell lines (ECLs) could only be established from cultivars SA281 and 296B. Large amounts of phenolics were produced from IEs of cultivars, SC49, Wary and Rio, and these compounds severely hindered callus formation and development. Cultivar SA281 also produced phenolics during regeneration. Attempts to suppress the production of these compounds in cultivars SA281 and SC49 using activated charcoal, PVP, ascorbic acid, citric acid and liquid filter paper bridge methods were either ineffective or had a detrimental effect on embryogenic callus formation, development and regeneration. Immature embryos sourced during summer were found to be far more responsive in vitro than those sourced during winter. In an attempt to overcome this problem, IEs were sourced from sorghum grown under summer conditions in either a temperature controlled glasshouse or a growth chamber. However, the performance of these explants was still inferior to that of natural summer-sourced explants. Leaf whorls, mature embryos, shoot tips and leaf primordia were found to be unsuitable as explants for establishing ECLs in sorghum cultivars SA281 and 296B. Using the florets of immature inflorescences (IFs) as explants, however, ECLs were established and regenerated for these cultivars, as well as for cultivar Tx430, using callus induction media, SCIM, and regeneration media, VWRM. The best in vitro responses, from the largest possible sized IFs, were obtained using plants at the FL-2 stage (where the last fully opened leaf was two leaves away from the flag leaf). Immature inflorescences could be stored at 25oC for up to three days without affecting their in vitro responses. Compared to IEs, the IFs were more robust in tissue culture and showed responses which were season and growth condition independent. A micropropagation protocol for sorghum was developed in this study. The optimum plant growth regulator (PGR) combination for the micropropagation of in vitro regenerated plantlets was found to be 1.0 mg/L BAP in combination with 0.5 mg/L NAA. With this protocol, cultivars 296B and SA281 produced an average of 57 and 13 off-shoots per plantlet, respectively. The plantlets were successfully acclimatised and developed into phenotypically normal plants that set seeds. A simplified acclimatisation protocol for in vitro regenerated plantlets was also developed. This protocol involved deflasking in vitro plantlets with at least 2 fully-opened healthy leaves and at least 3 roots longer than 1.5 cm, washing the media from the roots with running tap water, planting in 100 mm pots and placing in plastic trays covered with a clear plastic bag in a plant growth chamber. After seven days, the corners of the plastic cover were opened and the bags were completely removed after 10 days. All plantlets were successfully acclimatised regardless of whether 1:1 perlite:potting mix, potting mix, UC mix or vermiculite were used as potting substrates. Parameters were optimised for Agrobacterium-mediated transformation (AMT) of cultivars SA281, 296B and Tx430. The optimal conditions were the use of Agrobacterium strain LBA4404 at an inoculum density of 0.5 OD600nm, heat shock at 43oC for 3 min, use of the surfactant Pluronic F-68 (0.02% w/v) in the inoculation media with a pH of 5.2 and a 3 day co-cultivation period in dark at 22oC. Using these parameters, high frequencies of transient GFP expression was observed in IEs precultured on callus initiation media for 1-7 days as well as in four weeks old IE- and IF-derived callus. Cultivar SA281 appeared very sensitive to Agrobacterium since all tissue turned necrotic within two weeks post-exposure. For cultivar 296B, GFP expression was observed up to 20 days post co-cultivation but no stably transformed plants were regenerated. Using cultivar Tx430, GFP was expressed for up to 50 days post co-cultivation. Although no stably transformed plants of this cultivar were regenerated, this was most likely due to the use of unsuitable regeneration media. Parameters were optimised for transformation by particle bombardment (PB) of cultivars SA281, 296B and Tx430. The optimal conditions were use of 3-7 days old IEs and 4 weeks old IF callus, 4 hour pre- and post-bombardment osmoticum treatment, use of 0.6 µm gold microparticles, helium pressure of 1500 kPa and target distance of 15 cm. Using these parameters for PB, transient GFP expression was observed for up to 14, 30 and 50 days for cultivars SA281, 296B and Tx430, respectively. Further, the use of PB resulted in less tissue necrosis compared to AMT for the respective cultivars. Despite the presence of transient GFP expression, no stably transformed plants were regenerated. The establishment of regenerable ECLs and the optimization of AMT and PB parameters in this study provides a platform for future efforts to develop an efficient transformation protocol for sorghum. The development of GM sorghum will be an important step towards improving its agronomic properties as well as its exploitation for biofuel production.
Resumo:
In many applications, where encrypted traffic flows from an open (public) domain to a protected (private) domain, there exists a gateway that bridges the two domains and faithfully forwards the incoming traffic to the receiver. We observe that indistringuishability against (adaptive) chosen-ciphertext attacks (IND-CCA), which is a mandatory goal in face of active attacks in a public domain, can be essentially relaxed to indistinguishability against chosen-plaintext attacks (IND-CPA) for ciphertexts once they pass the gateway that acts as an IND-CCA/CPA filter by first checking the validity of an incoming IND-CCA ciphertext, then transforming it (if valid) into an IND-CPA ciphertext, and forwarding the latter to the receipient in the private domain. "Non-trivial filtering" can result in reduced decryption costs on the receivers' side. We identify a class of encryption schemes with publicaly verifiable ciphertexts that admit generic constructions of (non-trivial) IND-CCA/CPA filters. These schemes are characterized by existence of public algorithms that can distinguish between valid and invalid ciphertexts. To this end, we formally define (non-trivial) public verifiability of ciphertexts for general encryption schemes, key encapsulation mechanisms, and hybrid encryption schemes, encompassing public-key, identity-based, and tag-based encryption flavours. We further analyze the security impact of public verifiability and discuss generic transformations and concrete constructions that enjoy this property.
Resumo:
Appearance-based localization can provide loop closure detection at vast scales regardless of accumulated metric error. However, the computation time and memory requirements of current appearance-based methods scale not only with the size of the environment but also with the operation time of the platform. Additionally, repeated visits to locations will develop multiple competing representations, which will reduce recall performance over time. These properties impose severe restrictions on long-term autonomy for mobile robots, as loop closure performance will inevitably degrade with increased operation time. In this paper we present a graphical extension to CAT-SLAM, a particle filter-based algorithm for appearance-based localization and mapping, to provide constant computation and memory requirements over time and minimal degradation of recall performance during repeated visits to locations. We demonstrate loop closure detection in a large urban environment with capped computation time and memory requirements and performance exceeding previous appearance-based methods by a factor of 2. We discuss the limitations of the algorithm with respect to environment size, appearance change over time and applications in topological planning and navigation for long-term robot operation.
Resumo:
Despite its role in determining both indoor and outdoor human exposure to anthropogenic particles, there is limited information describing vertical profiles of particle concentrations in urban environments, especially for ultrafine particles. Furthermore, the results of the few studies performed have been inconsistent. As such, this study aimed to assess the influence of vehicle emissions and nucleation formation on particle characteristics (particle number size distribution-PNSD and PM 2.5 concentration) at different heights around three urban office buildings located next to busy roads in Brisbane, Australia, and place these results in the broader context of the existing literature. Two sets of instruments were used to simultaneously measure PNSD, particle number (PN) and PM 2.5 concentrations, respectively, for up to three weeks at each building. The results showed that both PNSD and PM 2.5 concentration around building envelopes were influenced by vehicle emissions and new particle formation, and that they exhibited variability across the three different office buildings. During nucleation events, PN concentration in size range of <30 nm and total PN concentration increased (7-65% and 5-46%, respectively), while PM 2.5 concentration decreased (36-52%) with height. This study has shown an under acknowledged role for nucleation in producing particles that can affect large numbers of people, due to the high density and occupancy of urban office buildings and the fact that the vast majority of people's time is spent indoors. These findings highlight important new information related to the previously overlooked role of particle formation in the urban atmosphere and its potential effects on selection of air intake locations and appropriate filter types when designing or upgrading mechanical ventilation systems in urban office buildings. The results also serve to better define particle behaviour and variability around building envelopes, which has implications for studies of both human exposure and particle dynamics. © 2012 Author(s).
Resumo:
Several track-before-detection approaches for image based aircraft detection have recently been examined in an important automated aircraft collision detection application. A particularly popular approach is a two stage processing paradigm which involves: a morphological spatial filter stage (which aims to emphasize the visual characteristics of targets) followed by a temporal or track filter stage (which aims to emphasize the temporal characteristics of targets). In this paper, we proposed new spot detection techniques for this two stage processing paradigm that fuse together raw and morphological images or fuse together various different morphological images (we call these approaches morphological reinforcement). On the basis of flight test data, the proposed morphological reinforcement operations are shown to offer superior signal to-noise characteristics when compared to standard spatial filter options (such as the close-minus-open and adaptive contour morphological operations). However, system operation characterised curves, which examine detection verses false alarm characteristics after both processing stages, illustrate that system performance is very data dependent.
Resumo:
For many years, computer vision has lured researchers with promises of a low-cost, passive, lightweight and information-rich sensor suitable for navigation purposes. The prime difficulty in vision-based navigation is that the navigation solution will continually drift with time unless external information is available, whether it be cues from the appearance of the scene, a map of features (whether built online or known a priori), or from an externally-referenced sensor. It is not merely position that is of interest in the navigation problem. Attitude (i.e. the angular orientation of a body with respect to a reference frame) is integral to a visionbased navigation solution and is often of interest in its own right (e.g. flight control). This thesis examines vision-based attitude estimation in an aerospace environment, and two methods are proposed for constraining drift in the attitude solution; one through a novel integration of optical flow and the detection of the sky horizon, and the other through a loosely-coupled integration of Visual Odometry and GPS position measurements. In the first method, roll angle, pitch angle and the three aircraft body rates are recovered though a novel method of tracking the horizon over time and integrating the horizonderived attitude information with optical flow. An image processing front-end is used to select several candidate lines in a image that may or may not correspond to the true horizon, and the optical flow is calculated for each candidate line. Using an Extended Kalman Filter (EKF), the previously estimated aircraft state is propagated using a motion model and a candidate horizon line is associated using a statistical test based on the optical flow measurements and location of the horizon in the image. Once associated, the selected horizon line, along with the associated optical flow, is used as a measurement to the EKF. To evaluate the accuracy of the algorithm, two flights were conducted, one using a highly dynamic Uninhabited Airborne Vehicle (UAV) in clear flight conditions and the other in a human-piloted Cessna 172 in conditions where the horizon was partially obscured by terrain, haze and smoke. The UAV flight resulted in pitch and roll error standard deviations of 0.42° and 0.71° respectively when compared with a truth attitude source. The Cessna 172 flight resulted in pitch and roll error standard deviations of 1.79° and 1.75° respectively. In the second method for estimating attitude, a novel integrated GPS/Visual Odometry (GPS/VO) navigation filter is proposed, using a structure similar to a classic looselycoupled GPS/INS error-state navigation filter. Under such an arrangement, the error dynamics of the system are derived and a Kalman Filter is developed for estimating the errors in position and attitude. Through similar analysis to the GPS/INS problem, it is shown that the proposed filter is capable of recovering the complete attitude (i.e. pitch, roll and yaw) of the platform when subjected to acceleration not parallel to velocity for both the monocular and stereo variants of the filter. Furthermore, it is shown that under general straight line motion (e.g. constant velocity), only the component of attitude in the direction of motion is unobservable. Numerical simulations are performed to demonstrate the observability properties of the GPS/VO filter in both the monocular and stereo camera configurations. Furthermore, the proposed filter is tested on imagery collected using a Cessna 172 to demonstrate the observability properties on real-world data. The proposed GPS/VO filter does not require additional restrictions or assumptions such as platform-specific dynamics, map-matching, feature-tracking, visual loop-closing, gravity vector or additional sensors such as an IMU or magnetic compass. Since no platformspecific dynamics are required, the proposed filter is not limited to the aerospace domain and has the potential to be deployed in other platforms such as ground robots or mobile phones.
Resumo:
This paper establishes practical stability results for an important range of approximate discrete-time filtering problems involving mismatch between the true system and the approximating filter model. Practical stability is established in the sense of an asymptotic bound on the amount of bias introduced by the model approximation. Our analysis applies to a wide range of estimation problems and justifies the common practice of approximating intractable infinite dimensional nonlinear filters by simpler computationally tractable filters.
Resumo:
In this paper a real-time vision based power line extraction solution is investigated for active UAV guidance. The line extraction algorithm starts from ridge points detected by steerable filters. A collinear line segments fitting algorithm is followed up by considering global and local information together with multiple collinear measurements. GPU boosted algorithm implementation is also investigated in the experiment. The experimental result shows that the proposed algorithm outperforms two baseline line detection algorithms and is able to fitting long collinear line segments. The low computational cost of the algorithm make suitable for real-time applications.
Resumo:
Here we present a sequential Monte Carlo approach to Bayesian sequential design for the incorporation of model uncertainty. The methodology is demonstrated through the development and implementation of two model discrimination utilities; mutual information and total separation, but it can also be applied more generally if one has different experimental aims. A sequential Monte Carlo algorithm is run for each rival model (in parallel), and provides a convenient estimate of the marginal likelihood (of each model) given the data, which can be used for model comparison and in the evaluation of utility functions. A major benefit of this approach is that it requires very little problem specific tuning and is also computationally efficient when compared to full Markov chain Monte Carlo approaches. This research is motivated by applications in drug development and chemical engineering.