214 resultados para Monocular Vision.
Resumo:
This paper proposes new metrics and a performance-assessment framework for vision-based weed and fruit detection and classification algorithms. In order to compare algorithms, and make a decision on which one to use fora particular application, it is necessary to take into account that the performance obtained in a series of tests is subject to uncertainty. Such characterisation of uncertainty seems not to be captured by the performance metrics currently reported in the literature. Therefore, we pose the problem as a general problem of scientific inference, which arises out of incomplete information, and propose as a metric of performance the(posterior) predictive probabilities that the algorithms will provide a correct outcome for target and background detection. We detail the framework through which these predicted probabilities can be obtained, which is Bayesian in nature. As an illustration example, we apply the framework to the assessment of performance of four algorithms that could potentially be used in the detection of capsicums (peppers).
Resumo:
Recovering the motion of a non-rigid body from a set of monocular images permits the analysis of dynamic scenes in uncontrolled environments. However, the extension of factorisation algorithms for rigid structure from motion to the low-rank non-rigid case has proved challenging. This stems from the comparatively hard problem of finding a linear “corrective transform” which recovers the projection and structure matrices from an ambiguous factorisation. We elucidate that this greater difficulty is due to the need to find multiple solutions to a non-trivial problem, casting a number of previous approaches as alleviating this issue by either a) introducing constraints on the basis, making the problems nonidentical, or b) incorporating heuristics to encourage a diverse set of solutions, making the problems inter-dependent. While it has previously been recognised that finding a single solution to this problem is sufficient to estimate cameras, we show that it is possible to bootstrap this partial solution to find the complete transform in closed-form. However, we acknowledge that our method minimises an algebraic error and is thus inherently sensitive to deviation from the low-rank model. We compare our closed-form solution for non-rigid structure with known cameras to the closed-form solution of Dai et al. [1], which we find to produce only coplanar reconstructions. We therefore make the recommendation that 3D reconstruction error always be measured relative to a trivial reconstruction such as a planar one.
Resumo:
This paper provides a comprehensive review of the vision-based See and Avoid problem for unmanned aircraft. The unique problem environment and associated constraints are detailed, followed by an in-depth analysis of visual sensing limitations. In light of such detection and estimation constraints, relevant human, aircraft and robot collision avoidance concepts are then compared from a decision and control perspective. Remarks on system evaluation and certification are also included to provide a holistic review approach. The intention of this work is to clarify common misconceptions, realistically bound feasible design expectations and offer new research directions. It is hoped that this paper will help us to unify design efforts across the aerospace and robotics communities.
Resumo:
Purpose To provide a summary of the classic paper "Differences in the accommodation stimulus response curves of adult myopes and emmetropes" published in Ophthalmic and Physiological Optics in 1998 and to provide an update on the topic of accommodation errors in myopia. Summary The accommodation responses of 33 participants (10 emmetropes, 11 early onset myopes and 12 late onset myopes) aged 18-31 years were measured using the Canon Autoref R-1 free space autorefractor using three methods to vary the accommodation demand: decreasing distance (4 m to 0.25 cm), negative lenses (0 to -4 D at 4 m) and positive lenses (+4 to 0 D at 0.25 m). We observed that the greatest accommodation errors occurred for the negative lens method whereas minimal errors were observed using positive lenses. Adult progressing myopes had greater lags of accommodation than stable myopes at higher demands induced by negative lenses. Progressing myopes had shallower response gradients than the emmetropes and stable myopes; however the reduced gradient was much less than that observed in children using similar methods. Recent Findings This paper has been often cited as evidence that accommodation responses at near may be primarily reduced in adults with progressing myopia and not in stable myopes and/or that challenging accommodation stimuli (negative lenses with monocular viewing) are required to generate larger accommodation errors. As an analogy, animals reared with hyperopic errors develop axial elongation and myopia. Retinal defocus signals are presumably passed to the retinal pigment epithelium and choroid and then ultimately the sclera to modify eye length. A number of lens treatments that act to slow myopia progression may partially work through reducing accommodation errors.
Resumo:
The aim was to investigate the effects of the GABAB receptor antagonist, CGP46381, on form-deprivation myopia (FDM) in guinea pigs. Twenty-four guinea pigs had monocular visual deprivation induced using a diffuser for 11 days (day 14 to 25). The deprived eyes were treated with daily subconjunctival injections (100 μl) of either 2% CGP46381, 0.2% CGP46381, or saline or received no injection. The fellow eyes were left untreated. Another six animals received no treatment. At the start and end of the treatment period, ocular refractions were measured using retinoscopy and vitreous chamber depth (VCD) and axial length (AL) using A-scan ultrasound. All of the deprived eyes developed relative myopia (treated versus untreated eyes, P < 0.05). The amount of myopia was significantly affected by the drug treatment (one-way ANOVA, P < 0.0001). The highest dose tested, 2% CGP46381, significantly inhibited myopia development compared to saline (2% CGP46381: -1.08 ± 0.40 D, saline: -4.33 ± 0.67 D, P < 0.01). The majority of these effects were due to less AL (2% CGP46381: 0.03 ± 0.01 mm, saline: 0.13 ± 0.02 mm, P < 0.01) and VCD (2% CGP46381: 0.02 ± 0.01 mm, saline: 0.08 ± 0.01 mm, P < 0.01) elongation. The lower dose tested, 0.2% CGP46381, did not significantly inhibit FDM (P > 0.05). Subconjunctival injections of CGP46381 inhibit FDM development in guinea pigs in a dose-dependent manner.
Resumo:
Recent interest in affect and the body have mobilized a contemporary review of aesthetics and phenomenology within architecture to unpack how environments affect spatial experience. Emerging spatial studies within the neuro-sciences, and their implications for architectural research as raised by architectural theorists Juhani Pallasmaa (2014) and Harry Mallgrave (2013) has been well supported by a raft of scientists and institutions including the prestigious Salk Institute. Although there has been some headway in spatial studies of the vision impaired (Cattaneo et al, 2011) to understand the role of their non-visual systems in assisting navigation and location, little is discussed in terms of their other abilities in sensing particular qualities of space which impinge upon emotion. This paper reviews a collection of studies exploring face vision and echo-location, amongst others, which provide insight into what might be termed affective perception of the vision impaired, and how further interplay between this research and the architectural field can contribute new knowledge regarding space and affect. By engaging with themes from the Aesthetics, Phenomenology and indeed Neuro-science fields, the paper provides background of current and potential cross disciplinary research, and highlights the role wearable technologies can play in enhancing knowledge of affective spatial experience.
Resumo:
The vision sense of standalone robots is limited by line of sight and onboard camera capabilities, but processing video from remote cameras puts a high computational burden on robots. This paper describes the Distributed Robotic Vision Service, DRVS, which implements an on-demand distributed visual object detection service. Robots specify visual information requirements in terms of regions of interest and object detection algorithms. DRVS dynamically distributes the object detection computation to remote vision systems with processing capabilities, and the robots receive high-level object detection information. DRVS relieves robots of managing sensor discovery and reduces data transmission compared to image sharing models of distributed vision. Navigating a sensorless robot from remote vision systems is demonstrated in simulation as a proof of concept.
Resumo:
Background: Optometry students are taught the process of subjective refraction through lectures and laboratory based practicals before progressing to supervised clinical practice. Simulated learning environments (SLEs) are an emerging technology that are used in a range of health disciplines, however, there is limited evidence regarding the effectiveness of clinical simulators as an educational tool. Methods: Forty optometry students (20 fourth year and 20 fifth year) were assessed twice by a qualified optometrist (two examinations separated by 4-8 weeks) while completing a monocular non-cycloplegic subjective refraction on the same patient with an unknown refractive error simulated using contact lenses. Half of the students were granted access to an online SLE, The Brien Holden Vision Institute (BHVI®) Virtual Refractor, and the remaining students formed a control group. The primary outcome measures at each visit were; accuracy of the clinical refraction compared to a qualified optometrist and relative to the Optometry Council of Australia and New Zealand (OCANZ) subjective refraction examination criteria. Secondary measures of interest included descriptors of student SLE engagement, student self-reported confidence levels and correlations between performance in the simulated and real world clinical environment. Results: Eighty percent of students in the intervention group interacted with the SLE (for an average of 100 minutes); however, there was no correlation between measures of student engagement with the BHVI® Virtual Refractor and speed or accuracy of clinical subjective refractions. Fifth year students were typically more confident and refracted more accurately and quickly than fourth year students. A year group by experimental group interaction (p = 0.03) was observed for accuracy of the spherical component of refraction, and post hoc analysis revealed that less experienced students exhibited greater gains in clinical accuracy following exposure to the SLE intervention. Conclusions: Short-term exposure to a SLE can positively influence clinical subjective refraction outcomes for less experienced optometry students and may be of benefit in increasing the skills of novice refractionists to levels appropriate for commencing supervised clinical interactions.
Resumo:
Purpose This review assessed the effectiveness of diabetic retinopathy (DR) screening programs, using retinal photography in Australian urban and rural settings, and considered implications for public health strategy and policy. Methods An electronic search of MEDLINE, PubMed, and Embase for studies published between 1 January 1996 and the 30 June 2013 was undertaken. Key search terms were “diabetic retinopathy,” “screening,” “retinal photography” and “Australia.” Results Twelve peer-reviewed publications were identified. The 14 DR screening programs identified from the 12 publications were successfully undertaken in urban, rural and remote communities across Australia. Locations included a pathology collection center, and Indigenous primary health care and Aboriginal community controlled organizations. Each intervention using retinal photography was highly effective at increasing the number of people who underwent screening for DR. The review identified that prior to commencement of the screening programs a median of 48% (range 16–85%) of those screened had not undergone a retinal examination within the recommended time frame (every year for Indigenous people and every 2 years for non-Indigenous people in Australia). A median of 16% (range 0–45%) of study participants had evidence of DR. Conclusions This review has shown there have been many pilot and demonstration projects in rural and urban Australia that confirm the effectiveness of retinal photography-based screening for DR
Resumo:
Regenerative sustainability is emerging as an alternative discourse around the transition from a ‘mechanistic’ to an ‘ecological’ or living systems worldview. This view helps us to re-conceptualize relationships among humans’ technological, ecological, economic, social and political systems. Through exploration of ‘net positive’ or ‘regenerative’ development lenses and the traditional sustainability literature, the conceptualization and approaches to achieve sustainable development and ecological modernization are expanded to articulate and to explore the evolving sustainability discourse, ‘regenerative sustainability’. This Special Volume of Journal of Cleaner Production (SV) is focused upon various dimensions of regenerative sustainability (e.g. regenerative design, regenerative development, and positive development) applied to the urban built environment at scales, which range from individual buildings, neighborhoods, and urban developments to integrated regional sustainable development. The main focus is on how these approaches and developments are evolving, how they can help us to prevent or adapt to climate change and how these approaches are likely to evolve in the next two to three decades. These approaches are addressed in four themes: (1) reviewing the theoretical development of the discourse of regenerative sustainability, its emerging principles and practices, (2) explaining how it can be measured and monitored, (3) providing encouraging practical pathways and examples of its implementation in multiple cultural and climatic contexts, and (4) mapping obstacles and enablers that must be addressed to help to ensure that more rapid progress is made in implementing the transitions towards an urban built environment that supports genuinely sustainable societies.
Resumo:
In recent years more and more complex humanoid robots have been developed. On the other hand programming these systems has become more difficult. There is a clear need for such robots to be able to adapt and perform certain tasks autonomously, or even learn by themselves how to act. An important issue to tackle is the closing of the sensorimotor loop. Especially when talking about humanoids the tight integration of perception with actions will allow for improved behaviours, embedding adaptation on the lower-level of the system.
Resumo:
There is an increased interest on the use of Unmanned Aerial Vehicles (UAVs) for wildlife and feral animal monitoring around the world. This paper describes a novel system which uses a predictive dynamic application that places the UAV ahead of a user, with a low cost thermal camera, a small onboard computer that identifies heat signatures of a target animal from a predetermined altitude and transmits that target’s GPS coordinates. A map is generated and various data sets and graphs are displayed using a GUI designed for easy use. The paper describes the hardware and software architecture and the probabilistic model for downward facing camera for the detection of an animal. Behavioral dynamics of target movement for the design of a Kalman filter and Markov model based prediction algorithm are used to place the UAV ahead of the user. Geometrical concepts and Haversine formula are applied to the maximum likelihood case in order to make a prediction regarding a future state of the user, thus delivering a new way point for autonomous navigation. Results show that the system is capable of autonomously locating animals from a predetermined height and generate a map showing the location of the animals ahead of the user.
Resumo:
This paper introduces a machine learning based system for controlling a robotic manipulator with visual perception only. The capability to autonomously learn robot controllers solely from raw-pixel images and without any prior knowledge of configuration is shown for the first time. We build upon the success of recent deep reinforcement learning and develop a system for learning target reaching with a three-joint robot manipulator using external visual observation. A Deep Q Network (DQN) was demonstrated to perform target reaching after training in simulation. Transferring the network to real hardware and real observation in a naive approach failed, but experiments show that the network works when replacing camera images with synthetic images.
Resumo:
Recent interest in affect and the body have mobilised a contemporary review of aesthetics and phenomenology within architecture to unpack how environments affect spatial experience. Emerging spatial studies within the neurosciences, and their implications for architectural research as raised by architectural theorists has been well supported by a raft of scientists and institutions. Although there has been some headway in spatial studies of the vision impaired (Cattaneo et al., 2011) to understand the role of their non-visual systems in assisting navigation and location, little is discussed in terms of their other abilities in sensing particular qualities of space which impinge upon emotion and wellbeing. This research explores, through published studies and constructed spatial interviews, the affective perception of the vision impaired and how further interplay between this research and the architectural field can contribute new knowledge regarding space and affect. The research aims to provide background of current and potential cross disciplinary research and highlight the role wearable technologies can play in enhancing knowledge of affective spatial experience.
Resumo:
Evidence has accumulated that rod activation under mesopic and scotopic light levels alters visual perception and performance. Here we review the most recent developments in the measurement of rod and cone contributions to mesopic color perception and temporal processing, with a focus on data measured using the four-primary photostimulator method that independently controls rod and cone excitations. We discuss the findings in the context of rod inputs to the three primary retinogeniculate pathways to understand rod contributions to mesopic vision. Additionally, we present evidence that hue perception is possible under scotopic, pure rod-mediated conditions that involves cortical mechanisms.