995 resultados para Human Errors


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Noradrenergic neurotransmission has been associated with the modulation of higher cognitive functions mediated by the prefrontal cortex. In the present study, the impact of noradrenergic stimulation on the human action-monitoring system, as indexed by eventrelated brain potentials, was examined. After the administration of a placebo or the selective 2 -adrenoceptor antagonist yohimbine, which stimulates firing in the locus ceruleus and noradrenaline release, electroencephalograpic recordings were obtained from healthy volunteers performing a letter flanker task. Yohimbine led to an increase in the amplitude of the error-related negativity in conjunction with a significant reduction of action errors. Reaction times were unchanged, and the drug did not modify the N2 in congruent versus incongruent trials, a measure of preresponse conflict, or posterror adjustments as measured by posterror slowing of reaction time. The present findings suggest that the locus ceruleusnoradrenaline system exerts a rather specific effect on human action monitoring.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

There are many ways to generate geometrical models for numerical simulation, and most of them start with a segmentation step to extract the boundaries of the regions of interest. This paper presents an algorithm to generate a patient-specific three-dimensional geometric model, based on a tetrahedral mesh, without an initial extraction of contours from the volumetric data. Using the information directly available in the data, such as gray levels, we built a metric to drive a mesh adaptation process. The metric is used to specify the size and orientation of the tetrahedral elements everywhere in the mesh. Our method, which produces anisotropic meshes, gives good results with synthetic and real MRI data. The resulting model quality has been evaluated qualitatively and quantitatively by comparing it with an analytical solution and with a segmentation made by an expert. Results show that our method gives, in 90% of the cases, as good or better meshes as a similar isotropic method, based on the accuracy of the volume reconstruction for a given mesh size. Moreover, a comparison of the Hausdorff distances between adapted meshes of both methods and ground-truth volumes shows that our method decreases reconstruction errors faster. Copyright © 2015 John Wiley & Sons, Ltd.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The present study is an attempt to highlight the problem of typographical errors in OPACS. The errors made while typing catalogue entries as well as importing bibliographical records from other libraries exist unnoticed by librarians resulting the non-retrieval of available records and affecting the quality of OPACs. This paper follows previous research on the topic mainly by Jeffrey Beall and Terry Ballard. The word “management” was chosen from the list of likely to be misspelled words identified by previous research. It was found that the word is wrongly entered in several forms in local, national and international OPACs justifying the observations of Ballard that typos occur in almost everywhere. Though there are lots of corrective measures proposed and are in use, the study asserts the fact that human effort is needed to get rid of the problem. The paper is also an invitation to the library professionals and system designers to construct a strategy to solve the issue

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The present study is an attempt to highlight the problem of typographical errors in OPACS. The errors made while typing catalogue entries as well as importing bibliographical records from other libraries exist unnoticed by librarians resulting the non-retrieval of available records and affecting the quality of OPACs. This paper follows previous research on the topic mainly by Jeffrey Beall and Terry Ballard. The word “management” was chosen from the list of likely to be misspelled words identified by previous research. It was found that the word is wrongly entered in several forms in local, national and international OPACs justifying the observations of Ballard that typos occur in almost everywhere. Though there are lots of corrective measures proposed and are in use, the study asserts the fact that human effort is needed to get rid of the problem. The paper is also an invitation to the library professionals and system designers to construct a strategy to solve the issue

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In an immersive virtual environment, observers fail to notice the expansion of a room around them and consequently make gross errors when comparing the size of objects. This result is difficult to explain if the visual system continuously generates a 3-D model of the scene based on known baseline information from interocular separation or proprioception as the observer walks. An alternative is that observers use view-based methods to guide their actions and to represent the spatial layout of the scene. In this case, they may have an expectation of the images they will receive but be insensitive to the rate at which images arrive as they walk. We describe the way in which the eye movement strategy of animals simplifies motion processing if their goal is to move towards a desired image and discuss dorsal and ventral stream processing of moving images in that context. Although many questions about view-based approaches to scene representation remain unanswered, the solutions are likely to be highly relevant to understanding biological 3-D vision.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In an immersive virtual reality environment, subjects fail to notice when a scene expands or contracts around them, despite correct and consistent information from binocular stereopsis and motion parallax, resulting in gross failures of size constancy (A. Glennerster, L. Tcheang, S. J. Gilson, A. W. Fitzgibbon, & A. J. Parker, 2006). We determined whether the integration of stereopsis/motion parallax cues with texture-based cues could be modified through feedback. Subjects compared the size of two objects, each visible when the room was of a different size. As the subject walked, the room expanded or contracted, although subjects failed to notice any change. Subjects were given feedback about the accuracy of their size judgments, where the “correct” size setting was defined either by texture-based cues or (in a separate experiment) by stereo/motion parallax cues. Because of feedback, observers were able to adjust responses such that fewer errors were made. For texture-based feedback, the pattern of responses was consistent with observers weighting texture cues more heavily. However, for stereo/motion parallax feedback, performance in many conditions became worse such that, paradoxically, biases moved away from the point reinforced by the feedback. This can be explained by assuming that subjects remap the relationship between stereo/motion parallax cues and perceived size or that they develop strategies to change their criterion for a size match on different trials. In either case, subjects appear not to have direct access to stereo/motion parallax cues.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this chapter we described how the inclusion of a model of a human arm, combined with the measurement of its neural input and a predictor, can provide to a previously proposed teleoperator design robustness under time delay. Our trials gave clear indications of the superiority of the NPT scheme over traditional as well as the modified Yokokohji and Yoshikawa architectures. Its fundamental advantages are: the time-lead of the slave, the more efficient, and providing a more natural feeling manipulation, and the fact that incorporating an operator arm model leads to more credible stability results. Finally, its simplicity allows less likely to fail local control techniques to be employed. However, a significant advantage for the enhanced Yokokohji and Yoshikawa architecture results from the very fact that it’s a conservative modification of current designs. Under large prediction errors, it can provide robustness through directing the master and slave states to their means and, since it relies on the passivity of the mechanical part of the system, it would not confuse the operator. An experimental implementation of the techniques will provide further evidence for the performance of the proposed architectures. The employment of neural networks and fuzzy logic, which will provide an adaptive model of the human arm and robustifying control terms, is scheduled for the near future.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We perform a multimodel detection and attribution study with climate model simulation output and satellite-based measurements of tropospheric and stratospheric temperature change. We use simulation output from 20 climate models participating in phase 5 of the Coupled Model Intercomparison Project. This multimodel archive provides estimates of the signal pattern in response to combined anthropogenic and natural external forcing (the finger-print) and the noise of internally generated variability. Using these estimates, we calculate signal-to-noise (S/N) ratios to quantify the strength of the fingerprint in the observations relative to fingerprint strength in natural climate noise. For changes in lower stratospheric temperature between 1979 and 2011, S/N ratios vary from 26 to 36, depending on the choice of observational dataset. In the lower troposphere, the fingerprint strength in observations is smaller, but S/N ratios are still significant at the 1% level or better, and range from three to eight. We find no evidence that these ratios are spuriously inflated by model variability errors. After removing all global mean signals, model fingerprints remain identifiable in 70% of the tests involving tropospheric temperature changes. Despite such agreement in the large-scale features of model and observed geographical patterns of atmospheric temperature change, most models do not replicate the size of the observed changes. On average, the models analyzed underestimate the observed cooling of the lower stratosphere and overestimate the warming of the troposphere. Although the precise causes of such differences are unclear, model biases in lower stratospheric temperature trends are likely to be reduced by more realistic treatment of stratospheric ozone depletion and volcanic aerosol forcing.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We have considered a Bayesian approach for the nonlinear regression model by replacing the normal distribution on the error term by some skewed distributions, which account for both skewness and heavy tails or skewness alone. The type of data considered in this paper concerns repeated measurements taken in time on a set of individuals. Such multiple observations on the same individual generally produce serially correlated outcomes. Thus, additionally, our model does allow for a correlation between observations made from the same individual. We have illustrated the procedure using a data set to study the growth curves of a clinic measurement of a group of pregnant women from an obstetrics clinic in Santiago, Chile. Parameter estimation and prediction were carried out using appropriate posterior simulation schemes based in Markov Chain Monte Carlo methods. Besides the deviance information criterion (DIC) and the conditional predictive ordinate (CPO), we suggest the use of proper scoring rules based on the posterior predictive distribution for comparing models. For our data set, all these criteria chose the skew-t model as the best model for the errors. These DIC and CPO criteria are also validated, for the model proposed here, through a simulation study. As a conclusion of this study, the DIC criterion is not trustful for this kind of complex model.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A new form of composition of the indicators employed to generate the United Nations Human Development Index (HDI) is presented here. This form of composition is based on the assumption that random errors affect the measurement of each indicator. This assumption allows for replacing the vector of evaluations according to each indicator by vectors of probabilities of being the best or the worst according to such attribute. The probabilistic composition of such probabilities of preference according to each indicator into probabilities of being the best or the worst according to all of them generates indices that may unveil, on one hand, performances to be followed and, on the other hand, extreme conditions that an additive composition would hide. Differences between the results of application of the diverse forms of composition are examined in the case of the HDI and in the case of the districts version of the HDI employed to compare Brazilian municipalities. It is verified that the smallest correlation between the education.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

3D video-fluoroscopy is an accurate but cumbersome technique to estimate natural or prosthetic human joint kinematics. This dissertation proposes innovative methodologies to improve the 3D fluoroscopic analysis reliability and usability. Being based on direct radiographic imaging of the joint, and avoiding soft tissue artefact that limits the accuracy of skin marker based techniques, the fluoroscopic analysis has a potential accuracy of the order of mm/deg or better. It can provide fundamental informations for clinical and methodological applications, but, notwithstanding the number of methodological protocols proposed in the literature, time consuming user interaction is exploited to obtain consistent results. The user-dependency prevented a reliable quantification of the actual accuracy and precision of the methods, and, consequently, slowed down the translation to the clinical practice. The objective of the present work was to speed up this process introducing methodological improvements in the analysis. In the thesis, the fluoroscopic analysis was characterized in depth, in order to evaluate its pros and cons, and to provide reliable solutions to overcome its limitations. To this aim, an analytical approach was followed. The major sources of error were isolated with in-silico preliminary studies as: (a) geometric distortion and calibration errors, (b) 2D images and 3D models resolutions, (c) incorrect contour extraction, (d) bone model symmetries, (e) optimization algorithm limitations, (f) user errors. The effect of each criticality was quantified, and verified with an in-vivo preliminary study on the elbow joint. The dominant source of error was identified in the limited extent of the convergence domain for the local optimization algorithms, which forced the user to manually specify the starting pose for the estimating process. To solve this problem, two different approaches were followed: to increase the optimal pose convergence basin, the local approach used sequential alignments of the 6 degrees of freedom in order of sensitivity, or a geometrical feature-based estimation of the initial conditions for the optimization; the global approach used an unsupervised memetic algorithm to optimally explore the search domain. The performances of the technique were evaluated with a series of in-silico studies and validated in-vitro with a phantom based comparison with a radiostereometric gold-standard. The accuracy of the method is joint-dependent, and for the intact knee joint, the new unsupervised algorithm guaranteed a maximum error lower than 0.5 mm for in-plane translations, 10 mm for out-of-plane translation, and of 3 deg for rotations in a mono-planar setup; and lower than 0.5 mm for translations and 1 deg for rotations in a bi-planar setups. The bi-planar setup is best suited when accurate results are needed, such as for methodological research studies. The mono-planar analysis may be enough for clinical application when the analysis time and cost may be an issue. A further reduction of the user interaction was obtained for prosthetic joints kinematics. A mixed region-growing and level-set segmentation method was proposed and halved the analysis time, delegating the computational burden to the machine. In-silico and in-vivo studies demonstrated that the reliability of the new semiautomatic method was comparable to a user defined manual gold-standard. The improved fluoroscopic analysis was finally applied to a first in-vivo methodological study on the foot kinematics. Preliminary evaluations showed that the presented methodology represents a feasible gold-standard for the validation of skin marker based foot kinematics protocols.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Turfgrasses are ubiquitous in urban landscape and their role on carbon (C) cycle is increasing important also due to the considerable footprint related to their management practices. It is crucial to understand the mechanisms driving the C assimilation potential of these terrestrial ecosystems Several approaches have been proposed to assess C dynamics: micro-meteorological methods, small-chamber enclosure system (SC), chrono-sequence approach and various models. Natural and human-induced variables influence turfgrasses C fluxes. Species composition, environmental conditions, site characteristics, former land use and agronomic management are the most important factors considered in literature driving C sequestration potential. At the same time different approaches seem to influence C budget estimates. In order to study the effect of different management intensities on turfgrass, we estimated net ecosystem exchange (NEE) through a SC approach in a hole of a golf course in the province of Verona (Italy) for one year. The SC approach presented several advantages but also limits related to the measurement frequency, timing and duration overtime, and to the methodological errors connected to the measuring system. Daily CO2 fluxes changed according to the intensity of maintenance, likely due to different inputs and disturbances affecting biogeochemical cycles, combined also to the different leaf area index (LAI). The annual cumulative NEE decreased with the increase of the intensity of management. NEE was related to the seasonality of turfgrass, following temperatures and physiological activity. Generally on the growing season CO2 fluxes towards atmosphere exceeded C sequestered. The cumulative NEE showed a system near to a steady state for C dynamics. In the final part greenhouse gases (GHGs) emissions due to fossil fuel consumption for turfgrass upkeep were estimated, pinpointing that turfgrass may result a considerable C source. The C potential of trees and shrubs needs to be considered to obtain a complete budget.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Numerical simulations of eye globes often rely on topographies that have been measured in vivo using devices such as the Pentacam or OCT. The topographies, which represent the form of the already stressed eye under the existing intraocular pressure, introduce approximations in the analysis. The accuracy of the simulations could be improved if either the stress state of the eye under the effect of intraocular pressure is determined, or the stress-free form of the eye estimated prior to conducting the analysis. This study reviews earlier attempts to address this problem and assesses the performance of an iterative technique proposed by Pandolfi and Holzapfel [1], which is both simple to implement and promises high accuracy in estimating the eye's stress-free form. A parametric study has been conducted and demonstrated reliance of the error level on the level of flexibility of the eye model, especially in the cornea region. However, in all cases considered 3-4 analysis iterations were sufficient to produce a stress-free form with average errors in node location <10(-6)mm and a maximal error <10(-4)mm. This error level, which is similar to what has been achieved with other methods and orders of magnitude lower than the accuracy of current clinical topography systems, justifies the use of the technique as a pre-processing step in ocular numerical simulations.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In the anti-saccade paradigm, subjects are instructed not to make a reflexive saccade to an appearing lateral target but to make an intentional saccade to the opposite side instead. The inhibition of reflexive saccade triggering is under the control of the dorsolateral prefrontal cortex (DLPFC). The critical time interval at which this inhibition takes place during the paradigm, however, is not exactly known. In the present study, we used single-pulse transcranial magnetic stimulation (TMS) to interfere with DLPFC function in 15 healthy subjects. TMS was applied over the right DLPFC either 100 ms before the onset of the visual target (i.e. -100 ms), at target onset (i.e. 0 ms) or 100 ms after target onset (i.e. +100 ms). Stimulation 100 ms before target onset significantly increased the percentage of anti-saccade errors to both sides, while stimulation at, or after, target onset had no significant effect. All three stimulation conditions had no significant influence on saccade latency of correct or erroneous anti-saccades. These findings show that the critical time interval at which the DLPFC controls the suppression of a reflexive saccade in the anti-saccade paradigm is before target onset. In addition, the results suggest the view that the triggering of correct anti-saccades is not under direct control of the DLPFC.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The planning of refractive surgical interventions is a challenging task. Numerical modeling has been proposed as a solution to support surgical intervention and predict the visual acuity, but validation on patient specific intervention is missing. The purpose of this study was to validate the numerical predictions of the post-operative corneal topography induced by the incisions required for cataract surgery. The corneal topography of 13 patients was assessed preoperatively and postoperatively (1-day and 30-day follow-up) with a Pentacam tomography device. The preoperatively acquired geometric corneal topography – anterior, posterior and pachymetry data – was used to build patient-specific finite element models. For each patient, the effects of the cataract incisions were simulated numerically and the resulting corneal surfaces were compared to the clinical postoperative measurements at one day and at 30-days follow up. Results showed that the model was able to reproduce experimental measurements with an error on the surgically induced sphere of 0.38D one day postoperatively and 0.19D 30 days postoperatively. The standard deviation of the surgically induced cylinder was 0.54D at the first postoperative day and 0.38D 30 days postoperatively. The prediction errors in surface elevation and curvature were below the topography measurement device accuracy of ±5μm and ±0.25D after the 30-day follow-up. The results showed that finite element simulations of corneal biomechanics are able to predict post cataract surgery within topography measurement device accuracy. We can conclude that the numerical simulation can become a valuable tool to plan corneal incisions in cataract surgery and other ophthalmosurgical procedures in order to optimize patients' refractive outcome and visual function.