963 resultados para random search algorithms
Resumo:
Perceived depth was measured for three-types of stereograms with the colour/texture of half-occluded (monocular) regions either similar to or dissimilar to that of binocular regions or background. In a two-panel random dot stereogram the monocular region was filled with texture either similar or different to the far panel or left blank. In unpaired background stereograms the monocular region either matched the background or was different in colour or texture and in phantom stereograms the monocular region matched the partially occluded object or was a different colour or texture. In all three cases depth was considerably impaired when the monocular texture did not match either the background or the more distant surface. The content and context of monocular regions as well as their position are important in determining their role as occlusion cues and thus in three-dimensional layout. We compare coincidence and accidental view accounts of these effects. (C) 2002 Elsevier Science Ltd. All rights reserved.
Resumo:
A two-component survival mixture model is proposed to analyse a set of ischaemic stroke-specific mortality data. The survival experience of stroke patients after index stroke may be described by a subpopulation of patients in the acute condition and another subpopulation of patients in the chronic phase. To adjust for the inherent correlation of observations due to random hospital effects, a mixture model of two survival functions with random effects is formulated. Assuming a Weibull hazard in both components, an EM algorithm is developed for the estimation of fixed effect parameters and variance components. A simulation study is conducted to assess the performance of the two-component survival mixture model estimators. Simulation results confirm the applicability of the proposed model in a small sample setting. Copyright (C) 2004 John Wiley Sons, Ltd.
Resumo:
Ohman and colleagues provided evidence for preferential processing of pictures depicting fear-relevant animals by showing that pictures of snakes and spiders are found faster among pictures of fiowers and mushrooms than vice versa and that the speed of detecting fear-relevant animals was not affected by set size whereas the speed of detecting fiowers/mushrooms was. Experiment 1 replicated this finding. Experiment 2, however, found similar search advantages when pictures of cats and horses or of wolves and big cats were to be found among pictures of flowers and mushrooms. Moreover, Experiment 3, in a within subject comparison, failed to find faster identification of snakes and spiders than of cats and horses among flowers and mushrooms. The present findings seem to indicate that previous reports of preferential processing of pictures of snakes and spiders in a visual search task may reflect a processing advantage for animal pictures in general rather than fear-relevance.
Resumo:
The image reconstruction using the EIT (Electrical Impedance Tomography) technique is a nonlinear and ill-posed inverse problem which demands a powerful direct or iterative method. A typical approach for solving the problem is to minimize an error functional using an iterative method. In this case, an initial solution close enough to the global minimum is mandatory to ensure the convergence to the correct minimum in an appropriate time interval. The aim of this paper is to present a new, simple and low cost technique (quadrant-searching) to reduce the search space and consequently to obtain an initial solution of the inverse problem of EIT. This technique calculates the error functional for four different contrast distributions placing a large prospective inclusion in the four quadrants of the domain. Comparing the four values of the error functional it is possible to get conclusions about the internal electric contrast. For this purpose, initially we performed tests to assess the accuracy of the BEM (Boundary Element Method) when applied to the direct problem of the EIT and to verify the behavior of error functional surface in the search space. Finally, numerical tests have been performed to verify the new technique.
Resumo:
The main focus of this essay is the first American round-the-world scientific voyage, the U. S Exploring Expedition, which took place between 1838 and 1841 and was lead by Lieutenant Charles Wilkes. Here, I discuss the purposes of this expedition in the context of the voyages of circumnavigation accomplished by the various European powers during the same period.
Resumo:
Chronic beryllium disease (CBD) is clinically similar to other granulomatous diseases such as sarcoidosis. It is often misdiagnosed if a thorough occupational history is not taken. When appropriate, a beryllium lymphocyte proliferation tests (BeLPT) need to be performed. We aimed to search for CBD among currently diagnosed pulmonary sarcoidosis patients and to identify the occupations and exposures in Ontario leading to CBD. Questionnaire items included work history and details of possible exposure to beryllium. Participants who provided a history of previous work with metals underwent BeLPTs and an ELISPOT on the basis of having a higher pretest probability of CBD. Among 121 sarcoid patients enrolled, 87 (72%) reported no known previous metal dust or fume exposure, while 34 (28%) had metal exposure, including 17 (14%) with beryllium exposure at work or home. However, none of these 34 who underwent testing had positive test results. Self-reported exposure to beryllium or metals was relatively common in these patients with clinical sarcoidosis, but CBD was not confirmed using blood assays in this population.
Resumo:
Minimal perfect hash functions are used for memory efficient storage and fast retrieval of items from static sets. We present an infinite family of efficient and practical algorithms for generating order preserving minimal perfect hash functions. We show that almost all members of the family construct space and time optimal order preserving minimal perfect hash functions, and we identify the one with minimum constants. Members of the family generate a hash function in two steps. First a special kind of function into an r-graph is computed probabilistically. Then this function is refined deterministically to a minimal perfect hash function. We give strong theoretical evidence that the first step uses linear random time. The second step runs in linear deterministic time. The family not only has theoretical importance, but also offers the fastest known method for generating perfect hash functions.
Resumo:
A robust semi-implicit central partial difference algorithm for the numerical solution of coupled stochastic parabolic partial differential equations (PDEs) is described. This can be used for calculating correlation functions of systems of interacting stochastic fields. Such field equations can arise in the description of Hamiltonian and open systems in the physics of nonlinear processes, and may include multiplicative noise sources. The algorithm can be used for studying the properties of nonlinear quantum or classical field theories. The general approach is outlined and applied to a specific example, namely the quantum statistical fluctuations of ultra-short optical pulses in chi((2)) parametric waveguides. This example uses a non-diagonal coherent state representation, and correctly predicts the sub-shot noise level spectral fluctuations observed in homodyne detection measurements. It is expected that the methods used wilt be applicable for higher-order correlation functions and other physical problems as well. A stochastic differencing technique for reducing sampling errors is also introduced. This involves solving nonlinear stochastic parabolic PDEs in combination with a reference process, which uses the Wigner representation in the example presented here. A computer implementation on MIMD parallel architectures is discussed. (C) 1997 Academic Press.
Resumo:
Objective: The study we assessed how often patients who are manifesting a myocardial infarction (MI) would not be considered candidates for intensive lipid-lowering therapy based on the current guidelines. Methods: In 355 consecutive patients manifesting ST elevation MI (STEMI), admission plasma C-reactive protein (CRP) was measured and Framingham risk score (FRS), PROCAM risk score, Reynolds risk score, ASSIGN risk score, QRISK, and SCORE algorithms were applied. Cardiac computed tomography and carotid ultrasound were performed to assess the coronary artery calcium score (CAC), carotid intima-media thickness (cIMT) and the presence of carotid plaques. Results: Less than 50% of STEMI patients would be identified as having high risk before the event by any of these algorithms. With the exception of FRS (9%), all other algorithms would assign low risk to about half of the enrolled patients. Plasma CRP was <1.0 mg/L in 70% and >2 mg/L in 14% of the patients. The average cIMT was 0.8 +/- 0.2 mm and only in 24% of patients was >= 1.0 mm. Carotid plaques were found in 74% of patients. CAC > 100 was found in 66% of patients. Adding CAC >100 plus the presence of carotid plaque, a high-risk condition would be identified in 100% of the patients using any of the above mentioned algorithms. Conclusion: More than half of patients manifesting STEMI would not be considered as candidates for intensive preventive therapy by the current clinical algorithms. The addition of anatomical parameters such as CAC and the presence of carotid plaques can substantially reduce the CVD risk underestimation. (C) 2010 Elsevier Ireland Ltd. All rights reserved.
Resumo:
A new conceptual model for soil pore-solid structure is formalized. Soil pore-solid structure is proposed to comprise spatially abutting elements each with a value which is its membership to the fuzzy set ''pore,'' termed porosity. These values have a range between zero (all solid) and unity (all pore). Images are used to represent structures in which the elements are pixels and the value of each is a porosity. Two-dimensional random fields are generated by allocating each pixel a porosity by independently sampling a statistical distribution. These random fields are reorganized into other pore-solid structural types by selecting parent points which have a specified local region of influence. Pixels of larger or smaller porosity are aggregated about the parent points and within the region of interest by controlled swapping of pixels in the image. This creates local regions of homogeneity within the random field. This is similar to the process known as simulated annealing. The resulting structures are characterized using one-and two-dimensional variograms and functions describing their connectivity. A variety of examples of structures created by the model is presented and compared. Extension to three dimensions presents no theoretical difficulties and is currently under development.
Resumo:
The concept of parameter-space size adjustment is pn,posed in order to enable successful application of genetic algorithms to continuous optimization problems. Performance of genetic algorithms with six different combinations of selection and reproduction mechanisms, with and without parameter-space size adjustment, were severely tested on eleven multiminima test functions. An algorithm with the best performance was employed for the determination of the model parameters of the optical constants of Pt, Ni and Cr.
Resumo:
The present study was carried out to evaluate the effectiveness of a specific program regarding the occurrence of vocal attrition symptoms in telemarketers. A total of 71 subjects participated in this study: 28 completed the Vocal Symptoms questionnaire to test its reliability, and 43 were randomly assigned to two groups: an 8-week vocal training group (n = 14) and a no-training control group (n = 29), to evaluate the effectiveness of the training program with this tool. The voice training group also filled in the posttraining questionnaire `Benefits Obtained with Voice Training` (BVT). The vocal training program was not considered effective with regard to the occurrence of vocal symptoms. However, due to a probable increase in symptoms in untrained telemarketers, it can work as a protective factor. According to BVT answers, the vocal training contributed to an improvement in vocal use as a communication tool for telemarketers. Copyright (C) 2009 S. Karger AG, Basel