992 resultados para Distance convex simple graphs
Resumo:
Finding an adequate paraphrase representation formalism is a challenging issue in Natural Language Processing. In this paper, we analyse the performance of Tree Edit Distance as a paraphrase representation baseline. Our experiments using Edit Distance Textual Entailment Suite show that, as Tree Edit Distance consists of a purely syntactic approach, paraphrase alternations not based on structural reorganizations do not find an adequate representation. They also show that there is much scope for better modelling of the way trees are aligned.
Resumo:
A new technique capable of obtaining quantitative values of the rotation angle of the polarization vector by using holography is presented. This is a two-stage holographic process; during the recording stage a hologram of the object of interest is obtained. The reference beam is composed of two beams that form a small angle between them and keep their polarization states at right angles to each other. In the reconstruction stage of the hologram, two images from the hologram are obtained along two different angles. As a result of the interference between these two images, a set of parallel fringes is formed at the image plane. The fringe contrast on the reconstruction is related to the angle of the polarization vector of the light at each position on the image plane. Measurements of the rotation of the polarization angle of a fraction of a degree were obtained. The main application of this technique is in the study of transient phenomena, where single-shot measurements are the only means of obtaining reliable data.
Resumo:
Even though much improvement has been made in plant transformation methods, the screening of transgenic plants is often a laborious work. Most approaches for detecting the transgene in transformed plants are still timeconsuming, and can be quite expensive. The objective of this study was to search for a simpler method to screen for transgenic plants. The infiltration of kanamycin (100 mg/mL) into tobacco leaves resulted in conspicuous chlorotic spots on the non-transgenic plant leaves, while no spots were seen on the leaves of transformed plants. This reaction occurred regardless of age of the tested plants, and the method has proven to be simple, fast, non-destructive, relatively cheap, and reliable. These results were comparable to those obtained by the polymerase chain reaction (PCR) amplification of the transgene using specific primers.
Resumo:
A simple method determining airborne monoethanolamine has been developed. Monoethanolamine determination has traditionally been difficult due to analytical separation problems. Even in recent sophisticated methods, this difficulty remains as the major issue often resulting in time-consuming sample preparations. Impregnated glass fiber filters were used for sampling. Desorption of monoethanolamine was followed by capillary GC analysis and nitrogen phosphorous selective detection. Separation was achieved using a specific column for monoethanolamines (35% diphenyl and 65% dimethyl polysiloxane). The internal standard was quinoline. Derivatization steps were not needed. The calibration range was 0.5-80 μg/mL with a good correlation (R(2) = 0.996). Averaged overall precisions and accuracies were 4.8% and -7.8% for intraday (n = 30), and 10.5% and -5.9% for interday (n = 72). Mean recovery from spiked filters was 92.8% for the intraday variation, and 94.1% for the interday variation. Monoethanolamine on stored spiked filters was stable for at least 4 weeks at 5°C. This newly developed method was used among professional cleaners and air concentrations (n = 4) were 0.42 and 0.17 mg/m(3) for personal and 0.23 and 0.43 mg/m(3) for stationary measurements. The monoethanolamine air concentration method described here was simple, sensitive, and convenient both in terms of sampling and analytical analysis.
Resumo:
[spa] Se presenta un nuevo modelo para la toma de decisiones basado en el uso de medidas de distancia y de operadores de agregación inducidos. Se introduce la distancia media ponderada ordenada inducida (IOWAD). Es un nuevo operador de agregación que extiende el operador OWA a través del uso de distancias y un proceso de reordenación de los argumentos basado en variables de ordenación inducidas. La principal ventaja el operador IOWAD es la posibilidad de utilizar una familia parametrizada de operadores de agregación entre la distancia individual máxima y la mínima. Se estudian algunas de sus principales propiedades y algunos casos particulares. Se desarrolla un ejemplo numérico en un problema de toma de decisiones sobre selección de inversiones. Se observa que la principal ventaja de este modelo en la toma de decisiones es la posibilidad de mostrar una visión más completa del proceso, de forma que el decisor está capacitado para seleccionar la alternativa que está más cerca de sus intereses.
Resumo:
[spa] Se presenta un nuevo modelo para la toma de decisiones basado en el uso de medidas de distancia y de operadores de agregación inducidos. Se introduce la distancia media ponderada ordenada inducida (IOWAD). Es un nuevo operador de agregación que extiende el operador OWA a través del uso de distancias y un proceso de reordenación de los argumentos basado en variables de ordenación inducidas. La principal ventaja el operador IOWAD es la posibilidad de utilizar una familia parametrizada de operadores de agregación entre la distancia individual máxima y la mínima. Se estudian algunas de sus principales propiedades y algunos casos particulares. Se desarrolla un ejemplo numérico en un problema de toma de decisiones sobre selección de inversiones. Se observa que la principal ventaja de este modelo en la toma de decisiones es la posibilidad de mostrar una visión más completa del proceso, de forma que el decisor está capacitado para seleccionar la alternativa que está más cerca de sus intereses.
Resumo:
Background: Modelling epidemiological knowledge in validated clinical scores is a practical mean of integrating EBM to usual care. Existing scores about cardiovascular disease have been largely developed in emergency settings, but few in primary care. Such a toll is needed for general practitioners (GP) to evaluate the probability of ischemic heart disease (IHD) in patients with non-traumatic chest pain. Objective: To develop a predictive model to use as a clinical score for detecting IHD in patients with non-traumatic chest-pain in primary care. Methods: A post-hoc secondary analysis on data from an observational study including 672 patients with chest pain of which 85 had IHD diagnosed by their GP during the year following their inclusion. Best subset method was used to select 8 predictive variables from univariate analysis and fitted in a multivariate logistic regression model to define the score. Reliability of the model was assessed using split-group method. Results: Significant predictors were: age (0-3 points), gender (1 point), having at least one cardiovascular risks factor (hypertension, dyslipidemia, diabetes, smoking, family history of CVD; 3 points), personal history of cardiovascular disease (1 point), duration of chest pain from 1 to 60 minutes (2 points), substernal chest pain (1 point), pain increasing with exertion (1 point) and absence of tenderness at palpation (1 point). Area under the ROC curve for the score was of 0.95 (IC95% 0.93; 0.97). Patients were categorised in three groups, low risk of IHD (score under 6; n = 360), moderate risk of IHD (score from 6 to 8; n = 187) and high risk of IHD (score from 9-13; n = 125). Prevalence of IHD in each group was respectively of 0%, 6.7%, 58.5%. Reliability of the model seems satisfactory as the model developed from the derivation set predicted perfectly (p = 0.948) the number of patients in each group in the validation set. Conclusion: This clinical score based only on history and physical exams can be an important tool in the practice of the general physician for the prediction of ischemic heart disease in patients complaining of chest pain. The score below 6 points (in more than half of our population) can avoid demanding complementary exams for selected patients (ECG, laboratory tests) because of the very low risk of IHD. Score above 6 points needs investigation to detect or rule out IHD. Further external validation is required in ambulatory settings.
Resumo:
In the early 1900s, the wolf (Canis lupus) was extirpated from France and Switzerland. There is growing evidence that the species is presently recolonizing these countries in the western Alps. By sequencing the mitochondrial DNA (mtDNA) control region of various samples mainly collected in the field (scats, hairs, regurgitates, blood or tissue; n = 292), we could (1) develop a non-invasive method enabling the unambiguous attribution of these samples to wolf, fox (Vulpes vulpes) or dog (Canis familiaris), among others; (2) demonstrate that Italian, French and Swiss wolves share the same mtDNA haplotype, a haplotype that has never been found in any other wolf population world-wide. Combined together, field and genetic data collected over 10 years corroborate the scenario of a natural expansion of wolves from the Italian source population. Furthermore, such a genetic approach is of conservation significance, since it has important consequences for management decisions. This first long-term report using non-invasive sampling demonstrates that long-distance dispersers are common, supporting the hypothesis that individuals may often attempt to colonize far from their native pack, even in the absence of suitable corridors across habitats characterized by intense human activities.
Resumo:
We propose a compressive sensing algorithm that exploits geometric properties of images to recover images of high quality from few measurements. The image reconstruction is done by iterating the two following steps: 1) estimation of normal vectors of the image level curves, and 2) reconstruction of an image fitting the normal vectors, the compressed sensing measurements, and the sparsity constraint. The proposed technique can naturally extend to nonlocal operators and graphs to exploit the repetitive nature of textured images to recover fine detail structures. In both cases, the problem is reduced to a series of convex minimization problems that can be efficiently solved with a combination of variable splitting and augmented Lagrangian methods, leading to fast and easy-to-code algorithms. Extended experiments show a clear improvement over related state-of-the-art algorithms in the quality of the reconstructed images and the robustness of the proposed method to noise, different kind of images, and reduced measurements.
Resumo:
Collection : Les archives de la Révolution française ; 8.450
Resumo:
Blowing and drifting of snow is a major concern for transportation efficiency and road safety in regions where their development is common. One common way to mitigate snow drift on roadways is to install plastic snow fences. Correct design of snow fences is critical for road safety and maintaining the roads open during winter in the US Midwest and other states affected by large snow events during the winter season and to maintain costs related to accumulation of snow on the roads and repair of roads to minimum levels. Of critical importance for road safety is the protection against snow drifting in regions with narrow rights of way, where standard fences cannot be deployed at the recommended distance from the road. Designing snow fences requires sound engineering judgment and a thorough evaluation of the potential for snow blowing and drifting at the construction site. The evaluation includes site-specific design parameters typically obtained with semi-empirical relations characterizing the local transport conditions. Among the critical parameters involved in fence design and assessment of their post-construction efficiency is the quantification of the snow accumulation at fence sites. The present study proposes a joint experimental and numerical approach to monitor snow deposits around snow fences, quantitatively estimate snow deposits in the field, asses the efficiency and improve the design of snow fences. Snow deposit profiles were mapped using GPS based real-time kinematic surveys (RTK) conducted at the monitored field site during and after snow storms. The monitored site allowed testing different snow fence designs under close to identical conditions over four winter seasons. The study also discusses the detailed monitoring system and analysis of weather forecast and meteorological conditions at the monitored sites. A main goal of the present study was to assess the performance of lightweight plastic snow fences with a lower porosity than the typical 50% porosity used in standard designs of such fences. The field data collected during the first winter was used to identify the best design for snow fences with a porosity of 50%. Flow fields obtained from numerical simulations showed that the fence design that worked the best during the first winter induced the formation of an elongated area of small velocity magnitude close to the ground. This information was used to identify other candidates for optimum design of fences with a lower porosity. Two of the designs with a fence porosity of 30% that were found to perform well based on results of numerical simulations were tested in the field during the second winter along with the best performing design for fences with a porosity of 50%. Field data showed that the length of the snow deposit away from the fence was reduced by about 30% for the two proposed lower-porosity (30%) fence designs compared to the best design identified for fences with a porosity of 50%. Moreover, one of the lower-porosity designs tested in the field showed no significant snow deposition within the bottom gap region beneath the fence. Thus, a major outcome of this study is to recommend using plastic snow fences with a porosity of 30%. It is expected that this lower-porosity design will continue to work well for even more severe snow events or for successive snow events occurring during the same winter. The approach advocated in the present study allowed making general recommendations for optimizing the design of lower-porosity plastic snow fences. This approach can be extended to improve the design of other types of snow fences. Some preliminary work for living snow fences is also discussed. Another major contribution of this study is to propose, develop protocols and test a novel technique based on close range photogrammetry (CRP) to quantify the snow deposits trapped snow fences. As image data can be acquired continuously, the time evolution of the volume of snow retained by a snow fence during a storm or during a whole winter season can, in principle, be obtained. Moreover, CRP is a non-intrusive method that eliminates the need to perform man-made measurements during the storms, which are difficult and sometimes dangerous to perform. Presently, there is lots of empiricism in the design of snow fences due to lack of data on fence storage capacity on how snow deposits change with the fence design and snow storm characteristics and in the estimation of the main parameters used by the state DOTs to design snow fences at a given site. The availability of such information from CRP measurements should provide critical data for the evaluation of the performance of a certain snow fence design that is tested by the IDOT. As part of the present study, the novel CRP method is tested at several sites. The present study also discusses some attempts and preliminary work to determine the snow relocation coefficient which is one of the main variables that has to be estimated by IDOT engineers when using the standard snow fence design software (Snow Drift Profiler, Tabler, 2006). Our analysis showed that standard empirical formulas did not produce reasonable values when applied at the Iowa test sites monitored as part of the present study and that simple methods to estimate this variable are not reliable. The present study makes recommendations for the development of a new methodology based on Large Scale Particle Image Velocimetry that can directly measure the snow drift fluxes and the amount of snow relocated by the fence.
Resumo:
Introduction: Ankle arthrodesis (AD) and total ankle replacement (TAR) are typical treatments for ankle osteoarthritis (AO). Despite clinical interest, there is a lack of their outcome evaluation using objective criteria. Gait analysis and plantar pressure assessment are appropriate to detect pathologies in orthopaedics but they are mostly used in lab with few gait cycles. In this study, we propose an ambulatory device based on inertial and plantar pressure sensors to compare the gait during long-distance trials between healthy subjects (H) and patients with AO or treated by AD and TAR. Methods: Our study included four groups: 11 patients with AO, 9 treated by TAR, 7 treated by AD and 6 control subjects. An ambulatory system (Physilog®, CH) was used for gait analysis; plantar pressure measurements were done using a portable insole (Pedar®-X, DE). The subjects were asked to walk 50 meters in two trials. Mean value and coefficient of variation of spatio-temporal gait parameters were calculated for each trial. Pressure distribution was analyzed in ten subregions of foot. All parameters were compared among the four groups using multi-level model-based statistical analysis. Results: Significant difference (p <0.05) with control was noticed for AO patients in maximum force in medial hindfoot and forefoot and in central forefoot. These differences were no longer significant in TAR and AD groups. Cadence and speed of all pathologic groups showed significant difference with control. Both treatments showed a significant improvement in double support and stance. TAR decreased variability in speed, stride length and knee ROM. Conclusions: In spite of a small sample size, this study showed that ankle function after AO treatments can be evaluated objectively based on plantar pressure and spatio-temporal gait parameters measured during unconstrained walking outside the lab. The combination of these two ambulatory techniques provides a promising way to evaluate foot function in clinics.
Resumo:
Predicting progeny performance from parental genetic divergence can potentially enhance the efficiency of supportive breeding programmes and facilitate risk assessment. Yet, experimental testing of the effects of breeding distance on offspring performance remains rare, especially in wild populations of vertebrates. Recent studies have demonstrated that embryos of salmonid fish are sensitive indicators of additive genetic variance for viability traits. We therefore used gametes of wild brown trout (Salmo trutta) from five genetically distinct populations of a river catchment in Switzerland, and used a full factorial design to produce over 2,000 embryos in 100 different crosses with varying genetic distances (FST range 0.005-0.035). Customized egg capsules allowed recording the survival of individual embryos until hatching under natural field conditions. Our breeding design enabled us to evaluate the role of the environment, of genetic and nongenetic parental contributions, and of interactions between these factors, on embryo viability. We found that embryo survival was strongly affected by maternal environmental (i.e. non-genetic) effects and by the microenvironment, i.e. by the location within the gravel. However, embryo survival was not predicted by population divergence, parental allelic dissimilarity, or heterozygosity, neither in the field nor under laboratory conditions. Our findings suggest that the genetic effects of inter-population hybridization within a genetically differentiated meta-population can be minor in comparison to environmental effects.