364 resultados para Dataset


Relevância:

10.00% 10.00%

Publicador:

Resumo:

BACKGROUND: The tendency to conceive dizygotic (DZ) twins is a complex trait influenced by genetic and environmental factors. To search for new candidate loci for twinning, we conducted a genome-wide linkage scan in 525 families using microsatellite and single nucleotide polymorphism marker panels. METHODS AND RESULTS: Non-parametric linkage analyses, including 523 families containing a total of 1115 mothers of DZ twins (MODZT) from Australia and New Zealand (ANZ) and The Netherlands (NL), produced four linkage peaks above the threshold for suggestive linkage, including a highly suggestive peak at the extreme telomeric end of chromosome 6 with an exponential logarithm of odds \[(exp)LOD] score of 2.813 (P = 0.0002). Since the DZ twinning rate increases steeply with maternal age independent of genetic effects, we also investigated linkage including only families where at least one MODZT gave birth to her first set of twins before the age of 30. These analyses produced a maximum expLOD score of 2.718 (P = 0.0002), largely due to linkage signal from the ANZ cohort, however, ordered subset analyses indicated this result is most likely a chance finding in the combined dataset. Linkage analyses were also performed for two large DZ twinning families from the USA, one of which produced a peak on chromosome 2 in the region of two potential candidate genes. Sequencing of FSHR and FIGLA, along with INHBB in MODZTs from two large NL families with family specific linkage peaks directly over this gene, revealed a potentially functional variant in the 5' untranslated region of FSHR that segregated with the DZ twinning phenotype in the Utah family. CONCLUSION: Our data provide further evidence for complex inheritance of familial DZ twinning.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Most information in linkage analysis for quantitative traits comes from pairs of relatives that are phenotypically most discordant or concordant. Confounding this, within-family outliers from non-genetic causes may create false positives and negatives. We investigated the influence of within-family outliers empirically, using one of the largest genome-wide linkage scans for height. The subjects were drawn from Australian twin cohorts consisting of 8447 individuals in 2861 families, providing a total of 5815 possible pairs of siblings in sibships. A variance component linkage analysis was performed, either including or excluding the within-family outliers. Using the entire dataset, the largest LOD scores were on chromosome 15q (LOD 2.3) and 11q (1.5). Excluding within-family outliers increased the LOD score for most regions, but the LOD score on chromosome 15 decreased from 2.3 to 1.2, suggesting that the outliers may create false negatives and false positives, although rare alleles of large effect may also be an explanation. Several regions suggestive of linkage to height were found after removing the outliers, including 1q23.1 (2.0), 3q22.1 (1.9) and 5q32 (2.3). We conclude that the investigation of the effect of within-family outliers, which is usually neglected, should be a standard quality control measure in linkage analysis for complex traits and may reduce the noise for the search of common variants of modest effect size as well as help identify rare variants of large effect and clinical significance. We suggest that the effect of within-family outliers deserves further investigation via theoretical and simulation studies.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In industrial and organizational psychology, there is a long tradition of studying personality as an antecedent of work outcomes. Recently, however, scholars have suggested that personality characteristics may not only predict, but also change due to certain work experiences, a notion that is depicted in the dynamic developmental model (DDM) of personality and work. Upward job changes are an important part of employees’ careers and career success in particular, and we argue that these career transitions can shape personality over time. In this study, we investigate the Big Five personality characteristics as both predictors and outcomes of upward job changes into managerial and professional positions. We tested our hypotheses by applying event history analyses and propensity score matching to a longitudinal dataset collected over five years from employees in Australia. Results indicated that participants’ openness to experience not only predicted, but that changes in openness to experience also followed from upward job changes into managerial and professional positions. Our findings thus provide support for a dynamic perspective on personality characteristics in the context of work and careers.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Assessing blood concentration of persistent organic pollutants (POPs) in infants is difficult due to the ethical and practical difficulties in obtaining sufficient quantities of blood. To determine whether measuring POPs in faeces might reflect blood concentration during infancy, we measured the concentrations of a range of POPs (i.e. polychlorinated biphenyls (PCBs), polybrominated diphenyl ethers (PBDEs) and organochlorine pesticides (OCPs)) in a pilot study using matched breast milk and infant faecal samples obtained from ten mother-child pairs. All infants were breast fed, with 8 of them also receiving solid food at the time of faecal sampling. In this small dataset faecal concentrations (range 0.01-41ngg-1 lipid) are strongly associated with milk concentrations (range 0.02-230ngg-1 lipid). Associations with other factors generally could not be detected in this dataset, with the exception of a small effect of age or growth. Different sources (external or internal) of exposure appeared to directly influence faecal concentrations of different chemicals based on different inter-individual variability in the faeces-to-milk concentration ratio Rfm. Overall, the matrix of faeces as an external measure of internal exposure in infants looks promising for some chemicals and is worth assessing further in larger datasets.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Video surveillance infrastructure has been widely installed in public places for security purposes. However, live video feeds are typically monitored by human staff, making the detection of important events as they occur difficult. As such, an expert system that can automatically detect events of interest in surveillance footage is highly desirable. Although a number of approaches have been proposed, they have significant limitations: supervised approaches, which can detect a specific event, ideally require a large number of samples with the event spatially and temporally localised; while unsupervised approaches, which do not require this demanding annotation, can only detect whether an event is abnormal and not specific event types. To overcome these problems, we formulate a weakly-supervised approach using Kullback-Leibler (KL) divergence to detect rare events. The proposed approach leverages the sparse nature of the target events to its advantage, and we show that this data imbalance guarantees the existence of a decision boundary to separate samples that contain the target event from those that do not. This trait, combined with the coarse annotation used by weakly supervised learning (that only indicates approximately when an event occurs), greatly reduces the annotation burden while retaining the ability to detect specific events. Furthermore, the proposed classifier requires only a decision threshold, simplifying its use compared to other weakly supervised approaches. We show that the proposed approach outperforms state-of-the-art methods on a popular real-world traffic surveillance dataset, while preserving real time performance.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper discusses the use of observational video recordings to document young children’s use of technology in their homes. Although observational research practices have been used for decades, often with video-based techniques, the participant group in this study (i.e., very young children) and the setting (i.e., private homes), provide a rich space for exploring the benefits and limitations of qualitative observation. The data gathered in this study point to a number of key decisions and issues that researchers must face in designing observational research, particularly where non-researchers (in this case, parents) act as surrogates for the researcher at the data collection stage. The involvement of parents and children as research videographers in the home resulted in very rich and detailed data about children’s use of technology in their daily lives. However, limitations noted in the dataset (e.g., image quality) provide important guidance for researchers developing projects using similar methods in future. The paper provides recommendations for future observational designs in similar settings and/or with similar participant groups.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Data-driven approaches such as Gaussian Process (GP) regression have been used extensively in recent robotics literature to achieve estimation by learning from experience. To ensure satisfactory performance, in most cases, multiple learning inputs are required. Intuitively, adding new inputs can often contribute to better estimation accuracy, however, it may come at the cost of a new sensor, larger training dataset and/or more complex learning, some- times for limited benefits. Therefore, it is crucial to have a systematic procedure to determine the actual impact each input has on the estimation performance. To address this issue, in this paper we propose to analyse the impact of each input on the estimate using a variance-based sensitivity analysis method. We propose an approach built on Analysis of Variance (ANOVA) decomposition, which can characterise how the prediction changes as one or more of the input changes, and also quantify the prediction uncertainty as attributed from each of the inputs in the framework of dependent inputs. We apply the proposed approach to a terrain-traversability estimation method we proposed in prior work, which is based on multi-task GP regression, and we validate this implementation experimentally using a rover on a Mars-analogue terrain.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper presents an effective classification method based on Support Vector Machines (SVM) in the context of activity recognition. Local features that capture both spatial and temporal information in activity videos have made significant progress recently. Efficient and effective features, feature representation and classification plays a crucial role in activity recognition. For classification, SVMs are popularly used because of their simplicity and efficiency; however the common multi-class SVM approaches applied suffer from limitations including having easily confused classes and been computationally inefficient. We propose using a binary tree SVM to address the shortcomings of multi-class SVMs in activity recognition. We proposed constructing a binary tree using Gaussian Mixture Models (GMM), where activities are repeatedly allocated to subnodes until every new created node contains only one activity. Then, for each internal node a separate SVM is learned to classify activities, which significantly reduces the training time and increases the speed of testing compared to popular the `one-against-the-rest' multi-class SVM classifier. Experiments carried out on the challenging and complex Hollywood dataset demonstrates comparable performance over the baseline bag-of-features method.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Early detection of (pre-)signs of ulceration on a diabetic foot is valuable for clinical practice. Hyperspectral imaging is a promising technique for detection and classification of such (pre-)signs. However, the number of the spectral bands should be limited to avoid overfitting, which is critical for pixel classification with hyperspectral image data. The goal was to design a detector/classifier based on spectral imaging (SI) with a small number of optical bandpass filters. The performance and stability of the design were also investigated. The selection of the bandpass filters boils down to a feature selection problem. A dataset was built, containing reflectance spectra of 227 skin spots from 64 patients, measured with a spectrometer. Each skin spot was annotated manually by clinicians as "healthy" or a specific (pre-)sign of ulceration. Statistical analysis on the data set showed the number of required filters is between 3 and 7, depending on additional constraints on the filter set. The stability analysis revealed that shot noise was the most critical factor affecting the classification performance. It indicated that this impact could be avoided in future SI systems with a camera sensor whose saturation level is higher than 106, or by postimage processing.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This thesis introduced two novel reputation models to generate accurate item reputation scores using ratings data and the statistics of the dataset. It also presented an innovative method that incorporates reputation awareness in recommender systems by employing voting system methods to produce more accurate top-N item recommendations. Additionally, this thesis introduced a personalisation method for generating reputation scores based on users' interests, where a single item can have different reputation scores for different users. The personalised reputation scores are then used in the proposed reputation-aware recommender systems to enhance the recommendation quality.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

One of the most evident casualties of a natural disaster is the property market. The private and social costs from such events run into millions of dollars. In this paper, we use a unique dataset to examine the impact on residential house prices affected by natural disasters using a hedonic property (HP) values approach. For this purpose, we use data before and after a wildfire and floods from Rockhampton in central Queensland, Australia. The data is unique because one suburb was affected by wildfires and another was affected by floods. For the analysis, three suburbs namely Frenchville, Park Avenue and Norman Gardens are used. Frenchville was significantly affected by wildfires in the latter part of 2009 and to a lesser extent in 2012, while Park Avenue was affected by floods at the end of 2010, January 2011–2013. Norman Gardens, which was relatively unaffected, is used as a control site. This enables us to examine the before and after effects on property values in the three suburbs. The results confirm that soon after a natural disaster property prices in affected areas decrease even though the large majority of individual houses remain unaffected. Furthermore, the results indicate that the largely unaffected suburb may gain immediately after a natural disaster but this gain may disappear if natural disasters continue to occur in the area/region due to the stigma created. The results have several important policy decision and welfare implications which are briefly discussed in the paper.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The DAYCENT biogeochemical model was used to investigate how the use of fertilizers coated with nitrification inhibitors and the introduction of legumes in the crop rotation can affect subtropical cereal production and {N2O} emissions. The model was validated using comprehensive multi-seasonal, high-frequency dataset from two field investigations conducted on an Oxisol, which is the most common soil type in subtropical regions. Different N fertilizer rates were tested for each N management strategy and simulated under varying weather conditions. DAYCENT was able to reliably predict soil N dynamics, seasonal {N2O} emissions and crop production, although some discrepancies were observed in the treatments with low or no added N inputs and in the simulation of daily {N2O} fluxes. Simulations highlighted that the high clay content and the relatively low C levels of the Oxisol analyzed in this study limit the chances for significant amounts of N to be lost via deep leaching or denitrification. The application of urea coated with a nitrification inhibitor was the most effective strategy to minimize {N2O} emissions. This strategy however did not increase yields since the nitrification inhibitor did not substantially decrease overall N losses compared to conventional urea. Simulations indicated that replacing part of crop N requirements with N mineralized by legume residues is the most effective strategy to reduce {N2O} emissions and support cereal productivity. The results of this study show that legumes have significant potential to enhance the sustainable and profitable intensification of subtropical cereal cropping systems in Oxisols.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This paper presents 'vSpeak', the first initiative taken in Pakistan for ICT enabled conversion of dynamic Sign Urdu gestures into natural language sentences. To realize this, vSpeak has adopted a novel approach for feature extraction using edge detection and image compression which gives input to the Artificial Neural Network that recognizes the gesture. This technique caters for the blurred images as well. The training and testing is currently being performed on a dataset of 200 patterns of 20 words from Sign Urdu with target accuracy of 90% and above.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Context: Pheochromocytomas and paragangliomas (PPGLs) are heritable neoplasms that can be classified into gene-expression subtypes corresponding to their underlying specific genetic drivers. Objective: This study aimed to develop a diagnostic and research tool (Pheo-type) capable of classifying PPGL tumors into gene-expression subtypes that could be used to guide and interpret genetic testing, determine surveillance programs, and aid in elucidation of PPGL biology. Design: A compendium of published microarray data representing 205 PPGL tumors was used for the selection of subtype-specific genes that were then translated to the Nanostring gene-expression platform. A support vector machine was trained on the microarray dataset and then tested on an independent Nanostring dataset representing 38 familial and sporadic cases of PPGL of known genotype (RET, NF1, TMEM127, MAX, HRAS, VHL, and SDHx). Different classifier models involving between three and six subtypes were compared for their discrimination potential. Results: A gene set of 46 genes and six endogenous controls was selected representing six known PPGL subtypes; RTK1–3 (RET, NF1, TMEM127, and HRAS), MAX-like, VHL, and SDHx. Of 38 test cases, 34 (90%) were correctly predicted to six subtypes based on the known genotype to gene-expression subtype association. Removal of the RTK2 subtype from training, characterized by an admixture of tumor and normal adrenal cortex, improved the classification accuracy (35/38). Consolidation of RTK and pseudohypoxic PPGL subtypes to four- and then three-class architectures improved the classification accuracy for clinical application. Conclusions: The Pheo-type gene-expression assay is a reliable method for predicting PPGL genotype using routine diagnostic tumor samples.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Identifying unusual or anomalous patterns in an underlying dataset is an important but challenging task in many applications. The focus of the unsupervised anomaly detection literature has mostly been on vectorised data. However, many applications are more naturally described using higher-order tensor representations. Approaches that vectorise tensorial data can destroy the structural information encoded in the high-dimensional space, and lead to the problem of the curse of dimensionality. In this paper we present the first unsupervised tensorial anomaly detection method, along with a randomised version of our method. Our anomaly detection method, the One-class Support Tensor Machine (1STM), is a generalisation of conventional one-class Support Vector Machines to higher-order spaces. 1STM preserves the multiway structure of tensor data, while achieving significant improvement in accuracy and efficiency over conventional vectorised methods. We then leverage the theory of nonlinear random projections to propose the Randomised 1STM (R1STM). Our empirical analysis on several real and synthetic datasets shows that our R1STM algorithm delivers comparable or better accuracy to a state-of-the-art deep learning method and traditional kernelised approaches for anomaly detection, while being approximately 100 times faster in training and testing.