63 resultados para statistical data analysis
Resumo:
Pounamu (NZ jade), or nephrite, is a protected mineral in its natural form following thetransfer of ownership back to Ngai Tahu under the Ngai Tahu (Pounamu Vesting) Act 1997.Any theft of nephrite is prosecutable under the Crimes Act 1961. Scientific evidence isessential in cases where origin is disputed. A robust method for discrimination of thismaterial through the use of elemental analysis and compositional data analysis is required.Initial studies have characterised the variability within a given nephrite source. This hasincluded investigation of both in situ outcrops and alluvial material. Methods for thediscrimination of two geographically close nephrite sources are being developed.Key Words: forensic, jade, nephrite, laser ablation, inductively coupled plasma massspectrometry, multivariate analysis, elemental analysis, compositional data analysis
Resumo:
This article provides an in-depth study of long-term female unemployment in Catalonia.Long-term unemployment statistics reveal which social groups are most likely to experience difficulty re-entering the labour market. In this case, we found that women are mainly affected by this type of labour exclusion, in particular poorly qualified, working-class women who are aged over 45 and with family responsibilities.The article aims to explore how the overlapping of factors such as gender, age, social class, origin and the division of work based on gender are related to long-term female unemployment. Moreover, we were able to detect which conceptual tools provide us with the production/reproduction paradigm so as to be able to analyse the future of female unemployment. The methodology we used combines quantitative and qualitative approaches. On the one hand, the analysis of secondary statistical data focusing on Catalonia is useful in understanding the situation from a macro-social perspective. On the other hand, an exploratory discussion group enables us to investigate social imaginary practises among unemployed working class women aged over 45. This discussion group was held in Igualada -capital of the Anoia region - an area of Catalonia deeply affected by unemployment in the current economic crisis.
Resumo:
The Aitchison vector space structure for the simplex is generalized to a Hilbert space structure A2(P) for distributions and likelihoods on arbitrary spaces. Centralnotations of statistics, such as Information or Likelihood, can be identified in the algebraical structure of A2(P) and their corresponding notions in compositional data analysis, such as Aitchison distance or centered log ratio transform.In this way very elaborated aspects of mathematical statistics can be understoodeasily in the light of a simple vector space structure and of compositional data analysis. E.g. combination of statistical information such as Bayesian updating,combination of likelihood and robust M-estimation functions are simple additions/perturbations in A2(Pprior). Weighting observations corresponds to a weightedaddition of the corresponding evidence.Likelihood based statistics for general exponential families turns out to have aparticularly easy interpretation in terms of A2(P). Regular exponential families formfinite dimensional linear subspaces of A2(P) and they correspond to finite dimensionalsubspaces formed by their posterior in the dual information space A2(Pprior).The Aitchison norm can identified with mean Fisher information. The closing constant itself is identified with a generalization of the cummulant function and shown to be Kullback Leiblers directed information. Fisher information is the local geometry of the manifold induced by the A2(P) derivative of the Kullback Leibler information and the space A2(P) can therefore be seen as the tangential geometry of statistical inference at the distribution P.The discussion of A2(P) valued random variables, such as estimation functionsor likelihoods, give a further interpretation of Fisher information as the expected squared norm of evidence and a scale free understanding of unbiased reasoning
Resumo:
The aim of this talk is to convince the reader that there are a lot of interesting statisticalproblems in presentday life science data analysis which seem ultimately connected withcompositional statistics.Key words: SAGE, cDNA microarrays, (1D-)NMR, virus quasispecies
Resumo:
In this paper we present a Bayesian image reconstruction algorithm with entropy prior (FMAPE) that uses a space-variant hyperparameter. The spatial variation of the hyperparameter allows different degrees of resolution in areas of different statistical characteristics, thus avoiding the large residuals resulting from algorithms that use a constant hyperparameter. In the first implementation of the algorithm, we begin by segmenting a Maximum Likelihood Estimator (MLE) reconstruction. The segmentation method is based on using a wavelet decomposition and a self-organizing neural network. The result is a predetermined number of extended regions plus a small region for each star or bright object. To assign a different value of the hyperparameter to each extended region and star, we use either feasibility tests or cross-validation methods. Once the set of hyperparameters is obtained, we carried out the final Bayesian reconstruction, leading to a reconstruction with decreased bias and excellent visual characteristics. The method has been applied to data from the non-refurbished Hubble Space Telescope. The method can be also applied to ground-based images.
Resumo:
The present study proposes a modification in one of the most frequently applied effect size procedures in single-case data analysis the percent of nonoverlapping data. In contrast to other techniques, the calculus and interpretation of this procedure is straightforward and it can be easily complemented by visual inspection of the graphed data. Although the percent of nonoverlapping data has been found to perform reasonably well in N = 1 data, the magnitude of effect estimates it yields can be distorted by trend and autocorrelation. Therefore, the data correction procedure focuses on removing the baseline trend from data prior to estimating the change produced in the behavior due to intervention. A simulation study is carried out in order to compare the original and the modified procedures in several experimental conditions. The results suggest that the new proposal is unaffected by trend and autocorrelation and can be used in case of unstable baselines and sequentially related measurements.
Resumo:
The present study focuses on single-case data analysis and specifically on two procedures for quantifying differences between baseline and treatment measurements The first technique tested is based on generalized least squares regression analysis and is compared to a proposed non-regression technique, which allows obtaining similar information. The comparison is carried out in the context of generated data representing a variety of patterns (i.e., independent measurements, different serial dependence underlying processes, constant or phase-specific autocorrelation and data variability, different types of trend, and slope and level change). The results suggest that the two techniques perform adequately for a wide range of conditions and researchers can use both of them with certain guarantees. The regression-based procedure offers more efficient estimates, whereas the proposed non-regression procedure is more sensitive to intervention effects. Considering current and previous findings, some tentative recommendations are offered to applied researchers in order to help choosing among the plurality of single-case data analysis techniques.
Resumo:
Background: The aim of this study was to evaluate how hospital capacity was managed focusing on standardizing the admission and discharge processes. Methods: This study was set in a 900-bed university affiliated hospital of the National Health Service, near Barcelona (Spain). This is a cross-sectional study of a set of interventions which were gradually implemented between April and December 2008. Mainly, they were focused on standardizing the admission and discharge processes to improve patient flow. Primary administrative data was obtained from the 2007 and 2009 Hospital Database. Main outcome measures were median length of stay, percentage of planned discharges, number of surgery cancellations and median number of delayed emergency admissions at 8:00 am. For statistical bivariate analysis, we used a Chi-squared for linear trend for qualitative variables and a Wilcoxon signed ranks test and a Mann–Whitney test for non-normal continuous variables. Results: The median patients’ global length of stay was 8.56 days in 2007 and 7.93 days in 2009 (p<0.051). The percentage of patients admitted the same day as surgery increased from 64.87% in 2007 to 86.01% in 2009 (p<0.05). The number of cancelled interventions due to lack of beds was 216 patients in 2007 and 42 patients in 2009. The median number of planned discharges went from 43.05% in 2007 to 86.01% in 2009 (p<0.01). The median number of emergency patients waiting for an in-hospital bed at 8:00 am was 5 patients in 2007 and 3 patients in 2009 (p<0.01). Conclusions: In conclusion, standardization of admission and discharge processes are largely in our control. There is a significant opportunity to create important benefits for increasing bed capacity and hospital throughput.
Resumo:
Background: The aim of this study was to evaluate how hospital capacity was managed focusing on standardizing the admission and discharge processes. Methods: This study was set in a 900-bed university affiliated hospital of the National Health Service, near Barcelona (Spain). This is a cross-sectional study of a set of interventions which were gradually implemented between April and December 2008. Mainly, they were focused on standardizing the admission and discharge processes to improve patient flow. Primary administrative data was obtained from the 2007 and 2009 Hospital Database. Main outcome measures were median length of stay, percentage of planned discharges, number of surgery cancellations and median number of delayed emergency admissions at 8:00¿am. For statistical bivariate analysis, we used a Chi-squared for linear trend for qualitative variables and a Wilcoxon signed ranks test and a Mann¿Whitney test for non-normal continuous variables. Results:The median patients' global length of stay was 8.56 days in 2007 and 7.93 days in 2009 (p<0.051). The percentage of patients admitted the same day as surgery increased from 64.87% in 2007 to 86.01% in 2009 (p<0.05). The number of cancelled interventions due to lack of beds was 216 patients in 2007 and 42 patients in 2009. The median number of planned discharges went from 43.05% in 2007 to 86.01% in 2009 (p<0.01). The median number of emergency patients waiting for an in-hospital bed at 8:00¿am was 5 patients in 2007 and 3 patients in 2009 (p<0.01). Conclusions: In conclusion, standardization of admission and discharge processes are largely in our control. There is a significant opportunity to create important benefits for increasing bed capacity and hospital throughput.
Resumo:
This article provides an in-depth study of long-term female unemployment in Catalonia.Long-term unemployment statistics reveal which social groups are most likely to experience difficulty re-entering the labour market. In this case, we found that women are mainly affected by this type of labour exclusion, in particular poorly qualified, working-class women who are aged over 45 and with family responsibilities.The article aims to explore how the overlapping of factors such as gender, age, social class, origin and the division of work based on gender are related to long-term female unemployment. Moreover, we were able to detect which conceptual tools provide us with the production/reproduction paradigm so as to be able to analyse the future of female unemployment. The methodology we used combines quantitative and qualitative approaches. On the one hand, the analysis of secondary statistical data focusing on Catalonia is useful in understanding the situation from a macro-social perspective. On the other hand, an exploratory discussion group enables us to investigate social imaginary practises among unemployed working class women aged over 45. This discussion group was held in Igualada -capital of the Anoia region - an area of Catalonia deeply affected by unemployment in the current economic crisis.
Resumo:
The present paper advocates for the creation of a federated, hybrid database in the cloud, integrating law data from all available public sources in one single open access system - adding, in the process, relevant meta-data to the indexed documents, including the identification of social and semantic entities and the relationships between them, using linked open data techniques and standards such as RDF. Examples of potential benefits and applications of this approach are also provided, including, among others, experiences from of our previous research, in which data integration, graph databases and social and semantic networks analysis were used to identify power relations, litigation dynamics and cross-references patterns both intra and inter-institutionally, covering most of the World international economic courts.
Resumo:
N = 1 designs imply repeated registrations of the behaviour of the same experimental unit and the measurements obtained are often few due to time limitations, while they are also likely to be sequentially dependent. The analytical techniques needed to enhance statistical and clinical decision making have to deal with these problems. Different procedures for analysing data from single-case AB designs are discussed, presenting their main features and revising the results reported by previous studies. Randomization tests represent one of the statistical methods that seemed to perform well in terms of controlling false alarm rates. In the experimental part of the study a new simulation approach is used to test the performance of randomization tests and the results suggest that the technique is not always robust against the violation of the independence assumption. Moreover, sensitivity proved to be generally unacceptably low for series lengths equal to 30 and 40. Considering the evidence available, there does not seem to be an optimal technique for single-case data analysis
Resumo:
ABSTRACT Dual-trap optical tweezers are often used in high-resolution measurements in single-molecule biophysics. Such measurements can be hindered by the presence of extraneous noise sources, the most prominent of which is the coupling of fluctuations along different spatial directions, which may affect any optical tweezers setup. In this article, we analyze, both from the theoretical and the experimental points of view, the most common source for these couplings in dual-trap optical-tweezers setups: the misalignment of traps and tether. We give criteria to distinguish different kinds of misalignment, to estimate their quantitative relevance and to include them in the data analysis. The experimental data is obtained in a, to our knowledge, novel dual-trap optical-tweezers setup that directly measures forces. In the case in which misalignment is negligible, we provide a method to measure the stiffness of traps and tether based on variance analysis. This method can be seen as a calibration technique valid beyond the linear trap region. Our analysis is then employed to measure the persistence length of dsDNA tethers of three different lengths spanning two orders of magnitude. The effective persistence length of such tethers is shown to decrease with the contour length, in accordance with previous studies.
Resumo:
The agricultural sector has always been characterized by a predominance of small firms. International competition and the consequent need for restraining costs are permanent challenges for farms. This paper performs an empirical investigation of cost behavior in agriculture using panel data analysis. Our results show that transactions caused by complexity influence farm costs with opposite effects for specific and indirect costs. While transactions allow economies of scale in specific costs, they significantly increase indirect costs. However, the main driver for farm costs is volume. In addition, important differences exist for small and big farms, since transactional variables significantly influence the former but not the latter. While sophisticated management tools, such ABC, could provide only limited complementary useful information but no essential allocation bases for farms, they seem inappropriate for small farms
Resumo:
The agricultural sector has always been characterized by a predominance of small firms. International competition and the consequent need for restraining costs are permanent challenges for farms. This paper performs an empirical investigation of cost behavior in agriculture using panel data analysis. Our results show that transactions caused by complexity influence farm costs with opposite effects for specific and indirect costs. While transactions allow economies of scale in specific costs, they significantly increase indirect costs. However, the main driver for farm costs is volume. In addition, important differences exist for small and big farms, since transactional variables significantly influence the former but not the latter. While sophisticated management tools, such ABC, could provide only limited complementary useful information but no essential allocation bases for farms, they seem inappropriate for small farms