12 resultados para STATISTICAL DATA INTERPRETATION
em Digital Commons at Florida International University
Resumo:
The present study examines the extent to which blacks are segregated in the suburban community of Coconut Grove, Florida. Hypersegregation, or the general tendency for blacks and whites to live apart, was examined in terms of four distinct dimensions: evenness, exposure, clustering, and concentration. Together, these dimensions define the geographic traits of the target area. Alone these indices can not capture the multi-dimensional levels of segregation and, therefore, by themselves underestimate the severity of segregation and isolation in this community. This study takes a contemporary view of segregation in a Dade County community to see if segregation is the catalyst to the sometime cited violent response of blacks. This study yields results that support the information in the literature review and the thesis research questions sections namely, that the blacks within the Grove do respond violently to the negative effects that racial segregation causes. This thesis is unique in two ways. It examines segregation in a suburban environment rather than an urban inner city, and it presents a responsive analysis of the individuals studied, rather than relying only on demographic and statistical data. ^
Resumo:
Higher education is a distribution center of knowledge and economic, social, and cultural power (Cervero & Wilson, 2001). A critical approach to understanding a higher education classroom begins with recognizing the instructor's position of power and authority (Tisdell, Hanley, & Taylor, 2000). The power instructors wield exists mostly unquestioned, allowing for teaching practices that reproduce the existing societal patterns of inequity in the classroom (Brookfield, 2000). ^ The purpose of this hermeneutic phenomenological study was to explore students' experiences with the power of their instructors in a higher education classroom. A hermeneutic phenomenological study intertwines the interpretations of both the participants and the researcher about a lived experience to uncover layers of meaning because the meanings of lived experiences are usually not readily apparent (van Manen, 1990). Fifteen participants were selected using criterion, convenience, and snowball sampling. The primary data gathering method were semi-structured interviews guided by an interview protocol (Creswell, 2003). Data were interpreted using thematic reflection (van Manen, 1990). ^ Three themes emerged from data interpretation: (a) structuring of instructor-student relationships, (b) connecting power to instructor personality, and (c) learning to navigate the terrains of higher education. How interpersonal relationships were structured in a higher education classroom shaped how students perceived power in that higher education classroom. Positive relationships were described using the metaphor of family and a perceived ethic of caring and nurturing by the instructor. As participants were consistently exposed to exercises of instructor power in a higher education classroom, they attributed those exercises of power to particular instructor traits rather than systemic exercises of power. As participants progressed from undergraduate to graduate studies, they perceived the benefits of expertise in content or knowledge development as secondary to expertise in successfully navigating the social, cultural, political, and interpersonal terrains of higher education. Ultimately, participants expressed that higher education is not about what you know; it is about learning how to play the game. Implications for teaching in higher education and considerations for future research conclude the study.^
Resumo:
This dissertation examines local governments' efforts to promote economic development in Latin America. The research uses a mixed method to explore how cities make decisions to innovate, develop, and finance economic development programs. First, this study provides a comparative analysis of decentralization policies in Argentina and Mexico as a means to gain a better understanding of the degree of autonomy exercised by local governments. Then, it analyzes three local governments each within the province of Santa Fe, Argentina and the State of Guanajuato, Mexico. The principal hypothesis of this dissertation is that if local governments collect more own-source tax revenue, they are more likely to promote economic development and thus, in turn, promote growth for their region. ^ By examining six cities, three of which are in Santa Fe—Rosario, Santa Fe (capital) and Rafaela—and three in Guanajuato—Leon, Guanajuato (capital) and San Miguel de Allende, this dissertation provides a better understanding of public finances and tax collection efforts of local governments in Latin America. Specific attention is paid to each city's budget authority to raise new revenue and efforts to promote economic development. The research also includes a large statistical dataset of Mexico's 2,454 municipalities and a regression analysis that evaluates local tax efforts on economic growth, controlling for population, territorial size, and the professional development. In order to generalize these results, the research tests these discoveries by using statistical data gathered from a survey administered to Latin American municipal officials. ^ The dissertation demonstrates that cities, which experience greater fiscal autonomy measured by the collection of more own-source revenue, are better able to stimulate effective economic development programs, and ultimately, create jobs within their communities. The results are bolstered by a large number of interviews, which were conducted with over 100 finance specialists, municipal presidents, and local authorities. The dissertation also includes an in-depth literature review on fiscal federalism, decentralization, debt financing and local development. It concludes with a discussion of the findings of the study and applications for the practice of public administration.^
Resumo:
The dissertation documented the degree of Turkey's involvement in the promotion of democracy in the Arab Middle East (ME). Initially, I investigated why and under what conditions Turkey promotes democracy in the ME, and then I explained strategies through which Turkey promotes democracy in the region. I applied the neo-classical realist theoretical framework and a mixed methodology in the research, and I provided evidence from two sources: face-to-face interviews with the Turkish and foreign officials and common citizens, and the statistical data from institutions, such as the OECD, Turkish Statistical Institute, and World Bank.^ My research indicates that Turkey promotes democracy through seven channels. These channels are official development assistance (ODA), mentoring, demonstrative effect, normative pressure, conditionality, military power, enlargement, and civil society organizations. Turkey promotes democracy in the ME for three substantial reasons: first, to advance its security and economic interests; second, to improve the political, social, and economic conditions of people living in the region; and third, to create long-term regional stability, crucial for cooperation in economic and security realms.^ I attempted to engage in debates with two distinct, but interrelated fields of comparative politics and international relations. My most important contribution to the field is that I documented Turkey's case of democracy promotion regarding the degree of Turkey's involvement in this endeavor, its strategies, specificities, and effectiveness in the region. I also contribute to the field as I explained the difference between democracy promotion policies of a regional power, such as Turkey, and global powers, such as the US. I further engaged in discussions that illuminate some aspects of the interplay between the identity and strategic interests in states' foreign policy decisions.^
Resumo:
There is an increasing demand for DNA analysis because of the sensitivity of the method and the ability to uniquely identify and distinguish individuals with a high degree of certainty. But this demand has led to huge backlogs in evidence lockers since the current DNA extraction protocols require long processing time. The DNA analysis procedure becomes more complicated when analyzing sexual assault casework samples where the evidence contains more than one contributor. Additional processing to separate different cell types in order to simplify the final data interpretation further contributes to the existing cumbersome protocols. The goal of the present project is to develop a rapid and efficient extraction method that permits selective digestion of mixtures. ^ Selective recovery of male DNA was achieved with as little as 15 minutes lysis time upon exposure to high pressure under alkaline conditions. Pressure cycling technology (PCT) is carried out in a barocycler that has a small footprint and is semi-automated. Typically less than 10% male DNA is recovered using the standard extraction protocol for rape kits, almost seven times more male DNA was recovered from swabs using this novel method. Various parameters including instrument setting and buffer composition were optimized to achieve selective recovery of sperm DNA. Some developmental validation studies were also done to determine the efficiency of this method in processing samples exposed to various conditions that can affect the quality of the extraction and the final DNA profile. ^ Easy to use interface, minimal manual interference and the ability to achieve high yields with simple reagents in a relatively short time make this an ideal method for potential application in analyzing sexual assault samples.^
Resumo:
There is an increasing demand for DNA analysis because of the sensitivity of the method and the ability to uniquely identify and distinguish individuals with a high degree of certainty. But this demand has led to huge backlogs in evidence lockers since the current DNA extraction protocols require long processing time. The DNA analysis procedure becomes more complicated when analyzing sexual assault casework samples where the evidence contains more than one contributor. Additional processing to separate different cell types in order to simplify the final data interpretation further contributes to the existing cumbersome protocols. The goal of the present project is to develop a rapid and efficient extraction method that permits selective digestion of mixtures. Selective recovery of male DNA was achieved with as little as 15 minutes lysis time upon exposure to high pressure under alkaline conditions. Pressure cycling technology (PCT) is carried out in a barocycler that has a small footprint and is semi-automated. Typically less than 10% male DNA is recovered using the standard extraction protocol for rape kits, almost seven times more male DNA was recovered from swabs using this novel method. Various parameters including instrument setting and buffer composition were optimized to achieve selective recovery of sperm DNA. Some developmental validation studies were also done to determine the efficiency of this method in processing samples exposed to various conditions that can affect the quality of the extraction and the final DNA profile. Easy to use interface, minimal manual interference and the ability to achieve high yields with simple reagents in a relatively short time make this an ideal method for potential application in analyzing sexual assault samples.
Resumo:
This dissertation develops a new mathematical approach that overcomes the effect of a data processing phenomenon known as “histogram binning” inherent to flow cytometry data. A real-time procedure is introduced to prove the effectiveness and fast implementation of such an approach on real-world data. The histogram binning effect is a dilemma posed by two seemingly antagonistic developments: (1) flow cytometry data in its histogram form is extended in its dynamic range to improve its analysis and interpretation, and (2) the inevitable dynamic range extension introduces an unwelcome side effect, the binning effect, which skews the statistics of the data, undermining as a consequence the accuracy of the analysis and the eventual interpretation of the data. ^ Researchers in the field contended with such a dilemma for many years, resorting either to hardware approaches that are rather costly with inherent calibration and noise effects; or have developed software techniques based on filtering the binning effect but without successfully preserving the statistical content of the original data. ^ The mathematical approach introduced in this dissertation is so appealing that a patent application has been filed. The contribution of this dissertation is an incremental scientific innovation based on a mathematical framework that will allow researchers in the field of flow cytometry to improve the interpretation of data knowing that its statistical meaning has been faithfully preserved for its optimized analysis. Furthermore, with the same mathematical foundation, proof of the origin of such an inherent artifact is provided. ^ These results are unique in that new mathematical derivations are established to define and solve the critical problem of the binning effect faced at the experimental assessment level, providing a data platform that preserves its statistical content. ^ In addition, a novel method for accumulating the log-transformed data was developed. This new method uses the properties of the transformation of statistical distributions to accumulate the output histogram in a non-integer and multi-channel fashion. Although the mathematics of this new mapping technique seem intricate, the concise nature of the derivations allow for an implementation procedure that lends itself to a real-time implementation using lookup tables, a task that is also introduced in this dissertation. ^
Resumo:
This dissertation develops a new figure of merit to measure the similarity (or dissimilarity) of Gaussian distributions through a novel concept that relates the Fisher distance to the percentage of data overlap. The derivations are expanded to provide a generalized mathematical platform for determining an optimal separating boundary of Gaussian distributions in multiple dimensions. Real-world data used for implementation and in carrying out feasibility studies were provided by Beckman-Coulter. It is noted that although the data used is flow cytometric in nature, the mathematics are general in their derivation to include other types of data as long as their statistical behavior approximate Gaussian distributions. ^ Because this new figure of merit is heavily based on the statistical nature of the data, a new filtering technique is introduced to accommodate for the accumulation process involved with histogram data. When data is accumulated into a frequency histogram, the data is inherently smoothed in a linear fashion, since an averaging effect is taking place as the histogram is generated. This new filtering scheme addresses data that is accumulated in the uneven resolution of the channels of the frequency histogram. ^ The qualitative interpretation of flow cytometric data is currently a time consuming and imprecise method for evaluating histogram data. This method offers a broader spectrum of capabilities in the analysis of histograms, since the figure of merit derived in this dissertation integrates within its mathematics both a measure of similarity and the percentage of overlap between the distributions under analysis. ^
Resumo:
The microarray technology provides a high-throughput technique to study gene expression. Microarrays can help us diagnose different types of cancers, understand biological processes, assess host responses to drugs and pathogens, find markers for specific diseases, and much more. Microarray experiments generate large amounts of data. Thus, effective data processing and analysis are critical for making reliable inferences from the data. ^ The first part of dissertation addresses the problem of finding an optimal set of genes (biomarkers) to classify a set of samples as diseased or normal. Three statistical gene selection methods (GS, GS-NR, and GS-PCA) were developed to identify a set of genes that best differentiate between samples. A comparative study on different classification tools was performed and the best combinations of gene selection and classifiers for multi-class cancer classification were identified. For most of the benchmarking cancer data sets, the gene selection method proposed in this dissertation, GS, outperformed other gene selection methods. The classifiers based on Random Forests, neural network ensembles, and K-nearest neighbor (KNN) showed consistently god performance. A striking commonality among these classifiers is that they all use a committee-based approach, suggesting that ensemble classification methods are superior. ^ The same biological problem may be studied at different research labs and/or performed using different lab protocols or samples. In such situations, it is important to combine results from these efforts. The second part of the dissertation addresses the problem of pooling the results from different independent experiments to obtain improved results. Four statistical pooling techniques (Fisher inverse chi-square method, Logit method. Stouffer's Z transform method, and Liptak-Stouffer weighted Z-method) were investigated in this dissertation. These pooling techniques were applied to the problem of identifying cell cycle-regulated genes in two different yeast species. As a result, improved sets of cell cycle-regulated genes were identified. The last part of dissertation explores the effectiveness of wavelet data transforms for the task of clustering. Discrete wavelet transforms, with an appropriate choice of wavelet bases, were shown to be effective in producing clusters that were biologically more meaningful. ^
Resumo:
Flow Cytometry analyzers have become trusted companions due to their ability to perform fast and accurate analyses of human blood. The aim of these analyses is to determine the possible existence of abnormalities in the blood that have been correlated with serious disease states, such as infectious mononucleosis, leukemia, and various cancers. Though these analyzers provide important feedback, it is always desired to improve the accuracy of the results. This is evidenced by the occurrences of misclassifications reported by some users of these devices. It is advantageous to provide a pattern interpretation framework that is able to provide better classification ability than is currently available. Toward this end, the purpose of this dissertation was to establish a feature extraction and pattern classification framework capable of providing improved accuracy for detecting specific hematological abnormalities in flow cytometric blood data. ^ This involved extracting a unique and powerful set of shift-invariant statistical features from the multi-dimensional flow cytometry data and then using these features as inputs to a pattern classification engine composed of an artificial neural network (ANN). The contribution of this method consisted of developing a descriptor matrix that can be used to reliably assess if a donor’s blood pattern exhibits a clinically abnormal level of variant lymphocytes, which are blood cells that are potentially indicative of disorders such as leukemia and infectious mononucleosis. ^ This study showed that the set of shift-and-rotation-invariant statistical features extracted from the eigensystem of the flow cytometric data pattern performs better than other commonly-used features in this type of disease detection, exhibiting an accuracy of 80.7%, a sensitivity of 72.3%, and a specificity of 89.2%. This performance represents a major improvement for this type of hematological classifier, which has historically been plagued by poor performance, with accuracies as low as 60% in some cases. This research ultimately shows that an improved feature space was developed that can deliver improved performance for the detection of variant lymphocytes in human blood, thus providing significant utility in the realm of suspect flagging algorithms for the detection of blood-related diseases.^
Resumo:
Stable isotope analysis has become a standard ecological tool for elucidating feeding relationships of organisms and determining food web structure and connectivity. There remain important questions concerning rates at which stable isotope values are incorporated into tissues (turnover rates) and the change in isotope value between a tissue and a food source (discrimination values). These gaps in our understanding necessitate experimental studies to adequately interpret field data. Tissue turnover rates and discrimination values vary among species and have been investigated in a broad array of taxa. However, little attention has been paid to ectothermic top predators in this regard. We quantified the turnover rates and discrimination values for three tissues (scutes, red blood cells, and plasma) in American alligators (Alligator mississippiensis). Plasma turned over faster than scutes or red blood cells, but turnover rates of all three tissues were very slow in comparison to those in endothermic species. Alligator δ15N discrimination values were surprisingly low in comparison to those of other top predators and varied between experimental and control alligators. The variability of δ15N discrimination values highlights the difficulties in using δ15N to assign absolute and possibly even relative trophic levels in field studies. Our results suggest that interpreting stable isotope data based on parameter estimates from other species can be problematic and that large ectothermic tetrapod tissues may be characterized by unique stable isotope dynamics relative to species occupying lower trophic levels and endothermic tetrapods.
Resumo:
This dissertation develops a new mathematical approach that overcomes the effect of a data processing phenomenon known as "histogram binning" inherent to flow cytometry data. A real-time procedure is introduced to prove the effectiveness and fast implementation of such an approach on real-world data. The histogram binning effect is a dilemma posed by two seemingly antagonistic developments: (1) flow cytometry data in its histogram form is extended in its dynamic range to improve its analysis and interpretation, and (2) the inevitable dynamic range extension introduces an unwelcome side effect, the binning effect, which skews the statistics of the data, undermining as a consequence the accuracy of the analysis and the eventual interpretation of the data. Researchers in the field contended with such a dilemma for many years, resorting either to hardware approaches that are rather costly with inherent calibration and noise effects; or have developed software techniques based on filtering the binning effect but without successfully preserving the statistical content of the original data. The mathematical approach introduced in this dissertation is so appealing that a patent application has been filed. The contribution of this dissertation is an incremental scientific innovation based on a mathematical framework that will allow researchers in the field of flow cytometry to improve the interpretation of data knowing that its statistical meaning has been faithfully preserved for its optimized analysis. Furthermore, with the same mathematical foundation, proof of the origin of such an inherent artifact is provided. These results are unique in that new mathematical derivations are established to define and solve the critical problem of the binning effect faced at the experimental assessment level, providing a data platform that preserves its statistical content. In addition, a novel method for accumulating the log-transformed data was developed. This new method uses the properties of the transformation of statistical distributions to accumulate the output histogram in a non-integer and multi-channel fashion. Although the mathematics of this new mapping technique seem intricate, the concise nature of the derivations allow for an implementation procedure that lends itself to a real-time implementation using lookup tables, a task that is also introduced in this dissertation.