863 resultados para Data sources detection
Resumo:
Today, databases have become an integral part of information systems. In the past two decades, we have seen different database systems being developed independently and used in different applications domains. Today's interconnected networks and advanced applications, such as data warehousing, data mining & knowledge discovery and intelligent data access to information on the Web, have created a need for integrated access to such heterogeneous, autonomous, distributed database systems. Heterogeneous/multidatabase research has focused on this issue resulting in many different approaches. However, a single, generally accepted methodology in academia or industry has not emerged providing ubiquitous intelligent data access from heterogeneous, autonomous, distributed information sources. This thesis describes a heterogeneous database system being developed at Highperformance Database Research Center (HPDRC). A major impediment to ubiquitous deployment of multidatabase technology is the difficulty in resolving semantic heterogeneity. That is, identifying related information sources for integration and querying purposes. Our approach considers the semantics of the meta-data constructs in resolving this issue. The major contributions of the thesis work include: (i.) providing a scalable, easy-to-implement architecture for developing a heterogeneous multidatabase system, utilizing Semantic Binary Object-oriented Data Model (Sem-ODM) and Semantic SQL query language to capture the semantics of the data sources being integrated and to provide an easy-to-use query facility; (ii.) a methodology for semantic heterogeneity resolution by investigating into the extents of the meta-data constructs of component schemas. This methodology is shown to be correct, complete and unambiguous; (iii.) a semi-automated technique for identifying semantic relations, which is the basis of semantic knowledge for integration and querying, using shared ontologies for context-mediation; (iv.) resolutions for schematic conflicts and a language for defining global views from a set of component Sem-ODM schemas; (v.) design of a knowledge base for storing and manipulating meta-data and knowledge acquired during the integration process. This knowledge base acts as the interface between integration and query processing modules; (vi.) techniques for Semantic SQL query processing and optimization based on semantic knowledge in a heterogeneous database environment; and (vii.) a framework for intelligent computing and communication on the Internet applying the concepts of our work.
Resumo:
Biodiversity citizen science projects are growing in number, size, and scope, and are gaining recognition as valuable data sources that build public engagement. Yet publication rates indicate that citizen science is still infrequently used as a primary tool for conservation research and the causes of this apparent disconnect have not been quantitatively evaluated. To uncover the barriers to the use of citizen science as a research tool, we surveyed professional biodiversity scientists (n = 423) and citizen science project managers (n = 125). We conducted three analyses using non-parametric recursive modeling (random forest), using questions that addressed: scientists' perceptions and preferences regarding citizen science, scientists' requirements for their own data, and the actual practices of citizen science projects. For all three analyses we identified the most important factors that influence the probability of publication using citizen science data. Four general barriers emerged: a narrow awareness among scientists of citizen science projects that match their needs; the fact that not all biodiversity science is well-suited for citizen science; inconsistency in data quality across citizen science projects; and bias among scientists for certain data sources (institutions and ages/education levels of data collectors). Notably, we find limited evidence to suggest a relationship between citizen science projects that satisfy scientists' biases and data quality or probability of publication. These results illuminate the need for greater visibility of citizen science practices with respect to the requirements of biodiversity science and show that addressing bias among scientists could improve application of citizen science in conservation.
Resumo:
Recent discussion regarding whether the noise that limits 2AFC discrimination performance is fixed or variable has focused either on describing experimental methods that presumably dissociate the effects of response mean and variance or on reanalyzing a published data set with the aim of determining how to solve the question through goodness-of-fit statistics. This paper illustrates that the question cannot be solved by fitting models to data and assessing goodness-of-fit because data on detection and discrimination performance can be indistinguishably fitted by models that assume either type of noise when each is coupled with a convenient form for the transducer function. Thus, success or failure at fitting a transducer model merely illustrates the capability (or lack thereof) of some particular combination of transducer function and variance function to account for the data, but it cannot disclose the nature of the noise. We also comment on some of the issues that have been raised in recent exchange on the topic, namely, the existence of additional constraints for the models, the presence of asymmetric asymptotes, the likelihood of history-dependent noise, and the potential of certain experimental methods to dissociate the effects of response mean and variance.
Resumo:
La présente thèse vise à évaluer le degré d’implantation et d’utilisation de systèmes de mesure de la performance (SMP) par les décideurs des organisations de réadaptation et à comprendre les facteurs contextuels ayant influencé leur implantation. Pour ce faire, une étude de cas multiples a été réalisée comprenant deux sources de données: des entrevues individuelles avec des cadres supérieurs des organisations de réadaptation du Québec et des documents organisationnels. Le cadre conceptuel Consolidated Framework for Implementation Research a été utilisé pour guider la collecte et l’analyse des données. Une analyse intra-cas ainsi qu’une analyse inter-cas ont été réalisées. Nos résultats montrent que le niveau de préparation organisationnelle à l’implantation d’un SMP était élevé et que les SMP ont été implantés avec succès et utilisés de plusieurs façons. Les organisations les ont utilisés de façon passive (comme outil d’information), de façon ciblée (pour tenter d’améliorer des domaines sous-performants) et de façon politique (comme outil de négociation auprès des autorités gouvernementales). Cette utilisation diversifiée des SMP est suscitée par l’interaction complexe de facteurs provenant du contexte interne propre à chaque organisation, des caractéristiques du SMP, du processus d’implantation appliqué et du contexte externe dans lequel évoluent ces organisations. Au niveau du contexte interne, l’engagement continu et le leadership de la haute direction ont été décisifs dans l’implantation du SMP de par leur influence sur l’identification du besoin d’un SMP, l’engagement des utilisateurs visés dans le projet, la priorité organisationnelle accordée au SMP ainsi que les ressources octroyées à son implantation, la qualité des communications et le climat d’apprentissage organisationnel. Toutefois, même si certains de ces facteurs, comme les ressources octroyées à l’implantation, la priorité organisationnelle du SMP et le climat d’apprentissage se sont révélés être des barrières à l’implantation, ultimement, ces barrières n’étaient pas suffisamment importantes pour entraver l’utilisation du SMP. Cette étude a également confirmé l’importance des caractéristiques du SMP, particulièrement la perception de qualité et d’utilité de l’information. Cependant, à elles seules, ces caractéristiques sont insuffisantes pour assurer le succès d’implantation. Cette analyse d’implantation a également révélé que, même si le processus d’implantation ne suit pas des étapes formelles, un plan de développement du SMP, la participation et l’engagement des décideurs ainsi que la désignation d’un responsable de projet ont tous facilité son implantation. Cependant, l’absence d’évaluation et de réflexion collective sur le processus d’implantation a limité le potentiel d’apprentissage organisationnel, un prérequis à l’amélioration de la performance. Quant au contexte externe, le soutien d’un organisme externe s’est avéré un facilitateur indispensable pour favoriser l’implantation de SMP par les organisations de réadaptation malgré l’absence de politiques et incitatifs gouvernementaux à cet effet. Cette étude contribue à accroître les connaissances sur les facteurs contextuels ainsi que sur leurs interactions dans l’utilisation d’innovations tels les SMP et confirme l’importance d’aborder l’analyse de l’implantation avec une perspective systémique.
Resumo:
Objective: To assess the effects of selective cyclo-oxygenase-2 (COX 2) inhibitors and traditional non-steroidal anti-inflammatory drugs (NSAIDs) on the risk of vascular events. Design: Meta-analysis of published and unpublished tabular data from randomised trials, with indirect estimation of the effects of traditional NSAIDs. Data sources: Medline and Embase (January 1966 to April 2005); Food and Drug Administration records; and data on file from Novartis, Pfizer, and Merck. Review methods: Eligible studies were randomised trials that included a comparison of a selective COX 2 inhibitor versus placebo or a selective COX 2 inhibitor versus a traditional NSAID, of at least four weeks' duration, with information on serious vascular events (defined as myocardial infarction, stroke, or vascular death). Individual investigators and manufacturers provided information on the number of patients randomised, numbers of vascular events, and the person time of follow-up for each randomised group. Results: In placebo comparisons, allocation to a selective COX 2 inhibitor was associated with a 42% relative increase in the incidence of serious vascular events (1.2%/year v 0.9%/year; rate ratio 1.42, 95% confidence interval 1.13 to 1.78; P = 0.003), with no significant heterogeneity among the different selective COX 2 inhibitors. This was chiefly attributable to an increased risk of myocardial infarction (0.6%/year v 0.3%/year; 1.86, 1.33 to 2.59; P = 0.0003), with little apparent difference in other vascular outcomes. Among trials of at least one year's duration (mean 2.7 years), the rate ratio for vascular events was 1.45 (1.12 to 1.89; P = 0.005). Overall, the incidence of serious vascular events was similar between a selective COX 2 inhibitor and any traditional NSAID (1.0%/year v 0.9/%year; 1.16, 0.97 to 1.38; P = 0.1). However, statistical heterogeneity (P = 0.001) was found between trials of a selective COX 2 inhibitor versus naproxen (1.57, 1.21 to 2.03) and of a selective COX 2 inhibitor versus non-naproxen NSAIDs (0.88, 0.69 to 1.12). The summary rate ratio for vascular events, compared with placebo, was 0.92 (0.67 to 1.26) for naproxen, 1.51 (0.96 to 2.37) for ibuprofen, and 1.63 (1.12 to 2.37) for diclofenac. Conclusions: Selective COX 2 inhibitors are associated with a moderate increase in the risk of vascular events, as are high dose regimens of ibuprofen and diclofenac, but high dose naproxen is not associated with such an excess.
Resumo:
Empirical studies of education programs and systems, by nature, rely upon use of student outcomes that are measurable. Often, these come in the form of test scores. However, in light of growing evidence about the long-run importance of other student skills and behaviors, the time has come for a broader approach to evaluating education. This dissertation undertakes experimental, quasi-experimental, and descriptive analyses to examine social, behavioral, and health-related mechanisms of the educational process. My overarching research question is simply, which inside- and outside-the-classroom features of schools and educational interventions are most beneficial to students in the long term? Furthermore, how can we apply this evidence toward informing policy that could effectively reduce stark social, educational, and economic inequalities?
The first study of three assesses mechanisms by which the Fast Track project, a randomized intervention in the early 1990s for high-risk children in four communities (Durham, NC; Nashville, TN; rural PA; and Seattle, WA), reduced delinquency, arrests, and health and mental health service utilization in adolescence through young adulthood (ages 12-20). A decomposition of treatment effects indicates that about a third of Fast Track’s impact on later crime outcomes can be accounted for by improvements in social and self-regulation skills during childhood (ages 6-11), such as prosocial behavior, emotion regulation and problem solving. These skills proved less valuable for the prevention of mental and physical health problems.
The second study contributes new evidence on how non-instructional investments – such as increased spending on school social workers, guidance counselors, and health services – affect multiple aspects of student performance and well-being. Merging several administrative data sources spanning the 1996-2013 school years in North Carolina, I use an instrumental variables approach to estimate the extent to which local expenditure shifts affect students’ academic and behavioral outcomes. My findings indicate that exogenous increases in spending on non-instructional services not only reduce student absenteeism and disciplinary problems (important predictors of long-term outcomes) but also significantly raise student achievement, in similar magnitude to corresponding increases in instructional spending. Furthermore, subgroup analyses suggest that investments in student support personnel such as social workers, health services, and guidance counselors, in schools with concentrated low-income student populations could go a long way toward closing socioeconomic achievement gaps.
The third study examines individual pathways that lead to high school graduation or dropout. It employs a variety of machine learning techniques, including decision trees, random forests with bagging and boosting, and support vector machines, to predict student dropout using longitudinal administrative data from North Carolina. I consider a large set of predictor measures from grades three through eight including academic achievement, behavioral indicators, and background characteristics. My findings indicate that the most important predictors include eighth grade absences, math scores, and age-for-grade as well as early reading scores. Support vector classification (with a high cost parameter and low gamma parameter) predicts high school dropout with the highest overall validity in the testing dataset at 90.1 percent followed by decision trees with boosting and interaction terms at 89.5 percent.
Resumo:
This dissertation is a three-part analysis examining how the welfare state in advanced Western democracies has responded to recent demographic changes. Specifically, this dissertation investigates two primary relationships, beginning with the influence of government spending on poverty. I analyze two at-risk populations in particular: immigrants and children of single mothers. Next, attention is turned to the influence of individual and environmental traits on preferences for social spending. I focus specifically on religiosity, religious beliefs and religious identity. I pool data from a number of international macro- and micro-data sources including the Luxembourg Income Study (LIS), International Social Survey Program (ISSP), the World Bank Databank, and the OECD Databank. Analyses highlight the power of the welfare state to reduce poverty, but also the effectiveness of specific areas of spending focused on addressing new social risks. While previous research has touted the strength of the welfare state, my analyses highlight the need to consider new social risks and encourage closer attention to how social position affects preferences for the welfare state.
Resumo:
The purpose of this study is to examine the effects of agglomeration economies on the productivity of manufacturing local units in Ireland. Four types of agglomeration economies are considered in this study. These are internal economies of scale, localization economies, related variety and urbanization economies. This study makes a number of contributions to the literature. Firstly, this is the first study to conduct an investigation of the effects of agglomeration economies on the productivity of manufacturing local units operating in Ireland. Secondly, this study distinguishes between indigenous and foreign-owned local units which is important given the dual nature of the Irish economy (Krugman, 1997). Thirdly, in addition to considering the effects of agglomeration economies, this study examines the impact of spurious agglomeration on the productivity of foreign-owned local units. Using data from the Census of Industrial Local Units and a series of IV GMM estimators to control for endogeneity, the results of the analysis conducted in Chapter 6 reveal that there are differences in the effects of agglomeration economies on the productivity of indigenous and foreign-owned local units. In Chapter 7 the Census of Industrial Local Units is supplemented by additional data sources and more in-depth measures are generated to capture the features of each of the external agglomeration economies considered in this analysis. There is some evidence to suggest that the availability of local inputs has a negative and significant impact on productivity. The NACE based measures of related variety reveal that the availability of local inputs and knowledge spillovers for related sectors have a negative and significant impact on productivity. There is clear evidence to suggest that urbanization economies are important for increasing the productivity of indigenous local units. The findings reveal that a 1% increase in population density in the NUTS 3 region leads to an increase in the productivity of indigenous local units of approximately 0.07% to 0.08%. The results also reveal that there is a significant difference in the effects of agglomeration economies on the productivity of low-tech and medium/high-tech indigenous local units. The more in-depth measures of agglomeration economies used in Chapter 7 are also used in Chapter 8. A series of IV GMM regressions are estimated in order to identify the impact of agglomeration economies and spurious agglomeration on the productivity of foreign-owned local units operating in Ireland. There is some evidence found to suggest that the availability of a pool of skilled labour has a positive and significant on productivity of foreign-owned local units. There is also evidence to suggest that localization knowledge spillovers have a negative impact on the productivity of foreign-owned local units. There is strong evidence to suggest that the availability of local inputs has a negative impact on the productivity. The negative impact is not confined to the NACE 4-digit sector but also extends into related sectors as determined by Porter’s (2003) cluster classification. The cluster based skills measure of related variety has a positive and significant impact on the productivity of foreign-owned local units. Similar to Chapter 7, there is clear evidence to suggest that urbanization economies are important for increasing the productivity of foreign-owned local units. Both the summary measure and each of the more in-depth measures of agglomeration economies have a positive and significant impact on productivity. Spurious agglomeration has a positive and significant impact on the productivity of foreign-owned local units. The results indicate that the more foreign-owned local units of the same nationality in the country the greater the levels of productivity for the local unit. From a policy perspective, urbanization economies are clearly important for increasing the productivity of both indigenous and foreign-owned local units. Furthermore, the availability of a pool of skilled labour appears to be important for increasing the productivity of foreign-owned local units. Another policy implication that arises from these results relates to the differences observed between indigenous local units and foreign-owned local units and also between low-tech and medium/high-tech indigenous local units. These findings indicate that ‘one-size-fits-all’ type policies are not appropriate for increasing the productivity of local units operating in Ireland. Policies should be tailored to the needs of either indigenous or foreign-owned local units and also to specific sectors. This positive finding for own country spurious agglomeration is important from a policy perspective and is one that IDA Ireland should take on board.
Resumo:
Objectives: The objective of this systematic review was to synthesize the available qualitative evidence on the knowledge, attitudes and beliefs of adult patients, healthcare professionals and carers about oral dosage form modification. Design: A systematic review and synthesis of qualitative studies was undertaken, utilising the thematic synthesis approach. Data sources: The following databases were searched from inception to September 2015: PubMed, Medline (EBSCO), EMBASE, CINAHL, PsycINFO, Web of Science, ProQuest Databases, Scopus, Turning Research Into Practice (TRIP), Cochrane Central Register of Controlled Trials (CENTRAL) and the Cochrane Database of Systematic Reviews (CDSR). Citation tracking and searching the references lists of included studies was also undertaken. Grey literature was searched using the OpenGrey database, internet searching and personal knowledge. An updated search was undertaken in June 2016. Review methods: Studies meeting the following criteria were eligible for inclusion; (i) used qualitative data collection and analysis methods; (ii) full-text was available in English; (iii) included adult patients who require oral dosage forms to be modified to meet their needs or; (iv) carers or healthcare professionals of patients who require oral dosage forms to be modified. Two reviewers independently appraised the quality of the included studies using the Critical Appraisal Skills Programme Checklist. A thematic synthesis was conducted and analytical themes were generated. Results: Of 5455 records screened, seven studies were eligible for inclusion; three involved healthcare professionals and the remaining four studies involved patients. Four analytical themes emerged from the thematic synthesis: (i) patient-centred individuality and variability; (ii) communication; (iii) knowledge and uncertainty and; (iv) complexity. The variability of individual patient’s requirements, poor communication practices and lack of knowledge about oral dosage form modification, when combined with the complex and multi-faceted healthcare environment complicate decision making regarding oral dosage form modification and administration. Conclusions: This systematic review has highlighted the key factors influencing the knowledge, attitudes and beliefs of patients and healthcare professionals about oral dosage form modifications. The findings suggest that in order to optimise oral medicine modification practices the needs of individual patients should be routinely and systematically assessed and decision-making should be supported by evidence based recommendations with multidisciplinary input. Further research is needed to optimise oral dosage form modification practices and the factors identified in this review should be considered in the development of future interventions.
Resumo:
Sub-ice shelf circulation and freezing/melting rates in ocean general circulation models depend critically on an accurate and consistent representation of cavity geometry. Existing global or pan-Antarctic data sets have turned out to contain various inconsistencies and inaccuracies. The goal of this work is to compile independent regional fields into a global data set. We use the S-2004 global 1-minute bathymetry as the backbone and add an improved version of the BEDMAP topography for an area that roughly coincides with the Antarctic continental shelf. Locations of the merging line have been carefully adjusted in order to get the best out of each data set. High-resolution gridded data for upper and lower ice surface topography and cavity geometry of the Amery, Fimbul, Filchner-Ronne, Larsen C and George VI Ice Shelves, and for Pine Island Glacier have been carefully merged into the ambient ice and ocean topographies. Multibeam survey data for bathymetry in the former Larsen B cavity and the southeastern Bellingshausen Sea have been obtained from the data centers of Alfred Wegener Institute (AWI), British Antarctic Survey (BAS) and Lamont-Doherty Earth Observatory (LDEO), gridded, and again carefully merged into the existing bathymetry map. The global 1-minute dataset (RTopo-1 Version 1.0.5) has been split into two netCDF files. The first contains digital maps for global bedrock topography, ice bottom topography, and surface elevation. The second contains the auxiliary maps for data sources and the surface type mask. A regional subset that covers all variables for the region south of 50 deg S is also available in netCDF format. Datasets for the locations of grounding and coast lines are provided in ASCII format.
Resumo:
Since submission of the draft report to the OECD‐LEED Program on 30 June 2007, a unique seven‐years retrospective study of the unemployed high technology workers was released by Statistics Canada.1 Drawing upon Statistics Canada’s confidential Longitudinal Worker File – itself constructed from four administrative data sources that linked Records of Employment and tax filer information by the Social Insurance Number and firm‐level data by a company identifier – this study was able to identify and trace the re‐employment of those permanently laid off in the high‐tech industry by location. The findings are stunning.
Resumo:
Most essay rating research in language assessment has examined human raters’ essay rating as a cognitive process, thus overlooking or oversimplifying the interaction between raters and sociocultural contexts. Given that raters are social beings, their practices have social meanings and consequences. Hence it is important to situate essay rating within its sociocultural context for a more meaningful understanding. Drawing on Engeström’s (1987, 2001) cultural-historical activity theory (CHAT) framework with a sociocultural perspective, this study reconceptualized essay rating as a socially mediated activity with both cognitive (individual raters’ goal-directed decision-making actions) and social layers (raters’ collective object-oriented essay rating activity at related settings). In particular, this study explored raters’ essay rating at one provincial rating centre in China within the context of a high-stakes university entrance examination, the National Matriculation English Test (NMET). This study adopted a multiple-method multiple-perspective qualitative case study design. Think-aloud protocols, stimulated recalls, interviews, and documents served as the data sources. This investigation involved 25 participants at two settings (rating centre and high schools), including rating centre directors, team leaders, NMET essay raters who were high school teachers, and school principals and teaching colleagues of these essay raters. Data were analyzed using Strauss and Corbin’s (1990) open and axial coding techniques, and CHAT for data integration. The findings revealed the interaction between raters and the NMET sociocultural context. Such interaction can be understood through a surface structure (cognitive layer) and a deep structure (social layer) concerning how raters assessed NMET essays, where the surface structure reflected the “what” and the deep structure explained the “how” and “why” in raters’ decision-making. This study highlighted the roles of goals and rules in rater decision-making, rating tensions and raters’ solutions, and the relationship between essay rating and teaching. This study highlights the value of a sociocultural view to essay rating research, demonstrates CHAT as a sociocultural approach to investigate essay rating, and proposes a direction for future washback research on the effect of essay rating. This study also provides support for NMET rating practices that can potentially bring positive washback to English teaching in Chinese high schools.
Resumo:
This study explores the specific characteristics of teacher-educator professional development interventions that enhance their transformative learning towards stimulating the inquiry-based attitude of students. An educational design research method was followed. Firstly, in partnership with five experienced educators, a professional development programme was designed, tested and redesigned. Secondly, a qualitative multiple case study was conducted to examine the active ingredients of the designed interventions with regard to educators changes in beliefs and behaviour. The study was carried out in four different educational settings in which 20 educators participated during nine months. Data sources included videos, questionnaires, interviews and written personal theories of practice. The analyses indicated that aligned self-study interventions on a personal, peer and group level guided by a trained facilitator supported the intended leaning.
Resumo:
Algae biodiesel is a promising but expensive alternative fuel to petro-diesel. To overcome cost barriers, detailed cost analyses are needed. A decade-old cost analysis by the U.S. National Renewable Energy Laboratory indicated that the costs of algae biodiesel were in the range of $0.53–0.85/L (2012 USD values). However, the cost of land and transesterification were just roughly estimated. In this study, an updated comprehensive techno-economic analysis was conducted with optimized processes and improved cost estimations. Latest process improvement, quotes from vendors, government databases, and other relevant data sources were used to calculate the updated algal biodiesel costs, and the final costs of biodiesel are in the range of $0.42–0.97/L. Additional improvements on cost-effective biodiesel production around the globe to cultivate algae was also recommended. Overall, the calculated costs seem promising, suggesting that a single step biodiesel production process is close to commercial reality.
Resumo:
The application of custom classification techniques and posterior probability modeling (PPM) using Worldview-2 multispectral imagery to archaeological field survey is presented in this paper. Research is focused on the identification of Neolithic felsite stone tool workshops in the North Mavine region of the Shetland Islands in Northern Scotland. Sample data from known workshops surveyed using differential GPS are used alongside known non-sites to train a linear discriminant analysis (LDA) classifier based on a combination of datasets including Worldview-2 bands, band difference ratios (BDR) and topographical derivatives. Principal components analysis is further used to test and reduce dimensionality caused by redundant datasets. Probability models were generated by LDA using principal components and tested with sites identified through geological field survey. Testing shows the prospective ability of this technique and significance between 0.05 and 0.01, and gain statistics between 0.90 and 0.94, higher than those obtained using maximum likelihood and random forest classifiers. Results suggest that this approach is best suited to relatively homogenous site types, and performs better with correlated data sources. Finally, by combining posterior probability models and least-cost analysis, a survey least-cost efficacy model is generated showing the utility of such approaches to archaeological field survey.