970 resultados para Statistical Method
Resumo:
The Brazilian CAPES Journal Portal aims to provide Information in Science and Technology (IST) for academic users. Thus, it is considered a relevant instrument for post-graduation dynamics and the country´s Science and Technology (S&T) development. Despite its importance, there are still few studies that focus on the policy analysis and efficiency of these resources. This research aims to fill in this gap once it proposes an analysis of the use of the CAPES Journal Portal done on behalf of the master´s and doctoral alumni of the Post Graduate Program in Management (PPGA) at the Federal University of Rio Grande do Norte (UFRN). The operationalization of the research´s main objective was possible through the specific objectives: characterize graduate profile as CAPES Journal Portal users b) identify motivation for the use of CAPES Journal Portal c) detect graduate satisfaction degree in information seeking done at CAPES Journal Portal d) verify graduate satisfaction regarding the use of the CAPES Journal Portal e) verify the use of the information that is obtained by graduates in the development of their academic activities. The research is of descriptive nature employing a mixed methodological strategy in which quantitative approach predominates. Data collection was done through a web survey questionnaire. Quantitative data analysis was made possible through the use of a statistical method. As for qualitative analysis, there was use of the Brenda Dervin´s sense-making approach as well as content analysis in open ended questions. The research samples were composed by 90 graduate students who had defended their dissertation/thesis in the PPGA program at UFRN in the time span of 2010-2013. This represented by 88% of this population. As for user profile, the analysis has made evident that there are no quantitative differences related to gender. There is predominance of male graduates that were aged 26 to 30 years old. As for female graduates, the great majority were 31 o 35 years old. Most graduates had Master´s degree scholarship in order to support their study. It was also seen that the great majority claim to use the Portal during their post graduation studies. The main reasons responsible for non use was: preference for the use of other data bases and lack of knowledge regarding the Portal. It was observed that the most used information resources were theses and dissertations. Data also indicate preference for complete text. Those who have used the Portal also claimed to have used other electronic information fonts in order to fulfill their information needs. The information fonts that were researched outside in the Portal were monographs, dissertations and thesis. Scielo was the most used information font. Results reveal that access and use of the Portal has been done in a regular manner during post graduation studies. But on the other hand, graduates also make use of other electronic information fonts in order to meet their information needs. The study also confirmed the important mission performed by the Portal regarding Brazilian scientific communication production. This was seen even though users have reported the need for improvement in some aspects such as: periodic training in order to promote, encourage and teach more effective use of the portal; investment aiming the expansion of Social Sciences Collection in the Portal as well as the need to implement continuous evaluation process related to user satisfaction in regarding the services provided.
Resumo:
In this study we investigate the potential of organic-walled dinoflagellate cysts (dinocysts) as tools for quantifying past sea-surface temperatures (SST) in the Southern Ocean. For this purpose, a dinocyst reference dataset has been formed, based on 138 surface sediment samples from different circum-Antarctic environments. The dinocyst assemblages of these samples are composed of phototrophic (gonyaulacoid) and heterotrophic (protoperidinioid) species that provide a broad spectrum of palaeoenvironmental information. The relationship between the environmental parameters in the upper water column and the dinocyst distribution patterns of individual species has been established using the statistical method of Canonical Correspondence Analysis (CCA). Among the variables tested, summer SST appeared to correspond to the maximum variance represented in the dataset. To establish quantitative summer SST reconstructions, a Modern Analogue Technique (MAT) has been performed on data from three Late Quaternary dinocyst records recovered from locations adjacent to prominent oceanic fronts in the Atlantic sector of the Southern Ocean. These dinocyst time series exhibit periodic changes in the dinocyst assemblage during the last two glacial/interglacial-cycles. During glacial conditions the relative abundance of protoperidinioid cysts was highest, whereas interglacial conditions are characterised by generally lower cyst concentrations and increased relative abundance of gonyaulacoid cysts. The MAT palaeotemperature estimates show trends in summer SST changes following the global oxygen isotope signal and a strong correlation with past temperatures of the last 140,000 years based on other proxies. However, by comparing the dinocyst results to quantitative estimates of summer SSTs based on diatoms, radiolarians and foraminifer-derived stable isotope records it can be shown that in several core intervals the dinocyst-based summer SSTs appeared to be extremely high. In these intervals the dinocyst record seems to be highly influenced by selective degradation, leading to unusual temperature ranges and to unrealistic palaeotemperatures. We used the selective degradation index (kt-index) to determine those intervals that have been biased by selective degradation in order to correct the palaeotemperature estimates. We show that after correction the dinocyst based SSTs correspond reasonably well with other palaeotemperature estimates for this region, supporting the great potential of dinoflagellate cysts as a basis for quantitative palaeoenvironmental studies.
Resumo:
Les modèles incrémentaux sont des modèles statistiques qui ont été développés initialement dans le domaine du marketing. Ils sont composés de deux groupes, un groupe contrôle et un groupe traitement, tous deux comparés par rapport à une variable réponse binaire (le choix de réponses est « oui » ou « non »). Ces modèles ont pour but de détecter l’effet du traitement sur les individus à l’étude. Ces individus n’étant pas tous des clients, nous les appellerons : « prospects ». Cet effet peut être négatif, nul ou positif selon les caractéristiques des individus composants les différents groupes. Ce mémoire a pour objectif de comparer des modèles incrémentaux d’un point de vue bayésien et d’un point de vue fréquentiste. Les modèles incrémentaux utilisés en pratique sont ceux de Lo (2002) et de Lai (2004). Ils sont initialement réalisés d’un point de vue fréquentiste. Ainsi, dans ce mémoire, l’approche bayésienne est utilisée et comparée à l’approche fréquentiste. Les simulations sont e ectuées sur des données générées avec des régressions logistiques. Puis, les paramètres de ces régressions sont estimés avec des simulations Monte-Carlo dans l’approche bayésienne et comparés à ceux obtenus dans l’approche fréquentiste. L’estimation des paramètres a une influence directe sur la capacité du modèle à bien prédire l’effet du traitement sur les individus. Nous considérons l’utilisation de trois lois a priori pour l’estimation des paramètres de façon bayésienne. Elles sont choisies de manière à ce que les lois a priori soient non informatives. Les trois lois utilisées sont les suivantes : la loi bêta transformée, la loi Cauchy et la loi normale. Au cours de l’étude, nous remarquerons que les méthodes bayésiennes ont un réel impact positif sur le ciblage des individus composant les échantillons de petite taille.
Resumo:
Thesis (Ph.D.)--University of Washington, 2016-08
Resumo:
Dans ce projet de recherche, le dépôt des couches minces de carbone amorphe (généralement connu sous le nom de DLC pour Diamond-Like Carbon en anglais) par un procédé de dépôt chimique en phase vapeur assisté par plasma (ou PECVD pour Plasma Enhanced Chemical Vapor deposition en anglais) a été étudié en utilisant la Spectroscopie d’Émission Optique (OES) et l’analyse partielle par régression des moindres carrés (PLSR). L’objectif de ce mémoire est d’établir un modèle statistique pour prévoir les propriétés des revêtements DLC selon les paramètres du procédé de déposition ou selon les données acquises par OES. Deux séries d’analyse PLSR ont été réalisées. La première examine la corrélation entre les paramètres du procédé et les caractéristiques du plasma pour obtenir une meilleure compréhension du processus de dépôt. La deuxième série montre le potentiel de la technique d’OES comme outil de surveillance du procédé et de prédiction des propriétés de la couche déposée. Les résultats montrent que la prédiction des propriétés des revêtements DLC qui était possible jusqu’à maintenant en se basant sur les paramètres du procédé (la pression, la puissance, et le mode du plasma), serait envisageable désormais grâce aux informations obtenues par OES du plasma (particulièrement les indices qui sont reliées aux concentrations des espèces dans le plasma). En effet, les données obtenues par OES peuvent être utilisées pour surveiller directement le processus de dépôt plutôt que faire une étude complète de l’effet des paramètres du processus, ceux-ci étant strictement reliés au réacteur plasma et étant variables d’un laboratoire à l’autre. La perspective de l’application d’un modèle PLSR intégrant les données de l’OES est aussi démontrée dans cette recherche afin d’élaborer et surveiller un dépôt avec une structure graduelle.
Resumo:
The aim of this research is twofold: Firstly, to model and solve a complex nurse scheduling problem with an integer programming formulation and evolutionary algorithms. Secondly, to detail a novel statistical method of comparing and hence building better scheduling algorithms by identifying successful algorithm modifications. The comparison method captures the results of algorithms in a single figure that can then be compared using traditional statistical techniques. Thus, the proposed method of comparing algorithms is an objective procedure designed to assist in the process of improving an algorithm. This is achieved even when some results are non-numeric or missing due to infeasibility. The final algorithm outperforms all previous evolutionary algorithms, which relied on human expertise for modification.
Resumo:
Introduction Prediction of soft tissue changes following orthognathic surgery has been frequently attempted in the past decades. It has gradually progressed from the classic “cut and paste” of photographs to the computer assisted 2D surgical prediction planning; and finally, comprehensive 3D surgical planning was introduced to help surgeons and patients to decide on the magnitude and direction of surgical movements as well as the type of surgery to be considered for the correction of facial dysmorphology. A wealth of experience was gained and numerous published literature is available which has augmented the knowledge of facial soft tissue behaviour and helped to improve the ability to closely simulate facial changes following orthognathic surgery. This was particularly noticed following the introduction of the three dimensional imaging into the medical research and clinical applications. Several approaches have been considered to mathematically predict soft tissue changes in three dimensions, following orthognathic surgery. The most common are the Finite element model and Mass tensor Model. These were developed into software packages which are currently used in clinical practice. In general, these methods produce an acceptable level of prediction accuracy of soft tissue changes following orthognathic surgery. Studies, however, have shown a limited prediction accuracy at specific regions of the face, in particular the areas around the lips. Aims The aim of this project is to conduct a comprehensive assessment of hard and soft tissue changes following orthognathic surgery and introduce a new method for prediction of facial soft tissue changes. Methodology The study was carried out on the pre- and post-operative CBCT images of 100 patients who received their orthognathic surgery treatment at Glasgow dental hospital and school, Glasgow, UK. Three groups of patients were included in the analysis; patients who underwent Le Fort I maxillary advancement surgery; bilateral sagittal split mandibular advancement surgery or bimaxillary advancement surgery. A generic facial mesh was used to standardise the information obtained from individual patient’s facial image and Principal component analysis (PCA) was applied to interpolate the correlations between the skeletal surgical displacement and the resultant soft tissue changes. The identified relationship between hard tissue and soft tissue was then applied on a new set of preoperative 3D facial images and the predicted results were compared to the actual surgical changes measured from their post-operative 3D facial images. A set of validation studies was conducted. To include: • Comparison between voxel based registration and surface registration to analyse changes following orthognathic surgery. The results showed there was no statistically significant difference between the two methods. Voxel based registration, however, showed more reliability as it preserved the link between the soft tissue and skeletal structures of the face during the image registration process. Accordingly, voxel based registration was the method of choice for superimposition of the pre- and post-operative images. The result of this study was published in a refereed journal. • Direct DICOM slice landmarking; a novel technique to quantify the direction and magnitude of skeletal surgical movements. This method represents a new approach to quantify maxillary and mandibular surgical displacement in three dimensions. The technique includes measuring the distance of corresponding landmarks digitized directly on DICOM image slices in relation to three dimensional reference planes. The accuracy of the measurements was assessed against a set of “gold standard” measurements extracted from simulated model surgery. The results confirmed the accuracy of the method within 0.34mm. Therefore, the method was applied in this study. The results of this validation were published in a peer refereed journal. • The use of a generic mesh to assess soft tissue changes using stereophotogrammetry. The generic facial mesh played a major role in the soft tissue dense correspondence analysis. The conformed generic mesh represented the geometrical information of the individual’s facial mesh on which it was conformed (elastically deformed). Therefore, the accuracy of generic mesh conformation is essential to guarantee an accurate replica of the individual facial characteristics. The results showed an acceptable overall mean error of the conformation of generic mesh 1 mm. The results of this study were accepted for publication in peer refereed scientific journal. Skeletal tissue analysis was performed using the validated “Direct DICOM slices landmarking method” while soft tissue analysis was performed using Dense correspondence analysis. The analysis of soft tissue was novel and produced a comprehensive description of facial changes in response to orthognathic surgery. The results were accepted for publication in a refereed scientific Journal. The main soft tissue changes associated with Le Fort I were advancement at the midface region combined with widening of the paranasal, upper lip and nostrils. Minor changes were noticed at the tip of the nose and oral commissures. The main soft tissue changes associated with mandibular advancement surgery were advancement and downward displacement of the chin and lower lip regions, limited widening of the lower lip and slight reversion of the lower lip vermilion combined with minimal backward displacement of the upper lip were recorded. Minimal changes were observed on the oral commissures. The main soft tissue changes associated with bimaxillary advancement surgery were generalized advancement of the middle and lower thirds of the face combined with widening of the paranasal, upper lip and nostrils regions. In Le Fort I cases, the correlation between the changes of the facial soft tissue and the skeletal surgical movements was assessed using PCA. A statistical method known as ’Leave one out cross validation’ was applied on the 30 cases which had Le Fort I osteotomy surgical procedure to effectively utilize the data for the prediction algorithm. The prediction accuracy of soft tissue changes showed a mean error ranging between (0.0006mm±0.582) at the nose region to (-0.0316mm±2.1996) at the various facial regions.
Resumo:
This thesis presents an analysis of the largest catalog to date of infrared spectra of massive young stellar objects in the Large Magellanic Cloud. Evidenced by their very different spectral features, the luminous objects span a range of evolutionary states from those most embedded in their natal molecular material to those that have dissipated and ionized their surroundings to form compact HII regions and photodissociation regions. We quantify the contributions of the various spectral features using the statistical method of principal component analysis. Using this analysis, we classify the YSO spectra into several distinct groups based upon their dominant spectral features: silicate absorption (S Group), silicate absorption and fine-structure line emission (SE), polycyclic aromatic hydrocarbon (PAH) emission (P Group), PAH and fine-structure line emission (PE), and only fine-structure line emission (E). Based upon the relative numbers of sources in each category, we are able to estimate the amount of time massive YSOs spend in each evolutionary stage. We find that approximately 50% of the sources have ionic fine-structure lines, indicating that a compact HII region forms about half-way through the YSO lifetime probed in our study. Of the 277 YSOs we collected spectra for, 41 have ice absorption features, indicating they are surrounded by cold ice-bearing dust particles. We have decomposed the shape of the ice features to probe the composition and thermal history of the ice. We find that most the CO2 ice is embedded a polar ice matrix that has been thermally processed by the embedded YSO. The amount of thermal processing may be correlated with the luminosity of the YSO. Using the Australia Telescope Compact Array, we imaged the dense gas around a subsample of our sources in the HII complexes N44, N105, N113, and N159 using HCO+ and HCN as dense gas tracers. We find that the molecular material in star forming environments is highly clumpy, with clumps that range from subparsec to ~2 parsecs in size and with masses between 10^2 to 10^4 solar masses. We find that there are varying levels of star formation in the clumps, with the lower-mass clumps tending to be without massive YSOs. These YSO-less clumps could either represent an earlier stage of clump to the more massive YSO-bearing ones or clumps that will never form a massive star. Clumps with massive YSOs at their centers have masses larger than those with massive YSOs at their edges, and we suggest that the difference is evolutionary: edge YSO clumps are more advanced than those with YSOs at their centers. Clumps with YSOs at their edges may have had a significant fraction of their mass disrupted or destroyed by the forming massive star. We find that the strength of the silicate absorption seen in YSO IR spectra feature is well-correlated with the on-source HCO+ and HCN flux densities, such that the strength of the feature is indicative of the embeddedness of the YSO. We estimate that ~40% of the entire spectral sample has strong silicate absorption features, implying that the YSOs are embedded in circumstellar material for about 40% of the time probed in our study.
Resumo:
The aim of this research is twofold: Firstly, to model and solve a complex nurse scheduling problem with an integer programming formulation and evolutionary algorithms. Secondly, to detail a novel statistical method of comparing and hence building better scheduling algorithms by identifying successful algorithm modifications. The comparison method captures the results of algorithms in a single figure that can then be compared using traditional statistical techniques. Thus, the proposed method of comparing algorithms is an objective procedure designed to assist in the process of improving an algorithm. This is achieved even when some results are non-numeric or missing due to infeasibility. The final algorithm outperforms all previous evolutionary algorithms, which relied on human expertise for modification.
Resumo:
The aim of this research is twofold: Firstly, to model and solve a complex nurse scheduling problem with an integer programming formulation and evolutionary algorithms. Secondly, to detail a novel statistical method of comparing and hence building better scheduling algorithms by identifying successful algorithm modifications. The comparison method captures the results of algorithms in a single figure that can then be compared using traditional statistical techniques. Thus, the proposed method of comparing algorithms is an objective procedure designed to assist in the process of improving an algorithm. This is achieved even when some results are non-numeric or missing due to infeasibility. The final algorithm outperforms all previous evolutionary algorithms, which relied on human expertise for modification.
Resumo:
Les modèles incrémentaux sont des modèles statistiques qui ont été développés initialement dans le domaine du marketing. Ils sont composés de deux groupes, un groupe contrôle et un groupe traitement, tous deux comparés par rapport à une variable réponse binaire (le choix de réponses est « oui » ou « non »). Ces modèles ont pour but de détecter l’effet du traitement sur les individus à l’étude. Ces individus n’étant pas tous des clients, nous les appellerons : « prospects ». Cet effet peut être négatif, nul ou positif selon les caractéristiques des individus composants les différents groupes. Ce mémoire a pour objectif de comparer des modèles incrémentaux d’un point de vue bayésien et d’un point de vue fréquentiste. Les modèles incrémentaux utilisés en pratique sont ceux de Lo (2002) et de Lai (2004). Ils sont initialement réalisés d’un point de vue fréquentiste. Ainsi, dans ce mémoire, l’approche bayésienne est utilisée et comparée à l’approche fréquentiste. Les simulations sont e ectuées sur des données générées avec des régressions logistiques. Puis, les paramètres de ces régressions sont estimés avec des simulations Monte-Carlo dans l’approche bayésienne et comparés à ceux obtenus dans l’approche fréquentiste. L’estimation des paramètres a une influence directe sur la capacité du modèle à bien prédire l’effet du traitement sur les individus. Nous considérons l’utilisation de trois lois a priori pour l’estimation des paramètres de façon bayésienne. Elles sont choisies de manière à ce que les lois a priori soient non informatives. Les trois lois utilisées sont les suivantes : la loi bêta transformée, la loi Cauchy et la loi normale. Au cours de l’étude, nous remarquerons que les méthodes bayésiennes ont un réel impact positif sur le ciblage des individus composant les échantillons de petite taille.
Resumo:
Wydział Nauk Politycznych i Dziennikarstwa
Resumo:
Lernvideos haben sich als geeignetes Instrument erwiesen Lernprozesse von Studierenden effektiv zu unterstützen. Vorliegender Beitrag thematisiert die videobasierte Statistiklehre an Hochschulen und stellt die Ergebnisse einer Experimentalstudie vor. Darin zeigt sich, dass bivariate Zusammenhänge zwischen der Bewertung von Lernvideos und dem Lernergebnis bestehen und dass Studierende bei theoriebasierten Lernvideos die besten Lernergebnisse erzielen. (DIPF/Orig.)
Resumo:
Dissertação (mestrado)—Universidade de Brasília, Instituto de Física, Programa de Pós-Graduação em Física, 2016.
Resumo:
Dissertação (mestrado)—Universidade de Brasília, Faculdade de Economia, Administração e Contabilidade, Programa de Pós-Graduação em Administração, 2016.