915 resultados para Secondary Data Analysis


Relevância:

90.00% 90.00%

Publicador:

Resumo:

Le présent mémoire de maîtrise porte sur la médiation préventive en relations du travail. La question de recherche est la suivante : « Pourquoi et comment la médiation préventive du ministère du Travail améliore-t-elle les relations du travail dans les entreprises québécoises? ». La médiation préventive est un mode d’intervention administratif et un mécanisme de soutien additionnel à la législation québécoise offert volontairement depuis plusieurs années par le ministère du Travail aux parties patronale et syndicale qui en font la demande. La médiation préventive se réalise avec le soutien d’un médiateur ou d’une médiatrice du ministère hors de la période de la négociation collective. Son contenu appartient aux parties et les ententes s’effectuent par consensus. La médiation préventive se présente de quatre manières : diagnostic des relations du travail par consensus, diagnostic des relations du travail par entrevue, séminaire en relations du travail et comité de relations du travail. Le modèle conceptuel comporte une variable dépendante qui est les relations du travail, une variable indépendante correspondant à la médiation préventive, le changement organisationnel en tant que variable intermédiaire et une variable de contrôle se rapportant à l’environnement organisationnel. La première des trois hypothèses générales de cette recherche affirme que, toutes choses égales par ailleurs, la médiation préventive amène un changement organisationnel. La seconde hypothèse générale relate que, toutes choses égales par ailleurs, le changement organisationnel contribue à améliorer les relations du travail. La dernière hypothèse générale correspond à l’idée que, toutes choses égales par ailleurs, l’environnement organisationnel a un impact sur les relations du travail. Parmi les douze hypothèses suivantes, quatre hypothèses ont été développées à partir de la première hypothèse générale : 1) toutes choses égales par ailleurs, l’établissement d’un diagnostic des relations du travail par le médiateur va aider à la mise en place d’un plan d’action; 2) toutes choses égales par ailleurs, l'établissement d'un plan d'action par le médiateur contribue au changement organisationnel; 3) toutes choses égales par ailleurs, un suivi par le médiateur contribue au changement organisationnel; 4) toutes choses égales par ailleurs, la perception des deux parties (patronale et syndicale) que le médiateur a bien joué son rôle contribue au changement organisationnel. Les quatre hypothèses suivantes découlent de la seconde hypothèse générale : 1) toutes choses égales par ailleurs, un changement culturel dans l’organisation avec un réaménagement des points de vue de la direction contribue à améliorer les relations du travail; 2) toutes choses égales par ailleurs, un changement relationnel dans l’organisation vers des relations de coopération contribue à améliorer les relations du travail; 3) toutes choses égales par ailleurs, un changement du style de direction dans l’organisation avec l’introduction de valeurs humaines ainsi qu’un « espace de discussion » contribue à améliorer les relations du travail; 4) toutes choses égales par ailleurs, la perception des deux parties (patronale et syndicale) d’avoir bien joué leur rôle contribue à améliorer les relations du travail. Enfin, les quatre dernières hypothèses ont été réalisées à partir de la troisième hypothèse générale : 1) toutes choses égales par ailleurs, l’organisation a un impact sur les relations du travail; 2) toutes choses égales par ailleurs, l’histoire syndicale a un impact sur les relations du travail; 3) toutes choses égales par ailleurs, la perception des contraintes extérieures et des risques pour l’emploi contribue à améliorer les relations du travail; 4) toutes choses égales par ailleurs, la perception d’un équilibre dans les rapports de force entre les parties (patronale et syndicale) contribue à améliorer les relations du travail. La vérification des hypothèses s’est effectuée à l’aide d’entretiens semi-structurés auprès de neuf participants des parties patronale et syndicale d’une entreprise ayant employé une médiation préventive. Une cueillette de données secondaires a aussi eu lieu pour notamment mesurer le niveau de conflit au travail sur dix ans. L’analyse des résultats de cette étude qualitative et longitudinale a permis de confirmer l’ensemble des hypothèses puisque la médiation préventive a amené à un changement organisationnel qui a aidé à améliorer les relations du travail et où l’environnement a eu un impact sur celles-ci. Un modèle d’analyse a été développé et validé pour cette recherche afin de savoir pourquoi et comment la médiation préventive du ministère du Travail améliore les relations du travail dans les entreprises québécoises. La présente recherche démontre de manière empirique que la médiation préventive améliore les relations du travail dans les entreprises québécoises tout en expliquant comment celle-ci procède. De plus, cette recherche suggère à toutes les organisations québécoises syndiquées d’entreprendre une médiation préventive si elles sont aux prises avec de mauvaises relations du travail.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Triple quadrupole mass spectrometers coupled with high performance liquid chromatography are workhorses in quantitative bioanalyses. It provides substantial benefits including reproducibility, sensitivity and selectivity for trace analysis. Selected Reaction Monitoring allows targeted assay development but data sets generated contain very limited information. Data mining and analysis of non-targeted high-resolution mass spectrometry profiles of biological samples offer the opportunity to perform more exhaustive assessments, including quantitative and qualitative analysis. The objectives of this study was to test method precision and accuracy, statistically compare bupivacaine drug concentration in real study samples and verify if high resolution and accurate mass data collected in scan mode can actually permit retrospective data analysis, more specifically, extract metabolite related information. The precision and accuracy data presented using both instruments provided equivalent results. Overall, the accuracy was ranging from 106.2 to 113.2% and the precision observed was from 1.0 to 3.7%. Statistical comparisons using a linear regression between both methods reveal a coefficient of determination (R2) of 0.9996 and a slope of 1.02 demonstrating a very strong correlation between both methods. Individual sample comparison showed differences from -4.5% to 1.6% well within the accepted analytical error. Moreover, post acquisition extracted ion chromatograms at m/z 233.1648 ± 5 ppm (M-56) and m/z 305.2224 ± 5 ppm (M+16) revealed the presence of desbutyl-bupivacaine and three distinct hydroxylated bupivacaine metabolites. Post acquisition analysis allowed us to produce semiquantitative evaluations of the concentration-time profiles for bupicavaine metabolites.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

La présente thèse poursuit l'objectif général de mieux comprendre l'expérience de la transplantation d'organe de la perspective de dyades comprenant un donneur et un receveur d'une donation vivante de rein. Pour ce faire, nous proposons une exploration en profondeur de la trajectoire de transplantation et des enjeux relationnels des donneurs et des receveurs. Un second objectif, qui s'est développé au cours de la recherche, est de contribuer à une meilleure compréhension des enjeux entourant la recherche qualitative impliquant des dyades. Cette thèse est présentée sous la forme de trois articles. Le premier article vise, par le biais de la méthode du méta-résumé, à dresser un portrait le plus complet possible du processus de donation tel qu'il avait été examiné à ce jour par les écrits empiriques qualitatifs disponibles, et à mettre en lumière les aspects qui avaient encore peu été appréhendés par les chercheurs, notamment l'importance de considérer conjointement au sein d'une même recherche des donneurs et des receveurs d'une donation vivante de rein. Le deuxième article a pour but d'explorer en profondeur l'expérience vécue pour des dyades de donneurs et de receveurs en contexte de donation vivante de rein, et en particulier la trajectoire de transplantation ainsi que les enjeux relationnels. Des entrevues qualitatives ont été menées auprès de cinq dyades (donc cinq donneurs et cinq receveurs). L'analyse phénoménologique interprétative des données recueillies propose un portrait de la trajectoire de transplantation, soit (a) l'expérience de la maladie du receveur, (b) l'expérience d'offrir et d'accepter un rein, (c) la période des tests, (d) la transplantation d'organe à proprement parler et finalement (e) la période post-transplantation. Un apport particulièrement significatif de l'analyse et de l'interprétation de nos données est la mise en relief que le processus de transplantation est mieux compris lorsque l'on considère le contexte social et interpersonnel plus large dans lequel évolue chaque dyade donneur-receveur, et en ce sens, la décision de donner ou d'accepter un rein peut être appréhendée comme un prolongement du rôle social auquel la personne adhère par rapport à autrui dans sa vie quotidienne. Le troisième article, qui a émergé des leçons acquises en réalisant la recherche effectuée dans l'article précédent, vise à décrire et discuter des défis éthiques et des impacts des décisions méthodologiques dans le cadre de recherches qualitatives impliquant des dyades de personnes se connaissant. Sur la base de nos propres défis empiriques et des écrits scientifiques disponibles, des réflexions et recommandations sont suggérées à différents stades d'un processus typique de recherche qualitative, soit (a) le recrutement, (b) le consentement à participer, (c) la collecte de données, (d) la validation des transcriptions par les participants, (e) l'analyse de données, et (f) la dissémination des résultats. Une réflexion en amont des enjeux entourant ces étapes est susceptible de faciliter l'entreprise de recherches qualitatives impliquant des dyades. En conclusion, les résultats de notre analyse soulignent l'importance du rôle social et du contexte interpersonnel plus large dans lequel évoluent les donneurs et les receveurs dans la façon dont ils parviennent à donner une signification à leur expérience. Enfin, si la recherche qualitative impliquant des dyades est en émergence, des défis éthiques et des décisions méthodologiques rigoureuses doivent être considérés en amont.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Ce mémoire de maitrise vise à dresser un portrait des erreurs syntaxiques des élèves du secondaire en analysant un corpus de textes de cinq groupes du Québec, de la 1re à la 5e secondaire. Les résultats actuels aux épreuves ministérielles des élèves de 5e secondaire nous indiquent que les élèves éprouvent des difficultés avec l’écriture du français. Une analyse approfondie nous permet de comprendre que l’amélioration de la situation passe par une meilleure connaissance des erreurs syntaxiques des élèves. En nous appuyant sur la grille de Boivin et Pinsonneault (2014), nous avons analysé les données provenant du codage des textes d’élèves de la 1re à la 5e secondaire. L’analyse de ces données nous a permis de constater que parmi les sept grandes catégories d’erreurs de la grille, c’est en syntaxe que les élèves commettent le plus d’erreurs. Une incursion au cœur des six sous-catégories de la syntaxe a révélé que la ponctuation causait le plus de problème aux élèves, et ce, à tous les niveaux. Les erreurs liées à la détermination de la catégorie grammaticale des mots (homophones) arrivaient en deuxième place. Par la suite, nous avons précisé davantage l’analyse en déterminant, pour chacun des codes, l’évolution du nombre d’erreurs d’un niveau du secondaire à l’autre. Il est ressorti de cette étude que les deux principales erreurs, basées sur les sous-catégories syntaxiques, sont celles portant sur l’usage de la virgule et celles liées à la confusion qui existe encore un verbe terminant par «er» et un adjectif ou un participe passé terminant par «é-e-s».

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The study has wider policy implications as it identifies the possible variables which influence the sustainability of participatory productive sector projects. The method which is developed to study the sustainability of projects under People’s Planning in Chempu Panchayat could be used for studying the same in other panchayats also. Unlike the case of the standard features of sustainability identified, the independent variables vary according to the nature of the project. Hence, this needs to be modified accordingly while applying the method in a dissimilar domain. Selection of a single panchayat for the present study is relevant on the basis of a common package of inputs for decentralised planning which is forwarded by the State Planning Board respectively for the three-tier panchayat system in Kerala. The dynamic filed realities could be brought out in view of a comprehensive planning approach through an in depth study of specific cases.The assessment of the nature and pattern of productive sector projects in the selected Village Panchayat puts the projects under close scrutiny. The analysis has depended largely on secondary sources of information, especially from panchayat level plan documents, and also on the primary information obtained using direct observation and on-site inspection of project sites. An analysis of the nature and pattem of productive sector projects is important as it gives all necessary information regarding follow-up, monitoring/evaluation and even termination of a particular project. It has also revealed the tendencies of including infrastructure and service sector projects under ‘productive’ category, especially for maintaining the stipulated ratio (40:30:30) of grant-in-aid distribution. The study regarding the allocation and expenditure pattern of plan funds is vital in policy level as it reveals the under-noticed allocation and expenditure pattern of plan funds other than grant-in-aid. One major limitation of the study has been the limited availability of secondary data, especially regarding project-wise expenditure and monitoring/evaluation reports of various project committees.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Microarray data analysis is one of data mining tool which is used to extract meaningful information hidden in biological data. One of the major focuses on microarray data analysis is the reconstruction of gene regulatory network that may be used to provide a broader understanding on the functioning of complex cellular systems. Since cancer is a genetic disease arising from the abnormal gene function, the identification of cancerous genes and the regulatory pathways they control will provide a better platform for understanding the tumor formation and development. The major focus of this thesis is to understand the regulation of genes responsible for the development of cancer, particularly colorectal cancer by analyzing the microarray expression data. In this thesis, four computational algorithms namely fuzzy logic algorithm, modified genetic algorithm, dynamic neural fuzzy network and Takagi Sugeno Kang-type recurrent neural fuzzy network are used to extract cancer specific gene regulatory network from plasma RNA dataset of colorectal cancer patients. Plasma RNA is highly attractive for cancer analysis since it requires a collection of small amount of blood and it can be obtained at any time in repetitive fashion allowing the analysis of disease progression and treatment response.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

The country has witnessed tremendous increase in the vehicle population and increased axle loading pattern during the last decade, leaving its road network overstressed and leading to premature failure. The type of deterioration present in the pavement should be considered for determining whether it has a functional or structural deficiency, so that appropriate overlay type and design can be developed. Structural failure arises from the conditions that adversely affect the load carrying capability of the pavement structure. Inadequate thickness, cracking, distortion and disintegration cause structural deficiency. Functional deficiency arises when the pavement does not provide a smooth riding surface and comfort to the user. This can be due to poor surface friction and texture, hydro planning and splash from wheel path, rutting and excess surface distortion such as potholes, corrugation, faulting, blow up, settlement, heaves etc. Functional condition determines the level of service provided by the facility to its users at a particular time and also the Vehicle Operating Costs (VOC), thus influencing the national economy. Prediction of the pavement deterioration is helpful to assess the remaining effective service life (RSL) of the pavement structure on the basis of reduction in performance levels, and apply various alternative designs and rehabilitation strategies with a long range funding requirement for pavement preservation. In addition, they can predict the impact of treatment on the condition of the sections. The infrastructure prediction models can thus be classified into four groups, namely primary response models, structural performance models, functional performance models and damage models. The factors affecting the deterioration of the roads are very complex in nature and vary from place to place. Hence there is need to have a thorough study of the deterioration mechanism under varied climatic zones and soil conditions before arriving at a definite strategy of road improvement. Realizing the need for a detailed study involving all types of roads in the state with varying traffic and soil conditions, the present study has been attempted. This study attempts to identify the parameters that affect the performance of roads and to develop performance models suitable to Kerala conditions. A critical review of the various factors that contribute to the pavement performance has been presented based on the data collected from selected road stretches and also from five corporations of Kerala. These roads represent the urban conditions as well as National Highways, State Highways and Major District Roads in the sub urban and rural conditions. This research work is a pursuit towards a study of the road condition of Kerala with respect to varying soil, traffic and climatic conditions, periodic performance evaluation of selected roads of representative types and development of distress prediction models for roads of Kerala. In order to achieve this aim, the study is focused into 2 parts. The first part deals with the study of the pavement condition and subgrade soil properties of urban roads distributed in 5 Corporations of Kerala; namely Thiruvananthapuram, Kollam, Kochi, Thrissur and Kozhikode. From selected 44 roads, 68 homogeneous sections were studied. The data collected on the functional and structural condition of the surface include pavement distress in terms of cracks, potholes, rutting, raveling and pothole patching. The structural strength of the pavement was measured as rebound deflection using Benkelman Beam deflection studies. In order to collect the details of the pavement layers and find out the subgrade soil properties, trial pits were dug and the in-situ field density was found using the Sand Replacement Method. Laboratory investigations were carried out to find out the subgrade soil properties, soil classification, Atterberg limits, Optimum Moisture Content, Field Moisture Content and 4 days soaked CBR. The relative compaction in the field was also determined. The traffic details were also collected by conducting traffic volume count survey and axle load survey. From the data thus collected, the strength of the pavement was calculated which is a function of the layer coefficient and thickness and is represented as Structural Number (SN). This was further related to the CBR value of the soil and the Modified Structural Number (MSN) was found out. The condition of the pavement was represented in terms of the Pavement Condition Index (PCI) which is a function of the distress of the surface at the time of the investigation and calculated in the present study using deduct value method developed by U S Army Corps of Engineers. The influence of subgrade soil type and pavement condition on the relationship between MSN and rebound deflection was studied using appropriate plots for predominant types of soil and for classified value of Pavement Condition Index. The relationship will be helpful for practicing engineers to design the overlay thickness required for the pavement, without conducting the BBD test. Regression analysis using SPSS was done with various trials to find out the best fit relationship between the rebound deflection and CBR, and other soil properties for Gravel, Sand, Silt & Clay fractions. The second part of the study deals with periodic performance evaluation of selected road stretches representing National Highway (NH), State Highway (SH) and Major District Road (MDR), located in different geographical conditions and with varying traffic. 8 road sections divided into 15 homogeneous sections were selected for the study and 6 sets of continuous periodic data were collected. The periodic data collected include the functional and structural condition in terms of distress (pothole, pothole patch, cracks, rutting and raveling), skid resistance using a portable skid resistance pendulum, surface unevenness using Bump Integrator, texture depth using sand patch method and rebound deflection using Benkelman Beam. Baseline data of the study stretches were collected as one time data. Pavement history was obtained as secondary data. Pavement drainage characteristics were collected in terms of camber or cross slope using camber board (slope meter) for the carriage way and shoulders, availability of longitudinal side drain, presence of valley, terrain condition, soil moisture content, water table data, High Flood Level, rainfall data, land use and cross slope of the adjoining land. These data were used for finding out the drainage condition of the study stretches. Traffic studies were conducted, including classified volume count and axle load studies. From the field data thus collected, the progression of each parameter was plotted for all the study roads; and validated for their accuracy. Structural Number (SN) and Modified Structural Number (MSN) were calculated for the study stretches. Progression of the deflection, distress, unevenness, skid resistance and macro texture of the study roads were evaluated. Since the deterioration of the pavement is a complex phenomena contributed by all the above factors, pavement deterioration models were developed as non linear regression models, using SPSS with the periodic data collected for all the above road stretches. General models were developed for cracking progression, raveling progression, pothole progression and roughness progression using SPSS. A model for construction quality was also developed. Calibration of HDM–4 pavement deterioration models for local conditions was done using the data for Cracking, Raveling, Pothole and Roughness. Validation was done using the data collected in 2013. The application of HDM-4 to compare different maintenance and rehabilitation options were studied considering the deterioration parameters like cracking, pothole and raveling. The alternatives considered for analysis were base alternative with crack sealing and patching, overlay with 40 mm BC using ordinary bitumen, overlay with 40 mm BC using Natural Rubber Modified Bitumen and an overlay of Ultra Thin White Topping. Economic analysis of these options was done considering the Life Cycle Cost (LCC). The average speed that can be obtained by applying these options were also compared. The results were in favour of Ultra Thin White Topping over flexible pavements. Hence, Design Charts were also plotted for estimation of maximum wheel load stresses for different slab thickness under different soil conditions. The design charts showed the maximum stress for a particular slab thickness and different soil conditions incorporating different k values. These charts can be handy for a design engineer. Fuzzy rule based models developed for site specific conditions were compared with regression models developed using SPSS. The Riding Comfort Index (RCI) was calculated and correlated with unevenness to develop a relationship. Relationships were developed between Skid Number and Macro Texture of the pavement. The effort made through this research work will be helpful to highway engineers in understanding the behaviour of flexible pavements in Kerala conditions and for arriving at suitable maintenance and rehabilitation strategies. Key Words: Flexible Pavements – Performance Evaluation – Urban Roads – NH – SH and other roads – Performance Models – Deflection – Riding Comfort Index – Skid Resistance – Texture Depth – Unevenness – Ultra Thin White Topping

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Futures trading in Commodities has three specific economic functions viz. price discovery, hedging and reduction in volatility. Natural rubber possesses all the specifications required for futures trading. Commodity futures trading in India attained momentum after the starting of national level commodity exchanges in 2003. The success of futures trading depends upon effective price risk management, price discovery and reduced volatility which in turn depends upon the volume of trading. In the case of rubber futures market, the volume of trading depends upon the extent of participation by market players like growers, dealers, manufacturers, rubber marketing co-operative societies and Rubber Producer’s Societies (RPS). The extent of participation by market players has a direct bearing on their awareness level and their perception about futures trading. In the light of the above facts and the review of literature available on rubber futures market, it is felt that a study on rubber futures market is necessary to fill the research gap, with specific focus on (1) the awareness and perception of rubber futures market participants viz. (i) rubber growers, (ii) dealers, (iii) rubber product manufacturers, (iv) rubber marketing co-operative societies and Rubber Producer’s Societies (RPS) about futures trading and (2) whether the rubber futures market is fulfilling the economic functions of futures market viz. hedging, reduction in volatility and price discovery or not. The study is confined to growers, dealers, rubber goods manufacturers, rubber marketing co-operative societies and RPS in Kerala. In order to achieve the stated objectives, the study utilized secondary data for the period from 2003 to 2013 from different published sources like bulletins, newsletters, circulars from NMCE, Reserve Bank of India (RBI), Warehousing Corporation and traders. The primary data required for this study were collected from rubber growers, rubber dealers, RPS & Rubber Marketing Co-operative Societies and rubber goods manufacturers in Kerala. Data pertaining to the awareness and perception of futures trading, participation in the futures trading, use of spot and futures prices and source of price information by dealers, farmers, manufacturers and cooperative societies also were collected. Statistical tools used for analysis include percentage, standard deviation, Chi-square test, Mann – Whitney U test, Kruskal Wallis test, Augmented Dickey – Fuller test statistic, t- statistic, Granger causality test, F- statistic, Johansen co – integration test, Trace statistic and Max –Eigen statistic. The study found that 71.5 per cent of the total hedges are effective and 28.5 per cent are ineffective for the period under study. It implies that futures market in rubber reduced the impact of price risks by approximately 71.5 per cent. Further, it is observed that, on 54.4 per cent occasions, the futures market exercised a stabilizing effect on the spot market, and on 45.6 per cent occasions futures trading exercised a destabilizing effect on the spot market. It implies that elasticity of expectation of futures market in rubber has a predominant stabilizing effect on spot prices. The market, as a whole, exhibits a bias in favour of long hedges. Spot price volatility of rubber during futures suspension period is more than that of the pre suspension period and post suspension period. There is a bi-directional association-ship or bi-directional causality or pair- wise causality between spot price and futures price of rubber. From the results of the hedging efficiency, spot price volatility, and price discovery, it can be concluded that rubber futures market fulfils all the economic functions expected from a commodity futures market. Thus in India, the future of rubber futures is Bright…!!!

Relevância:

90.00% 90.00%

Publicador:

Resumo:

While most data analysis and decision support tools use numerical aspects of the data, Conceptual Information Systems focus on their conceptual structure. This paper discusses how both approaches can be combined.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Empowerment is one possible approach to increase political, social and economic participation of a target group and thus decrease food insecurity among vulnerable people. The objective of this research was to evaluate a socio-economic empowerment project implemented for 8 years in rural northern Bangladesh by using secondary data from 2000 and primary data from 2009. Three hundred women were primarily selected out of all existing and active 121 women groups formed during 8 years project implementation. Standardized interviews were conducted to collect quantitative data. In addition an overall qualitative analysis included focus group discussions, the review of project summaries and discussions with implementing staff. Data was collected on socio-economic status, rice storages and food consumption. In addition knowledge was gained on current activities of the women groups. The empowerment concept implemented in this setting was successful to increase the participants’ political, social and economic engagement and visibility. Apart from that the utilization of skills for income generation led to an improvement of the socio-economic status and food insecurity became less common. Recommendations for future projects include the increase of female members among the NGO staff as well as the further follow-up and strengthening of political acceptance and visibility of the empowered participants to sustain the outcome of the project.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Die zunehmende Vernetzung der Informations- und Kommunikationssysteme führt zu einer weiteren Erhöhung der Komplexität und damit auch zu einer weiteren Zunahme von Sicherheitslücken. Klassische Schutzmechanismen wie Firewall-Systeme und Anti-Malware-Lösungen bieten schon lange keinen Schutz mehr vor Eindringversuchen in IT-Infrastrukturen. Als ein sehr wirkungsvolles Instrument zum Schutz gegenüber Cyber-Attacken haben sich hierbei die Intrusion Detection Systeme (IDS) etabliert. Solche Systeme sammeln und analysieren Informationen von Netzwerkkomponenten und Rechnern, um ungewöhnliches Verhalten und Sicherheitsverletzungen automatisiert festzustellen. Während signatur-basierte Ansätze nur bereits bekannte Angriffsmuster detektieren können, sind anomalie-basierte IDS auch in der Lage, neue bisher unbekannte Angriffe (Zero-Day-Attacks) frühzeitig zu erkennen. Das Kernproblem von Intrusion Detection Systeme besteht jedoch in der optimalen Verarbeitung der gewaltigen Netzdaten und der Entwicklung eines in Echtzeit arbeitenden adaptiven Erkennungsmodells. Um diese Herausforderungen lösen zu können, stellt diese Dissertation ein Framework bereit, das aus zwei Hauptteilen besteht. Der erste Teil, OptiFilter genannt, verwendet ein dynamisches "Queuing Concept", um die zahlreich anfallenden Netzdaten weiter zu verarbeiten, baut fortlaufend Netzverbindungen auf, und exportiert strukturierte Input-Daten für das IDS. Den zweiten Teil stellt ein adaptiver Klassifikator dar, der ein Klassifikator-Modell basierend auf "Enhanced Growing Hierarchical Self Organizing Map" (EGHSOM), ein Modell für Netzwerk Normalzustand (NNB) und ein "Update Model" umfasst. In dem OptiFilter werden Tcpdump und SNMP traps benutzt, um die Netzwerkpakete und Hostereignisse fortlaufend zu aggregieren. Diese aggregierten Netzwerkpackete und Hostereignisse werden weiter analysiert und in Verbindungsvektoren umgewandelt. Zur Verbesserung der Erkennungsrate des adaptiven Klassifikators wird das künstliche neuronale Netz GHSOM intensiv untersucht und wesentlich weiterentwickelt. In dieser Dissertation werden unterschiedliche Ansätze vorgeschlagen und diskutiert. So wird eine classification-confidence margin threshold definiert, um die unbekannten bösartigen Verbindungen aufzudecken, die Stabilität der Wachstumstopologie durch neuartige Ansätze für die Initialisierung der Gewichtvektoren und durch die Stärkung der Winner Neuronen erhöht, und ein selbst-adaptives Verfahren eingeführt, um das Modell ständig aktualisieren zu können. Darüber hinaus besteht die Hauptaufgabe des NNB-Modells in der weiteren Untersuchung der erkannten unbekannten Verbindungen von der EGHSOM und der Überprüfung, ob sie normal sind. Jedoch, ändern sich die Netzverkehrsdaten wegen des Concept drif Phänomens ständig, was in Echtzeit zur Erzeugung nicht stationärer Netzdaten führt. Dieses Phänomen wird von dem Update-Modell besser kontrolliert. Das EGHSOM-Modell kann die neuen Anomalien effektiv erkennen und das NNB-Model passt die Änderungen in Netzdaten optimal an. Bei den experimentellen Untersuchungen hat das Framework erfolgversprechende Ergebnisse gezeigt. Im ersten Experiment wurde das Framework in Offline-Betriebsmodus evaluiert. Der OptiFilter wurde mit offline-, synthetischen- und realistischen Daten ausgewertet. Der adaptive Klassifikator wurde mit dem 10-Fold Cross Validation Verfahren evaluiert, um dessen Genauigkeit abzuschätzen. Im zweiten Experiment wurde das Framework auf einer 1 bis 10 GB Netzwerkstrecke installiert und im Online-Betriebsmodus in Echtzeit ausgewertet. Der OptiFilter hat erfolgreich die gewaltige Menge von Netzdaten in die strukturierten Verbindungsvektoren umgewandelt und der adaptive Klassifikator hat sie präzise klassifiziert. Die Vergleichsstudie zwischen dem entwickelten Framework und anderen bekannten IDS-Ansätzen zeigt, dass der vorgeschlagene IDSFramework alle anderen Ansätze übertrifft. Dies lässt sich auf folgende Kernpunkte zurückführen: Bearbeitung der gesammelten Netzdaten, Erreichung der besten Performanz (wie die Gesamtgenauigkeit), Detektieren unbekannter Verbindungen und Entwicklung des in Echtzeit arbeitenden Erkennungsmodells von Eindringversuchen.

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Modern methods of compositional data analysis are not well known in biomedical research. Moreover, there appear to be few mathematical and statistical researchers working on compositional biomedical problems. Like the earth and environmental sciences, biomedicine has many problems in which the relevant scienti c information is encoded in the relative abundance of key species or categories. I introduce three problems in cancer research in which analysis of compositions plays an important role. The problems involve 1) the classi cation of serum proteomic pro les for early detection of lung cancer, 2) inference of the relative amounts of di erent tissue types in a diagnostic tumor biopsy, and 3) the subcellular localization of the BRCA1 protein, and it's role in breast cancer patient prognosis. For each of these problems I outline a partial solution. However, none of these problems is \solved". I attempt to identify areas in which additional statistical development is needed with the hope of encouraging more compositional data analysts to become involved in biomedical research

Relevância:

90.00% 90.00%

Publicador:

Resumo:

This analysis was stimulated by the real data analysis problem of household expenditure data. The full dataset contains expenditure data for a sample of 1224 households. The expenditure is broken down at 2 hierarchical levels: 9 major levels (e.g. housing, food, utilities etc.) and 92 minor levels. There are also 5 factors and 5 covariates at the household level. Not surprisingly, there are a small number of zeros at the major level, but many zeros at the minor level. The question is how best to model the zeros. Clearly, models that try to add a small amount to the zero terms are not appropriate in general as at least some of the zeros are clearly structural, e.g. alcohol/tobacco for households that are teetotal. The key question then is how to build suitable conditional models. For example, is the sub-composition of spending excluding alcohol/tobacco similar for teetotal and non-teetotal households? In other words, we are looking for sub-compositional independence. Also, what determines whether a household is teetotal? Can we assume that it is independent of the composition? In general, whether teetotal will clearly depend on the household level variables, so we need to be able to model this dependence. The other tricky question is that with zeros on more than one component, we need to be able to model dependence and independence of zeros on the different components. Lastly, while some zeros are structural, others may not be, for example, for expenditure on durables, it may be chance as to whether a particular household spends money on durables within the sample period. This would clearly be distinguishable if we had longitudinal data, but may still be distinguishable by looking at the distribution, on the assumption that random zeros will usually be for situations where any non-zero expenditure is not small. While this analysis is based on around economic data, the ideas carry over to many other situations, including geological data, where minerals may be missing for structural reasons (similar to alcohol), or missing because they occur only in random regions which may be missed in a sample (similar to the durables)

Relevância:

90.00% 90.00%

Publicador:

Resumo:

As stated in Aitchison (1986), a proper study of relative variation in a compositional data set should be based on logratios, and dealing with logratios excludes dealing with zeros. Nevertheless, it is clear that zero observations might be present in real data sets, either because the corresponding part is completely absent –essential zeros– or because it is below detection limit –rounded zeros. Because the second kind of zeros is usually understood as “a trace too small to measure”, it seems reasonable to replace them by a suitable small value, and this has been the traditional approach. As stated, e.g. by Tauber (1999) and by Martín-Fernández, Barceló-Vidal, and Pawlowsky-Glahn (2000), the principal problem in compositional data analysis is related to rounded zeros. One should be careful to use a replacement strategy that does not seriously distort the general structure of the data. In particular, the covariance structure of the involved parts –and thus the metric properties– should be preserved, as otherwise further analysis on subpopulations could be misleading. Following this point of view, a non-parametric imputation method is introduced in Martín-Fernández, Barceló-Vidal, and Pawlowsky-Glahn (2000). This method is analyzed in depth by Martín-Fernández, Barceló-Vidal, and Pawlowsky-Glahn (2003) where it is shown that the theoretical drawbacks of the additive zero replacement method proposed in Aitchison (1986) can be overcome using a new multiplicative approach on the non-zero parts of a composition. The new approach has reasonable properties from a compositional point of view. In particular, it is “natural” in the sense that it recovers the “true” composition if replacement values are identical to the missing values, and it is coherent with the basic operations on the simplex. This coherence implies that the covariance structure of subcompositions with no zeros is preserved. As a generalization of the multiplicative replacement, in the same paper a substitution method for missing values on compositional data sets is introduced

Relevância:

90.00% 90.00%

Publicador:

Resumo:

Hydrogeological research usually includes some statistical studies devised to elucidate mean background state, characterise relationships among different hydrochemical parameters, and show the influence of human activities. These goals are achieved either by means of a statistical approach or by mixing models between end-members. Compositional data analysis has proved to be effective with the first approach, but there is no commonly accepted solution to the end-member problem in a compositional framework. We present here a possible solution based on factor analysis of compositions illustrated with a case study. We find two factors on the compositional bi-plot fitting two non-centered orthogonal axes to the most representative variables. Each one of these axes defines a subcomposition, grouping those variables that lay nearest to it. With each subcomposition a log-contrast is computed and rewritten as an equilibrium equation. These two factors can be interpreted as the isometric log-ratio coordinates (ilr) of three hidden components, that can be plotted in a ternary diagram. These hidden components might be interpreted as end-members. We have analysed 14 molarities in 31 sampling stations all along the Llobregat River and its tributaries, with a monthly measure during two years. We have obtained a bi-plot with a 57% of explained total variance, from which we have extracted two factors: factor G, reflecting geological background enhanced by potash mining; and factor A, essentially controlled by urban and/or farming wastewater. Graphical representation of these two factors allows us to identify three extreme samples, corresponding to pristine waters, potash mining influence and urban sewage influence. To confirm this, we have available analysis of diffused and widespread point sources identified in the area: springs, potash mining lixiviates, sewage, and fertilisers. Each one of these sources shows a clear link with one of the extreme samples, except fertilisers due to the heterogeneity of their composition. This approach is a useful tool to distinguish end-members, and characterise them, an issue generally difficult to solve. It is worth note that the end-member composition cannot be fully estimated but only characterised through log-ratio relationships among components. Moreover, the influence of each endmember in a given sample must be evaluated in relative terms of the other samples. These limitations are intrinsic to the relative nature of compositional data