858 resultados para Robust Probabilistic Model, Dyslexic Users, Rewriting, Question-Answering
Resumo:
The objective of this Master’s thesis is to create a calculation model for working capital management in value chains. The study has been executed using literature review and constructive research methods. Constructive research methods were mainly modeling. The theory in this thesis is founded in research articles and management literature. The model is developed for students and researchers. They can use the model for working capital management and comparing firms to each other. The model can also be used to cash management. The model tells who benefits and who suffers most in the value chain. Companies and value chains cash flows can be seen. By using the model can be seen are the set targets really achieved. The amount of operational working capital can be observed. The model enables user to simulate the amount of working capital. The created model is based on cash conversion cycle, return on investment and cash flow forecasting. The model is tested with carefully considered figures which seem to be though realistic. The modeled value chain is literally a chain. Implementing this model requires from the user that he/she have some kind of understanding about working capital management and some figures from balance sheet and income statement. By using this model users can improve their knowledge about working capital management in value chains.
Resumo:
Open innovation paradigm states that the boundaries of the firm have become permeable, allowing knowledge to flow inwards and outwards to accelerate internal innovations and take unused knowledge to the external environment; respectively. The successful implementation of open innovation practices in firms like Procter & Gamble, IBM, and Xerox, among others; suggest that it is a sustainable trend which could provide basis for achieving competitive advantage. However, implementing open innovation could be a complex process which involves several domains of management; and whose term, classification, and practices have not totally been agreed upon. Thus, with many possible ways to address open innovation, the following research question was formulated: How could Ericsson LMF assess which open innovation mode to select depending on the attributes of the project at hand? The research followed the constructive research approach which has the following steps: find a practical relevant problem, obtain general understanding of the topic, innovate the solution, demonstrate the solution works, show theoretical contributions, and examine the scope of applicability of the solution. The research involved three phases of data collection and analysis: Extensive literature review of open innovation, strategy, business model, innovation, and knowledge management; direct observation of the environment of the case company through participative observation; and semi-structured interviews based of six cases involving multiple and heterogeneous open innovation initiatives. Results from the cases suggest that the selection of modes depend on multiple reasons, with a stronger influence of factors related to strategy, business models, and resources gaps. Based on these and others factors found in the literature review and observations; it was possible to construct a model that supports approaching open innovation. The model integrates perspectives from multiple domains of the literature review, observations inside the case company, and factors from the six open innovation cases. It provides steps, guidelines, and tools to approach open innovation and assess the selection of modes. Measuring the impact of open innovation could take years; thus, implementing and testing entirely the model was not possible due time limitation. Nevertheless, it was possible to validate the core elements of the model with empirical data gathered from the cases. In addition to constructing the model, this research contributed to the literature by increasing the understanding of open innovation, providing suggestions to the case company, and proposing future steps.
Resumo:
Human activity recognition in everyday environments is a critical, but challenging task in Ambient Intelligence applications to achieve proper Ambient Assisted Living, and key challenges still remain to be dealt with to realize robust methods. One of the major limitations of the Ambient Intelligence systems today is the lack of semantic models of those activities on the environment, so that the system can recognize the speci c activity being performed by the user(s) and act accordingly. In this context, this thesis addresses the general problem of knowledge representation in Smart Spaces. The main objective is to develop knowledge-based models, equipped with semantics to learn, infer and monitor human behaviours in Smart Spaces. Moreover, it is easy to recognize that some aspects of this problem have a high degree of uncertainty, and therefore, the developed models must be equipped with mechanisms to manage this type of information. A fuzzy ontology and a semantic hybrid system are presented to allow modelling and recognition of a set of complex real-life scenarios where vagueness and uncertainty are inherent to the human nature of the users that perform it. The handling of uncertain, incomplete and vague data (i.e., missing sensor readings and activity execution variations, since human behaviour is non-deterministic) is approached for the rst time through a fuzzy ontology validated on real-time settings within a hybrid data-driven and knowledgebased architecture. The semantics of activities, sub-activities and real-time object interaction are taken into consideration. The proposed framework consists of two main modules: the low-level sub-activity recognizer and the high-level activity recognizer. The rst module detects sub-activities (i.e., actions or basic activities) that take input data directly from a depth sensor (Kinect). The main contribution of this thesis tackles the second component of the hybrid system, which lays on top of the previous one, in a superior level of abstraction, and acquires the input data from the rst module's output, and executes ontological inference to provide users, activities and their in uence in the environment, with semantics. This component is thus knowledge-based, and a fuzzy ontology was designed to model the high-level activities. Since activity recognition requires context-awareness and the ability to discriminate among activities in di erent environments, the semantic framework allows for modelling common-sense knowledge in the form of a rule-based system that supports expressions close to natural language in the form of fuzzy linguistic labels. The framework advantages have been evaluated with a challenging and new public dataset, CAD-120, achieving an accuracy of 90.1% and 91.1% respectively for low and high-level activities. This entails an improvement over both, entirely data-driven approaches, and merely ontology-based approaches. As an added value, for the system to be su ciently simple and exible to be managed by non-expert users, and thus, facilitate the transfer of research to industry, a development framework composed by a programming toolbox, a hybrid crisp and fuzzy architecture, and graphical models to represent and con gure human behaviour in Smart Spaces, were developed in order to provide the framework with more usability in the nal application. As a result, human behaviour recognition can help assisting people with special needs such as in healthcare, independent elderly living, in remote rehabilitation monitoring, industrial process guideline control, and many other cases. This thesis shows use cases in these areas.
Resumo:
Ensikielen jälkeen opittavan kielen tutkimusta ja suomi toisena kielenä alaa sen osana ovat koko niiden olemassaolon ajan hallinneet samat peruskysymykset: millaista oppiminen on eri vaiheissa ja eri ympäristöissä, sekä mikä oppimisessa on yleistä ja toisaalta mikä riippuu opittavasta kielestä ja oppijoiden kielitaustasta. Sähköisten oppijankielen tutkimusaineistojen eli korpusten lisääntymisen myötä tutkijat voivat aiempaa helpommin tutkia näitä ilmiöitä määrällisesti ja tarkastella oppijankielen sisäistä vaihtelua ja sen suhdetta tyypilliseen ensikieliseen kielenkäyttöön kielen eri osa-alueilla käyttöpohjaisesti eli todelliseen kielenkäyttöön pohjautuen. Tekninen kehitys on tuonut mukanaan aineisto- eli korpusvetoisuuden kaltaisia uusia tapoja lähestyä tutkimusaineistoa, jolloin tyypillisiä tutkimuskysymyksiä ”Miksi?” ja ”Miten?” edeltää kysymys: ”Mikä?”. Tässä väitöskirjassa tarkastellaan edistyneiden suomenoppijoiden kirjoitettua akateemista kieltä ja suhteutetaan suomen oppimiselle ominaisia seikkoja käyttöpohjaisen mallin perusolettamuksiin. Aineisto on suomea toisena kielenä käyttävien opiskelijoiden tenttivastauksia, ja se on osa Edistyneiden suomenoppijoiden korpusta. Tutkimus on osin metodologinen, sillä väitöskirjassa esitellään ja siinä sovelletaan uutta korpusvetoista avainrakenneanalyysi-menetelmää, jonka avulla aineistoa lähestytään ilman hypoteeseja siitä, mitkä kielen ilmiöt ovat ominaisia edistyneelle oppijansuomelle. Tutkimus kuuluu kieliopin tutkimuksen piiriin, ja se nojaa kognitiivisen konstruktiokieliopin ajatukseen abstraktiudeltaan vaihtelevista konstruktioista kielijärjestelmän perusyksiköinä. Tulokset puoltavat menetelmän sovellettavuutta kielen oppimisen tutkimukseen, sillä sen avulla kyettiin tunnistamaan konstruktioita, jotka erottavat edistyneitä oppijoita ensikielisistä kirjoittajista (esim. modaaliset verbiketjut), eri ensikieliä puhuvia suomenoppijoita (esim. konjunktiot) sekä konstruktioita, joiden käyttö muuttuu ajan kuluessa (esim. preteriti ja preesens). Monet havaitut erot ovat akateemisen kirjoittamisen erityispiirteitä, mikä tukee ajatusta kielen käyttö- ja kontekstikohtaisesta oppimisesta. Tuloksia voidaan yhtäältä soveltaa akateemisen kielitaidon opetuksessa. Toisaalta menetelmää voidaan käyttää kielenoppimisen tutkimuksen ohella uusien näkökulmien kartoittamiseksi erilaisten tai eri-ikäisten tekstien tyypillisten ominaisuuksien ja erojen tutkimuksessa.
Resumo:
Mass transfer kinetics in osmotic dehydration is usually modeled by Fick's law, empirical models and probabilistic models. The aim of this study was to determine the applicability of Peleg model to investigate the mass transfer during osmotic dehydration of mackerel (Scomber japonicus) slices at different temperatures. Osmotic dehydration was performed on mackerel slices by cooking-infusion in solutions with glycerol and salt (a w = 0.64) at different temperatures: 50, 70, and 90 ºC. Peleg rate constant (K1) (h(g/gdm)-1) varied with temperature variation from 0.761 to 0.396 for water loss, from 5.260 to 2.947 for salt gain, and from 0.854 to 0.566 for glycerol intake. In all cases, it followed the Arrhenius relationship (R²>0.86). The Ea (kJ / mol) values obtained were 16.14; 14.21, and 10.12 for water, salt, and glycerol, respectively. The statistical parameters that qualify the goodness of fit (R²>0.91 and RMSE<0.086) indicate promising applicability of Peleg model.
Resumo:
Phenomena in cyber domain, especially threats to security and privacy, have proven an increasingly heated topic addressed by different writers and scholars at an increasing pace – both nationally and internationally. However little public research has been done on the subject of cyber intelligence. The main research question of the thesis was: To what extent is the applicability of cyber intelligence acquisition methods circumstantial? The study was conducted in sequential a manner, starting with defining the concept of intelligence in cyber domain and identifying its key attributes, followed by identifying the range of intelligence methods in cyber domain, criteria influencing their applicability, and types of operatives utilizing cyber intelligence. The methods and criteria were refined into a hierarchical model. The existing conceptions of cyber intelligence were mapped through an extensive literature study on a wide variety of sources. The established understanding was further developed through 15 semi-structured interviews with experts of different backgrounds, whose wide range of points of view proved to substantially enhance the perspective on the subject. Four of the interviewed experts participated in a relatively extensive survey based on the constructed hierarchical model on cyber intelligence that was formulated in to an AHP hierarchy and executed in the Expert Choice Comparion online application. It was concluded that Intelligence in cyber domain is an endorsing, cross-cutting intelligence discipline that adds value to all aspects of conventional intelligence and furthermore that it bears a substantial amount of characteristic traits – both advantageous and disadvantageous – and furthermore that the applicability of cyber intelligence methods is partly circumstantially limited.
Resumo:
Tämä diplomityö arvioi hitsauksen laadunhallintaohjelmistomarkkinoiden kilpailijoita. Kilpailukenttä on uusi ja ei ole tarkkaa tietoa siitä minkälaisia kilpailijoita on markkinoilla. Hitsauksen laadunhallintaohjelmisto auttaa yrityksiä takaamaan korkean laadun. Ohjelmisto takaa korkean laadun varmistamalla, että hitsaaja on pätevä, hän noudattaa hitsausohjeita ja annettuja parametreja. Sen lisäksi ohjelmisto kerää kaiken tiedon hitsausprosessista ja luo siitä vaadittavat dokumentit. Diplomityön teoriaosuus muodostuu kirjallisuuskatsauksesta ratkaisuliike-toimintaan, kilpailija-analyysin ja kilpailuvoimien teoriaan sekä hitsauksen laadunhallintaan. Työn empiriaosuus on laadullinen tutkimus, jossa tutkitaan kilpailevia hitsauksen laadunhallintaohjelmistoja ja haastatellaan ohjelmistojen käyttäjiä. Diplomityön tuloksena saadaan uusi kilpailija-analyysimalli hitsauksen laadunhallintaohjelmistoille. Mallin avulla voidaan arvostella ohjelmistot niiden tarjoamien primääri- ja sekundääriominaisuuksien perusteella. Toiseksi tässä diplomityössä analysoidaan nykyinen kilpailijatilanne hyödyntämällä juuri kehitettyä kilpailija-analyysimallia.
Resumo:
The importance of industrial maintenance has been emphasized during the last decades; it is no longer a mere cost item, but one of the mainstays of business. Market conditions have worsened lately, investments in production assets have decreased, and at the same time competition has changed from taking place between companies to competition between networks. Companies have focused on their core functions and outsourced support services, like maintenance, above all to decrease costs. This new phenomenon has led to increasing formation of business networks. As a result, a growing need for new kinds of tools for managing these networks effectively has arisen. Maintenance costs are usually a notable part of the life-cycle costs of an item, and it is important to be able to plan the future maintenance operations for the strategic period of the company or for the whole life-cycle period of the item. This thesis introduces an itemlevel life-cycle model (LCM) for industrial maintenance networks. The term item is used as a common definition for a part, a component, a piece of equipment etc. The constructed LCM is a working tool for a maintenance network (consisting of customer companies that buy maintenance services and various supplier companies). Each network member is able to input their own cost and profit data related to the maintenance services of one item. As a result, the model calculates the net present values of maintenance costs and profits and presents them from the points of view of all the network members. The thesis indicates that previous LCMs for calculating maintenance costs have often been very case-specific, suitable only for the item in question, and they have also been constructed for the needs of a single company, without the network perspective. The developed LCM is a proper tool for the decision making of maintenance services in the network environment; it enables analysing the past and making scenarios for the future, and offers choices between alternative maintenance operations. The LCM is also suitable for small companies in building active networks to offer outsourcing services for large companies. The research introduces also a five-step constructing process for designing a life-cycle costing model in the network environment. This five-step designing process defines model components and structure throughout the iteration and exploitation of user feedback. The same method can be followed to develop other models. The thesis contributes to the literature of value and value elements of maintenance services. It examines the value of maintenance services from the perspective of different maintenance network members and presents established value element lists for the customer and the service provider. These value element lists enable making value visible in the maintenance operations of a networked business. The LCM added with value thinking promotes the notion of maintenance from a “cost maker” towards a “value creator”.
Resumo:
The meaning of information technology (IT) and information systems have increased during the last few years. This is mainly because business is nowadays seen more and more as a service business and IT is one of the key elements to support those business services. Since the meaning of IT services has increased also the meaning of IT service support should be a factor paid more attention to. Especially after a merger and acquisition (M&A) it is more important than ever to consider service support. The purpose of this study is to discover the best practices for choosing a suitable service support model. The research question is How to choose a service support organization model for the ERP service desk function after a merger? A qualitative method is selected as a research method. This thesis includes two parts: a literature review and a case study. Theoretical part compiles an integrated model of previous research on the topic. It consists a collection of academic articles, publications and reports. The empirical part focuses on the issues in the case organization. That part tries to answer the question: what would be the most suitable service support model for the case organization? The empirical part is conducted by interviewing the employees of the case organization. This study finds that even though there are many ways of selecting a service support model it is difficult to define an unambiguous guidelines. However, there are few main objectives that should be taken into account regardless the case. Especially by using ITIL processes it is possible to implement a comprehensive service support and raise overall awareness of the existing service support models. The main functions that need to be taken into account are nature, industry and size of the organization. Also the business strategy, goals and resources need to be considered. These are the same factors that are noticed in the case study as well. The suggestions for the case organization are presented based on the interviews and the literature review.
Resumo:
Most of the applications of airborne laser scanner data to forestry require that the point cloud be normalized, i.e., each point represents height from the ground instead of elevation. To normalize the point cloud, a digital terrain model (DTM), which is derived from the ground returns in the point cloud, is employed. Unfortunately, extracting accurate DTMs from airborne laser scanner data is a challenging task, especially in tropical forests where the canopy is normally very thick (partially closed), leading to a situation in which only a limited number of laser pulses reach the ground. Therefore, robust algorithms for extracting accurate DTMs in low-ground-point-densitysituations are needed in order to realize the full potential of airborne laser scanner data to forestry. The objective of this thesis is to develop algorithms for processing airborne laser scanner data in order to: (1) extract DTMs in demanding forest conditions (complex terrain and low number of ground points) for applications in forestry; (2) estimate canopy base height (CBH) for forest fire behavior modeling; and (3) assess the robustness of LiDAR-based high-resolution biomass estimation models against different field plot designs. Here, the aim is to find out if field plot data gathered by professional foresters can be combined with field plot data gathered by professionally trained community foresters and used in LiDAR-based high-resolution biomass estimation modeling without affecting prediction performance. The question of interest in this case is whether or not the local forest communities can achieve the level technical proficiency required for accurate forest monitoring. The algorithms for extracting DTMs from LiDAR point clouds presented in this thesis address the challenges of extracting DTMs in low-ground-point situations and in complex terrain while the algorithm for CBH estimation addresses the challenge of variations in the distribution of points in the LiDAR point cloud caused by things like variations in tree species and season of data acquisition. These algorithms are adaptive (with respect to point cloud characteristics) and exhibit a high degree of tolerance to variations in the density and distribution of points in the LiDAR point cloud. Results of comparison with existing DTM extraction algorithms showed that DTM extraction algorithms proposed in this thesis performed better with respect to accuracy of estimating tree heights from airborne laser scanner data. On the other hand, the proposed DTM extraction algorithms, being mostly based on trend surface interpolation, can not retain small artifacts in the terrain (e.g., bumps, small hills and depressions). Therefore, the DTMs generated by these algorithms are only suitable for forestry applications where the primary objective is to estimate tree heights from normalized airborne laser scanner data. On the other hand, the algorithm for estimating CBH proposed in this thesis is based on the idea of moving voxel in which gaps (openings in the canopy) which act as fuel breaks are located and their height is estimated. Test results showed a slight improvement in CBH estimation accuracy over existing CBH estimation methods which are based on height percentiles in the airborne laser scanner data. However, being based on the idea of moving voxel, this algorithm has one main advantage over existing CBH estimation methods in the context of forest fire modeling: it has great potential in providing information about vertical fuel continuity. This information can be used to create vertical fuel continuity maps which can provide more realistic information on the risk of crown fires compared to CBH.
Resumo:
Organizations often consider investing in a new Enterprise Resource Planning (ERP) system as a way to enhance their business processes, as it allows integrating information used by multiple different departments into a harmonized computing system. The hope of gaining significant business benefits, such as reducing operating costs, is the key reason why organizations have decided to invest in ERP systems since 1990’s. Still, all ERP projects do not end up in success, and deployment of ERP system does not necessarily guarantee the results people were waiting for. This research studies why organizations invest in ERP, but also what downsides ERP projects currently have. Additionally Enterprise Application Integrations (EAI) as next generation’s ERP solutions are studied to challenge and develop traditional ERP. The research questions are: What are the weaknesses in traditional ERP deployment in today’s business? How does the proposed next generation’s ERP answer to these weaknesses? At the beginning of the thesis, as an answer to the first research question, the basics of ERP implementation are introduced with both the pros and cons of investing in ERP. Key concepts such as IS integration and EAI are also studied. Empirical section of the thesis focuses on answering the second research question from the integration approach. A qualitative research is executed by interviewing five experienced IT professionals about EAI benefits, limitations, and problems. The thematic interview and questionnaire follow the presented ERP main elements from literature. The research shows that adopting traditional ERP includes multiple downsides, e.g. inflexibility and requiring big investments in terms of money. To avoid these critical issues, organizations could find a solution from integrations between their current IS. Based on the empirical study a new framework for the next generation’s ERP is created, consisting of a model and a framework that deal with various features regarding IS adoption. With this framework organizations can assess whether they should implement EAI or ERP. The model and framework suggest that there are multiple factors IT managers needs to consider when planning their IT investments, including their current IS, role of IT in the organization, as well as new system’s flexibility, investment level, and number of vendors. The framework created in the thesis encourages IT management to assess holistically their i) organization, ii) its IT, and iii) solution requirements in order to determine what kind of IS solution would suit their needs the best.
Resumo:
Organizations often consider investing in a new Enterprise Resource Planning (ERP) system as a way to enhance their business processes, as it allows integrating information used by multiple different departments into a harmonized computing system. The hope of gaining significant business benefits, such as reducing operating costs, is the key reason why organizations have decided to invest in ERP systems since 1990’s. Still, all ERP projects do not end up in success, and deployment of ERP system does not necessarily guarantee the results people were waiting for. This research studies why organizations invest in ERP, but also what downsides ERP projects currently have. Additionally Enterprise Application Integrations (EAI) as next generation’s ERP solutions are studied to challenge and develop traditional ERP. The research questions are: What are the weaknesses in traditional ERP deployment in today’s business? How does the proposed next generation’s ERP answer to these weaknesses? At the beginning of the thesis, as an answer to the first research question, the basics of ERP implementation are introduced with both the pros and cons of investing in ERP. Key concepts such as IS integration and EAI are also studied. Empirical section of the thesis focuses on answering the second research question from the integration approach. A qualitative research is executed by interviewing five experienced IT professionals about EAI benefits, limitations, and problems. The thematic interview and questionnaire follow the presented ERP main elements from literature. The research shows that adopting traditional ERP includes multiple downsides, e.g. inflexibility and requiring big investments in terms of money. To avoid these critical issues, organizations could find a solution from integrations between their current IS. Based on the empirical study a new framework for the next generation’s ERP is created, consisting of a model and a framework that deal with various features regarding IS adoption. With this framework organizations can assess whether they should implement EAI or ERP. The model and framework suggest that there are multiple factors IT managers needs to consider when planning their IT investments, including their current IS, role of IT in the organization, as well as new system’s flexibility, investment level, and number of vendors. The framework created in the thesis encourages IT management to assess holistically their i) organization, ii) its IT, and iii) solution requirements in order to determine what kind of IS solution would suit their needs the best.
Resumo:
Modern automobiles are no longer just mechanical tools. The electronics and computing services they are shipping with are making them not less than a computer. They are massive kinetic devices with sophisticated computing power. Most of the modern vehicles are made with the added connectivity in mind which may be vulnerable to outside attack. Researchers have shown that it is possible to infiltrate into a vehicle’s internal system remotely and control the physical entities such as steering and brakes. It is quite possible to experience such attacks on a moving vehicle and unable to use the controls. These massive connected computers can be life threatening as they are related to everyday lifestyle. First part of this research studied the attack surfaces in the automotive cybersecurity domain. It also illustrated the attack methods and capabilities of the damages. Online survey has been deployed as data collection tool to learn about the consumers’ usage of such vulnerable automotive services. The second part of the research portrayed the consumers’ privacy in automotive world. It has been found that almost hundred percent of modern vehicles has the capabilities to send vehicle diagnostic data as well as user generated data to their manufacturers, and almost thirty five percent automotive companies are collecting them already. Internet privacy has been studies before in many related domain but no privacy scale were matched for automotive consumers. It created the research gap and motivation for this thesis. A study has been performed to use well established consumers privacy scale – IUIPC to match with the automotive consumers’ privacy situation. Hypotheses were developed based on the IUIPC model for internet consumers’ privacy and they were studied by the finding from the data collection methods. Based on the key findings of the research, all the hypotheses were accepted and hence it is found that automotive consumers’ privacy did follow the IUIPC model under certain conditions. It is also found that a majority of automotive consumers use the services and devices that are vulnerable and prone to cyber-attacks. It is also established that there is a market for automotive cybersecurity services and consumers are willing to pay certain fees to avail that.
Resumo:
The purpose of this thesis is to examine various policy implementation models, and to determine what use they are to a government. In order to insure that governmental proposals are created and exercised in an effective manner, there roust be some guidelines in place which will assist in resolving difficult situations. All governments face the challenge of responding to public demand, by delivering the type of policy responses that will attempt to answer those demands. The problem for those people in positions of policy-making responsibility is to balance the competitive forces that would influence policy. This thesis examines provincial government policy in two unique cases. The first is the revolutionary recommendations brought forth in the Hall -Dennis Report. The second is the question of extending full -funding to the end of high school in the separate school system. These two cases illustrate how divergent and problematic the policy-making duties of any government may be. In order to respond to these political challenges decision-makers must have a clear understanding of what they are attempting to do. They must also have an assortment of policy-making models that will insure a policy response effectively deals with the issue under examination. A government must make every effort to insure that all policymaking methods are considered, and that the data gathered is inserted into the most appropriate model. Currently, there is considerable debate over the benefits of the progressive individualistic education approach as proposed by the Hall -Dennis Committee. This debate is usually intensified during periods of economic uncertainty. Periodically, the province will also experience brief yet equally intense debate on the question of separate school funding. At one level, this debate centres around the efficiency of maintaining two parallel education systems, but the debate frequently has undertones of the religious animosity common in Ontario's history. As a result of the two policy cases under study we may ask ourselves these questions: a) did the policies in question improve the general quality of life in the province? and b) did the policies unite the province? In the cases of educational instruction and finance the debate is ongoing and unsettling. Currently, there is a widespread belief that provincial students at the elementary and secondary levels of education are not being educated adequately to meet the challenges of the twenty-first century. The perceived culprit is individual education which sees students progressing through the system at their own pace and not meeting adequate education standards. The question of the finance of Catholic education occasionally rears its head in a painful fashion within the province. Some public school supporters tend to take extension as a personal religious defeat, rather than an opportunity to demonstrate that educational diversity can be accommodated within Canada's most populated province. This thesis is an attempt to analyze how successful provincial policy-implementation models were in answering public demand. A majority of the public did not demand additional separate school funding, yet it was put into place. The same majority did insist on an examination of educational methods, and the government did put changes in place. It will also demonstrate how policy if wisely created may spread additional benefits to the public at large. Catholic students currently enjoy a much improved financial contribution from the province, yet these additional funds were taken from somewhere. The public system had it funds reduced with what would appear to be minimal impact. This impact indicates that government policy is still sensitive to the strongly held convictions of those people in opposition to a given policy.
Resumo:
The purpose of this study was to replicate and extend a motivational model of problem drinking (Cooper, Frone, Russel, & Mudar, 1995; Read, Wood, Kahler, Maddock & Tibor, 2003), testing the notion that attachment is a common antecedent for both the affective and social paths to problem drinking. The model was tested with data from three samples, first-year university students (N=679), students about to graduate from university (N=206), and first-time clients at an addiction treatment facility (N=21 1). Participants completed a battery of questionnaires assessing alcohol use, alcohol-related consequences, drinking motives, peer models of alcohol use, positive and negative affect, attachment anxiety and attachment avoidance. Results underscored the importance of the affective path to problem drinking, while putting the social path to problem drinking into question. While drinking to cope was most prominent among the clinical sample, coping motives served as a risk factor for problem drinking for both individuals identified as problem drinkers and university students. Moreover, drinking for enhancement purposes appeared to be the strongest overall predictor of alcohol use. Results of the present study also supported the notion that attachment anxiety and avoidance are antecedents for the affective path to problem drinking, such that those with higher levels of attachment anxiety and avoidance were more vulnerable to experiencing adverse consequences related to their drinking, explained in terms of diminished affect regulation. Evidence that nonsecure attachment is a potent predictor of problem drinking was also demonstrated by the finding that attachment anxiety was directly related to alcohol-related consequences over and above its indirect relationship through affect regulation. However, results failed to show that attachment anxiety or attachment avoidance increased the risk of problem drinking via social influence.