819 resultados para Project 2001-004-A : Knowledge Management and Innovation Diffusion


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Résumé La mondialisation des marchés, les mutations du contexte économique et enfin l'impact des nouvelles technologies de l'information ont obligé les entreprises à revoir la façon dont elles gèrent leurs capitaux intellectuel (gestion des connaissances) et humain (gestion des compétences). II est communément admis aujourd'hui que ceux-ci jouent un rôle particulièrement stratégique dans l'organisation. L'entreprise désireuse de se lancer dans une politique gestion de ces capitaux devra faire face à différents problèmes. En effet, afin de gérer ces connaissances et ces compétences, un long processus de capitalisation doit être réalisé. Celui-ci doit passer par différentes étapes comme l'identification, l'extraction et la représentation des connaissances et des compétences. Pour cela, il existe différentes méthodes de gestion des connaissances et des compétences comme MASK, CommonKADS, KOD... Malheureusement, ces différentes méthodes sont très lourdes à mettre en oeuvre, et se cantonnent à certains types de connaissances et sont, par conséquent, plus limitées dans les fonctionnalités qu'elles peuvent offrir. Enfin, la gestion des compétences et la gestion des connaissances sont deux domaines dissociés alors qu'il serait intéressant d'unifier ces deux approches en une seule. En effet, les compétences sont très proches des connaissances comme le souligne la définition de la compétence qui suit : « un ensemble de connaissances en action dans un contexte donné ». Par conséquent, nous avons choisi d'appuyer notre proposition sur le concept de compétence. En effet, la compétence est parmi les connaissances de l'entreprise l'une des plus cruciales, en particulier pour éviter la perte de savoir-faire ou pour pouvoir prévenir les besoins futurs de l'entreprise, car derrière les compétences des collaborateurs, se trouve l'efficacité de l'organisation. De plus, il est possible de décrire grâce à la compétence de nombreux autres concepts de l'organisation, comme les métiers, les missions, les projets, les formations... Malheureusement, il n'existe pas réellement de consensus sur la définition de la compétence. D'ailleurs, les différentes définitions existantes, même si elles sont pleinement satisfaisantes pour les experts, ne permettent pas de réaliser un système opérationnel. Dans notre approche; nous abordons la gestion des compétences à l'aide d'une méthode de gestion des connaissances. En effet, de par leur nature même, connaissance et compétence sont intimement liées et donc une telle méthode est parfaitement adaptée à la gestion des compétences. Afin de pouvoir exploiter ces connaissances et ces compétences nous avons dû, dans un premier temps, définir les concepts organisationnels de façon claire et computationnelle. Sur cette base, nous proposons une méthodologie de construction des différents référentiels d'entreprise (référentiel de compétences, des missions, des métiers...). Pour modéliser ces différents référentiels, nous avons choisi l'ontologie, car elle permet d'obtenir des définitions cohérentes et consensuelles aux concepts tout en supportant les diversités langagières. Ensuite, nous cartographions les connaissances de l'entreprise (formations, missions, métiers...) sur ces différentes ontologies afin de pouvoir les exploiter et les diffuser. Notre approche de la gestion des connaissances et de la gestion des compétences a permis la réalisation d'un outil offrant de nombreuses fonctionnalités comme la gestion des aires de mobilités, l'analyse stratégique, les annuaires ou encore la gestion des CV. Abstract The globalization of markets, the easing of economical regulation and finally the impact of new information and communication technologies have obliged firms to re-examine the way they manage their knowledge capital (knowledge management) and their human capital (competence management). It is commonly admitted that knowledge plays a slightly strategical role in the organization. The firms who want to establish one politic of management of these capitals will have to face with different problems. To manage that knowledge, a long process of capitalization must be done. That one has different steps like identification, extraction and representation of knowledge and competences. There are some different methods of knowledge management like MASK, CommonKADS or KOD. Unfortunately, those methods are very difficult to implement and are using only some types of knowledge and are consequently more limited in the functionalities they can offer. Knowledge management and competence management are two different domain where it could be interesting to unify those to one. Indeed, competence is very close than knowledge as underline this definition: "a set of knowledge in action in a specified context". We choose in our approach to rely on the concept of competence. Indeed, the competence is one of crucial knowledge in the company, particularly to avoid the loss of know-how or to prevent future needs. Because behind collaborator's competence, we can find company efficiency. Unfortunately, there is no real consensus on the definition of the concept of competence. Moreover, existing different definitions don't permit to develop an operational system. Among other key concept, we can find jobs, mission, project, and training... Moreover, we approach different problems of the competence management under the angle of the knowledge management. Indeed, knowledge and competence are closely linked. Then, we propose a method to build different company repositories (competence, jobs, projects repositories). To model those different repositories we choose ontology because it permits to obtain coherent and consensual definitions of the concepts with support of linguistics diversities too. This building repositories method coupled with this knowledge and competence management approach permitted the realization of a tool offering functionalities like mobility management, strategical analysis, yellow pages or CV management.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Background¦The outcome after primary percutaneous coronary intervention (pPCI) for STElevation¦Myocardial Infarction (STEMI) is strongly affected by time delays. In thepresent study, we sought to identify the impact of specific socioeconomic factors on time delays, subsequent STEMI management and outcomes in STEMI patients from a well-defined region of the French part of Switzerland.¦Method¦A total of 402 consecutive patients undergoing pPCI for STEMI in a large tertiary hospital were retrospectively studied. Symptom-to-first-medical-contact time was analyzed for the following socioeconomic factors: level of education, gender, origin and marital status. Main exclusion criteria were: time delay beyond 12 hours, previous treatment by fibrinolysis or patients immediately referred for CABG.¦Therefore, 352 patients were finally included.¦Results¦At one year, there was no difference in mortality amongst the different socioeconomic groups. Furthermore, there was no difference in management characteristics between them. Symptom-to-first-medical-contact time was significantly higher for patients with a low level of education, Swiss citizens and non-married patients with median differences of 40 minutes, 48 minutes, and 60 minutes, respectively (p<0.05).¦Nevertheless, no difference was found regarding in-hospital management and clinical outcome.¦Conclusion¦This study demonstrates that symptom-to-first-medical-contact time is higher amongst people with a lower educational level, Swiss-citizens, and non-married people. Because of the low mortality rate in general, these differences in time delays did not affect clinical outcomes. Still, primary prevention measures should particularly focus on these vulnerable populations.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The availability of rich firm-level data sets has recently led researchers to uncover new evidence on the effects of trade liberalization. First, trade openness forces the least productive firms to exit the market. Secondly, it induces surviving firms to increase their innovation efforts and thirdly, it increases the degree of product market competition. In this paper we propose a model aimed at providing a coherent interpretation of these findings. We introducing firm heterogeneity into an innovation-driven growth model, where incumbent firms operating in oligopolistic industries perform cost-reducing innovations. In this framework, trade liberalization leads to higher product market competition, lower markups and higher quantity produced. These changes in markups and quantities, in turn, promote innovation and productivity growth through a direct competition effect, based on the increase in the size of the market, and a selection effect, produced by the reallocation of resources towards more productive firms. Calibrated to match US aggregate and firm-level statistics, the model predicts that a 10 percent reduction in variable trade costs reduces markups by 1:15 percent, firm surviving probabilities by 1 percent, and induces an increase in productivity growth of about 13 percent. More than 90 percent of the trade-induced growth increase can be attributed to the selection effect.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Part A:The formulation of local policy and choices

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Part B:Common elements

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Part C:Steam sterilization

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper analyzes the effect of firms’ innovation activities on their growth performance. In particular, we observe how important innovation is for high-growth firms (HGFs) for an extensive sample of Spanish manufacturing and services firms. The panel data used comprises diverse waves of Spanish CIS over the the period 2004-2008. First, a probit analysis determines whether innovation affects the probability of being a high-growth firm. And second, a quantile regression technique is applied to explore the determinants and characteristics of specific groups of firms (manufacturing versus service firms and high-tech versus low-tech firms). It is revealed that R&D plays a significant role in the probability of becoming a HGF. Investment in internal and external R&D per employee has a positive impact on firm growth (although internal R&D presents a significant impact in the last quantiles, external R&D is significant up to the median). Furthermore, we show evidence that there is a positive impact of employment (sales) growth on the sales (employment) growth. Keywords: high-growth firms, firm growth, innovation activity JEL Classifications: L11, L25, O30

Relevância:

100.00% 100.00%

Publicador:

Resumo:

BACKGROUND: DNA sequence integrity, mRNA concentrations and protein-DNA interactions have been subject to genome-wide analyses based on microarrays with ever increasing efficiency and reliability over the past fifteen years. However, very recently novel technologies for Ultra High-Throughput DNA Sequencing (UHTS) have been harnessed to study these phenomena with unprecedented precision. As a consequence, the extensive bioinformatics environment available for array data management, analysis, interpretation and publication must be extended to include these novel sequencing data types. DESCRIPTION: MIMAS was originally conceived as a simple, convenient and local Microarray Information Management and Annotation System focused on GeneChips for expression profiling studies. MIMAS 3.0 enables users to manage data from high-density oligonucleotide SNP Chips, expression arrays (both 3'UTR and tiling) and promoter arrays, BeadArrays as well as UHTS data using MIAME-compliant standardized vocabulary. Importantly, researchers can export data in MAGE-TAB format and upload them to the EBI's ArrayExpress certified data repository using a one-step procedure. CONCLUSION: We have vastly extended the capability of the system such that it processes the data output of six types of GeneChips (Affymetrix), two different BeadArrays for mRNA and miRNA (Illumina) and the Genome Analyzer (a popular Ultra-High Throughput DNA Sequencer, Illumina), without compromising on its flexibility and user-friendliness. MIMAS, appropriately renamed into Multiomics Information Management and Annotation System, is currently used by scientists working in approximately 50 academic laboratories and genomics platforms in Switzerland and France. MIMAS 3.0 is freely available via http://multiomics.sourceforge.net/.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Part D:Washer-disinfectors

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Part E:Alternatives to steam for the sterilization of reusable medical devices

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Les coûts de traitement de certains patients s'avèrent extrêmement élevés, et peuvent faire soupçonner une prise en charge médicale inadéquate. Comme I'évolution du remboursement des prestations hospitalières passe à des forfaits par pathologie, il est essentiel de vérifier ce point, d'essayer de déterminer si ce type de patients peut être identifié à leur admission, et de s'assurer que leur devenir soit acceptable. Pour les années 1995 et 1997. les coûts de traitement dépassant de 6 déviations standard le coût moyen de la catégorie diagnostique APDRG ont été identifiés, et les dossiers des 50 patients dont les coûts variables étaient les plus élevés ont été analysés. Le nombre total de patients dont I'hospitalisation a entraîné des coûts extrêmes a passé de 391 en 1995 à 328 patients en 1997 (-16%). En ce qui concerne les 50 patients ayant entraîné les prises en charge les plus chères de manière absolue, les longs séjours dans de multiples services sont fréquents, mais 90% des patients sont sortis de l'hôpital en vie, et près de la moitié directement à domicile. Ils présentaient une variabilité importante de diagnostics et d'interventions, mais pas d'évidence de prise en charge inadéquate. En conclusion, les patients qualifiés de cas extrêmes sur un plan économique, ne le sont pas sur un plan strictement médical, et leur devenir est bon. Face à la pression qu'exercera le passage à un mode de financement par pathologie, les hôpitaux doivent mettre au point un système de revue interne de I'adéquation des prestations fournies basées sur des caractéristiques cliniques, s'ils veulent garantir des soins de qualité. et identifier les éventuelles prestations sous-optimales qu'ils pourraient être amenés à délivrer. [Auteurs] Treatment costs for some patients are extremely high and might let think that medical care could have been inadequate. As hospital financing systems move towards reimbursement by diagnostic groups, it is essential to assess whether inadequate care is provided, to try to identify these patients upon admission, and make sure that their outcome is good. For the years 1995 and 1997, treatment costs exceeding by 6 standard deviations the average cost of their APDRG category were identified, and the charts of the 50 patients with the highest variable costs were analyzed. The total number of patients with such extreme costs diminished from 391 in 1995 to 328 in 1997 (-16%). For the 50 most expensive patients, long stays in several services were frequent, but 90% of these patients left the hospital alive, and about half directly to their home. They presented an important variation in diagnoses and operations, but no evidence for inadequate care. Thus, patients qualified as extreme from an economic perspective cannot be qualified as such from a medical perspective, and their outcome is good. To face the pressure linked with the change in financing system, hospitals must develop an internal review system for assessing the adequacy of care, based on clinical characteristics, if they want to guarantee good quality of care and identify potentially inadequate practice.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

A population-based telephone survey conducted in 2002 estimated that there were 3.2 million episodes of acute gastroenteritis on the island of Ireland each year (Scallon et al., 2004). It is often very dif ficult to definitively identify the source of illness. However, of the respondents in that study suspecting food as the reason for their illness, 74% blamed food consumed from commercial premises such as restaurants, cafés, takeaways, canteens and pubs. Within the food services industry, statistics show a significant level of prosecutions, prohibition and closure orders of restaurants for food hygiene offences. The Food Safety Authority of Ireland has identified the main contributory factors to foodborne infections to be: cross-contamination, inadequate cooking, inadequate storage, inadequate reheating, delayed serving and infected food handlers (FSAI, 2000). Development of appropriate training and education campaigns to target problem areas requires initial understanding of the current level of food safety knowledge and practices in the food services industry.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Summary Forests are key ecosystems of the earth and associated with a large range of functions. Many of these functions are beneficial to humans and are referred to as ecosystem services. Sustainable development requires that all relevant ecosystem services are quantified, managed and monitored equally. Natural resource management therefore targets the services associated with ecosystems. The main hypothesis of this thesis is that the spatial and temporal domains of relevant services do not correspond to a discrete forest ecosystem. As a consequence, the services are not quantified, managed and monitored in an equal and sustainable manner. The thesis aims were therefore to test this hypothesis, establish an improved conceptual approach and provide spatial applications for the relevant land cover and structure variables. The study was carried out in western Switzerland and based primarily on data from a countrywide landscape inventory. This inventory is part of the third Swiss national forest inventory and assesses continuous landscape variables based on a regular sampling of true colour aerial imagery. In addition, land cover variables were derived from Landsat 5 TM passive sensor data and land structure variables from active sensor data from a small footprint laserscanning system. The results confirmed the main hypothesis, as relevant services did not scale well with the forest ecosystem. Instead, a new conceptual approach for sustainable management of natural resources was described. This concept quantifies the services as a continuous function of the landscape, rather than a discrete function of the forest ecosystem. The explanatory landscape variables are therefore called continuous fields and the forest becomes a dependent and function-driven management unit. Continuous field mapping methods were established for land cover and structure variables. In conclusion, the discrete forest ecosystem is an adequate planning and management unit. However, monitoring the state of and trends in sustainability of services requires them to be quantified as a continuous function of the landscape. Sustainable natural resource management iteratively combines the ecosystem and gradient approaches. Résumé Les forêts sont des écosystèmes-clés de la terre et on leur attribue un grand nombre de fonctions. Beaucoup de ces fonctions sont bénéfiques pour l'homme et sont nommées services écosystémiques. Le développement durable exige que ces services écosystémiques soient tous quantifiés, gérés et surveillés de façon égale. La gestion des ressources naturelles a donc pour cible les services attribués aux écosystèmes. L'hypothèse principale de cette thèse est que les domaines spatiaux et temporels des services attribués à la forêt ne correspondent pas à un écosystème discret. Par conséquent, les services ne sont pas quantifiés, aménagés et surveillés d'une manière équivalente et durable. Les buts de la thèse étaient de tester cette hypothèse, d'établir une nouvelle approche conceptuelle de la gestion des ressources naturelles et de préparer des applications spatiales pour les variables paysagères et structurelles appropriées. L'étude a été menée en Suisse occidentale principalement sur la base d'un inventaire de paysage à l'échelon national. Cet inventaire fait partie du troisième inventaire forestier national suisse et mesure de façon continue des variables paysagères sur la base d'un échantillonnage régulier sur des photos aériennes couleur. En outre, des variables de couverture ? terrestre ont été dérivées des données d'un senseur passif Landsat 5 TM, ainsi que des variables structurelles, dérivées du laserscanning, un senseur actif. Les résultats confirment l'hypothèse principale, car l'échelle des services ne correspond pas à celle de l'écosystème forestier. Au lieu de cela, une nouvelle approche a été élaborée pour la gestion durable des ressources naturelles. Ce concept représente les services comme une fonction continue du paysage, plutôt qu'une fonction discrète de l'écosystème forestier. En conséquence, les variables explicatives de paysage sont dénommées continuous fields et la forêt devient une entité dépendante, définie par la fonction principale du paysage. Des méthodes correspondantes pour la couverture terrestre et la structure ont été élaborées. En conclusion, l'écosystème forestier discret est une unité adéquate pour la planification et la gestion. En revanche, la surveillance de la durabilité de l'état et de son évolution exige que les services soient quantifiés comme fonction continue du paysage. La gestion durable des ressources naturelles joint donc l'approche écosystémique avec celle du gradient de manière itérative.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

OBJECTIVE: Antitumor necrosis factor a agents have significantly improved the management of Crohn's disease (CD), but not all patients benefit from this therapy. We used data from the Swiss Inflammatory Bowel Disease Cohort Study and predefined appropriateness criteria to examine the appropriateness of use of infliximab (IFX) in CD patients. METHODS: EPACT II (European Panel on the Appropriateness of CD Therapy, 2007; www.epact.ch) appropriateness criteria have been developed using a formal explicit panel process combining evidence from the published literature and expert opinion. Questionnaires relating to EPACT II criteria were used at enrollment and follow-up of all Swiss Inflammatory Bowel Disease Cohort Study patients. A step-by-step analysis of all possible indications for IFX therapy in a given patient allowed identification of the most appropriate indication and final classification in a single appropriateness category (appropriate, uncertain, inappropriate). RESULTS: Eight hundred and twenty-one CD patients were prospectively enrolled between November 2006 and March 2009. IFX was administered to 146 patients (18%) at enrollment and was most frequently used for complex fistulizing disease and for the maintenance of remission induced by biological therapy. IFX therapy was considered appropriate in 44%, uncertain in 44%, and inappropriate in 10% of patients. CONCLUSION: In this cohort, 9 out of 10 indications for IFX therapy were clinically generally acceptable (appropriate or uncertain) according to EPACT II criteria. Uncertain indications resulted mainly from the current more liberal use of IFX in clinical practice as compared with the EPACT II criteria.