927 resultados para Endogenous Information Structure
Resumo:
Objective: Previous studies reported on the association of left ventricular mass index (LVMI) with urinary sodium or with circulating or urinary aldosterone.We investigated the independent associations of LVMI with the urinary excretion of both sodium and aldosterone. Design and method: We randomly recruited 317 untreated subjects from a White population (45.1%women; mean age 48.2 years).Measurements included echocardiographic left ventricular (LV) properties, the 24 h urinary excretion of sodium and aldosterone, plasma renin activity (PRA), and proximal (RNaprox) and distal (RNadist) renal sodium reabsorption, assessed fromthe endogenous lithium clearance. Inmultivariable-adjusted models,we expressed changes in LVMI per 1 SD increase in the explanatory variables, while accounting for sex, age, systolic blood pressure and the waist-to-hip ratio. Results: LVMI increased independentlywith the urinary excretion of both sodium (+2.48 g/m2; P=0.005) and aldosterone (+2.63 g/m2; P=0.004). Higher sodium excretion was associated with increased mean wall thickness (MWT: +0.126 mm, P=0.054), but with no change in LV end-diastolic diameter (LVID: +0.12mm, P=0.64). In contrast, higher aldosterone excretion was associated with higher LVID (+0.54 mm; P=0.017), but with no change in MWT (+0.070mm; P=0.28).Higher RNadistwas associatedwith lower relativewall thickness (−0.81×10−2, P=0.017), because of opposite trends in LVID(+0.33 mm; P=0.13) and MWT (−0.130mm; P=0.040). LVMI was not associated with PRA or RNaprox. Conclusions: LVMI independently increased with both urinary sodium and aldosterone excretion. IncreasedMWT explained the association of LVMI with urinary sodium and increased LVID the association of LVMI with urinary aldosterone.
Resumo:
This work consists of three essays investigating the ability of structural macroeconomic models to price zero coupon U.S. government bonds. 1. A small scale 3 factor DSGE model implying constant term premium is able to provide reasonable a fit for the term structure only at the expense of the persistence parameters of the structural shocks. The test of the structural model against one that has constant but unrestricted prices of risk parameters shows that the exogenous prices of risk-model is only weakly preferred. We provide an MLE based variance-covariance matrix of the Metropolis Proposal Density that improves convergence speeds in MCMC chains. 2. Affine in observable macro-variables, prices of risk specification is excessively flexible and provides term-structure fit without significantly altering the structural parameters. The exogenous component of the SDF is separating the macro part of the model from the term structure and the good term structure fit has as a driving force an extremely volatile SDF and an implied average short rate that is inexplicable. We conclude that the no arbitrage restrictions do not suffice to temper the SDF, thus there is need for more restrictions. We introduce a penalty-function methodology that proves useful in showing that affine prices of risk specifications are able to reconcile stable macro-dynamics with good term structure fit and a plausible SDF. 3. The level factor is reproduced most importantly by the preference shock to which it is strongly and positively related but technology and monetary shocks, with negative loadings, are also contributing to its replication. The slope factor is only related to the monetary policy shocks and it is poorly explained. We find that there are gains in in- and out-of-sample forecast of consumption and inflation if term structure information is used in a time varying hybrid prices of risk setting. In-sample yield forecast are better in models with non-stationary shocks for the period 1982-1988. After this period, time varying market price of risk models provide better in-sample forecasts. For the period 2005-2008, out of sample forecast of consumption and inflation are better if term structure information is incorporated in the DSGE model but yields are better forecasted by a pure macro DSGE model.
Resumo:
EXECUTIVE SUMMARY : Evaluating Information Security Posture within an organization is becoming a very complex task. Currently, the evaluation and assessment of Information Security are commonly performed using frameworks, methodologies and standards which often consider the various aspects of security independently. Unfortunately this is ineffective because it does not take into consideration the necessity of having a global and systemic multidimensional approach to Information Security evaluation. At the same time the overall security level is globally considered to be only as strong as its weakest link. This thesis proposes a model aiming to holistically assess all dimensions of security in order to minimize the likelihood that a given threat will exploit the weakest link. A formalized structure taking into account all security elements is presented; this is based on a methodological evaluation framework in which Information Security is evaluated from a global perspective. This dissertation is divided into three parts. Part One: Information Security Evaluation issues consists of four chapters. Chapter 1 is an introduction to the purpose of this research purpose and the Model that will be proposed. In this chapter we raise some questions with respect to "traditional evaluation methods" as well as identifying the principal elements to be addressed in this direction. Then we introduce the baseline attributes of our model and set out the expected result of evaluations according to our model. Chapter 2 is focused on the definition of Information Security to be used as a reference point for our evaluation model. The inherent concepts of the contents of a holistic and baseline Information Security Program are defined. Based on this, the most common roots-of-trust in Information Security are identified. Chapter 3 focuses on an analysis of the difference and the relationship between the concepts of Information Risk and Security Management. Comparing these two concepts allows us to identify the most relevant elements to be included within our evaluation model, while clearing situating these two notions within a defined framework is of the utmost importance for the results that will be obtained from the evaluation process. Chapter 4 sets out our evaluation model and the way it addresses issues relating to the evaluation of Information Security. Within this Chapter the underlying concepts of assurance and trust are discussed. Based on these two concepts, the structure of the model is developed in order to provide an assurance related platform as well as three evaluation attributes: "assurance structure", "quality issues", and "requirements achievement". Issues relating to each of these evaluation attributes are analysed with reference to sources such as methodologies, standards and published research papers. Then the operation of the model is discussed. Assurance levels, quality levels and maturity levels are defined in order to perform the evaluation according to the model. Part Two: Implementation of the Information Security Assurance Assessment Model (ISAAM) according to the Information Security Domains consists of four chapters. This is the section where our evaluation model is put into a welldefined context with respect to the four pre-defined Information Security dimensions: the Organizational dimension, Functional dimension, Human dimension, and Legal dimension. Each Information Security dimension is discussed in a separate chapter. For each dimension, the following two-phase evaluation path is followed. The first phase concerns the identification of the elements which will constitute the basis of the evaluation: ? Identification of the key elements within the dimension; ? Identification of the Focus Areas for each dimension, consisting of the security issues identified for each dimension; ? Identification of the Specific Factors for each dimension, consisting of the security measures or control addressing the security issues identified for each dimension. The second phase concerns the evaluation of each Information Security dimension by: ? The implementation of the evaluation model, based on the elements identified for each dimension within the first phase, by identifying the security tasks, processes, procedures, and actions that should have been performed by the organization to reach the desired level of protection; ? The maturity model for each dimension as a basis for reliance on security. For each dimension we propose a generic maturity model that could be used by every organization in order to define its own security requirements. Part three of this dissertation contains the Final Remarks, Supporting Resources and Annexes. With reference to the objectives of our thesis, the Final Remarks briefly analyse whether these objectives were achieved and suggest directions for future related research. Supporting resources comprise the bibliographic resources that were used to elaborate and justify our approach. Annexes include all the relevant topics identified within the literature to illustrate certain aspects of our approach. Our Information Security evaluation model is based on and integrates different Information Security best practices, standards, methodologies and research expertise which can be combined in order to define an reliable categorization of Information Security. After the definition of terms and requirements, an evaluation process should be performed in order to obtain evidence that the Information Security within the organization in question is adequately managed. We have specifically integrated into our model the most useful elements of these sources of information in order to provide a generic model able to be implemented in all kinds of organizations. The value added by our evaluation model is that it is easy to implement and operate and answers concrete needs in terms of reliance upon an efficient and dynamic evaluation tool through a coherent evaluation system. On that basis, our model could be implemented internally within organizations, allowing them to govern better their Information Security. RÉSUMÉ : Contexte général de la thèse L'évaluation de la sécurité en général, et plus particulièrement, celle de la sécurité de l'information, est devenue pour les organisations non seulement une mission cruciale à réaliser, mais aussi de plus en plus complexe. A l'heure actuelle, cette évaluation se base principalement sur des méthodologies, des bonnes pratiques, des normes ou des standards qui appréhendent séparément les différents aspects qui composent la sécurité de l'information. Nous pensons que cette manière d'évaluer la sécurité est inefficiente, car elle ne tient pas compte de l'interaction des différentes dimensions et composantes de la sécurité entre elles, bien qu'il soit admis depuis longtemps que le niveau de sécurité globale d'une organisation est toujours celui du maillon le plus faible de la chaîne sécuritaire. Nous avons identifié le besoin d'une approche globale, intégrée, systémique et multidimensionnelle de l'évaluation de la sécurité de l'information. En effet, et c'est le point de départ de notre thèse, nous démontrons que seule une prise en compte globale de la sécurité permettra de répondre aux exigences de sécurité optimale ainsi qu'aux besoins de protection spécifiques d'une organisation. Ainsi, notre thèse propose un nouveau paradigme d'évaluation de la sécurité afin de satisfaire aux besoins d'efficacité et d'efficience d'une organisation donnée. Nous proposons alors un modèle qui vise à évaluer d'une manière holistique toutes les dimensions de la sécurité, afin de minimiser la probabilité qu'une menace potentielle puisse exploiter des vulnérabilités et engendrer des dommages directs ou indirects. Ce modèle se base sur une structure formalisée qui prend en compte tous les éléments d'un système ou programme de sécurité. Ainsi, nous proposons un cadre méthodologique d'évaluation qui considère la sécurité de l'information à partir d'une perspective globale. Structure de la thèse et thèmes abordés Notre document est structuré en trois parties. La première intitulée : « La problématique de l'évaluation de la sécurité de l'information » est composée de quatre chapitres. Le chapitre 1 introduit l'objet de la recherche ainsi que les concepts de base du modèle d'évaluation proposé. La maniéré traditionnelle de l'évaluation de la sécurité fait l'objet d'une analyse critique pour identifier les éléments principaux et invariants à prendre en compte dans notre approche holistique. Les éléments de base de notre modèle d'évaluation ainsi que son fonctionnement attendu sont ensuite présentés pour pouvoir tracer les résultats attendus de ce modèle. Le chapitre 2 se focalise sur la définition de la notion de Sécurité de l'Information. Il ne s'agit pas d'une redéfinition de la notion de la sécurité, mais d'une mise en perspectives des dimensions, critères, indicateurs à utiliser comme base de référence, afin de déterminer l'objet de l'évaluation qui sera utilisé tout au long de notre travail. Les concepts inhérents de ce qui constitue le caractère holistique de la sécurité ainsi que les éléments constitutifs d'un niveau de référence de sécurité sont définis en conséquence. Ceci permet d'identifier ceux que nous avons dénommés « les racines de confiance ». Le chapitre 3 présente et analyse la différence et les relations qui existent entre les processus de la Gestion des Risques et de la Gestion de la Sécurité, afin d'identifier les éléments constitutifs du cadre de protection à inclure dans notre modèle d'évaluation. Le chapitre 4 est consacré à la présentation de notre modèle d'évaluation Information Security Assurance Assessment Model (ISAAM) et la manière dont il répond aux exigences de l'évaluation telle que nous les avons préalablement présentées. Dans ce chapitre les concepts sous-jacents relatifs aux notions d'assurance et de confiance sont analysés. En se basant sur ces deux concepts, la structure du modèle d'évaluation est développée pour obtenir une plateforme qui offre un certain niveau de garantie en s'appuyant sur trois attributs d'évaluation, à savoir : « la structure de confiance », « la qualité du processus », et « la réalisation des exigences et des objectifs ». Les problématiques liées à chacun de ces attributs d'évaluation sont analysées en se basant sur l'état de l'art de la recherche et de la littérature, sur les différentes méthodes existantes ainsi que sur les normes et les standards les plus courants dans le domaine de la sécurité. Sur cette base, trois différents niveaux d'évaluation sont construits, à savoir : le niveau d'assurance, le niveau de qualité et le niveau de maturité qui constituent la base de l'évaluation de l'état global de la sécurité d'une organisation. La deuxième partie: « L'application du Modèle d'évaluation de l'assurance de la sécurité de l'information par domaine de sécurité » est elle aussi composée de quatre chapitres. Le modèle d'évaluation déjà construit et analysé est, dans cette partie, mis dans un contexte spécifique selon les quatre dimensions prédéfinies de sécurité qui sont: la dimension Organisationnelle, la dimension Fonctionnelle, la dimension Humaine, et la dimension Légale. Chacune de ces dimensions et son évaluation spécifique fait l'objet d'un chapitre distinct. Pour chacune des dimensions, une évaluation en deux phases est construite comme suit. La première phase concerne l'identification des éléments qui constituent la base de l'évaluation: ? Identification des éléments clés de l'évaluation ; ? Identification des « Focus Area » pour chaque dimension qui représentent les problématiques se trouvant dans la dimension ; ? Identification des « Specific Factors » pour chaque Focus Area qui représentent les mesures de sécurité et de contrôle qui contribuent à résoudre ou à diminuer les impacts des risques. La deuxième phase concerne l'évaluation de chaque dimension précédemment présentées. Elle est constituée d'une part, de l'implémentation du modèle général d'évaluation à la dimension concernée en : ? Se basant sur les éléments spécifiés lors de la première phase ; ? Identifiant les taches sécuritaires spécifiques, les processus, les procédures qui auraient dû être effectués pour atteindre le niveau de protection souhaité. D'autre part, l'évaluation de chaque dimension est complétée par la proposition d'un modèle de maturité spécifique à chaque dimension, qui est à considérer comme une base de référence pour le niveau global de sécurité. Pour chaque dimension nous proposons un modèle de maturité générique qui peut être utilisé par chaque organisation, afin de spécifier ses propres exigences en matière de sécurité. Cela constitue une innovation dans le domaine de l'évaluation, que nous justifions pour chaque dimension et dont nous mettons systématiquement en avant la plus value apportée. La troisième partie de notre document est relative à la validation globale de notre proposition et contient en guise de conclusion, une mise en perspective critique de notre travail et des remarques finales. Cette dernière partie est complétée par une bibliographie et des annexes. Notre modèle d'évaluation de la sécurité intègre et se base sur de nombreuses sources d'expertise, telles que les bonnes pratiques, les normes, les standards, les méthodes et l'expertise de la recherche scientifique du domaine. Notre proposition constructive répond à un véritable problème non encore résolu, auquel doivent faire face toutes les organisations, indépendamment de la taille et du profil. Cela permettrait à ces dernières de spécifier leurs exigences particulières en matière du niveau de sécurité à satisfaire, d'instancier un processus d'évaluation spécifique à leurs besoins afin qu'elles puissent s'assurer que leur sécurité de l'information soit gérée d'une manière appropriée, offrant ainsi un certain niveau de confiance dans le degré de protection fourni. Nous avons intégré dans notre modèle le meilleur du savoir faire, de l'expérience et de l'expertise disponible actuellement au niveau international, dans le but de fournir un modèle d'évaluation simple, générique et applicable à un grand nombre d'organisations publiques ou privées. La valeur ajoutée de notre modèle d'évaluation réside précisément dans le fait qu'il est suffisamment générique et facile à implémenter tout en apportant des réponses sur les besoins concrets des organisations. Ainsi notre proposition constitue un outil d'évaluation fiable, efficient et dynamique découlant d'une approche d'évaluation cohérente. De ce fait, notre système d'évaluation peut être implémenté à l'interne par l'entreprise elle-même, sans recourir à des ressources supplémentaires et lui donne également ainsi la possibilité de mieux gouverner sa sécurité de l'information.
Resumo:
The objective of this work was to evaluate the population structure and the genetic and phenotypic progress of Nelore cattle in Northern Brazil. Pedigree information concerning animals born between 1942 and 2006 were analyzed. Population structure was performed using the Endog program. Out of the 140,628 animals studied, 67.7, 14.52 and 3.18% had complete pedigree record of the first, second and third parental generation, respectively. Inbreeding and average relatedness coefficients were low: 0.2 and 0.13%, respectively. However, these parameters may have been underestimated, since information on pedigree was incomplete. The effective number of founders was 370 and the genetic contribution of 10, 50 and 448 most influent ancestors explained 13.2, 28 and 50% of the genetic variability in the population, respectively. The genetic variability for growth traits and population structure demonstrates high probability of increasing productivity through selective breeding. Moreover, management strategies to reduce the currently observed age at first calving and generation intervals are important for Nelore cattle genetic improvement.
Resumo:
We propose a procedure for analyzing and characterizing complex networks. We apply this to the social network as constructed from email communications within a medium sized university with about 1700 employees. Email networks provide an accurate and nonintrusive description of the flow of information within human organizations. Our results reveal the self-organization of the network into a state where the distribution of community sizes is self-similar. This suggests that a universal mechanism, responsible for emergence of scaling in other self-organized complex systems, as, for instance, river networks, could also be the underlying driving force in the formation and evolution of social networks.
Resumo:
This article introduces a new interface for T-Coffee, a consistency-based multiple sequence alignment program. This interface provides an easy and intuitive access to the most popular functionality of the package. These include the default T-Coffee mode for protein and nucleic acid sequences, the M-Coffee mode that allows combining the output of any other aligners, and template-based modes of T-Coffee that deliver high accuracy alignments while using structural or homology derived templates. These three available template modes are Expresso for the alignment of protein with a known 3D-Structure, R-Coffee to align RNA sequences with conserved secondary structures and PSI-Coffee to accurately align distantly related sequences using homology extension. The new server benefits from recent improvements of the T-Coffee algorithm and can align up to 150 sequences as long as 10,000 residues and is available from both http://www.tcoffee.org and its main mirror http://tcoffee.crg.cat.
Resumo:
The Internet is becoming more and more popular among drug users. The use of websites and forums to obtain illicit drugs and relevant information about the means of consumption is a growing phenomenon mainly for new synthetic drugs. Gamma Butyrolactone (GBL), a chemical precursor of Gamma Hydroxy Butyric acid (GHB), is used as a "club drug" and also in drug facilitated sexual assaults. Its market takes place mainly on the Internet through online websites but the structure of the market remains unknown. This research aims to combine digital, physical and chemical information to help understand the distribution routes and the structure of the GBL market. Based on an Internet monitoring process, thirty-nine websites selling GBL, mainly in the Netherlands, were detected between January 2010 and December 2011. Seventeen websites were categorized into six groups based on digital traces (e.g. IP addresses and contact information). In parallel, twenty-five bulk GBL specimens were purchased from sixteen websites for packaging comparisons and carbon isotopic measurements. Packaging information showed a high correlation with digital data confirming the links previously established whereas chemical information revealed undetected links and provided complementary information. Indeed, while digital and packaging data give relevant information about the retailers, the supply routes and the distribution close to the consumer, the carbon isotopic data provides upstream information about the production level and in particular the synthesis pathways and the chemical precursors. A three-level structured market has been thereby identified with a production level mainly located in China and in Germany, an online distribution level mainly hosted in the Netherlands and the customers who order on the Internet.
Resumo:
This paper studies fiscal federalism when regions differ in voters' ability to monitor publicofficials. We develop a model of political agency in which rent-seeking politicians providepublic goods to win support from heterogeneously informed voters. In equilibrium, voterinformation increases government accountability but displays decreasing returns. Therefore,political centralization reduces aggregate rent extraction when voter information varies acrossregions. It increases welfare as long as the central government is required to provide publicgoods uniformly across regions. The need for uniformity implies an endogenous trade off between reducing rents through centralization and matching idiosyncratic preferences throughdecentralization. We find that a federal structure with overlapping levels of government canbe optimal only if regional differences in accountability are sufficiently large. The modelpredicts that less informed regions should reap greater benefits when the central governmentsets a uniform policy. Consistent with our theory, we present empirical evidence that lessinformed states enjoyed faster declines in pollution after the 1970 Clean Air Act centralizedenvironmental policy at the federal level.
Resumo:
Printed electronics is an emerging concept in electronics manufacturing and it is in very early development stage. The technology is not stable, design kits are not developed, and flows and Computer Aided Design (CAD) tools are not fixed yet. The European project TDK4PE addresses all this issues and this PFC has been realized on this context. The goal is to develop an XML-based information system for the collection and management of information from the technology and cell libraries developed in TDK4PE. This system will ease the treatment of that information for a later generation of specific Design Kits (DK) and the corresponding documentation. This work proposes a web application to generate technology files and design kits in a formatted way; it also proposes a structure for them and a database implementation for storing the needed information. The application will allow its users to redefine the structure of those files, as well as export and import XML files, between other formats.
Resumo:
Diplomityön tavoitteena on tutkia mitä uusia tiedonhallinnallisia ongelmia ilmenee, kun massaräätälöidyn tuotteen tuotetieto hallitaan läpi tuotteen elinkaaren, sekä miten nämä ongelmat voitaisiin ratkaista. Ongelmat ja haasteet kerätään kirjallisuuslähteistä ja massaräätälöintiprosessi yhdistetään PLM-vaiheisiin. Ratkaisua tutkitaan testaamalla kuinka standardit STEP ja PLCS sekä standardeja tukeva PLM järjestelmä voisivat tukea massaräätälöidyn tuotteen elinkaaren tiedonhallintaa. MC tuotteiden ongelmia ovat tuoterakenteen monimutkaisuus, jäljitettävyys ja muutosten hallinta läpi elinkaaren. STEP ja PLCS pystyvät kummatkin tahollaan tukemaan tiedonhallintaa. MC-tuotteen geneerinen tuoterakenne on kuitenkin manuaalisesti liittettävä elinkaaritiedon tukemiseen. PLM-järjestelmä pystyy tukemaan MC-tuotteiden elinkaarta, mutta koska toiminto ei ole järjestelmään sisäänrakennettuna, MC-tuotteiden tukemisen parantamisessa on edelleen haasteita.
Resumo:
Thisthesis supplements the systematic approach to competitive intelligence and competitor analysis by introducing an information-processing perspective on management of the competitive environment and competitors therein. The cognitive questions connected to the intelligence process and also the means that organizational actors use in sharing information are discussed. The ultimate aim has been to deepen knowledge of the different intraorganizational processes that are used in acorporate organization to manage and exploit the vast amount of competitor information that is received from the environment. Competitor information and competitive knowledge management is examined as a process, where organizational actorsidentify and perceive the competitive environment by using cognitive simplification, make interpretations resulting in learning and finally utilize competitor information and competitive knowledge in their work processes. The sharing of competitive information and competitive knowledge is facilitated by intraorganizational networks that evolve as a means of developing a shared, organizational level knowledge structure and ensuring that the right information is in the right place at the right time. This thesis approaches competitor information and competitive knowledge management both theoretically and empirically. Based on the conceptual framework developed by theoretical elaboration, further understanding of the studied phenomena is sought by an empirical study. The empirical research was carried out in a multinationally operating forest industry company. This thesis makes some preliminary suggestions of improving the competitive intelligence process. It is concluded that managing competitor information and competitive knowledge is not simply a question of managing information flow or improving sophistication of competitor analysis, but the crucial question to be solved is rather, how to improve the cognitive capabilities connected to identifying and making interpretations of the competitive environment and how to increase learning. It is claimed that competitive intelligence can not be treated like an organizational function or assigned solely to a specialized intelligence unit.
Resumo:
This article introduces a new interface for T-Coffee, a consistency-based multiple sequence alignment program. This interface provides an easy and intuitive access to the most popular functionality of the package. These include the default T-Coffee mode for protein and nucleic acid sequences, the M-Coffee mode that allows combining the output of any other aligners, and template-based modes of T-Coffee that deliver high accuracy alignments while using structural or homology derived templates. These three available template modes are Expresso for the alignment of protein with a known 3D-Structure, R-Coffee to align RNA sequences with conserved secondary structures and PSI-Coffee to accurately align distantly related sequences using homology extension. The new server benefits from recent improvements of the T-Coffee algorithm and can align up to 150 sequences as long as 10 000 residues and is available from both http://www.tcoffee.org and its main mirror http://tcoffee.crg.cat.
Resumo:
In the present research we have set forth a new, simple, Trade-Off model that would allow us to calculate how much debt and, by default, how much equity a company should have, using easily available information and calculating the cost of debt dynamically on the basis of the effect that the capital structure of the company has on the risk of bankruptcy; in an attempt to answer this question. The proposed model has been applied to the companies that make up the Dow Jones Industrial Average (DJIA) in 2007. We have used consolidated financial data from 1996 to 2006, published by Bloomberg. We have used simplex optimization method to find the debt level that maximizes firm value. Then, we compare the estimated debt with real debt of companies using statistical nonparametric Mann-Whitney. The results indicate that 63% of companies do not show a statistically significant difference between the real and the estimated debt.
Resumo:
Introduction : Décrire les patients d'une structure gériatrique offrant des hospitalisations de courte durée, dans un contexte ambulatoire, pour des situations gériatriques courantes dans le canton de Genève (Suisse). Mesurer les performances de cette structure en termes de qualité des soins et de coûts. Méthodes : Des données relatives au profil des 100 premiers patients ont été collectées (huit mois), ainsi qu'aux prestations, aux ressources et aux effets (réadmissions, décès, satisfaction, complications) de manière à mesurer différents indicateurs de qualité et de coûts. Les valeurs observées ont été systématiquement comparées aux valeurs attendues, calculées à partir du profil des patients. Résultats : Des critères d'admission ont été fixés pour exclure les situations dans lesquelles d'autres structures offrent des soins mieux adaptés. La spécificité de cette structure intermédiaire a été d'assurer une continuité des soins et d'organiser d'emblée le retour à domicile par des prestations de liaison ambulatoire. La faible occurrence des réadmissions potentiellement évitables, une bonne satisfaction des patients, l'absence de décès prématurés et le faible nombre de complications suggèrent que les soins médicaux et infirmiers ont été délivrés avec une bonne qualité. Le coût s'est révélé nettement plus économique que des séjours hospitaliers après ajustement pour la lourdeur des cas. Conclusion : L'expérience-pilote a démontré la faisabilité et l'utilité d'une unité d'hébergement et d'hospitalisation de court séjour en toute sécurité. Le suivi du patient par le médecin traitant assure une continuité des soins et évite la perte d'information lors des transitions ainsi que les examens non pertinents. INTRODUCTION: To describe patients admitted to a geriatric institution, providing short-term hospitalizations in the context of ambulatory care in the canton of Geneva. To measure the performances of this structure in terms of quality ofcare and costs. METHOD: Data related to the clinical,functioning and participation profiles of the first 100 patients were collected. Data related to effects (readmission, deaths, satisfaction, complications), services and resources were also documented over an 8-month period to measure various quality and costindicators. Observed values were systematically compared to expected values, adjusted for case mix. RESULTS: Explicit criteria were proposed to focus on the suitable patients, excluding situations in which other structures were considered to be more appropriate. The specificity of this intermediate structure was to immediately organize, upon discharge, outpatient services at home. The low rate of potentially avoidable readmissions, the high patient satisfaction scores, the absence of premature death and the low number of iatrogenic complications suggest that medical and nursing care delivered reflect a good quality of services. The cost was significantly lower than expected, after adjusting for case mix. CONCLUSION: The pilot experience showed that a short-stay hospitalization unit was feasible with acceptable security conditions. The attending physician's knowledge of the patients allowed this system tofocus on essential issues without proposing inappropriate services.
Resumo:
Cette thèse comprend trois essais qui abordent l'information le processus d'ap-prentissage ainsi que le risque dans les marchés finances. Elle se concentre d'abord sur les implications à l'équilibre de l'hétérogénéité des agents à travers un processus d'apprentissage comprtemental et de mise à jour de l'information. De plus, elle examine les effets du partage des risques dans un reseau entreprise-fournisseur. Le premier chapitre étudie les effets du biais de disponibili sur l'évaluation des actifs. Ce biais décrit le fait que les agents surestiment l'importance de l'information acquise via l'expérience personnelle. L'hétérogénéité restante des différentes perceptions individuelles amène à une volonté d'échanges. Conformé¬ment aux données empiriques, les jeunes agents échangent plus mais en même temps souffrent d'une performance inférieure. Le deuxième chapitre se penche sur l'impact qu'ont les différences de modelisation entre les agents sur leurs percevons individuelles du processus de prix, dans le contexte des projections de modèles. Les agents sujets à un biais de projection pensent être représentatifs et interprètent les opinions des autres agents comme du bruit. Les agents, avec des modèles plus persistants, perçoivent que les prix réagissent de façon excessive lors des périodes de turbulence. Le troisième chapitre analyse l'impact du partage des risques dans la relation entreprise-fournisseur sur la décision optimale de financement de l'entreprise. Il étudie l'impact sur l'optimisation de la structure du capital ainsi que sur le coût du capital. Les résultats indiquent en particulier qu'un fournisseur avec un effet de levier faible est utile pour le financement d'un nouveau projet d'investissement. Pour des projets très rentables et des fournisseurs à faible effet de levier, le coût des capitaux propres de l'entreprise peut diminuer.