11 resultados para Unité de réanimation
em DigitalCommons@The Texas Medical Center
Resumo:
Utilizing advanced information technology, Intensive Care Unit (ICU) remote monitoring allows highly trained specialists to oversee a large number of patients at multiple sites on a continuous basis. In the current research, we conducted a time-motion study of registered nurses’ work in an ICU remote monitoring facility. Data were collected on seven nurses through 40 hours of observation. The results showed that nurses’ essential tasks were centered on three themes: monitoring patients, maintaining patients’ health records, and managing technology use. In monitoring patients, nurses spent 52% of the time assimilating information embedded in a clinical information system and 15% on monitoring live vitals. System-generated alerts frequently interrupted nurses in their task performance and redirected them to manage suddenly appearing events. These findings provide insight into nurses’ workflow in a new, technology-driven critical care setting and have important implications for system design, work engineering, and personnel selection and training.
Resumo:
Coronary artery bypass graft (CABG) surgery is among the most common operations performed in the United States and accounts for more resources expended in cardiovascular medicine than any other single procedure. CABG surgery patients initially recover in the Cardiovascular Intensive Care Unit (CVICU). The post-procedure CVICU length of stay (LOS) goal is two days or less. A longer ICU LOS is associated with a prolonged hospital LOS, poor health outcomes, greater use of limited resources, and increased medical costs. ^ Research has shown that experienced clinicians can predict LOS no better than chance. Current CABG surgery LOS risk models differ greatly in generalizability and ease of use in the clinical setting. A predictive model that identified modifiable pre- and intra-operative risk factors for CVICU LOS greater than two days could have major public health implications as modification of these identified factors could decrease CVICU LOS and potentially minimize morbidity and mortality, optimize use of limited health care resources, and decrease medical costs. ^ The primary aim of this study was to identify modifiable pre-and intra-operative predictors of CVICU LOS greater than two days for CABG surgery patients with cardiopulmonary bypass (CPB). A secondary aim was to build a probability equation for CVICU LOS greater than two days. Data were extracted from 416 medical records of CABG surgery patients with CPB, 50 to 80 years of age, recovered in the CVICU of a large teaching, referral hospital in southeastern Texas, during the calendar year 2004 and the first quarter of 2005. Exclusion criteria included Diagnosis Related Group (DRG) 106, CABG surgery without CPB, CABG surgery with other procedures, and operative deaths. The data were analyzed using multivariate logistic regression for an alpha=0.05, power=0.80, and correlation=0.26. ^ This study found age, history of peripheral arterial disease, and total operative time equal to and greater than four hours to be independent predictors of CVICU LOS greater than two days. The probability of CVICU LOS greater than two days can be calculated by the following equation: -2.872941 +.0323081 (age in years) + .8177223 (history of peripheral arterial disease) + .70379 (operative time). ^
Resumo:
Introduction. Selectively manned units have a long, international history, both military and civilian. Some examples include SWAT teams, firefighters, the FBI, the DEA, the CIA, and military Special Operations. These special duty operators are individuals who perform a highly skilled and dangerous job in a unique environment. A significant amount of money is spent by the Department of Defense (DoD) and other federal agencies to recruit, select, train, equip and support these operators. When a critical incident or significant life event occurs, that jeopardizes an operator's performance; there can be heavy losses in terms of training, time, money, and potentially, lives. In order to limit the number of critical incidents, selection processes have been developed over time to “select out” those individuals most likely to perform below desired performance standards under pressure or stress and to "select in" those with the "right stuff". This study is part of a larger program evaluation to assess markers that identify whether a person will fail under the stresses in a selectively manned unit. The primary question of the study is whether there are indicators in the selection process that signify potential negative performance at a later date. ^ Methods. The population being studied included applicants to a selectively manned DoD organization between 1993 and 2001 as part of a unit assessment and selection process (A&S). Approximately 1900 A&S records were included in the analysis. Over this nine year period, seventy-two individuals were determined to have had a critical incident. A critical incident can come in the form of problems with the law, personal, behavioral or family problems, integrity issues, and skills deficit. Of the seventy-two individuals, fifty-four of these had full assessment data and subsequent supervisor performance ratings which assessed how an individual performed while on the job. This group was compared across a variety of variables including demographics and psychometric testing with a group of 178 individuals who did not have a critical incident and had been determined to be good performers with positive ratings by their supervisors.^ Results. In approximately 2004, an online pre-screen survey was developed in the hopes of preselecting out those individuals with items that would potentially make them ineligible for selection to this organization. This survey has aided the organization to increase its selection rates and save resources in the process. (Patterson, Howard Smith, & Fisher, Unit Assessment and Selection Project, 2008) When the same prescreen was used on the critical incident individuals, it was found that over 60% of the individuals would have been flagged as unacceptable. This would have saved the organization valuable resources and heartache.^ There were some subtle demographic differences between the two groups (i.e. those with critical incidents were almost twice as likely to be divorced compared with the positive performers). Upon comparison of Psychometric testing several items were noted to be different. The two groups were similar when their IQ levels were compared using the Multidimensional Aptitude Battery (MAB). When looking at the Minnesota Multiphasic Personality Inventory (MMPI), there appeared to be a difference on the MMPI Social Introversion; the Critical Incidence group scored somewhat higher. When analysis was done, the number of MMPI Critical Items between the two groups was similar as well. When scores on the NEO Personality Inventory (NEO) were compared, the critical incident individuals tended to score higher on Openness and on its subscales (Ideas, Actions, and Feelings). There was a positive correlation between Total Neuroticism T Score and number of MMPI critical items.^ Conclusions. This study shows that the current pre-screening process is working and would have saved the organization significant resources. ^ If one was to develop a profile of a candidate who potentially could suffer a critical incident and subsequently jeopardize the unit, mission and the safety of the public they would look like the following: either divorced or never married, score high on the MMPI in Social Introversion, score low on MMPI with an "excessive" amount of MMPI critical items; and finally scores high on the NEO Openness and subscales Ideas, Feelings, and Actions.^ Based on the results gleaned from the analysis in this study there seems to be several factors, within psychometric testing, that when taken together, will aid the evaluators in selecting only the highest quality operators in order to save resources and to help protect the public from unfortunate critical incidents which may adversely affect our health and safety.^
Resumo:
Background. Nosocomial infections are a source of concern for many hospitals in the United States and worldwide. These infections are associated with increased morbidity, mortality and hospital costs. Nosocomial infections occur in ICUs at a rate which is five times greater than those in general wards. Understanding the reasons for the higher rates can ultimately help reduce these infections. The literature has been weak in documenting a direct relationship between nosocomial infections and non-traditional risk factors, such as unit staffing or patient acuity.^ Objective. To examine the relationship, if any, between nosocomial infections and non-traditional risk factors. The potential non-traditional risk factors we studied were the patient acuity (which comprised of the mortality and illness rating of the patient), patient days for patients hospitalized in the ICU, and the patient to nurse ratio.^ Method. We conducted a secondary data analysis on patients hospitalized in the Medical Intensive Care Unit (MICU) of the Memorial Hermann- Texas Medical Center in Houston during the months of March 2008- May 2009. The average monthly values for the patient acuity (mortality and illness Diagnostic Related Group (DRG) scores), patient days for patients hospitalized in the ICU and average patient to nurse ratio were calculated during this time period. Active surveillance of Bloodstream Infections (BSIs), Urinary Tract Infections (UTIs) and Ventilator Associated Pneumonias (VAPs) was performed by Infection Control practitioners, who visited the MICU and performed a personal infection record for each patient. Spearman's rank correlation was performed to determine the relationship between these nosocomial infections and the non-traditional risk factors.^ Results. We found weak negative correlations between BSIs and two measures (illness and mortality DRG). We also found a weak negative correlation between UTI and unit staffing (patient to nurse ratio). The strongest positive correlation was found between illness DRG and mortality DRG, validating our methodology.^ Conclusion. From this analysis, we were able to infer that non-traditional risk factors do not appear to play a significant role in transmission of infection in the units we evaluated.^
Resumo:
Risk factors for Multi-Drug Resistant Acinetobacter (MDRA) acquisition were studied in patients in a burn intensive care unit (ICU) where there was an outbreak of MDRA. Forty cases were matched with eighty controls based on length of stay in the Burn ICU and statistical analysis was performed on data for several different variables. Matched analysis showed that mechanical ventilation, transport ventilation, number of intubations, number of bronchoscopy procedures, total body surface area burn, and prior Methicillin Resistant Staphylococcus aureus colonization were all significant risk factors for MDRA acquisition. ^ MDRA remains a significant threat to the burn population. Treatment for burn patients with MDRA is challenging as resistance to antibiotics continues to increase. This study underlined the need to closely monitor the most critically ill ventilated patients during an outbreak of MDRA as they are the most at risk for MDRA acquisition.^
Resumo:
Much attention has been given to treating Operation Iraqi Freedom/Operation Enduring (OIF/OEF) Veterans with posttraumatic stress disorder (PTSD). However, little attention is given to those Veterans who do not meet diagnostic criteria for PTSD but who may still benefit from intervention. Research is needed to investigate the impact of how different racial/ethnic backgrounds, different levels of social support and comorbid mental health disorders impact OIF/OEF Veterans with varying levels of PTSD. The purpose of this dissertation is to examine the association of comorbid Axis I disorders, race/ethnicity, different levels of postdeployment social support and unit support on OIF/OEF Veterans with varying levels of PTSD. Data for this dissertation were from postdeployment screenings of OIF/OEF Veterans from a large Veterans Affairs hospital in southeast Texas. To examine the study hypotheses, we conducted multinomial logistic regressions of the clinician reported data. ^ The first article examined the prevalence of subthreshold and full levels of PTSD and compared Axis I and alcohol use comorbidity rates among 1,362 OIF/OEF Veterans with varying levels of PTSD. Results suggest that OIF/OEF Veterans with subthreshold PTSD experience similar levels of psychological distress as those with full PTSD and highlight the need to provide timely and appropriate mental health services to individuals who may not meet the diagnostic criteria for full PTSD. ^ These results suggest that OIF/OEF Veterans of all race/ethnicities can benefit from strong social support systems. Postdeployment social support was found to be a protective factor against the development of PTSD among White, Black and Hispanic veterans while deployment unit support was a protective factor only among Black Veterans. The second article investigated the association between postdeployment social support and unit support with varying levels of PTSD by race/ethnicity among 1,115 OIF/OEF Veterans. ^ The results of this study can help to formulate treatment and interventions for OIF/OEF Veterans with varying levels of PTSD and social support systems.^
Resumo:
Sepsis is a significant cause for multiple organ failure and death in the burn patient, yet identification in this population is confounded by chronic hypermetabolism and impaired immune function. The purpose of this study was twofold: 1) determine the ability of the systemic inflammatory response syndrome (SIRS) and American Burn Association (ABA) criteria to predict sepsis in the burn patient; and 2) develop a model representing the best combination of clinical predictors associated with sepsis in the same population. A retrospective, case-controlled, within-patient comparison of burn patients admitted to a single intensive care unit (ICU) was conducted for the period January 2005 to September 2010. Blood culture results were paired with clinical condition: "positive-sick"; "negative-sick", and "screening-not sick". Data were collected for the 72 hours prior to each blood culture. The most significant predictors were evaluated using logistic regression, Generalized Estimating Equations (GEE) and ROC area under the curve (AUC) analyses to assess model predictive ability. Bootstrapping methods were employed to evaluate potential model over-fitting. Fifty-nine subjects were included, representing 177 culture periods. SIRS criteria were not found to be associated with culture type, with an average of 98% of subjects meeting criteria in the 3 days prior. ABA sepsis criteria were significantly different among culture type only on the day prior (p = 0.004). The variables identified for the model included: heart rate>130 beats/min, mean blood pressure<60 mmHg, base deficit<-6 mEq/L, temperature>36°C, use of vasoactive medications, and glucose>150 mg/d1. The model was significant in predicting "positive culture-sick" and sepsis state, with AUC of 0.775 (p < 0.001) and 0.714 (p < .001), respectively; comparatively, the ABA criteria AUC was 0.619 (p = 0.028) and 0.597 (p = .035), respectively. SIRS criteria are not appropriate for identifying sepsis in the burn population. The ABA criteria perform better, but only for the day prior to positive blood culture results. The time period useful to diagnose sepsis using clinical criteria may be limited to 24 hours. A combination of predictors is superior to individual variable trends, yet algorithms or computer support will be necessary for the clinician to find such models useful. ^
Resumo:
Background: Poor communication among health care providers is cited as the most common cause of sentinel events involving patients. Sign-out of patient data at the change of clinician shifts is a component of communication that is especially vulnerable to errors. Sign-outs are particularly extensive and complex in intensive care units (ICUs). There is a paucity of validated tools to assess ICU sign-outs. ^ Objective: To design a valid and reliable survey tool to assess the perceptions of Pediatric ICU (PICU) clinicians about sign-out. ^ Design: Cross-sectional, web-based survey ^ Setting: Academic hospital, 31-bed PICU ^ Subjects: Attending faculty, fellows, nurse practitioners and physician assistants. ^ Interventions: A survey was designed with input from a focus group and administered to PICU clinicians. Test-retest reliability, internal consistency and validity of the survey tool were assessed. ^ Measurements and Main Results: Forty-eight PICU clinicians agreed to participate. We had 42(88%) and 40(83%) responses in the test and retest phases. The mean scores for the ten survey items ranged from 2.79 to 3.67 on a five point Likert scale with no significant test-retest difference and a Pearson correlation between pre and post answers of 0.65. The survey item scores showed internal consistency with a Cronbach's Alpha of 0.85. Exploratory factor analysis revealed three constructs: efficacy of sign-out process, recipient satisfaction and content applicability. Seventy eight % clinicians affirmed the need for improvement of the sign-out process and 83% confirmed the need for face- to-face verbal sign-out. A system-based sign-out format was favored by fellows and advanced level practitioners while attendings preferred a problem-based format (p=0.003). ^ Conclusions: We developed a valid and reliable survey to assess clinician perceptions about the ICU sign-out process. These results can be used to design a verbal template to improve and standardize the sign-out process.^
Resumo:
The first manuscript, entitled "Time-Series Analysis as Input for Clinical Predictive Modeling: Modeling Cardiac Arrest in a Pediatric ICU" lays out the theoretical background for the project. There are several core concepts presented in this paper. First, traditional multivariate models (where each variable is represented by only one value) provide single point-in-time snapshots of patient status: they are incapable of characterizing deterioration. Since deterioration is consistently identified as a precursor to cardiac arrests, we maintain that the traditional multivariate paradigm is insufficient for predicting arrests. We identify time series analysis as a method capable of characterizing deterioration in an objective, mathematical fashion, and describe how to build a general foundation for predictive modeling using time series analysis results as latent variables. Building a solid foundation for any given modeling task involves addressing a number of issues during the design phase. These include selecting the proper candidate features on which to base the model, and selecting the most appropriate tool to measure them. We also identified several unique design issues that are introduced when time series data elements are added to the set of candidate features. One such issue is in defining the duration and resolution of time series elements required to sufficiently characterize the time series phenomena being considered as candidate features for the predictive model. Once the duration and resolution are established, there must also be explicit mathematical or statistical operations that produce the time series analysis result to be used as a latent candidate feature. In synthesizing the comprehensive framework for building a predictive model based on time series data elements, we identified at least four classes of data that can be used in the model design. The first two classes are shared with traditional multivariate models: multivariate data and clinical latent features. Multivariate data is represented by the standard one value per variable paradigm and is widely employed in a host of clinical models and tools. These are often represented by a number present in a given cell of a table. Clinical latent features derived, rather than directly measured, data elements that more accurately represent a particular clinical phenomenon than any of the directly measured data elements in isolation. The second two classes are unique to the time series data elements. The first of these is the raw data elements. These are represented by multiple values per variable, and constitute the measured observations that are typically available to end users when they review time series data. These are often represented as dots on a graph. The final class of data results from performing time series analysis. This class of data represents the fundamental concept on which our hypothesis is based. The specific statistical or mathematical operations are up to the modeler to determine, but we generally recommend that a variety of analyses be performed in order to maximize the likelihood that a representation of the time series data elements is produced that is able to distinguish between two or more classes of outcomes. The second manuscript, entitled "Building Clinical Prediction Models Using Time Series Data: Modeling Cardiac Arrest in a Pediatric ICU" provides a detailed description, start to finish, of the methods required to prepare the data, build, and validate a predictive model that uses the time series data elements determined in the first paper. One of the fundamental tenets of the second paper is that manual implementations of time series based models are unfeasible due to the relatively large number of data elements and the complexity of preprocessing that must occur before data can be presented to the model. Each of the seventeen steps is analyzed from the perspective of how it may be automated, when necessary. We identify the general objectives and available strategies of each of the steps, and we present our rationale for choosing a specific strategy for each step in the case of predicting cardiac arrest in a pediatric intensive care unit. Another issue brought to light by the second paper is that the individual steps required to use time series data for predictive modeling are more numerous and more complex than those used for modeling with traditional multivariate data. Even after complexities attributable to the design phase (addressed in our first paper) have been accounted for, the management and manipulation of the time series elements (the preprocessing steps in particular) are issues that are not present in a traditional multivariate modeling paradigm. In our methods, we present the issues that arise from the time series data elements: defining a reference time; imputing and reducing time series data in order to conform to a predefined structure that was specified during the design phase; and normalizing variable families rather than individual variable instances. The final manuscript, entitled: "Using Time-Series Analysis to Predict Cardiac Arrest in a Pediatric Intensive Care Unit" presents the results that were obtained by applying the theoretical construct and its associated methods (detailed in the first two papers) to the case of cardiac arrest prediction in a pediatric intensive care unit. Our results showed that utilizing the trend analysis from the time series data elements reduced the number of classification errors by 73%. The area under the Receiver Operating Characteristic curve increased from a baseline of 87% to 98% by including the trend analysis. In addition to the performance measures, we were also able to demonstrate that adding raw time series data elements without their associated trend analyses improved classification accuracy as compared to the baseline multivariate model, but diminished classification accuracy as compared to when just the trend analysis features were added (ie, without adding the raw time series data elements). We believe this phenomenon was largely attributable to overfitting, which is known to increase as the ratio of candidate features to class examples rises. Furthermore, although we employed several feature reduction strategies to counteract the overfitting problem, they failed to improve the performance beyond that which was achieved by exclusion of the raw time series elements. Finally, our data demonstrated that pulse oximetry and systolic blood pressure readings tend to start diminishing about 10-20 minutes before an arrest, whereas heart rates tend to diminish rapidly less than 5 minutes before an arrest.
Resumo:
Over the last 2 decades, survival rates in critically ill cancer patients have improved. Despite the increase in survival, the intensive care unit (ICU) continues to be a location where end-of-life care takes place. More than 20% of deaths in the United States occur after admission to an ICU, and as baby boomers reach the seventh and eighth decades of their lives, the volume of patients in the ICU is predicted to rise. The aim of this study was to evaluate intensive care unit utilization among patients with cancer who were at the end of life. End of life was defined using decedent and high-risk cohort study designs. The decedent study evaluated characteristics and ICU utilization during the terminal hospital stay among patients who died at The University of Texas MD Anderson Cancer Center during 2003-2007. The high-risk cohort study evaluated characteristics and ICU utilization during the index hospital stay among patients admitted to MD Anderson during 2003-2007 with a high risk of in-hospital mortality. Factors associated with higher ICU utilization in the decedent study included non-local residence, hematologic and non-metastatic solid tumor malignancies, malignancy diagnosed within 2 months, and elective admission to surgical or pediatric services. Having a palliative care consultation on admission was associated with dying in the hospital without ICU services. In the cohort of patients with high risk of in-hospital mortality, patients who went to the ICU were more likely to be younger, male, with newly diagnosed non-metastatic solid tumor or hematologic malignancy, and admitted from the emergency center to one of the surgical services. A palliative care consultation on admission was associated with a decreased likelihood of having an ICU stay. There were no differences in ethnicity, marital status, comorbidities, or insurance status between patients who did and did not utilize ICU services. Inpatient mortality probability models developed for the general population are inadequate in predicting in-hospital mortality for patients with cancer. The following characteristics that differed between the decedent study and high-risk cohort study can be considered in future research to predict risk of in-hospital mortality for patients with cancer: ethnicity, type and stage of malignancy, time since diagnosis, and having advance directives. Identifying those at risk can precipitate discussions in advance to ensure care remains appropriate and in accordance with the wishes of the patient and family.^