978 resultados para Venous Insufficiency
Resumo:
OBJECTIVE Whether or not a high risk of falls increases the risk of bleeding in patients receiving anticoagulants remains a matter of debate. METHODS We conducted a prospective cohort study involving 991 patients ≥ 65 years of age who received anticoagulants for acute venous thromboembolism (VTE) at nine Swiss hospitals between September 2009 and September 2012. The study outcomes were as follows: the time to a first major episode of bleeding; and clinically relevant nonmajor bleeding. We determined the associations between the risk of falls and the time to a first episode of bleeding using competing risk regression, accounting for death as a competing event. We adjusted for known bleeding risk factors and anticoagulation as a time-varying covariate. RESULTS Four hundred fifty-eight of 991 patients (46%) were at high risk of falls. The mean duration of follow-up was 16.7 months. Patients at high risk of falls had a higher incidence of major bleeding (9.6 vs. 6.6 events/100 patient-years; P = 0.05) and a significantly higher incidence of clinically relevant nonmajor bleeding (16.7 vs. 8.3 events/100 patient-years; P < 0.001) than patients at low risk of falls. After adjustment, a high risk of falls was associated with clinically relevant nonmajor bleeding [subhazard ratio (SHR) = 1.74, 95% confidence interval (CI) = 1.23-2.46], but not with major bleeding (SHR = 1.24, 95% CI = 0.83-1.86). CONCLUSION In elderly patients who receive anticoagulants because of VTE, a high risk of falls is significantly associated with clinically relevant nonmajor bleeding, but not with major bleeding. Whether or not a high risk of falls is a reason against providing anticoagulation beyond 3 months should be based on patient preferences and the risk of VTE recurrence.
Resumo:
OBJECTIVES The purpose of this study was to investigate the survival effects of inferior vena cava filters in patients with venous thromboembolism (VTE) who had a significant bleeding risk. BACKGROUND The effectiveness of inferior vena cava filter use among patients with acute symptomatic VTE and known significant bleeding risk remains unclear. METHODS In this prospective cohort study of patients with acute VTE identified from the RIETE (Computerized Registry of Patients With Venous Thromboembolism), we assessed the association between inferior vena cava filter insertion for known significant bleeding risk and the outcomes of all-cause mortality, pulmonary embolism (PE)-related mortality, and VTE rates through 30 days after the initiation of VTE treatment. Propensity score matching was used to adjust for the likelihood of receiving a filter. RESULTS Of the 40,142 eligible patients who had acute symptomatic VTE, 371 underwent filter placement because of known significant bleeding risk. A total of 344 patients treated with a filter were matched with 344 patients treated without a filter. Propensity score-matched pairs showed a nonsignificant trend toward lower risk of all-cause death for filter insertion compared with no insertion (6.6% vs. 10.2%; p = 0.12). The risk-adjusted PE-related mortality rate was lower for filter insertion than no insertion (1.7% vs. 4.9%; p = 0.03). Risk-adjusted recurrent VTE rates were higher for filter insertion than for no insertion (6.1% vs. 0.6%; p < 0.001). CONCLUSIONS In patients presenting with VTE and with a significant bleeding risk, inferior vena cava filter insertion compared with anticoagulant therapy was associated with a lower risk of PE-related death and a higher risk of recurrent VTE. However, study design limitations do not imply a causal relationship between filter insertion and outcome.
Resumo:
Venous thromboembolism (VTE) is a potentially lethal clinical condition that is suspected in patients with common clinical complaints, in many and varied, clinical care settings. Once VTE is diagnosed, optimal therapeutic management (thrombolysis, IVC filters, type and duration of anticoagulants) and ideal therapeutic management settings (outpatient, critical care) are also controversial. Clinical prediction tools, including clinical decision rules and D-Dimer, have been developed, and some validated, to assist clinical decision making along the diagnostic and therapeutic management paths for VTE. Despite these developments, practice variation is high and there remain many controversies in the use of the clinical prediction tools. In this narrative review, we highlight challenges and controversies in VTE diagnostic and therapeutic management with a focus on clinical decision rules and D-Dimer.
Resumo:
OBJECTIVE The purpose of this study was to evaluate the prevalence of mesenteric venous thrombosis (MVT) in the Swiss Inflammatory Bowel Disease Cohort Study and to correlate MVT with clinical outcome. MATERIALS AND METHODS Abdominal portal phase CT was used to examine patients with inflammatory bowel disease (IBD). Two experienced abdominal radiologists retrospectively analyzed the images, focusing on the superior and inferior mesenteric vein branches and looking for signs of acute or chronic thrombosis. The location of abnormalities was registered. The presence of MVT was correlated with IBD-related radiologic signs and complications. RESULTS The cases of 160 patients with IBD (89 women, 71 men; Crohn disease [CD], 121 patients; ulcerative colitis [UC], 39 patients; median age at diagnosis, 27 years for patients with CD, 32 years for patients with UC) were analyzed. MVT was detected in 43 patients with IBD (26.8%). One of these patients had acute MVT; 38, chronic MVT; and four, both. The prevalence of MVT did not differ between CD (35/121 [28.9%]) and UC (8/39 [20.5%]) (p = 0.303). The location of thrombosis was different between CD and UC (CD, jejunal or ileal veins only [p = 0.005]; UC, rectocolic veins only [p = 0.001]). Almost all (41/43) cases of thrombosis were peripheral. MVT in CD patients was more frequently associated with bowel wall thickening (p = 0.013), mesenteric fat hypertrophy (p = 0.005), ascites (p = 0.002), and mesenteric lymph node enlargement (p = 0.036) and was associated with higher rate of bowel stenosis (p < 0.001) and more intestinal IBD-related surgery (p = 0.016) in the outcome. Statistical analyses for patients with UC were not relevant because of the limited population (n = 8). CONCLUSION MVT is frequently found in patients with IBD. Among patients with CD, MVT is associated with bowel stenosis and CD-related intestinal surgery.
Resumo:
PURPOSE Blood loss and blood substitution are associated with higher morbidity after major abdominal surgery. During major liver resection, low local venous pressure, has been shown to reduce blood loss. Ambiguity persists concerning the impact of local venous pressure on blood loss during open radical cystectomy. We aimed to determine the association between intraoperative blood loss and pelvic venous pressure (PVP) and determine factors affecting PVP. MATERIAL AND METHODS In the frame of a single-center, double-blind, randomized trial, PVP was measured in 82 patients from a norepinephrine/low-volume group and in 81 from a control group with liberal hydration. For this secondary analysis, patients from each arm were stratified into subgroups with PVP <5 mmHg or ≥5 mmHg measured after cystectomy (optimal cut-off value for discrimination of patients with relevant blood loss according to the Youden's index). RESULTS Median blood loss was 800 ml [range: 300-1600] in 55/163 patients (34%) with PVP <5 mmHg and 1200 ml [400-3000] in 108/163 patients (66%) with PVP ≥5 mmHg; (P<0.0001). A PVP <5 mmHg was measured in 42/82 patients (51%) in the norepinephrine/low-volume group and 13/81 (16%) in the control group (P<0.0001). PVP dropped significantly after removal of abdominal packing and abdominal lifting in both groups at all time points (at begin and end of pelvic lymph node dissection, end of cystectomy) (P<0.0001). No correlation between PVP and central venous pressure could be detected. CONCLUSIONS Blood loss was significantly reduced in patients with low PVP. Factors affecting PVP were fluid management and abdominal packing.
Resumo:
Renal cell carcinoma (RCC) extension into the renal vein or the inferior vena cava occurs in 4%-10% of all kidney cancer cases. This entity shows a wide range of different clinical and surgical scenarios, making natural history and oncological outcomes variable and poorly characterized. Infrequency and variability make it necessary to share the experience from different institutions to properly analyze surgical outcomes in this setting. The International Renal Cell Carcinoma-Venous Tumor Thrombus Consortium was created to answer the questions generated by competing results from different retrospective studies in RCC with venous extension on current controversial topics. The aim of this article is to summarize the experience gained from the analysis of the world's largest cohort of patients in this unique setting to date.
Resumo:
BACKGROUND Liver regeneration is of crucial importance for patients undergoing living liver transplantations or extended liver resections and can be associated with elevated portal venous pressure, impaired hepatic regeneration, and postoperative morbidity. The aim of this study was to assess whether reduction of portal venous pressure by terlipressin improves postoperative liver regeneration in normal and steatotic livers after partial hepatectomy in a rodent model. METHODS Portal venous pressure was assessed after minor (30%), standard (60%), or extended (80%) partial hepatectomy (PH) in mice with and without liver steatosis. Liver regeneration was assessed by BrdU incorporation and Ki-67 immunostaining. RESULTS Portal venous pressure was significantly elevated post-PH in mice with normal and steatotic livers compared to sham-operated mice. Reduction of elevated portal pressure after 80% PH by terlipressin was associated with an increase of hepatocellular proliferation. In steatotic livers, animals treated with terlipressin had an increase in liver regeneration after 30% PH and increased survival after 60% PH. Mechanistically, terlipressin alleviated IL-6 mRNA expression following PH and down-regulated p21 and GADD45 mRNA suggesting a reduction of cell cycle inhibition and cellular stress. CONCLUSIONS Reduction of elevated portal pressure post-PH by the use of terlipressin improves liver regeneration after PH in lean and steatotic mouse livers.
Resumo:
Primary adrenal insufficiency (PAI) is a rare condition in childhood which is either inherited (mostly) or acquired. It is characterized by glucocorticoid and maybe mineralocorticoid deficiency. The most common form in children is 21-hydroxylase deficiency, which belongs to the steroid biosynthetic defects causing PAI. Newer forms of complex defects of steroid biosynthesis are P450 oxidoreductase deficiency and (apparent) cortisone reductase deficiency. Other forms of PAI include metabolic disorders, autoimmune disorders and adrenal dysgenesis, e.g. the IMAGe syndrome, for which the underlying genetic defect has been recently identified. Newer work has also expanded the genetic causes underlying isolated, familial glucocorticoid deficiency (FGD). Mild mutations of CYP11A1 or StAR have been identified in patients with FGD. MCM4 mutations were found in a variant of FGD in an Irish travelling community manifesting with PAI, short stature, microcephaly and recurrent infections. Finally, mutations in genes involved in the detoxification of reactive oxygen species were identified in patients with unsolved FGD. Most mutations were found in the enzyme nicotinamide nucleotide transhydrogenase, which uses the mitochondrial proton pump gradient to produce NADPH. NADPH is essential in maintaining high levels of reduced forms of antioxidant enzymes for the reduction of hydrogen peroxide. Similarly, mutations in the gene for TXNRD2 involved in this system were found in FGD patients, suggesting that the adrenal cortex is particularly susceptible to oxidative stress.
Resumo:
Abstract PURPOSE: Reliable animal models are essential to evaluate future therapeutic options like cell-based therapies for external anal sphincter insufficiency. The goal of our study was to describe the most reliable model for external sphincter muscle insufficiency by comparing three different methods to create sphincter muscle damage. METHODS: In an experimental animal study, female Lewis rats (200-250 g) were randomly assigned to three treatment groups (n = 5, each group). The external sphincter muscle was weakened in the left dorsal quadrant by microsurgical excision, cryosurgery, or electrocoagulation by diathermy. Functional evaluation included in vivo measurements of resting pressure, spontaneous muscle contraction, and contraction in response to electrical stimulation of the afferent nerve at baseline and at 2, 4, and 6 weeks after sphincter injury. Masson's trichrome staining and immunofluorescence for skeletal muscle markers was performed for morphological analysis. RESULTS: Peak contraction after electrical stimulation was significantly decreased after sphincter injury in all groups. Contraction forces recovered partially after cryosurgery and electrocoagulation but not after microsurgical excision. Morphological analysis revealed an incomplete destruction of the external sphincter muscle in the cryosurgery and electrocoagulation groups compared to the microsurgery group. CONCLUSIONS: For the first time, three different models of external sphincter muscle insufficiency were directly compared. The animal model using microsurgical sphincter destruction offers the highest level of consistency regarding tissue damage and sphincter insufficiency, and therefore represents the most reliable model to evaluate future therapeutic options. In addition, this study represents a novel model to specifically test the external sphincter muscle function.
Resumo:
OBJECTIVES To evaluate the impact of preoperative sepsis on risk of postoperative arterial and venous thromboses. DESIGN Prospective cohort study using the National Surgical Quality Improvement Program database of the American College of Surgeons (ACS-NSQIP). SETTING Inpatient and outpatient procedures in 374 hospitals of all types across the United States, 2005-12. PARTICIPANTS 2,305,380 adults who underwent surgical procedures. MAIN OUTCOME MEASURES Arterial thrombosis (myocardial infarction or stroke) and venous thrombosis (deep venous thrombosis or pulmonary embolism) in the 30 days after surgery. RESULTS Among all surgical procedures, patients with preoperative systemic inflammatory response syndrome or any sepsis had three times the odds of having an arterial or venous postoperative thrombosis (odds ratio 3.1, 95% confidence interval 3.0 to 3.1). The adjusted odds ratios were 2.7 (2.5 to 2.8) for arterial thrombosis and 3.3 (3.2 to 3.4) for venous thrombosis. The adjusted odds ratios for thrombosis were 2.5 (2.4 to 2.6) in patients with systemic inflammatory response syndrome, 3.3 (3.1 to 3.4) in patients with sepsis, and 5.7 (5.4 to 6.1) in patients with severe sepsis, compared with patients without any systemic inflammation. In patients with preoperative sepsis, both emergency and elective surgical procedures had a twofold increased odds of thrombosis. CONCLUSIONS Preoperative sepsis represents an important independent risk factor for both arterial and venous thromboses. The risk of thrombosis increases with the severity of the inflammatory response and is higher in both emergent and elective surgical procedures. Suspicion of thrombosis should be higher in patients with sepsis who undergo surgery.
Resumo:
Portal vein embolization (PVE) may be performed before hemihepatectomy to increase the volume of future liver remnant (FLR) and to reduce the risk of postoperative liver insufficiency. We report the case of a 71-year-old patient with hilar cholangiocarcinoma undergoing PVE with access from the right portal vein using a mixture of n-butyl-2-cyanoacrylate and ethiodized oil. During the procedure, nontarget embolization of the left portal vein occurred. An aspiration maneuver of the polymerized plug failed; however, the embolus obstructing portal venous flow in the FLR was successfully relocated into the right portal vein while carefully bypassing the plug with a balloon catheter, inflating the balloon, and pulling the plug into the main right portal vein.
Resumo:
BACKGROUND: In clinical practise the high dose ACTH stimulation test (HDT) is frequently used in the assessment of adrenal insufficiency (AI). However, there is uncertainty regarding optimal time-points and number of blood samplings. The present study compared the utility of a single cortisol value taken either 30 or 60 minutes after ACTH stimulation with the traditional interpretation of the HDT. METHODS: Retrospective analysis of 73 HDT performed at a single tertiary endocrine centre. Serum cortisol was measured at baseline, 30 and 60 minutes after intravenous administration of 250 µg synthetic ACTH1-24. Adrenal insufficiency (AI) was defined as a stimulated cortisol level <550 nmol/l. RESULTS: There were twenty patients (27.4%) who showed an insufficient rise in serum cortisol using traditional HDT criteria and were diagnosed to suffer from AI. There were ten individuals who showed insufficient cortisol values after 30 minutes, rising to sufficient levels at 60 minutes. All patients revealing an insufficient cortisol response result after 60 minutes also had an insufficient result after 30 minutes. The cortisol value taken after 30 minutes did not add incremental diagnostic value in any of the cases under investigation compared with the 60 minutes' sample. CONCLUSIONS: Based on the findings of the present analysis the utility of a cortisol measurement 30 minutes after high dose ACTH injection was low and did not add incremental diagnostic value to a single measurement after 60 minutes.
Resumo:
BACKGROUND AND PURPOSE Currently one of the most widely used models for the development of endovascular techniques and coiling devices for treatment of aneurysm is the elastase-induced aneurysm model in the rabbit carotid artery. Microsurgical techniques for creating an aneurysm with a venous pouch have also been established, although both techniques usually result in aneurysms less than 1 cm in diameter. We investigated whether an increase in blood flow toward the neck would produce larger aneurysms in a microsurgical venous pouch model. MATERIALS AND METHODS Microsurgical operations were performed on 11 New Zealand white rabbits. Both carotid arteries and the right jugular vein were dissected, and the right carotid artery was temporarily clipped followed by an arteriotomy. The left carotid artery was also clipped proximally, ligated distally, and sutured onto the proximal half of the arteriotomy in the right carotid artery. The venous graft was sutured onto the distal half of the arteriotomy. Digital subtraction angiography was also performed. RESULTS Angiography showed patent anastomosed vessels and aneurysms in the seven surviving rabbits. Mean aneurysm measurements among surviving rabbits with patent vessels were: 13.9 mm length, 9.3 mm width, and neck diameter 4.7 mm. The resulting mean aspect ratio was 3.35 and the mean bottleneck ratio was 3.05. CONCLUSION A large venous graft and increased blood flow toward the base of the aneurysm seem to be key factors in the creation of large venous pouch aneurysms. These large aneurysms allow testing of endovascular devices designed for large and giant aneurysms.