567 resultados para 001
Resumo:
Background There are minimal reports of seasonal variations in chronic heart failure (CHF)-related morbidity and mortality beyond the northern hemisphere. Aims and methods We examined potential seasonal variations with respect to morbidity and all-cause mortality over more than a decade in a cohort of 2961 patients with CHF from a tertiary referral hospital in South Australia subject to mild winters and hot summers. Results Seasonal variation across all event-types was observed. CHF-related morbidity peaked in winter (July) and was lowest in summer (February): 70 (95% CI: 65 to 76) vs. 33 (95% CI: 30 to 37) admissions/1000 at risk (p<0.005). All-cause admissions (113 (95% CI: 107 to 120) vs. 73 (95% CI 68 to 79) admissions/1000 at risk, p<0.001) and concurrent respiratory disease (21% vs. 12%,p<0.001) were consistently higher in winter. 2010 patients died, mortality was highest in August relative to February: 23 (95% CI: 20 to 27) vs. 12 (95% CI: 10 to 15) deaths per 1000 at risk, p<0.001. Those aged 75 years or older were most at risk of seasonal variations in morbidity and mortality. Conclusion Seasonal variations in CHF-related morbidity and mortality occur in the hot climate of South Australia, suggesting that relative (rather than absolute) changes in temperature drive this global phenomenon.
Resumo:
Skeletal muscle displays enormous plasticity to respond to contractile activity with muscle from strength- (ST) and endurance-trained (ET) athletes representing diverse states of the adaptation continuum. Training adaptation can be viewed as the accumulation of specific proteins. Hence, the altered gene expression that allows for changes in protein concentration is of major importance for any training adaptation. Accordingly, the aim of the present study was to quantify acute subcellular responses in muscle to habitual and unfamiliar exercise. After 24-h diet/exercise control, 13 male subjects (7 ST and 6 ET) performed a random order of either resistance (8 × 5 maximal leg extensions) or endurance exercise (1 h of cycling at 70% peak O2 uptake). Muscle biopsies were taken from vastus lateralis at rest and 3 h after exercise. Gene expression was analyzed using real-time PCR with changes normalized relative to preexercise values. After cycling exercise, peroxisome proliferator-activated receptor-γ coactivator-1α (ET ∼8.5-fold, ST ∼10-fold, P < 0.001), pyruvate dehydrogenase kinase-4 (PDK-4; ET ∼26-fold, ST ∼39-fold), vascular endothelial growth factor (VEGF; ET ∼4.5-fold, ST ∼4-fold), and muscle atrophy F-box protein (MAFbx) (ET ∼2-fold, ST ∼0.4-fold) mRNA increased in both groups, whereas MyoD (∼3-fold), myogenin (∼0.9-fold), and myostatin (∼2-fold) mRNA increased in ET but not in ST (P < 0.05). After resistance exercise PDK-4 (∼7-fold, P < 0.01) and MyoD (∼0.7-fold) increased, whereas MAFbx (∼0.7-fold) and myostatin (∼0.6-fold) decreased in ET but not in ST. We conclude that prior training history can modify the acute gene responses in skeletal muscle to subsequent exercise.
Resumo:
Objective: To determine whether primary care management of chronic heart failure (CHF) differed between rural and urban areas in Australia. Design: A cross-sectional survey stratified by Rural, Remote and Metropolitan Areas (RRMA) classification. The primary source of data was the Cardiac Awareness Survey and Evaluation (CASE) study. Setting: Secondary analysis of data obtained from 341 Australian general practitioners and 23 845 adults aged 60 years or more in 1998. Main outcome measures: CHF determined by criteria recommended by the World Health Organization, diagnostic practices, use of pharmacotherapy, and CHF-related hospital admissions in the 12 months before the study. Results: There was a significantly higher prevalence of CHF among general practice patients in large and small rural towns (16.1%) compared with capital city and metropolitan areas (12.4%) (P < 0.001). Echocardiography was used less often for diagnosis in rural towns compared with metropolitan areas (52.0% v 67.3%, P < 0.001). Rates of specialist referral were also significantly lower in rural towns than in metropolitan areas (59.1% v 69.6%, P < 0.001), as were prescribing rates of angiotensin-converting enzyme inhibitors (51.4% v 60.1%, P < 0.001). There was no geographical variation in prescribing rates of β-blockers (12.6% [rural] v 11.8% [metropolitan], P = 0.32). Overall, few survey participants received recommended “evidence-based practice” diagnosis and management for CHF (metropolitan, 4.6%; rural, 3.9%; and remote areas, 3.7%). Conclusions: This study found a higher prevalence of CHF, and significantly lower use of recommended diagnostic methods and pharmacological treatment among patients in rural areas.
Resumo:
Background: Chronic diseases including type 2 diabetes are a leading cause of morbidity and mortality in midlife and older Australian women. There are a number of modifiable risk factors for type 2 diabetes and other chronic diseases including smoking, nutrition, physical activity and overweight and obesity. Little research has been conducted in the Australian context to explore the perceived barriers to health promotion activities in midlife and older Australian women with a chronic disease. Aims: The primary aim of this study was to explore women’s perceived barriers to health promotion activities to reduce modifiable risk factors, and the relationship of perceived barriers to smoking behaviour, fruit and vegetable intake, physical activity and body mass index. A secondary aim of this study was to investigate nurses’ perceptions of the barriers to action for women with a chronic disease, and to compare those perceptions with those of the women. Methods: The study was divided into two phases where Phase 1 was a cross sectional survey of women, aged over 45 years with type 2 diabetes who were attending Diabetes clinics in the Primary and Community Health Service of the Metro North Health Service District of Queensland Health (N = 22). The women were a subsample of women participating in a multi-model lifestyle intervention, the ‘Reducing Chronic Disease among Adult Australian Women’ project. Phase 2 of the study was a cross sectional online survey of nurses working in Primary and Community Health Service in the Metro North Health Service District of Queensland Health (N = 46). Pender’s health promotion model was used as the theoretical framework for this study. Results: Women in this study had an average total barriers score of 32.18 (SD = 9.52) which was similar to average scores reported in the literature for women with a range of physical disabilities and illnesses. The leading five barriers for this group of women were: concern about safety; too tired; not interested; lack of information about what to do; with lack of time and feeling I can’t do things correctly the equal fifth ranked barriers. In this study there was no statistically significant difference in average total barriers scores between women in the intervention group and those is the usual care group of the parent study. There was also no significant relationship between the women’s socio-demographic variables and lifestyle risk factors and their level of perceived barriers. Nurses in the study had an average total barriers score of 44.48 (SD = 6.24) which was higher than all other average scores reported in the literature. The leading five barriers that nurses perceived were an issue for women with a chronic disease were: lack of time and interferes with other responsibilities the leading barriers; embarrassment about appearance; lack of money; too tired and lack of support from family and friends. There was no significant relationship between the nurses’ sociodemographic and nursing variables and the level of perceived barriers. When comparing the results of women and nurses in the study there was a statistically significant difference in the median total barriers score between the groups (p < 0.001), where the nurses perceived the barriers to be higher (Md = 43) than the women (Md = 33). There was also a significant difference in the responses to the individual barriers items in fifteen of the eighteen items (p < 0.002). Conclusion: Although this study is limited by a small sample size, it contributes to understanding the perception of midlife and older women with a chronic disease and also the perception of nurses, about the barriers to healthy lifestyle activities that women face. The study provides some evidence that the perceptions of women and nurses may differ and argues that these differences may have significant implications for clinical practice. The study recommends a greater emphasis on assessing and managing perceived barriers to health promotion activities in health education and policy development and proposes a conceptual model for understanding perceived barriers to action.
Resumo:
We compared changes in markers of muscle damage and systemic inflammation after submaximal and maximal lengthening muscle contractions of the elbow flexors. Using a cross-over design, 10 healthy young men not involved in resistance training completed a submaximal trial (10 sets of 60 lengthening contractions at 10% maximum isometric strength, 1 min rest between sets), followed by a maximal trial (10 sets of three lengthening contractions at 100% maximum isometric strength, 3 min rest between sets). Lengthening contractions were performed on an isokinetic dynamometer. Opposite arms were used for the submaximal and maximal trials, and the trials were separated by a minimum of two weeks. Blood was sampled before, immediately after, 1 h, 3 h, and 1-4 d after each trial. Total leukocyte and neutrophil numbers, and the serum concentration of soluble tumor necrosis factor-alpha receptor 1 were elevated after both trials (P < 0.01), but there were no differences between the trials. Serum IL-6 concentration was elevated 3 h after the submaximal contractions (P < 0.01). The concentrations of serum tumor necrosis factor-alpha, IL-1 receptor antagonist, IL-10, granulocyte-colony stimulating factor and plasma C-reactive protein remained unchanged following both trials. Maximum isometric strength and range of motion decreased significantly (P < 0.001) after both trials, and were lower from 1-4 days after the maximal contractions compared to the submaximal contractions. Plasma myoglobin concentration and creatine kinase activity, muscle soreness and upper arm circumference all increased after both trials (P < 0.01), but were not significantly different between the trials. Therefore, there were no differences in markers of systemic inflammation, despite evidence of greater muscle damage following maximal versus submaximal lengthening contractions of the elbow flexors.
Resumo:
Lower fruit and vegetable intake among socioeconomically disadvantaged groups has been well documented, and may be a consequence of a higher consumption of take-out foods. This study examined whether, and to what extent, take-out food consumption mediated (explained) the association between socioeconomic position and fruit and vegetable intake. A cross-sectional postal survey was conducted among 1500 randomly selected adults aged 25–64 years in Brisbane, Australia in 2009 (response rate = 63.7%, N = 903). A food frequency questionnaire assessed usual daily servings of fruits and vegetables (0 to 6), overall take-out consumption (times/week) and the consumption of 22 specific take-out items (never to ≥once/day). These specific take-out items were grouped into “less healthy” and “healthy” choices and indices were created for each type of choice (0 to 100). Socioeconomic position was ascertained by education. The analyses were performed using linear regression, and a bootstrap re-sampling approach estimated the statistical significance of the mediated effects. Mean daily serves of fruits and vegetables was 1.89 (SD 1.05) and 2.47 (SD 1.12) respectively. The least educated group were more likely to consume fewer serves of fruit (B= –0.39, p<0.001) and vegetables (B= –0.43, p<0.001) compared with the highest educated. The consumption of “less healthy” take-out food partly explained (mediated) education differences in fruit and vegetable intake; however, no mediating effects were observed for overall and “healthy” take-out consumption. Regular consumption of “less healthy” take-out items may contribute to socioeconomic differences in fruit and vegetable intake, possibly by displacing these foods.
Resumo:
Objective: To assess the recall of media reports about vitamin D and associated factors. Methods: Analysis of cross-sectional telephone interview data (2,001 Queensland adults, 18-70 years) on vitamin D and personal sun protection, recall of media reports and participant characteristics. Results: 83.7% of participants had heard of vitamin D, 47.5% through the media. Only 513 (25.6%) participants recalled the media content within four main themes: vitamin D is beneficial/comes from the sun (47.0%); some people aren’t getting enough vitamin D, need more sun (27.9%); need to balance sun exposure and skin protection (11.5%); or other (13.6%). Only 65 of the 950 participants (6.8%) reported a change to their behaviour(s) due to the media report. Conclusion: Although the media were the main source of information about vitamin D for almost 50% of participants, recall of the content and direct effect on behaviour was low. Only a small minority recalled a balanced media report of beneficial and harmful aspects of sun exposure. Implications Health professionals often supply media with background information. To achieve best public health practice for sun protection and vitamin D, information to foster balanced media reports should be provided.
Resumo:
Background: Queensland men aged 50 years and older are at high risk for melanoma. Early detection via skin self examination (SSE) (particularly whole-body SSE) followed by presentation to a doctor with suspicious lesions, may decrease morbidity and mortality from melanoma. Prevalence of whole-body SSE (wbSSE) is lower in Queensland older men compared to other population subgroups. With the exception of the present study no previous research has investigated the determinants of wbSSE in older men, or interventions to increase the behaviour in this population. Furthermore, although past SSE intervention studies for other populations have cited health behaviour models in the development of interventions, no study has tested these models in full. The Skin Awareness Study: A recent randomised trial, called the Skin Awareness Study, tested the impact of a video-delivered intervention compared to written materials alone on wbSSE in men aged 50 years or older (n=930). Men were recruited from the general population and interviewed over the telephone at baseline and 13 months. The proportion of men who reported wbSSE rose from 10% to 31% in the control group, and from 11% to 36% in the intervention group. Current research: The current research was a secondary analysis of data collected for the Skin Awareness Study. The objectives were as follows: • To describe how men who did not take up any SSE during the study period differed from those who did take up examining their skin. • To determine whether the intervention program was successful in affecting the constructs of the Health Belief Model it was aimed at (self-efficacy, perceived threat, and outcome expectations); and whether this in turn influenced wbSSE. • To determine whether the Health Action Process Approach (HAPA) was a better predictor of wbSSE behaviour compared to the Health Belief Model (HBM). Methods: For objective 1, men who did not report any past SSE at baseline (n=308) were categorised as having ‘taken up SSE’ (reported SSE at study end) or ‘resisted SSE’ (reported no SSE at study end). Bivariate logistic regression, followed by multivariable regression, investigated the association between participant characteristics measured at baseline and resisting SSE. For objective 2 proxy measures of self-efficacy, perceived threat, and outcome expectations were selected. To determine whether these mediated the effect of the intervention on the outcome, a mediator analysis was performed with all participants who completed interviews at both time points (n=830) following the Baron and Kenny approach, modified for use with structural equation modelling (SEM). For objective 3, control group participants only were included (n=410). Proxy measures of all HBM and HAPA constructs were selected and SEM was used to build up models and test the significance of each hypothesised pathway. A likelihood ratio test compared the HAPA to the HBM. Results: Amongst men who did not report any SSE at baseline, 27% did not take up any SSE by the end of the study. In multivariable analyses, resisting SSE was associated with having more freckly skin (p=0.027); being unsure about the statement ‘if I saw something suspicious on my skin, I’d go to the doctor straight away’ (p=0.028); not intending to perform SSE (p=0.015), having lower SSE self-efficacy (p<0.001), and having no recommendation for SSE from a doctor (p=0.002). In the mediator analysis none of the tested variables mediated the relationship between the intervention and wbSSE. In regards to health behaviour models, the HBM did not predict wbSSE well overall. Only the construct of self-efficacy was a significant predictor of future wbSSE (p=0.001), while neither perceived threat (p=0.584) nor outcome expectations (p=0.220) were. By contrast, when the HAPA constructs were added, all three HBM variables predicted intention to perform SSE, which in turn predicted future behaviour (p=0.015). The HAPA construct of volitional self-efficacy was also associated with wbSSE (p=0.046). The HAPA was a significantly better model compared to the HBM (p<0.001). Limitations: Items selected to measure HBM and HAPA model constructs for objectives 2 and 3 may not have accurately reflected each construct. Conclusions: This research added to the evidence base on how best to target interventions to older men; and on the appropriateness of particular health behaviour models to guide interventions. Findings indicate that to overcome resistance those men with more negative pre-existing attitudes to SSE (not intending to do it, lower initial self-efficacy) may need to be targeted with more intensive interventions in the future. Involving general practitioners in recommending SSE to their patients in this population, alongside disseminating an intervention, may increase its success. Comparison of the HBM and HAPA showed that while two of the three HBM variables examined did not directly predict future wbSSE, all three were associated with intention to self-examine skin. This suggests that in this population, intervening on these variables may increase intention to examine skin, but not necessarily the behaviour itself. Future interventions could potentially focus on increasing both the motivational variables of perceived threat and outcome expectations as well as a combination of both action and volitional self-efficacy; with the aim of increasing intention as well as its translation to taking up and maintaining regular wbSSE.
Resumo:
A healthy human would be expected to show periodic blinks, making a brief closure of the eyelids. Most blinks are spontaneous, occurring regularly with no external stimulus. However a reflex blink can occur in response to external stimuli such as a bright light, a sudden loud noise, or an object approaching toward the eyes. A voluntary or forced blink is another type of blink in which the person deliberately closes the eyes and the lower eyelid raises to meet the upper eyelid. A complete blink, in which the upper eyelid touches the lower eyelid, contributes to the health of ocular surface by providing a fresh layer of tears as well as maintaining optical integrity by providing a smooth tear film over the cornea. The rate of blinking and its completeness vary depending on the task undertaken during blink assessment, the direction of gaze, the emotional state of the subjects and the method under which the blink was measured. It is also well known that wearing contact lenses (both rigid and soft lenses) can induce significant changes in blink rate and completeness. It is been established that efficient blinking plays an important role in ocular surface health during contact lens wear and for improving contact lens performance and comfort. Inefficient blinking during contact lens wear may be related to a low blink rate or incomplete blinking and can often be a reason for dry eye symptoms or ocular surface staining. It has previously been shown that upward gaze can affect blink rate, causing it to become faster. In the first experiment, it was decided to expand on previous studies in this area by examining the effect of various gaze directions (i.e. upward gaze, primary gaze, downward gaze and lateral gaze) as well as head angle (recumbent position) on normal subjects’ blink rate and completeness through the use of filming with a high-speed camera. The results of this experiment showed that as the open palpebral aperture (and exposed ocular surface area) increased from downward gaze to upward gaze, the number of blinks significantly increased (p<0.04). Also, the size of closed palpebral aperture significantly increased from downward gaze to upward gaze (p<0.005). A weak positive correlation (R² = 0.18) between the blink rate and ocular surface area was found in this study. Also, it was found that the subjects showed 81% complete blinks, 19% incomplete blinks and 2% of twitch blinks in primary gaze, consistent with previous studies. The difference in the percentage of incomplete blinks between upward gaze and downward gaze was significant (p<0.004), showing more incomplete blinks in upward gaze. The findings of this experiment suggest that while blink rate becomes slower in downward gaze, the completeness of blinking is typically better, thereby potentially reducing the risk of tear instability. On the other hand, in upward gaze while the completeness of blinking becomes worse, this is potentially offset by increased blink frequency. In addition, blink rate and completeness were not affected by lateral gaze or head angle, possibly because these conditions have similar size of the open palpebral aperture compared with primary gaze. In the second experiment, an investigation into the changes in blink rate and completeness was carried out in primary gaze and downward gaze with soft and rigid contact lenses in unadapted wearers. Not surprisingly, rigid lens wear caused a significant increase in the blink rate in both primary (p<0.001) and downward gaze (p<0.02). After fitting rigid contact lenses, the closed palpebral aperture (blink completeness) did not show any changes but the open palpebral aperture showed a significant narrowing (p<0.04). This might occur from the subjects’ attempt to avoid interaction between the upper eyelid and the edge of the lens to minimize discomfort. After applying topical anaesthetic eye drops in the eye fitted with rigid lenses, the increased blink rate dropped to values similar to that before lens insertion and the open palpebral aperture returned to baseline values, suggesting that corneal and/or lid margin sensitivity was mediating the increased blink rate and narrowed palpebral aperture. We also investigated the changes in the blink rate and completeness with soft contact lenses including a soft sphere, double slab-off toric design and periballast toric design. Soft contact lenses did not cause any significant changes in the blink rate, closed palpebral aperture, open palpebral aperture and the percentage of incomplete blinks in either primary gaze or downward gaze. After applying anaesthetic eye drops, the blink rate reduced in both primary gaze and downward gaze, however this difference was not statistically significant. The size of the closed palpebral aperture and open palpebral aperture did not show any significant changes after applying anaesthetic eye drops. However it should be noted that the effects of rigid and soft contact lenses that we observed in these studies were only the immediate reaction to contact lenses and in the longer term, it is likely that these responses will vary as the eye adapts to the presence of the lenses.