790 resultados para Research into design
Resumo:
Only recently, during the past five years, consumer electronics has been evolving rapidly. Many products have started to include “smart home” capabilities, enabling communication and interoperability of various smart devices. Even more devices and sensors can be remote controlled and monitored through cloud services. While the smart home systems have become very affordable to average consumer compared to the early solutions decades ago, there are still many issues and things that need to be fixed or improved upon: energy efficiency, connectivity with other devices and applications, security and privacy concerns, reliability, and response time. This paper focuses on designing Internet of Things (IoT) node and platform architectures that take these issues into account, notes other currently used solutions, and selects technologies in order to provide better solution. The node architecture aims for energy efficiency and modularity, while the platform architecture goals are in scalability, portability, maintainability, performance, and modularity. Moreover, the platform architecture attempts to improve user experience by providing higher reliability and lower response time compared to the alternative platforms. The architectures were developed iteratively using a development process involving research, planning, design, implementation, testing, and analysis. Additionally, they were documented using Kruchten’s 4+1 view model, which is used to describe the use cases and different views of the architectures. The node architecture consisted of energy efficient hardware, FC3180 microprocessor and CC2520 RF transceiver, modular operating system, Contiki, and a communication protocol, AllJoyn, used for providing better interoperability with other IoT devices and applications. The platform architecture provided reliable low response time control, monitoring, and initial setup capabilities by utilizing web technologies on various devices such as smart phones, tablets, and computers. Furthermore, an optional cloud service was provided in order to control devices and monitor sensors remotely by utilizing scalable high performance technologies in the backend enabling low response time and high reliability.
Resumo:
Abstract Scheduling problems are generally NP-hard combinatorial problems, and a lot of research has been done to solve these problems heuristically. However, most of the previous approaches are problem-specific and research into the development of a general scheduling algorithm is still in its infancy. Mimicking the natural evolutionary process of the survival of the fittest, Genetic Algorithms (GAs) have attracted much attention in solving difficult scheduling problems in recent years. Some obstacles exist when using GAs: there is no canonical mechanism to deal with constraints, which are commonly met in most real-world scheduling problems, and small changes to a solution are difficult. To overcome both difficulties, indirect approaches have been presented (in [1] and [2]) for nurse scheduling and driver scheduling, where GAs are used by mapping the solution space, and separate decoding routines then build solutions to the original problem. In our previous indirect GAs, learning is implicit and is restricted to the efficient adjustment of weights for a set of rules that are used to construct schedules. The major limitation of those approaches is that they learn in a non-human way: like most existing construction algorithms, once the best weight combination is found, the rules used in the construction process are fixed at each iteration. However, normally a long sequence of moves is needed to construct a schedule and using fixed rules at each move is thus unreasonable and not coherent with human learning processes. When a human scheduler is working, he normally builds a schedule step by step following a set of rules. After much practice, the scheduler gradually masters the knowledge of which solution parts go well with others. He can identify good parts and is aware of the solution quality even if the scheduling process is not completed yet, thus having the ability to finish a schedule by using flexible, rather than fixed, rules. In this research we intend to design more human-like scheduling algorithms, by using ideas derived from Bayesian Optimization Algorithms (BOA) and Learning Classifier Systems (LCS) to implement explicit learning from past solutions. BOA can be applied to learn to identify good partial solutions and to complete them by building a Bayesian network of the joint distribution of solutions [3]. A Bayesian network is a directed acyclic graph with each node corresponding to one variable, and each variable corresponding to individual rule by which a schedule will be constructed step by step. The conditional probabilities are computed according to an initial set of promising solutions. Subsequently, each new instance for each node is generated by using the corresponding conditional probabilities, until values for all nodes have been generated. Another set of rule strings will be generated in this way, some of which will replace previous strings based on fitness selection. If stopping conditions are not met, the Bayesian network is updated again using the current set of good rule strings. The algorithm thereby tries to explicitly identify and mix promising building blocks. It should be noted that for most scheduling problems the structure of the network model is known and all the variables are fully observed. In this case, the goal of learning is to find the rule values that maximize the likelihood of the training data. Thus learning can amount to 'counting' in the case of multinomial distributions. In the LCS approach, each rule has its strength showing its current usefulness in the system, and this strength is constantly assessed [4]. To implement sophisticated learning based on previous solutions, an improved LCS-based algorithm is designed, which consists of the following three steps. The initialization step is to assign each rule at each stage a constant initial strength. Then rules are selected by using the Roulette Wheel strategy. The next step is to reinforce the strengths of the rules used in the previous solution, keeping the strength of unused rules unchanged. The selection step is to select fitter rules for the next generation. It is envisaged that the LCS part of the algorithm will be used as a hill climber to the BOA algorithm. This is exciting and ambitious research, which might provide the stepping-stone for a new class of scheduling algorithms. Data sets from nurse scheduling and mall problems will be used as test-beds. It is envisaged that once the concept has been proven successful, it will be implemented into general scheduling algorithms. It is also hoped that this research will give some preliminary answers about how to include human-like learning into scheduling algorithms and may therefore be of interest to researchers and practitioners in areas of scheduling and evolutionary computation. References 1. Aickelin, U. and Dowsland, K. (2003) 'Indirect Genetic Algorithm for a Nurse Scheduling Problem', Computer & Operational Research (in print). 2. Li, J. and Kwan, R.S.K. (2003), 'Fuzzy Genetic Algorithm for Driver Scheduling', European Journal of Operational Research 147(2): 334-344. 3. Pelikan, M., Goldberg, D. and Cantu-Paz, E. (1999) 'BOA: The Bayesian Optimization Algorithm', IlliGAL Report No 99003, University of Illinois. 4. Wilson, S. (1994) 'ZCS: A Zeroth-level Classifier System', Evolutionary Computation 2(1), pp 1-18.
Resumo:
The research investigates the feasibility of using web-based project management systems for dredging. To achieve this objective the research assessed both the positive and negative aspects of using web-based technology for the management of dredging projects. Information gained from literature review and prior investigations of dredging projects revealed that project performance, social, political, technical, and business aspects of the organization were important factors in deciding to use web-based systems for the management of dredging projects. These factors were used to develop the research assumptions. An exploratory case study methodology was used to gather the empirical evidence and perform the analysis. An operational prototype of the system was developed to help evaluate developmental and functional requirements, as well as the influence on performance, and on the organization. The evidence gathered from three case study projects, and from a survey of 31 experts, were used to validate the assumptions. Baselines, representing the assumptions, were created as a reference to assess the responses and qualitative measures. The deviation of the responses was used to evaluate for the analysis. Finally, the conclusions were assessed by validating the assumptions with the evidence, derived from the analysis. The research findings are as follows: 1. The system would help improve project performance. 2. Resistance to implementation may be experienced if the system is implemented. Therefore, resistance to implementation needs to be investigated further and more R&D work is needed in order to advance to the final design and implementation. 3. System may be divided into standalone modules in order to simplify the system and facilitate incremental changes. 4. The QA/QC conceptual approach used by this research needs to be redefined during future R&D to satisfy both owners and contractors. Yin (2009) Case Study Research Design and Methods was used to develop the research approach, design, data collection, and analysis. Markus (1983) Resistance Theory was used during the assumptions definition to predict potential problems to the implementation of web-based project management systems for the dredging industry. Keen (1981) incremental changes and facilitative approach tactics were used as basis to classify solutions, and how to overcome resistance to implementation of the web-based project management system. Davis (1989) Technology Acceptance Model (TAM) was used to assess the solutions needed to overcome the resistances to the implementation of web-base management systems for dredging projects.
Resumo:
Abstract Scheduling problems are generally NP-hard combinatorial problems, and a lot of research has been done to solve these problems heuristically. However, most of the previous approaches are problem-specific and research into the development of a general scheduling algorithm is still in its infancy. Mimicking the natural evolutionary process of the survival of the fittest, Genetic Algorithms (GAs) have attracted much attention in solving difficult scheduling problems in recent years. Some obstacles exist when using GAs: there is no canonical mechanism to deal with constraints, which are commonly met in most real-world scheduling problems, and small changes to a solution are difficult. To overcome both difficulties, indirect approaches have been presented (in [1] and [2]) for nurse scheduling and driver scheduling, where GAs are used by mapping the solution space, and separate decoding routines then build solutions to the original problem. In our previous indirect GAs, learning is implicit and is restricted to the efficient adjustment of weights for a set of rules that are used to construct schedules. The major limitation of those approaches is that they learn in a non-human way: like most existing construction algorithms, once the best weight combination is found, the rules used in the construction process are fixed at each iteration. However, normally a long sequence of moves is needed to construct a schedule and using fixed rules at each move is thus unreasonable and not coherent with human learning processes. When a human scheduler is working, he normally builds a schedule step by step following a set of rules. After much practice, the scheduler gradually masters the knowledge of which solution parts go well with others. He can identify good parts and is aware of the solution quality even if the scheduling process is not completed yet, thus having the ability to finish a schedule by using flexible, rather than fixed, rules. In this research we intend to design more human-like scheduling algorithms, by using ideas derived from Bayesian Optimization Algorithms (BOA) and Learning Classifier Systems (LCS) to implement explicit learning from past solutions. BOA can be applied to learn to identify good partial solutions and to complete them by building a Bayesian network of the joint distribution of solutions [3]. A Bayesian network is a directed acyclic graph with each node corresponding to one variable, and each variable corresponding to individual rule by which a schedule will be constructed step by step. The conditional probabilities are computed according to an initial set of promising solutions. Subsequently, each new instance for each node is generated by using the corresponding conditional probabilities, until values for all nodes have been generated. Another set of rule strings will be generated in this way, some of which will replace previous strings based on fitness selection. If stopping conditions are not met, the Bayesian network is updated again using the current set of good rule strings. The algorithm thereby tries to explicitly identify and mix promising building blocks. It should be noted that for most scheduling problems the structure of the network model is known and all the variables are fully observed. In this case, the goal of learning is to find the rule values that maximize the likelihood of the training data. Thus learning can amount to 'counting' in the case of multinomial distributions. In the LCS approach, each rule has its strength showing its current usefulness in the system, and this strength is constantly assessed [4]. To implement sophisticated learning based on previous solutions, an improved LCS-based algorithm is designed, which consists of the following three steps. The initialization step is to assign each rule at each stage a constant initial strength. Then rules are selected by using the Roulette Wheel strategy. The next step is to reinforce the strengths of the rules used in the previous solution, keeping the strength of unused rules unchanged. The selection step is to select fitter rules for the next generation. It is envisaged that the LCS part of the algorithm will be used as a hill climber to the BOA algorithm. This is exciting and ambitious research, which might provide the stepping-stone for a new class of scheduling algorithms. Data sets from nurse scheduling and mall problems will be used as test-beds. It is envisaged that once the concept has been proven successful, it will be implemented into general scheduling algorithms. It is also hoped that this research will give some preliminary answers about how to include human-like learning into scheduling algorithms and may therefore be of interest to researchers and practitioners in areas of scheduling and evolutionary computation. References 1. Aickelin, U. and Dowsland, K. (2003) 'Indirect Genetic Algorithm for a Nurse Scheduling Problem', Computer & Operational Research (in print). 2. Li, J. and Kwan, R.S.K. (2003), 'Fuzzy Genetic Algorithm for Driver Scheduling', European Journal of Operational Research 147(2): 334-344. 3. Pelikan, M., Goldberg, D. and Cantu-Paz, E. (1999) 'BOA: The Bayesian Optimization Algorithm', IlliGAL Report No 99003, University of Illinois. 4. Wilson, S. (1994) 'ZCS: A Zeroth-level Classifier System', Evolutionary Computation 2(1), pp 1-18.
Resumo:
Diversity has become a buzz word in public discourse and in educational circles. Higher education institutions in the US have increasingly used this word as a cornerstone of their mission statements and have made increasing efforts to attract students from different backgrounds. As part of the increase in diversity efforts among US colleges, is a significant rise in the number of international students. Attracting international students has become a priority for U.S. universities regardless of size or location. This study examines the intersection between the structure of American educational environment and the blended identities of African Graduate Student Mothers. Within the context of contemporary diversity efforts in US educational institutions, this study examines both the structural environments and the socio-cultural constructs that affect the experiences of African graduate student mothers. Based on a qualitative research interview design, a total of nineteen African graduate student mothers at a Mid-Western University in the US were interviewed individually and in groups over a six weeks period. Results from this study show that apart from the difficult and often dehumanizing treatment African student mothers endure from immigration and consular officials in their various countries and ports of entry, they often find themselves at the margins of their various programs and departments with very little support if any. This is because most of them enroll into graduate programs after arriving as dependants of their spouses; a process that does not allow them to negotiate for departmental commitments and support prior to their arrival. Not only do these women face racial discrimination from white professors, staff and fellow students, but they also experience discrimination and hostilities from African Americans and other minority groups who see them as threats to the limited resources that are often set aside for minority groups in such institutions.
Resumo:
Background: An extensive research literature has documented the impact of caring for an individual with acquired brain injury (ABI) on caregivers and family members, including role adjustment, psychological distress, social isolation, family tension and coping with the cognitive and behavioural difficulties of the injured person. Given these findings it is important this population have access to services and supports. Acceptance and Commitment Therapy (ACT) is an intervention that helps individuals to accept difficult experiences and commit to behaviour that is consistent with their values. Research into the effectiveness of ACT to support caregivers is at a preliminary stage. Aim: To investigate the feasibility of using ACT to reduce psychological distress and increase psychological flexibility in ABI caregivers. A secondary aim was to gain an understanding of the experience of caregivers in this context and how this can inform the development and delivery of interventions for this population. Method: Phase one was a randomised controlled feasibility trial of an ACT intervention for use with ABI caregivers. The parameters of this study were formulated around the PICO (population, intervention, control, and outcome) framework. Eighteen carers were recruited and randomised to ACT or an enhanced treatment as usual (ETAU) group. ACT was implemented over 3 sessions; and ETAU was implemented over 2 sessions. The General Health Questionnaire, Valuing Questionnaire, Acceptance and Action Questionnaire, Experiential Avoidance of Caregiving Questionnaire and the Flexibility of Responses to Self-Critical Thoughts Scale were administered to both groups at baseline and following the final session. Phase two used a retrospective qualitative design that involved conducting semi-structured interviews with four participants from phase one. Results: ACT and control participants were successfully recruited. Positive feedback was obtained from ACT participants suggesting that the intervention was acceptable. There were no significant differences between the ACT and ETAU groups on outcome measures. However, there were challenges retaining participants and the overall attrition rate was high (44.44%). Therefore a number of participants did not complete the full complement of sessions, which may have impacted on this result. Qualitative results illustrated the challenges this population face including significant adjustments in their life, the emotional impact of having a loved one with a brain injury and trying to adapt to the changes in the injured person. In addition, findings elucidated the types of support that this population would find helpful and the barriers to accessing same. Conclusions: Findings from this study highlight factors that will help the development of this intervention further for a caring population. Recommendations for future implementation include completing some preparatory work with carers before beginning the intervention, consideration of a larger sample and wider recruitment strategy from local services, barriers to attending interventions and the possibility of holding groups in local venues.
Resumo:
La notion de chez-soi est porteuse d’un imaginaire foisonnant et génère un grand intérêt dans notre culture et société. Bien qu’elle soit une considération importante pour la plupart d’entre nous, l’architecte occupe une position privilégiée qui lui permette d’agir sur le chez-soi de manière significative et tangible. La présente recherche explore le concept du chez-soi tel qu’étudié par les architectes et non-architectes de manière à comprendre son impact sur la création du lieu et sur la construction des environnements domestiques en Amérique du nord. Un regard porté sur les connotations entre espace et lieu, à travers la temporalité, les comportements et les perspectives sociales, supporte l’épistémologie du chez-soi comme un élément important des théories et pratiques de design contemporain. Le démantèlement hypothétique d’un espace en ses composantes suppose que les dispositifs architecturaux peuvent être modelés de manière à ce qu’ils opèrent un transfert depuis la maison, en tant qu’élément physique, vers le domaine psychologique du chez-soi. Afin d’élargir la maniabilité des éléments constitutifs du bâtiment et de son environnement, six thèmes sont appliqués à trois champs de données. Les six thèmes, qui incluent l’entre-deux, la limite, la voie, le nœud, le détail et la représentation, illustrent des moments architecturaux déterminants, potentiellement présents à l’intérieur et à l’extérieur du projet domestique et qui transforment les comportements physiques et psychiques. Depuis la pratique normalisée du logement social et abordable au Canada, une analyse de photographies de maisons abordables existantes, du discours critique sur cette typologie et de projets de recherche-création conduits par des étudiants en architecture, révèle le caractère opérationnel de la notion de chez-soi et consolide les valeurs de communauté et de frontière. L’objectif premier de la recherche est d’avancer la production de connaissances en architecture par l’exploration de la notion de chezsoi dans l’enseignement, la recherche et le design. L’approche fonctionnaliste vis-à-vis le < penser > en design, place l’usager au centre de l’environnement domestique, soutient la proposition que le chezsoi donne sens et utilité au logement, et renforce la responsabilité éthique de l’architecte à faire de cette notion une partie intégrante de la réalité quotidienne.
Resumo:
High quality, well designed medical devices are necessary to provide safe and effective clinical care for patients as well as to ensure the health and safety of professional and lay device users. Capturing the user requirements of users and incorporating these into design is an essential component of this. The field of ergonomics has an opportunity to assist, not only with this area, but also to encourage a more general consideration of the user during medical device development. A review of the literature on methods for assessing user requirements in engineering and ergonomics found that little published work exists on the ergonomics aspects of medical device development. In particular there is little advice available to developers on which issues to consider during design and development or recommendations for good practice in terms of the methods and approaches needed to capture the full range of user requirements. The Multidisciplinary Assessment of Technology Centre for Healthcare (MATCH) is a research collaboration that is working in conjunction with industrial collaborators to apply ergonomics methods to real case study projects with the ultimate aim of producing an industry-focused guide to applying ergonomics principles in medical device development.
Resumo:
Crossing the Franco-Swiss border, the Large Hadron Collider (LHC), designed to collide 7 TeV proton beams, is the world's largest and most powerful particle accelerator the operation of which was originally intended to commence in 2008. Unfortunately, due to an interconnect discontinuity in one of the main dipole circuit's 13 kA superconducting busbars, a catastrophic quench event occurred during initial magnet training, causing significant physical system damage. Furthermore, investigation into the cause found that such discontinuities were not only present in the circuit in question, but throughout the entire LHC. This prevented further magnet training and ultimately resulted in the maximum sustainable beam energy being limited to approximately half that of the design nominal, 3.5-4 TeV, for the first three years of operation (Run 1, 2009-2012) and a major consolidation campaign being scheduled for the first long shutdown (LS 1, 2012-2014). Throughout Run 1, a series of studies attempted to predict the amount of post-installation training quenches still required to qualify each circuit to nominal-energy current levels. With predictions in excess of 80 quenches (each having a recovery time of 8-12+ hours) just to achieve 6.5 TeV and close to 1000 quenches for 7 TeV, it was decided that for Run 2, all systems be at least qualified for 6.5 TeV operation. However, even with all interconnect discontinuities scheduled to be repaired during LS 1, numerous other concerns regarding circuit stability arose. In particular, observations of an erratic behaviour of magnet bypass diodes and the degradation of other potentially weak busbar sections, as well as observations of seemingly random millisecond spikes in beam losses, known as unidentified falling object (UFO) events, which, if persist at 6.5 TeV, may eventually deposit sufficient energy to quench adjacent magnets. In light of the above, the thesis hypothesis states that, even with the observed issues, the LHC main dipole circuits can safely support and sustain near-nominal proton beam energies of at least 6.5 TeV. Research into minimising the risk of magnet training led to the development and implementation of a new qualification method, capable of providing conclusive evidence that all aspects of all circuits, other than the magnets and their internal joints, can safely withstand a quench event at near-nominal current levels, allowing for magnet training to be carried out both systematically and without risk. This method has become known as the Copper Stabiliser Continuity Measurement (CSCM). Results were a success, with all circuits eventually being subject to a full current decay from 6.5 TeV equivalent current levels, with no measurable damage occurring. Research into UFO events led to the development of a numerical model capable of simulating typical UFO events, reproducing entire Run 1 measured event data sets and extrapolating to 6.5 TeV, predicting the likelihood of UFO-induced magnet quenches. Results provided interesting insights into the involved phenomena as well as confirming the possibility of UFO-induced magnet quenches. The model was also capable of predicting that such events, if left unaccounted for, are likely to be commonplace or not, resulting in significant long-term issues for 6.5+ TeV operation. Addressing the thesis hypothesis, the following written works detail the development and results of all CSCM qualification tests and subsequent magnet training as well as the development and simulation results of both 4 TeV and 6.5 TeV UFO event modelling. The thesis concludes, post-LS 1, with the LHC successfully sustaining 6.5 TeV proton beams, but with UFO events, as predicted, resulting in otherwise uninitiated magnet quenches and being at the forefront of system availability issues.
Resumo:
Dissertação para obtenção do grau de Doutor em Design, apresentada na Universidade de Lisboa - Faculdade de Arquitetura.
Resumo:
Objective: The burden of sexually transmitted infections (STIs) rests with young people, yet in Ireland there has been very little research into this population. The purpose of this study was to determine the incidence rate and establish risk factors that predict STI occurrence among adolescents in Ireland. Design: Routine diagnostic, demographic and behavioural data from first-time visits to three screening centres in the southwest of Ireland were obtained. Univariate and multivariable logistic regression models were used to assess risk factors that predict STI occurrence among adolescents. Results: A total of 2784 first-time patients, aged 13–19 years, received 3475 diagnoses between January 1999 and September 2009; 1168 (42%) of adolescents had notifiable STIs. The incidence rate of STIs is 225/100 000 person-years. Univariate analysis identified eligible risk factors (p<0.2) for inclusion in the multivariable model. Multivariable logistic regression showed the dominant risk factors for STI diagnosis to be: males who sometimes [odds ratio (OR) 2.02] or never (OR 1.83) use condoms; and females 18–19 years (OR 2.26) and 16–18 years (OR 1.8), with 2 (OR 1.33) or 3+ (OR 1.56) partners in the last 12 months, who are non-intravenous drug users (OR 0.72), are most likely to receive a positive STI diagnosis. Conclusions: STI diagnosis has become increasingly common in Ireland. The proportion of notifications among those aged under 20 years is increasing. These data illustrate the significance of age, condom use and number of sexual partners as risk factors for STI diagnosis. Furthermore, providing data for the first time, we report on the high incidence rate of STIs among adolescents in Ireland. The high levels of risk-taking behaviour and STI acquisition are highlighted and suggest that there is a need for an integrated public health approach to combat this phenomenon in the adolescent population.
Resumo:
The relationship between school belongingness and mental health functioning before and after the primary-secondary school transition has not been previously investigated in students with and without disabilities. This study used a prospective longitudinal design to test the bi-directional relationships between these constructs, by surveying 266 students with and without disabilities and their parents, 6-months before and after the transition to secondary school. Cross-lagged multi-group analyses found student perception of belongingness in the final year of primary school to contribute to change in their mental health functioning a year later. The beneficial longitudinal effects of school belongingness on subsequent mental health functioning were evident in all student subgroups; even after accounting for prior mental health scores and the cross-time stability in mental health functioning and school belongingness scores. Findings of the current study substantiate the role of school contextual influences on early adolescent mental health functioning. They highlight the importance for primary and secondary schools to assess students' school belongingness and mental health functioning and transfer these records as part of the transition process, so that appropriate scaffolds are in place to support those in need. Longer term longitudinal studies are needed to increase the understanding of the temporal sequencing between school belongingness and mental health functioning of all mainstream students.
Resumo:
Students negotiate the transition to secondary school in different ways. While some thrive on the opportunity, others are challenged. A prospective longitudinal design was used to determine the contribution of personal background and school contextual factors on academic competence (AC) and mental health functioning (MHF) of 266 students, 6-months before and after the transition to secondary school. Data from 197 typically developing students and 69 students with a disability were analysed using hierarchical linear regression modelling. Both in primary and secondary school, students with a disability and from socially disadvantaged backgrounds gained poorer scores for AC and MHF than their typically developing and more affluent counterparts. Students who attended independent and mid-range sized primary schools had the highest concurrent AC. Those from independent primary schools had the lowest MHF. The primary school organisational model significantly influenced post-transition AC scores; with students from Kindergarten--Year 7 schools reporting the lowest scores, while those from the Kindergarten--Year 12 structure without middle school having the highest scores. Attending a school which used the Kindergarten--Year 12 with middle school structure was associated with a reduction in AC scores across the transition. Personal background factors accounted for the majority of the variability in post-transition AC and MHF. The contribution of school contextual factors was relatively minor. There is a potential opportunity for schools to provide support to disadvantaged students before the transition to secondary school, as they continue to be at a disadvantage after the transition.
Resumo:
Background: After a myocardial infarction and discharge from the hospital a recovery process follows for the women. In order to facilitate their recovery, both a preventive and promotive perspective should be taken into consideration. Despite this, today´s healthcare focuses more on prevention and thus research into the promotion of women’s recovery process is needed. Aim: To explore how women’s recovery processes are promoted after a first myocardial infarction Methods: The study had an explorative and descriptive design based on qualitative content analysis. Findings: The women’s recovery process was promoted through using external and internal resources as well as embracing behaviour, social and psychological dimensions. The women embraced these dimensions to a varying extent and this process led to them being able to take in a new perspective on life. Conclusions: The women’s personal recovery is a multidirectional process with a desire to develop and approach a new perspective on life. It is important for cardiac rehabilitation nurses to not only focus on lifestyle changes and social support, but also on working actively with the women’s inner strength in order to promote the personal recovery of the women. Furthermore, it would be interesting to investigate men’s experiences of how their recovery process after a first Myocardial infarction is promoted.
Resumo:
An employee's inability to balance work and family responsibilities has resulted in an increase in stress related illnesses. Historically, research into the nexus between work and family has primarily focused on the work/family conflict relationship, predominately investigating the impact of this conflict on parents, usually mothers. To date research has not sufficiently examined the human resource management practices that enable all parents to achieve a balance between their work and family lives. This paper explores the relationship between contemporary family friendly HRM policies and employed parents perceptions of work/family enhancement, work/family satisfaction, propensity to turnover, and work/family conflict. Self-report questionnaire data from 326 men and women is analysed and discussed to enable organisations to consider the use of family friendly policies and thus create a convergence between the well-being of employees and the effectiveness of the organisation.