31 resultados para Organisation-Level Performance, Business Value,

em Helda - Digital Repository of University of Helsinki


Relevância:

100.00% 100.00%

Publicador:

Resumo:

ERP system implementations have evolved so rapidly that now they represent a must-have within industries. ERP systems are viewed as the cost of doing business. Yet, the research that adopted the resource-based view on the business value of ERP systems concludes that companies may gain competitive advantage when they successfully manage their ERP projects, when they carefully reengineer the organization and when they use the system in line with the organizational strategies. This thesis contributes to the literature on ERP business value by examining key drivers of ERP business value in organizations. The first research paper investigates how ERP systems with different degrees of system functionality are correlated with the development of the business performance after the completion of the ERP projects. The companies with a better perceived system functionality obtained efficiency benefits in the first two years of post-implementation. However, in the third year there is no significant difference in efficiency benefits between successfully and less successfully managed ERP projects. The second research paper examines what business process changes occur in companies implementing ERP for different motivations and how these changes impact the business performance. The findings show that companies reported process changes mainly in terms of workflow changes. In addition, the companies having a business-led motivation focused more on observing average costs of each increase in the input unit. Companies having a technological-led motivation focused more on the benefits coming from the fit of the system with the organizational processes. The third research paper considers the role of alignment between ERP and business strategies for the realization of business value from ERP use. These findings show that strategic alignment and business process changes are significantly correlated with the perceived benefits of ERP at three levels: internal efficiency, customers and financial. Overall, by combining quantitative and qualitative research methods, this thesis puts forward a model that illustrates how successfully managed ERP projects, aligned with the business strategy, have automate and informate effects on processes that ultimately improve the customer service and reduce the companies’ costs.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Most of the existing research within the business network approach is based on companies that are operating on different levels within the same value chain, as a buyer and a supplier. Intercompetitor cooperation, i.e. cooperation between companies occupying the same level within different value chains, has not been studied to the same extent. Moreover scholars within the business network approach have usually described industrial relationships as long term, consisting of mutual commitment and trust. Industrial relationships are not static, but dynamic, and they contain situations of both harmony and conflict. There is consequently a need for more research both concerning intercompetitor cooperation and conflicts. The purpose of this study is to develop our theoretical and empirical understanding of the nature of conflicts in intercompetitor cooperation from a business network perspective. The focus of the study lies on issue and intensity of conflict. The issue of a conflict can be divided into cause and topic, while the intensity comprises the importance and outcome of a conflict. The empirical part of the study is based on two case studies of groups of cooperating competitors from two different industries. The applied research method is interviews. According to the findings of this study causes of conflicts in intercompetitor cooperation can be divided into three groups: focus, awareness and capacity. Topics of conflict can be related to domain, delivery, advertising or cooperation. Moreover the findings show that conflict situations may be grouped into not important, important or very important. Some conflicts may also be of varying importance, meaning that the importance varies from one point of time to another. Based on the findings of the study the outcome or status of a conflict can be analyzed both on a concrete and general level. The findings also indicate that several conflicts are partly hidden, which means that only one or some of the involved actors perceive the conflict. Furthermore several conflict situations can be related to external network actors.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The driving force behind this study is the gap between the reality of the firms engaged in project business and the available studies covering project management and business process development. Previous studies show that project-based organizations were ‘immature’ in terms of the project-management ‘maturity model’, as few firms were found to be optimizing processes. Even within those, very little attention was paid to combine inter-organizational and intra-organizational perspectives. In this study an effort is made to elaborate some thoughts and views on project management, which interrelate firms’ external and internal activities. In line with the integration, the dissertation uses an approach to the management of project-business interdependencies in the networks of actors, activities and resources. Firstly, the study develops an understanding for inter-organizational perspectives by exploring the complementarities of process activities in the basic development of project business. It presents a framework that is elaborated on the basis of the reciprocal interactions of activities within and outside the organization—thus providing a coherent basis for continuous business-process improvement. In addition, the study presents new tools that can be used to develop project-business processes in each of its functional areas. The research demonstrates how project-business activities can be optimized using the right resources at the right time with the right actors and the right actions. The selected five articles included in this dissertation explain the basic framework for the development of project business. Each paper covers various aspects of inter-organizational and intra-organizational perspectives for project management. The study develops a valuable and procedural model for business-process improvement using the Delphi method that can be used not only in academia but also as a guide for practitioners that takes them through a series of well-defined steps when making informed, consistent and efficient changes to their business processes.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The forest simulator is a computerized model for predicting forest growth and future development as well as effects of forest harvests and treatments. The forest planning system is a decision support tool, usually including a forest simulator and an optimisation model, for finding the optimal forest management actions. The information produced by forest simulators and forest planning systems is used for various analytical purposes and in support of decision making. However, the quality and reliability of this information can often be questioned. Natural variation in forest growth and estimation errors in forest inventory, among other things, cause uncertainty in predictions of forest growth and development. This uncertainty stemming from different sources has various undesirable effects. In many cases outcomes of decisions based on uncertain information are something else than desired. The objective of this thesis was to study various sources of uncertainty and their effects in forest simulators and forest planning systems. The study focused on three notable sources of uncertainty: errors in forest growth predictions, errors in forest inventory data, and stochastic fluctuation of timber assortment prices. Effects of uncertainty were studied using two types of forest growth models, individual tree-level models and stand-level models, and with various error simulation methods. New method for simulating more realistic forest inventory errors was introduced and tested. Also, three notable sources of uncertainty were combined and their joint effects on stand-level net present value estimates were simulated. According to the results, the various sources of uncertainty can have distinct effects in different forest growth simulators. The new forest inventory error simulation method proved to produce more realistic errors. The analysis on the joint effects of various sources of uncertainty provided interesting knowledge about uncertainty in forest simulators.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Naked oat (Avena sativa f.sp. nuda L.) is the highest quality cereal in northern growing conditions. However the cultivation area of naked oat is remarkably small. Major challenges for naked oat production are to observe its nakedness. The caryopsis of naked oat is sensitive to mechanical damage at harvest, especially at high grain moisture content. The greater the grain moisture content of naked oat at harvest, the more loses of germination capacity was caused by threshing. For producing high quality naked oat seed, it is recommended that harvesting be done at as low grain moisture content as possible. However, if this is not possible, better germination can be ensure with gentle harvest by reducing the cylinder speed. In spite of conventional oat s excellent fat and amino acid composition in animal feed use, as far as nutritional value is concerned, the total energy yield of oat is weaker than other cereals because of the hulls. Also with naked oat the dehulling is not complete, while hull content on different cultivars mostly varied between one to six percent. In addition to genotype, environmental conditions markedly control the expression of nakedness. Thresher settings had only limited effects on hull content. The function of hulls is to protect the groat, but this was confirmed only for Finnish, small grain, cultivar Lisbeth. The oat kernel is generally covered with fine silky hairs termed trichomes. The trichomes of naked oat are partly lost during threshing and handling of grains. Trichomes can cause itchiness in those handling the grains and also accumulate and form fine dust and can block-up machinery. The cultivars differed considerably in pubescence. Some thresher settings, including increased cylinder speed, slightly increased grain polishing such that grains had some areas completely free of trichomes. Adjusting thresher settings was generally not an efficient means of solving the problems associated with naked oat trichomes. The main differences in cultivation costs between naked and conventional oat lie in the amount of seeds required and the drying costs. The main differences affecting the economic result lie in market prices, yield level and feed value. The results indicate that naked oat is financially more profitable than conventional oat, when the crop is sold at a specific price at all yield levels and when the crop is used as feed at highest yield level. At lower yield levels, conventional oat is, in spite of its lower feed value, the more profitable option for feed use. Dehulled oat did not achieve the same economic result as naked oat, as the cost of dehulling, including the hull waste, was considerable. According to this study naked oat can be cultivated successfully under northern conditions, when taking into consideration the soft, naked grain through cultivation chain.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The aim of this thesis is to analyse the key ecumenical dialogues between Methodists and Lutherans from the perspective of Arminian soteriology and Methodist theology in general. The primary research question is defined as: "To what extent do the dialogues under analysis relate to Arminian soteriology?" By seeking an answer to this question, new knowledge is sought on the current soteriological position of the Methodist-Lutheran dialogues, the contemporary Methodist theology and the commonalities between the Lutheran and Arminian understanding of soteriology. This way the soteriological picture of the Methodist-Lutheran discussions is clarified. The dialogues under analysis were selected on the basis of versatility. Firstly, the sole world organisation level dialogue was chosen: The Church – Community of Grace. Additionally, the document World Methodist Council and the Joint Declaration on the Doctrine of Justification is analysed as a supporting document. Secondly, a document concerning the discussions between two main-line churches in the United States of America was selected: Confessing Our Faith Together. Thirdly, two dialogues between non-main-line Methodist churches and main-line Lutheran national churches in Europe were chosen: Fellowship of Grace from Norway and Kristuksesta osalliset from Finland. The theoretical approach to the research conducted in this thesis is systematic analysis. The Remonstrant articles of Arminian soteriology are utilised as an analysis tool to examine the soteriological positions of the dialogues. New knowledge is sought by analysing the stances of the dialogues concerning the doctrines of partial depravity, conditional election, universal atonement, resistible grace and conditional perseverance of saints. This way information is also provided for approaching the Calvinist-Arminian controversy from new perspectives. The results of this thesis show that the current soteriological position of the Methodist-Lutheran dialogues is closer to Arminianism than Calvinism. The dialogues relate to Arminian soteriology especially concerning the doctrines of universal atonement, resistible grace and conditional perseverance of saints. The commonalities between the Lutheran and Arminian understanding of soteriology exist mainly in these three doctrines as they are uniformly favoured in the dialogues. The most discussed area of soteriology is human depravity, in which the largest diversity of stances occurs as well. On the other hand, divine election is the least discussed topic. The overall perspective, which the results of the analysis provide, indicates that the Lutherans could approach the Calvinist churches together with the Methodists with a wider theological perspective and understanding when the soteriological issues are considered as principal. Human depravity is discovered as the area of soteriology which requires most work in future ecumenical dialogues. However, the detected Lutheran hybrid notion on depravity (a Calvinist-Arminian mixture) appears to provide a useful new perspective for Calvinist-Arminian ecumenism and offers potentially fruitful considerations to future ecumenical dialogues.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Purpose – This research paper studies how the strategy of repositioning enables marketers to communicate CSR as their brand’s differentiating factor. It aims at understanding how consumer perceptions can be managed to generate brand value through corporate brand repositioning when CSR is the differentiating factor. The purpose of this paper is to answer the following research question: How can consumer perceptions be managed to generate brand value through corporate brand repositioning when CSR is the differentiating factor? The two research objectives were: 1. to build a model, which describes the different components of consumer perceptions involved in generation of brand value through repositioning when CSR is the differentiating factor, 2. to identify the most critical components in the context of the case company, IKEA for generation of brand value during the process of corporate brand repositioning Design/methodology/approach – This paper is based on the literature review covering the logic of brand value generation, repositioning strategy and consumer perceptions connected to CSR activities. A key concept of the positioning theory, the brand’s differentiating factor, was explored. Previous studies have concluded that desirability of the differentiating factor largely determines the level of brand value-creation for the target customers. The criterion of desirability is based on three dimensions: relevance, distinctiveness and believability. A model was built in terms of these desirability dimensions. This paper takes a case study approach where the predefined theoretical framework is tested using IKEA as the case company. When developing insights on the multifaceted nature of brand perceptions, personal interviews and individual probing are vital. They enable the interviewees to reflect on their feelings and perceptions with their own words. This is why the data collection was based on means-end type of questioning. Qualitative interviews were conducted with 12 consumers. Findings – The paper highlights five critical components that may determine whether IKEA will fail in its repositioning efforts. The majority of the critical components involved believability perceptions. Hence, according to the findings, establishing credibility and trustworthiness for the brand in the context of CSR seems primary. The most critical components identified of the believability aspect were: providing proof of responsible codes of conduct via conducting specific and concrete CSR actions, connecting the company’s products and the social cause, and building a linkage between the initial and new positioning while also weakening the old positioning. Originality/value – Marketers’ obligation is to prepare the company for future demands. Companies all over the globe have recognized the durable trend of responsibility and sustainability. Consumer´s worry about the environmental and social impact of modern lifestyles is growing. This is why Corporate Social Responsibility (CSR) provides brands an important source of differentiation and strength in the future. The strategy of repositioning enables marketers to communicate CSR as their brand’s differentiating factor. This study aimed at understanding how consumer perceptions can be managed to generate brand value through corporate brand repositioning when CSR is the differentiating factor.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

Suvi Nenonen Customer asset management in action: using customer portfolios for allocating resources across business-to-business relationships for improved shareholder value Customers are crucial assets to all firms as customers are the ultimate source of all cash flows. Regardless this financial importance of customer relationships, for decades there has been a lack of suitable frameworks explaining how customer relationships contribute to the firm financial performance and how this contribution can be actively managed. In order to facilitate a better understanding of the customer asset, contemporary marketing has investigated the use of financial theories and asset management practices in the customer relationship context. Building on this, marketing academics have promoted the customer lifetime value concept as a solution for valuating and managing customer relationships for optimal financial outcomes. However, the empirical investigation of customer asset management lags behind the conceptual development steps taken. Additionally, the practitioners have not embraced the use of customer lifetime value in guiding managerial decisions - especially in the business-to-business context. The thesis points out that there are fundamental differences between customer relationships and investment instruments as investment targets, effectively eliminating the possibility to use financial theories in a customer relationships context or to optimize the customer base as a single investment portfolio. As an alternative, the thesis proposes the use of customer portfolio approach for allocating resources across the customer base for improved shareholder value. In the customer portfolio approach, the customer base of a firm is divided into multiple portfolios based on customer relationships’ potential to contribute to the shareholder value creation. After this, customer management concepts are tailored to each customer portfolio, designed to improve the shareholder value in their own respect. Therefore, effective customer asset management with the customer portfolio approach necessitates that firms are able to manage multiple parallel customer management concepts, or business models, simultaneously. The thesis is one of the first empirical studies on customer asset management, bringing empirical evidence from multiple business-to-business case studies on how customer portfolio models can be formed, how customer portfolios can be managed, and how customer asset management has contributed to the firm financial performance.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

Suvi Nenonen Customer asset management in action: using customer portfolios for allocating resources across business-to-business relationships for improved shareholder value Customers are crucial assets to all firms as customers are the ultimate source of all cash flows. Regardless this financial importance of customer relationships, for decades there has been a lack of suitable frameworks explaining how customer relationships contribute to the firm financial performance and how this contribution can be actively managed. In order to facilitate a better understanding of the customer asset, contemporary marketing has investigated the use of financial theories and asset management practices in the customer relationship context. Building on this, marketing academics have promoted the customer lifetime value concept as a solution for valuating and managing customer relationships for optimal financial outcomes. However, the empirical investigation of customer asset management lags behind the conceptual development steps taken. Additionally, the practitioners have not embraced the use of customer lifetime value in guiding managerial decisions - especially in the business-to-business context. The thesis points out that there are fundamental differences between customer relationships and investment instruments as investment targets, effectively eliminating the possibility to use financial theories in a customer relationships context or to optimize the customer base as a single investment portfolio. As an alternative, the thesis proposes the use of customer portfolio approach for allocating resources across the customer base for improved shareholder value. In the customer portfolio approach, the customer base of a firm is divided into multiple portfolios based on customer relationships’ potential to contribute to the shareholder value creation. After this, customer management concepts are tailored to each customer portfolio, designed to improve the shareholder value in their own respect. Therefore, effective customer asset management with the customer portfolio approach necessitates that firms are able to manage multiple parallel customer management concepts, or business models, simultaneously. The thesis is one of the first empirical studies on customer asset management, bringing empirical evidence from multiple business-to-business case studies on how customer portfolio models can be formed, how customer portfolios can be managed, and how customer asset management has contributed to the firm financial performance.

Relevância:

50.00% 50.00%

Publicador:

Resumo:

This paper examines the association between corporate governance attributes and firm performance of Finnish firms during 1990 – 2000. The empirical results suggest that corporate governance matters for firm performance. First, univariate test results indicate that firms characterized by a high (efficient) level of corporate governance have delivered greater stock returns, are higher valued based on the measure of Tobin’s Q, and exhibit higher ratios of cash flow to assets, on average, in comparison to their counterparts characterized by a low (inefficient) level of corporate governance. Second, controlling for a number of well-known determinants of stock returns, we find evidence that firms categorized by inefficient corporate governance have delivered inferior returns to shareholders during the investigation period. Finally, after controlling for several common determinants of firm value, we find that firms characterized by efficient corporate governance have been valued higher during the investigation period, measured by Tobin’s Q.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

A randomised and population-based screening design with new technologies has been applied to the organised cervical cancer screening programme in Finland. In this experiment the women invited to routine five-yearly screening are individually randomised to be screened with automation-assisted cytology, human papillomavirus (HPV) test or conventional cytology. By using the randomised design, the ultimate aim is to assess and compare the long-term outcomes of the different screening regimens. The primary aim of the current study was to evaluate, based on the material collected during the implementation phase of the Finnish randomised screening experiment, the cross-sectional performance and validity of automation-assisted cytology (Papnet system) and primary HPV DNA testing (Hybrid Capture II assay for 13 oncogenic HPV types) within service screening, in comparison to conventional cytology. The parameters of interest were test positivity rate, histological detection rate, relative sensitivity, relative specificity and positive predictive value. Also, the effect of variation in performance by screening laboratory on age-adjusted cervical cancer incidence was assessed. Based on the cross-sectional results, almost no differences were observed in the performance of conventional and automation-assisted screening. Instead, primary HPV screening found 58% (95% confidence interval 19-109%) more cervical lesions than conventional screening. However, this was mainly due to overrepresentation of mild- and moderate-grade lesions and, thus, is likely to result in overtreatment since a great deal of these lesions would never progress to invasive cancer. Primary screening with an HPV DNA test alone caused substantial loss in specificity in comparison to cytological screening. With the use of cytology triage test, the specificity of HPV screening improved close to the level of conventional cytology. The specificity of primary HPV screening was also increased by increasing the test positivity cutoff from the level recommended for clinical use, but the increase was more modest than the one gained with the use of cytology triage. The performance of the cervical cancer screening programme varied widely between the screening laboratories, but the variation in overall programme effectiveness between respective populations was more marginal from the very beginning of the organised screening activity. Thus, conclusive interpretations on the quality or success of screening should not be based on performance parameters only. In the evaluation of cervical cancer screening the outcome should be selected as closely as possible to the true measure of programme effectiveness, which is the number of invasive cervical cancers and subsequent deaths prevented in the target population. The evaluation of benefits and adverse effects of each new suggested screening technology should be performed before the technology becomes an accepted routine in the existing screening programme. At best, the evaluation is performed randomised, within the population and screening programme in question, which makes the results directly applicable to routine use.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

The adequacy of anesthesia has been studied since the introduction of balanced general anesthesia. Commercial monitors based on electroencephalographic (EEG) signal analysis have been available for monitoring the hypnotic component of anesthesia from the beginning of the 1990s. Monitors measuring the depth of anesthesia assess the cortical function of the brain, and have gained acceptance during surgical anesthesia with most of the anesthetic agents used. However, due to frequent artifacts, they are considered unsuitable for monitoring consciousness in intensive care patients. The assessment of analgesia is one of the cornerstones of general anesthesia. Prolonged surgical stress may lead to increased morbidity and delayed postoperative recovery. However, no validated monitoring method is currently available for evaluating analgesia during general anesthesia. Awareness during anesthesia is caused by an inadequate level of hypnosis. This rare but severe complication of general anesthesia may lead to marked emotional stress and possibly posttraumatic stress disorder. In the present series of studies, the incidence of awareness and recall during outpatient anesthesia was evaluated and compared with that of in inpatient anesthesia. A total of 1500 outpatients and 2343 inpatients underwent a structured interview. Clear intraoperative recollections were rare the incidence being 0.07% in outpatients and 0.13% in inpatients. No significant differences emerged between outpatients and inpatients. However, significantly smaller doses of sevoflurane were administered to outpatients with awareness than those without recollections (p<0.05). EEG artifacts in 16 brain-dead organ donors were evaluated during organ harvest surgery in a prospective, open, nonselective study. The source of the frontotemporal biosignals in brain-dead subjects was studied, and the resistance of bispectral index (BIS) and Entropy to the signal artifacts was compared. The hypothesis was that in brain-dead subjects, most of the biosignals recorded from the forehead would consist of artifacts. The original EEG was recorded and State Entropy (SE), Response Entropy (RE), and BIS were calculated and monitored during solid organ harvest. SE differed from zero (inactive EEG) in 28%, RE in 29%, and BIS in 68% of the total recording time (p<0.0001 for all). The median values during the operation were SE 0.0, RE 0.0, and BIS 3.0. In four of the 16 organ donors, EEG was not inactive, and unphysiologically distributed, nonreactive rhythmic theta activity was present in the original EEG signal. After the results from subjects with persistent residual EEG activity were excluded, SE, RE, and BIS differed from zero in 17%, 18%, and 62% of the recorded time, respectively (p<0.0001 for all). Due to various artifacts, the highest readings in all indices were recorded without neuromuscular blockade. The main sources of artifacts were electrocauterization, electromyography (EMG), 50-Hz artifact, handling of the donor, ballistocardiography, and electrocardiography. In a prospective, randomized study of 26 patients, the ability of Surgical Stress Index (SSI) to differentiate patients with two clinically different analgesic levels during shoulder surgery was evaluated. SSI values were lower in patients with an interscalene brachial plexus block than in patients without an additional plexus block. In all patients, anesthesia was maintained with desflurane, the concentration of which was targeted to maintain SE at 50. Increased blood pressure or heart rate (HR), movement, and coughing were considered signs of intraoperative nociception and treated with alfentanil. Photoplethysmographic waveforms were collected from the contralateral arm to the operated side, and SSI was calculated offline. Two minutes after skin incision, SSI was not increased in the brachial plexus block group and was lower (38 ± 13) than in the control group (58 ± 13, p<0.005). Among the controls, one minute prior to alfentanil administration, SSI value was higher than during periods of adequate antinociception, 59 ± 11 vs. 39 ± 12 (p<0.01). The total cumulative need for alfentanil was higher in controls (2.7 ± 1.2 mg) than in the brachial plexus block group (1.6 ± 0.5 mg, p=0.008). Tetanic stimulation to the ulnar region of the hand increased SSI significantly only among patients with a brachial plexus block not covering the site of stimulation. Prognostic value of EEG-derived indices was evaluated and compared with Transcranial Doppler Ultrasonography (TCD), serum neuron-specific enolase (NSE) and S-100B after cardiac arrest. Thirty patients resuscitated from out-of-hospital arrest and treated with induced mild hypothermia for 24 h were included. Original EEG signal was recorded, and burst suppression ratio (BSR), RE, SE, and wavelet subband entropy (WSE) were calculated. Neurological outcome during the six-month period after arrest was assessed with the Glasgow-Pittsburgh Cerebral Performance Categories (CPC). Twenty patients had a CPC of 1-2, one patient had a CPC of 3, and nine patients died (CPC 5). BSR, RE, and SE differed between good (CPC 1-2) and poor (CPC 3-5) outcome groups (p=0.011, p=0.011, p=0.008, respectively) during the first 24 h after arrest. WSE was borderline higher in the good outcome group between 24 and 48 h after arrest (p=0.050). All patients with status epilepticus died, and their WSE values were lower (p=0.022). S-100B was lower in the good outcome group upon arrival at the intensive care unit (p=0.010). After hypothermia treatment, NSE and S-100B values were lower (p=0.002 for both) in the good outcome group. The pulsatile index was also lower in the good outcome group (p=0.004). In conclusion, the incidence of awareness in outpatient anesthesia did not differ from that in inpatient anesthesia. Outpatients are not at increased risk for intraoperative awareness relative to inpatients undergoing general anesthesia. SE, RE, and BIS showed non-zero values that normally indicate cortical neuronal function, but were in these subjects mostly due to artifacts after clinical brain death diagnosis. Entropy was more resistant to artifacts than BIS. During general anesthesia and surgery, SSI values were lower in patients with interscalene brachial plexus block covering the sites of nociceptive stimuli. In detecting nociceptive stimuli, SSI performed better than HR, blood pressure, or RE. BSR, RE, and SE differed between the good and poor neurological outcome groups during the first 24 h after cardiac arrest, and they may be an aid in differentiating patients with good neurological outcomes from those with poor outcomes after out-of-hospital cardiac arrest.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

As companies become more efficient with respect to their internal processes, they begin to shift the focus beyond their corporate boundaries. Thus, the recent years have witnessed an increased interest by practitioners and researchers in interorganizational collaboration, which promises better firm performance through more effective supply chain management. It is no coincidence that this interest comes in parallel with the recent advancements in Information and Communication Technologies, which offer many new collaboration possibilities for companies. However, collaboration, or any other type of supply chain integration effort, relies heavily on information sharing. Hence, this study focuses on information sharing, in particular on the factors that determine it and on its value. The empirical evidence from Finnish and Swedish companies suggests that uncertainty (both demand and environmental) and dependency in terms of switching costs and asset specific investments are significant determinants of information sharing. Results also indicate that information sharing improves company performance regarding resource usage, output, and flexibility. However, companies share information more intensely at the operational rather than the strategic level. The use of supply chain practices and technologies is substantial but varies across the two countries. This study sheds light on a common trend in supply chains today. Whereas the results confirm the value of information sharing, the contingent factors help to explain why the intensity of information shared across companies differ. In the future, competitive pressures and uncertainty are likely to intensify. Therefore, companies may want to continue with their integration efforts by focusing on the determinants discussed in this study. However, at the same time, the possibility of opportunistic behavior by the exchange partner cannot be disregarded.