783 resultados para Fundamentals of computing theory


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Lignocellulosic biomasses (e.g., wood and straws) are a potential renewable source for the production of a wide variety of chemicals that could be used to replace those currently produced by petrochemical industry. This would lead to lower greenhouse gas emissions and waste amounts, and to economical savings. There are many possible pathways available for the manufacturing of chemicals from lignocellulosic biomasses. One option is to hydrolyze the cellulose and hemicelluloses of these biomasses into monosaccharides using concentrated sulfuric acid as catalyst. This process is an efficient method for producing monosaccharides which are valuable platforn chemicals. Also other valuable products are formed in the hydrolysis. Unfortunately, the concentrated acid hydrolysis has been deemed unfeasible mainly due to high chemical consumption resulting from the need to remove sulfuric acid from the obtained hydrolysates prior to the downstream processing of the monosaccharides. Traditionally, this has been done by neutralization with lime. This, however, results in high chemical consumption. In addition, the by-products formed in the hydrolysis are not removed and may, thus, hinder the monosaccharide processing. In order to improve the feasibility of the concentrated acid hydrolysis, the chemical consumption should be decreased by recycling of sulfuric acid without neutralization. Furthermore, the monosaccharides and the other products formed in the hydrolysis should be recovered selectively for efficient downstream processing. The selective recovery of the hydrolysis by-products would have additional economical benefits on the process due to their high value. In this work, the use of chromatographic fractionation for the recycling of sulfuric acid and the selective recovery of the main components from the hydrolysates formed in the concentrated acid hydrolysis was investigated. Chromatographic fractionation based on the electrolyte exclusion with gel type strong acid cation exchange resins in acid (H+) form as a stationary phase was studied. A systematic experimental and model-based study regarding the separation task at hand was conducted. The phenomena affecting the separation were determined and their effects elucidated. Mathematical models that take accurately into account these phenomena were derived and used in the simulation of the fractionation process. The main components of the concentrated acid hydrolysates (sulfuric acid, monosaccharides, and acetic acid) were included into this model. Performance of the fractionation process was investigated experimentally and by simulations. Use of different process options was also studied. Sulfuric acid was found to have a significant co-operative effect on the sorption of the other components. This brings about interesting and beneficial effects in the column operations. It is especially beneficial for the separation of sulfuric acid and the monosaccharides. Two different approaches for the modelling of the sorption equilibria were investigated in this work: a simple empirical approach and a thermodynamically consistent approach (the Adsorbed Solution theory). Accurate modelling of the phenomena observed in this work was found to be possible using the simple empirical models. The use of the Adsorbed Solution theory is complicated by the nature of the theory and the complexity of the studied system. In addition to the sorption models, a dynamic column model that takes into account the volume changes of the gel type resins as changing resin bed porosity was also derived. Using the chromatography, all the main components of the hydrolysates can be recovered selectively, and the sulfuric acid consumption of the hydrolysis process can be lowered considerably. Investigation of the performance of the chromatographic fractionation showed that the highest separation efficiency in this separation task is obtained with a gel type resin with a high crosslinking degree (8 wt. %); especially when the hydrolysates contain high amounts of acetic acid. In addition, the concentrated acid hydrolysis should be done with as low sulfuric acid concentration as possible to obtain good separation performance. The column loading and flow rate also have large effects on the performance. In this work, it was demonstrated that when recycling of the fractions obtained in the chromatographic fractionation are recycled to preceding unit operations these unit operations should included in the performance evaluation of the fractionation. When this was done, the separation performance and the feasibility of the concentrated acid hydrolysis process were found to improve considerably. Use of multi-column chromatographic fractionation processes, the Japan Organo process and the Multi-Column Recycling Chromatography process, was also investigated. In the studied case, neither of these processes could compete with the single-column batch process in the productivity. However, due to internal recycling steps, the Multi-Column Recycling Chromatography was found to be superior to the batch process when the product yield and the eluent consumption were taken into account.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The safety of shipping operations in the Baltic Sea is an extensively studied issue due to the density of traffic and the ecological sensitivity of the area. The focus has, however, mainly been on ship technology or on traffic control measures and the operative safety aspect of shipping is in a minor position in maritime safety studies and is lacking in terms of solutions. Self-regulatory and voluntary measures could be effective ways to improve the operational safety of shipping. Corporate social responsibility, or CSR, is one example of a voluntary measure that the shipping companies can take. CSR can enhance maritime safety and improve the shipping companies’ competitiveness. The aim of this study is to increase the knowledge of CSR in the maritime sector and study its applicability and benefits to the shipping companies. The research comprises of a theory part and a questionnaire study, which examine the significance of corporate social responsibility in shipping companies’ maritime safety and competitiveness. The aim of the questionnaire study is to find out how corporate social responsibility is implemented in the shipping companies. According to the literature review, responsible actions can produce financial and time costs, but due to these actions cost savings in the form of learning and increased efficiency can be achieved. Responsible actions can also produce concrete improvements and a reputation of responsibility that can lead to both cost savings and increase in the company’s income. CSR is recognised as having real business benefits in terms of attracting customers and high-quality employees. In shipping, CSR usually focuses on environmental issues. Environmental social responsibility in shipping is mainly motivated by the need to comply with existing and forthcoming regulation. Shipping companies engage in CSR to gain competitive advantage and to increase maritime safety. The social aspects of CSR take into account the well-being and skills of the employees, corporation and other stakeholders of the company. The questionnaire study revealed that the most common CSR measures in shipping companies are environmental measures, and that environmental concerns are considered to be the most important reason to engage in CSR. From the preliminary question about the concept of CSR it can also be seen that safety issues are commonly considered to be a part of CSR and safety gains are the second most important reason to engage in CSR. From the questionnaire, it can also be extrapolated that gaining a better reputation is one of the most important reasons to engage in CSR in the first place. For example, the main economic benefit was seen to be the increase of customer numbers as a result of a better reputation. Based on the study, it would seem that companies are starting to realise that they might gain competitive advantage and be favoured as shippers if they engage in sustainable, responsible operations and present themselves as “green”.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Video transcoding refers to the process of converting a digital video from one format into another format. It is a compute-intensive operation. Therefore, transcoding of a large number of simultaneous video streams requires a large amount of computing resources. Moreover, to handle di erent load conditions in a cost-e cient manner, the video transcoding service should be dynamically scalable. Infrastructure as a Service Clouds currently offer computing resources, such as virtual machines, under the pay-per-use business model. Thus the IaaS Clouds can be leveraged to provide a coste cient, dynamically scalable video transcoding service. To use computing resources e ciently in a cloud computing environment, cost-e cient virtual machine provisioning is required to avoid overutilization and under-utilization of virtual machines. This thesis presents proactive virtual machine resource allocation and de-allocation algorithms for video transcoding in cloud computing. Since users' requests for videos may change at di erent times, a check is required to see if the current computing resources are adequate for the video requests. Therefore, the work on admission control is also provided. In addition to admission control, temporal resolution reduction is used to avoid jitters in a video. Furthermore, in a cloud computing environment such as Amazon EC2, the computing resources are more expensive as compared with the storage resources. Therefore, to avoid repetition of transcoding operations, a transcoded video needs to be stored for a certain time. To store all videos for the same amount of time is also not cost-e cient because popular transcoded videos have high access rate while unpopular transcoded videos are rarely accessed. This thesis provides a cost-e cient computation and storage trade-o strategy, which stores videos in the video repository as long as it is cost-e cient to store them. This thesis also proposes video segmentation strategies for bit rate reduction and spatial resolution reduction video transcoding. The evaluation of proposed strategies is performed using a message passing interface based video transcoder, which uses a coarse-grain parallel processing approach where video is segmented at group of pictures level.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In some literature variations in photosynthetic rates are considered to be of little relevance for individual fitness. This depends among other things on how one defines fitness, i.e. if one takes strictly Darwinian fitness as seed production or if one needs to evaluate particular traits and consider plant establishment. It also matters if one takes the Darwinian "organism individual" as the central entity in evolution ("individual fitness") or the "species individual" in a modified "Structure of Evolutionary Theory" sensu Stephen Jay Gould. A phenotypically expressed trait like photosynthetic rate, even if intra- and interspecific differences may be small, can matter in habitat performance and niche acquisition. Light dependence curves (LCs) of photosynthetic rates are now readily measured under field conditions using miniaturized equipment of pulse amplitude modulated fluorometers. In contrast to actual momentary measurements of quantum yield of photosynthesis under actually prevailing ambient conditions, LC measurements reflect the expressed intrinsic capacity of photosynthesis. In this review we explore the power of LC measurements yielding cardinal points such as maximum apparent electron transport rate of photosystem II (ETRmax) and saturating photosynthetically active radiation (PARsat) in making intra- and interspecific comparisons of plant performance and synecological fingerprinting in ecophysiological studies across species, sites, habitats and ecosystems.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Distillation is a unit operation of process industry, which is used to separate a liquid mixture into two or more products and to concentrate liquid mixtures. A drawback of the distillation is its high energy consumption. An increase in energy and raw material prices has led to seeking ways to improve the energy efficiency of distillation. In this Master's Thesis, these ways are studied in connection with the concentration of hydrogen peroxide at the Solvay Voikkaa Plant. The aim of this thesis is to improve the energy efficiency of the concentration of the Voikkaa Plant. The work includes a review of hydrogen peroxide and its manufacturing. In addition, the fundamentals of distillation and its energy efficiency are reviewed. An energy analysis of the concentration unit of Solvay Voikkaa Plant is presented in the process development study part. It consists of the current and past information of energy and utility consumptions, balances, and costs. After that, the potential ways to improve the energy efficiency of the distillation unit at the factory are considered and their feasibility is evaluated technically and economically. Finally, proposals to improve the energy efficiency are suggested. Advanced process control, heat integration and energy efficient equipment are the most potential ways to carry out the energy efficient improvements of the concentration at the Solvay Voikkaa factory. Optimization of the reflux flow and the temperatures of the overhead condensers can offer immediate savings in the energy and utility costs without investments. Replacing the steam ejector system with a vacuum pump would result in savings of tens of thousands of euros per year. The heat pump solutions, such as utilizing a mechanical vapor recompression or thermal vapor recompression, are not feasible due to the high investment costs and long pay back times.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Nykyaikaiset pilvipalvelut tarjoavat suurille yrityksille mahdollisuuden tehostaa laskennallista tietojenkäsittelyä. Pilvipalveluiden käyttöönotto tuo mukanaan kuitenkin esimerkiksi useita tietoturvakysymyksiä, joiden vuoksi käyttöönoton tulee olla tarkasti suunniteltua. Tämä tutkimus esittelee kirjallisuuskatsaukseen perustuvan, asteittaisen suunnitelman pilvipalveluiden käyttöönotolle energialiiketoimintaympäristössä. Kohdeyrityksen sisäiset haastattelut ja katsaus nykyisiin energiateollisuuden pilviratkaisuihin muodostavat kokonaiskuvan käyttöönoton haasteista ja mahdollisuuksista. Tutkimuksen päätavoitteena on esittää ratkaisut tyypillisiin pilvipalvelun käyttöönotossa esiintyviin ongelmiin käyttöönottomallin avulla. Tutkimuksessa rakennettu käyttöönottomalli testattiin esimerkkitapauksen avulla ja malli todettiin toimivaksi. Ulkoisten palveluiden herättämien tietoturvakysymysten takia käyttöönoton ensimmäiset osiot, kuten lopputuotteen määrittely ja huolellinen suunnittelu, ovat koko käyttöönottoprosessin ydin. Lisäksi pilvipalveluiden käyttöönotto vaatii nykyiseltä käyttöympäristöltä uusia teknisiä ja hallinnollisia taitoja. Tutkimuksen tulokset osoittavat pilvipalveluiden monipuolisen hyödyn erityisesti laskentatehon tarpeen vaihdellessa. Käyttöönottomallin rinnalle luotu kustannusvertailu tukee kirjallisuuskatsauksessa esille tuotuja hyötyjä ja tarjoaa kohdeyritykselle perusteen tutkimuksen eteenpäin viemiselle.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Tämä työ vastaa tarpeeseen hallita korkeapainevesisumusuuttimen laatua virtausmekaniikan työkalujen avulla. Työssä tutkitaan suutinten testidatan lisäksi virtauksen käyttäytymistä suuttimen sisällä CFD-laskennan avulla. Virtausmallinnus tehdään Navier-Stokes –pohjaisella laskentamenetelmällä. Työn teoriaosassa käsitellään virtaustekniikkaa ja sen kehitystä yleisesti. Lisäksi esitetään suuttimen laskennassa käytettävää perusteoriaa sekä teknisiä ratkaisuja. Teoriaosassa käydään myös läpi laskennalliseen virtausmekaniikkaan (CFD-laskenta) liittyvää perusteoriaa. Tutkimusosiossa esitetään käsitellyt suutintestitulokset sekä mallinnetaan suutinvirtausta ajasta riippumattomaan virtauslaskentaan perustuvalla laskentamenetelmällä. Virtauslaskennassa käytetään OpenFOAM-laskentaohjelmiston SIMPLE-virtausratkaisijaa sekä k-omega SST –turbulenssimallia. Tehtiin virtausmallinnus kaikilla paineilla, joita suuttimen testauksessa myös todellisuudessa käytetään. Lisäksi selvitettiin mahdolliset kavitaatiokohdat suuttimessa ja suunniteltiin kavitaatiota ehkäisevä suutingeometria. Todettiin myös lämpötilan ja epäpuhtauksien vaikuttavan kavitaatioon sekä mallinnettiin lämpötilan vaikutusta. Luotiin malli, jolla suuttimen suunnitteluun liittyviin haasteisiin voidaan vastata numeerisella laskennalla.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Arsenic is a toxic substance. The amount of arsenic in waste water is a raising problem because of increasing mining industry. Arsenic is connected to cancers in areas where arsenic concentration in drinking water is higher than recommendations. The main object in this master’s thesis was to research how ferrous hydroxide waste material is adsorbed arsenic from ammonia containing waste water. In this master’s thesis there is two parts: theoretical and experimental part. In theoretical part harmful effects of arsenic, theory of adsorption, isotherms modeling of adsorption and analysis methods of arsenic are described. In experimental part adsorption capacity of ferrous hydroxide waste material and adsorption time with different concentrations of arsenic were studied. Waste material was modified with two modification methods. Based on experimental results the adsorption capacity of waste material was high. The problem with waste material was that at same time with arsenic adsorption sulfur was dissolving in solution. Waste material was purified from sulfur but purification methods were not efficient enough. Purification methods require more research.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The aim of the study and research questions: The aim of this study is to illuminate how caring communion can aid in promoting health as becoming in elderly people in the context of natural caring. The target group of the study consists of elderly citizens living at home. The focus of this thesis is on the concept of communion and how caring communion can affect the inner health resources in a patient’s inner health domain, as well as how caring communion can support health as becoming and inner health resources in the elderly. The main research questions of this study are the following: 1) what does communion mean? 2) what does caring communion mean? 3) what is the connection between caring communion and health? Theoretical perspective: The theoretical perspective of this qualitative study relies on the caritative caring theory as developed by scholars of caring sciences at the Åbo Academi University’s Vasa unit. The caritas motive is based on an ethos built on a consideration of togetherness, i.e. caring communion, a place where one feels at home and where one can be the person one was meant to be. Methodology: A hermeneutic research approach based on Gadamer (1997) permeates the study. This entails that understanding and interpretation become central. The study conducted in the thesis is divided into three sub-studies. Sub-study one and two are based on ontological determination whereas the third sub-study is carried out by contextual determination. The first sub-study is conducted by etymological and semantic analysis of the concept of communion (gemenskap) based on Koort (1975) and the second sub-study by determining the basic epistemological category of the concept based on Eriksson (2010b). Sub-study three is conducted through content analysis of 18 multidisciplinary and 13 caring science articles and dissertations based on Kvale (2009). The aim in the third sub-study is to define caring communions in various contexts of meaning based on Eriksson´s model of conceptual determination (2010b). All studies are interpreted through hermeneutic interpretation where the continuous movement from a part of a whole, to the whole, to part again, leads to new understanding. Finally, the findings from all the three sub-studies are compared to the concepts of pre-understanding and the inner-health-domain model of Wärnås (2002). Results: The results of the study offer a description of the dimensions of caring communion and a model that illuminates how caring communion can further health as becoming. The fundamentals of caring communion rest on the idea of a human being’s absolute right to dignity as a base for communion. The concept of communion contains a moral, an ethical, and a spiritual component. In communion, there exists a moral and ethical responsibility and a willingness to commit oneself. The individual is part of a connection or relation and knows the aim and course for the communion. A caring connection, a caring culture, a caring atmosphere and caring listening are characteristics of caring communion. In caring communion, the elderly feel trusting and see themselves as unique, powerful, and valuable. The model demonstrates that when the elderly are able to rest in caring communion, the virtues of courage and faith become strong and desire for life awaken within the elderly and health as becoming becomes possible. Conclusions: The outcome of the study is that all communion is not necessarily caring communion. In order for communion to be caring and for the elderly to achieve health as becoming, there are certain criteria that must be met. This is especially important when designing activities for the elderly in the context of natural caring.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This qualitative study has started from the interest to examine how the reality of crosscultural encounters is presented in the global business press. The research paper emphasizes different ways to classify culture and cross-cultural competency, both from the point of view of individuals and organizations. The analysis consists of public discourses, where cross-cultural realities are created through different persons, stories and contexts For data collection, a comprehensive database search was performed and 10 articles from the widely known worldwide business magazine The Financial Times were chosen as the data for the study paper. For the functions of addressing the research study questions, Thematic Content Analysis (TCA) and also Discourse Analysis (DA) are utilized, added with the continuous comparison method of grounded theory in the formation of the data.The academic references consist of literary works and articles presenting relevant concepts, creating a cross-cultural framework, and it is designed to assist the reader in the navigation through the topics of culture and cross-cultural competency. The repertoires were formed from the data and following, the first repertoire is contrast difference between home and target culture that the individual was able to discern. As a consequence of the first repertoire, the companies then offer cultural training to their employees to prepare them to situations of increasing levels of cultural variation. The third repertoire is increased awareness of other cultures, which is conveyed as a result of cultural training and contextual work experience. The fourth repertoire is globalization as an international business environment, where the people in the articles perform their job functions. It is stated in the conclusions that the representations emphasize Western values and personal traits in leadership.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This work investigates theoretical properties of symmetric and anti-symmetric kernels. First chapters give an overview of the theory of kernels used in supervised machine learning. Central focus is on the regularized least squares algorithm, which is motivated as a problem of function reconstruction through an abstract inverse problem. Brief review of reproducing kernel Hilbert spaces shows how kernels define an implicit hypothesis space with multiple equivalent characterizations and how this space may be modified by incorporating prior knowledge. Mathematical results of the abstract inverse problem, in particular spectral properties, pseudoinverse and regularization are recollected and then specialized to kernels. Symmetric and anti-symmetric kernels are applied in relation learning problems which incorporate prior knowledge that the relation is symmetric or anti-symmetric, respectively. Theoretical properties of these kernels are proved in a draft this thesis is based on and comprehensively referenced here. These proofs show that these kernels can be guaranteed to learn only symmetric or anti-symmetric relations, and they can learn any relations relative to the original kernel modified to learn only symmetric or anti-symmetric parts. Further results prove spectral properties of these kernels, central result being a simple inequality for the the trace of the estimator, also called the effective dimension. This quantity is used in learning bounds to guarantee smaller variance.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This thesis concerns the analysis of epidemic models. We adopt the Bayesian paradigm and develop suitable Markov Chain Monte Carlo (MCMC) algorithms. This is done by considering an Ebola outbreak in the Democratic Republic of Congo, former Zaïre, 1995 as a case of SEIR epidemic models. We model the Ebola epidemic deterministically using ODEs and stochastically through SDEs to take into account a possible bias in each compartment. Since the model has unknown parameters, we use different methods to estimate them such as least squares, maximum likelihood and MCMC. The motivation behind choosing MCMC over other existing methods in this thesis is that it has the ability to tackle complicated nonlinear problems with large number of parameters. First, in a deterministic Ebola model, we compute the likelihood function by sum of square of residuals method and estimate parameters using the LSQ and MCMC methods. We sample parameters and then use them to calculate the basic reproduction number and to study the disease-free equilibrium. From the sampled chain from the posterior, we test the convergence diagnostic and confirm the viability of the model. The results show that the Ebola model fits the observed onset data with high precision, and all the unknown model parameters are well identified. Second, we convert the ODE model into a SDE Ebola model. We compute the likelihood function using extended Kalman filter (EKF) and estimate parameters again. The motivation of using the SDE formulation here is to consider the impact of modelling errors. Moreover, the EKF approach allows us to formulate a filtered likelihood for the parameters of such a stochastic model. We use the MCMC procedure to attain the posterior distributions of the parameters of the SDE Ebola model drift and diffusion parts. In this thesis, we analyse two cases: (1) the model error covariance matrix of the dynamic noise is close to zero , i.e. only small stochasticity added into the model. The results are then similar to the ones got from deterministic Ebola model, even if methods of computing the likelihood function are different (2) the model error covariance matrix is different from zero, i.e. a considerable stochasticity is introduced into the Ebola model. This accounts for the situation where we would know that the model is not exact. As a results, we obtain parameter posteriors with larger variances. Consequently, the model predictions then show larger uncertainties, in accordance with the assumption of an incomplete model.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The aim of this master's thesis is to develop a two-dimensional drift-di usion model, which describes charge transport in organic solar cells. The main bene t of a two-dimensional model compared to a one-dimensional one is the inclusion of the nanoscale morphology of the active layer of a bulk heterojunction solar cell. The developed model was used to study recombination dynamics at the donor-acceptor interface. In some cases, it was possible to determine e ective parameters, which reproduce the results of the two-dimensional model in the one-dimensional case. A summary of the theory of charge transport in semiconductors was presented and discussed in the context of organic materials. Additionally, the normalization and discretization procedures required to nd a numerical solution to the charge transport problem were outlined. The charge transport problem was solved by implementing an iterative scheme called successive over-relaxation. The obtained solution is given as position-dependent electric potential, free charge carrier concentrations and current densities in the active layer. An interfacial layer, separating the pure phases, was introduced in order to describe charge dynamics occurring at the interface between the donor and acceptor. For simplicity, an e ective generation of free charge carriers in the interfacial layer was implemented. The pure phases simply act as transport layers for the photogenerated charges. Langevin recombination was assumed in the two-dimensional model and an analysis of the apparent recombination rate in the one-dimensional case is presented. The recombination rate in a two-dimensional model is seen to e ectively look like reduced Langevin recombination at open circuit. Replicating the J-U curves obtained in the two-dimensional model is, however, not possible by introducing a constant reduction factor in the Langevin recombination rate. The impact of an acceptor domain in the pure donor phase was investigated. Two cases were considered, one where the acceptor domain is isolated and another where it is connected to the bulk of the acceptor. A comparison to the case where no isolated domains exist was done in order to quantify the observed reduction in the photocurrent. The results show that all charges generated at the isolated domain are lost to recombination, but the domain does not have a major impact on charge transport. Trap-assisted recombination at interfacial trap states was investigated, as well as the surface dipole caused by the trapped charges. A theoretical expression for the ideality factor n_id as a function of generation was derived and shown to agree with simulation data. When the theoretical expression was fitted to simulation data, no interface dipole was observed.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The objective of this study was to find out how third party influencers can facilitate value-based selling in a network and how suppliers should aim to impact on these third party influencers to facilitate value-based selling. The study considers construction industry, selling the column connection solution and third party influencers. Third party influencers examined in this study were structural designers. The study also aims to find out structural designers’ value drivers and the differences between the market areas that this study related to. The theoretical part of the study focuses on two separate areas. The first part of the theory focuses on a value-based selling concept: what it is, what it requires and what are the main barriers for value-based selling. The second part of the theory examines value creation in networks. The present knowledge over value creation in networks and different network actors are presented. Project marketing is also discussed briefly because this study’s topic, which is highly related to project business. The results reveal structural designers’ value drivers considering the usage of the column connection solution and present ways how suppliers should aim to impact structural designers to facilitate value-based selling. The main result of the study indicates that third party influencers can have a positive impact on facilitating value-based selling. Structural designers are communicating more or less with all the salient actors in different project phases and they can act as sponsors to support the sales of Peikko’s column connection solution and promote solution to other actors involved to the project. This requires that structural designers can understand the actual benefits of how the solution can improve their and their customers’ business.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The study focuses on five lower secondary school pupils’ daily use of their one-toone computers, the overall aim being to investigate literacy in this form of computing. Theoretically, the study is rooted in the New Literacy tradition with an ecological perspective, in combination with socio-semiotic theory in a multimodal perspective. New Literacy in the ecological perspective focuses on literacy practices and place/space and on the links between them. Literacy is viewed as socially based, in specific situations and in recurring social practices. Socio-semiotic theory embodying the multimodal perspective is used for the text analysis. The methodology is known as socio-semiotic ethnography. The ethnographic methods encompass just over two years of fieldwork with participating observations of the five participants’ computing activities at home, at school and elsewhere. The participants, one boy and two girls from the Blue (Anemone) School and two girls from the White (Anemone) School, were chosen to reflect a broad spectrum in terms of sociocultural and socioeconomic background. The study shows the existence of a both broad and deep variation in the way digital literacy features in the participants’ one-to-one computing. These variations are associated with experience in relation to the home, the living environment, place, personal qualities and school. The more varied computer usage of the Blue School participants is connected with the interests they developed in their homes and living environments and in the computing practices undertaken in school. Their more varied usage of the computer is reflected in their broader digital literacy repertoires and their greater number and variety of digital literacy abilities. The Blue School participants’ text production is more multifaceted, covers a wider range of subjects and displays a broader palette of semiotic resources. It also combines more text types and the texts are generally longer than those of the White School participants. The Blue School girls have developed a text culture that is close to that of the school. In their case, there is clear linkage between school-initiated and self-initiated computing activities, while other participants do not have the same opportunities to link and integrate self-initiated computing activities into the school context. It also becomes clear that the Blue School girls can relate and adapt their texts to different communicative practices and recipients. In addition, the study shows that the Blue School girls have some degree of scope in their school practice as a result of incorporating into it certain communicative practices that they have developed in nonschool contexts. Quite contrary to the hopes expressed that one-to-one computing would reduce digital inequality, it has increased between these participants. Whether the same or similar results apply in a larger perspective, on a more structural level, is a question that this study cannot answer. It can only draw attention to the need to investigate the matter. The study shows in a variety of ways that the White School participants do not have the same opportunity to develop their digital literacy as the Blue School participants. In an equivalence perspective, schools have a compensational task to perform. It is abundantly clear from the study that investing in one-to-one projects is not enough to combat digital inequality and achieve the digitisation goals established for school education. Alongside their investments in technology, schools need to develop a didactic that legitimises and compensates for the different circumstances of different pupils. The compensational role of schools in this connection is important not only for the present participants but also for the community at large, in that it can help to secure a cohesive, open and democratic society.