925 resultados para 700103 Information processing services


Relevância:

100.00% 100.00%

Publicador:

Resumo:

Tavoitteena diplomityössä oli kuvata projektiliiketoimintaa harjoittavan yrityksen informaatio- ja materiaalivirrat toimitusprojekteissa. Selvitys haluttiin tehdä tulevia kehitysprojekteja varten. Diplomityö toteutettiin kahessa osassa: laatimalla aiheeseen liittyvä teoriakatsaus ja kartoittamalla todellinen tilanne yrityksen sisällä.Teoriaosassa kuvattiin tarkastelualue ja projektiliiketoiminnan erityispiirteitä sekä käsiteltiin tietojärjestelmiä ja materiaalivirran hallintaan liittyviä asioita. Yrityksen informaatio- ja materiaalivirtojen nykytila kartoitettiin laatimalla vuokaaviot kolmen esimerkkituotteen toimitusprojekteista. Kaaviot laadittiin yrityksen sisäisten dokumenttien ja kvalitatiivisten haastattelujen pohjalta. Kaavioiden sisältöä täydennettiin taulukoimalla kaavioissa esiintyvien dokumettien sisältöä. Työn tuloksena muodostui kuva tiedonkäsittelyn nykytilasta myynti- ja toimitusprojekteissa, pelkistetty visio tulevaisuuden tietojärjestelmien rooleista sekä ehdotuksia tulevia kehitysprojekteja varten.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Information processing in groups has long been seen as a cooperative process. In contrast with this assumption, group members were rarely found to behave cooperatively: They withhold unshared information and stick to initial incorrect decisions. In the present article, we examined how group members' cooperative and competitivemotives impact on group information processing and propose that information sharing and use in groups could be seen as strategic behavior. We reviewed the latest developments in the literature investigating different forms of strategic information processing and their underlying mechanisms. This review suggests that explicit cooperative goals are needed for effective group decision-making.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Tapahtumat ovat perusta monille nykyaikaisille tietoteknisille palveluille. Yksittäinen tapahtuma voidaan nähdä työtehtävänä, joka käsittelevän järjestelmän tulee suorittaa. Tapahtumankäsittely pyrkii pitämään järjestelmän tunnetussa ja ristiriidattomassa tilassa. Tämä toteutetaan pitämällä huolta, että jokainen tapahtuma joko onnistuu tai epäonnistuu kokonaisuudessaan. Tapahtumankäsittelyjärjestelmät ovat kasvaneet ja yhtäaikaisten käsiteltävien tapahtumien määrä noussut palveluiden siirtyessä yhä enemmän tietoverkkoihin. Samalla järjestelmien kehittäminen ja ylläpito vaikeutuvat, jolloin kehittäjät tarvitsevat parempia työkaluja järjestelmän valvontaan. Tapahtumankäsittelyn seuranta pyrkii seuraamaan järjestelmän sisäistä toimintaa yksittäisen tapahtuman tai osatapahtuman tarkkuudella. Riippuen toteutuksesta kehittäjä voi joko tarkkailla järjestelmää reaaliaikaisesti tai jälkikäteen suorituksen perusteella tallennetun seurantatiedon avulla. Työssä esitellään tapahtumanseurantakomponentin suunnitteluprosessi ratkaisuineen, joka mahdollistaa tapahtumien suorituksen tarkkailun, seurantatiedon tallentamisen sekä tulosten tarkastelun jälkikäteen. Työ on toteutettu osaksi Syncron Tech Oy:n Syncware-ohjelmistoalustaa.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Many cognitive deficits after TBI (traumatic brain injury) are well known, such as memory and concentration problems, as well as reduced information-processing speed. What happens to patients and cognitive functioning after immediate recovery is poorly known. Cognitive functioning is flexible and may be influenced by genetic, psychological and environmental factors decades after TBI. The general aim of this thesis was to describe the long-term cognitive course after TBI, to find variables that may contribute to it, and how the cognitive functions after TBI are associated with specific medical factors and reduced survival. The original study group consisted of 192 patients with TBI who were originally assessed with the Mild Deterioration Battery (MDB) on average two years after the injury, during the years 1966 – 1972. During a 30-year follow-up, we studied the risks for reduced survival, and the mortality of the patients was compared with the general population using the Standardized Mortality Ratio (SMR). Sixty-one patients were re-assessed during 1998-2000. These patients were evaluated with the MDB, computerized testing, and with various other neuropsychological methods for attention and executive functions. Apolipoprotein-E (ApoE) genotyping and magnetic resonance imaging (MRI) based on volumetric analysis of the hippocampus and lateral ventricles were performed. Depressive symptoms were evaluated with the short form of the Beck depression inventory. The cognitive performance at follow-up was compared with a control group that was similar to the study group in regard to age and education. The cognitive outcome of the patients with TBI varied after three decades. The majority of the patients showed a decline in their cognitive level, the rest either improved or stayed at the same level. Male gender and higher age at injury were significant risk factors for the decline. Whereas most cognitive domains declined during the follow-up, semantic memory behaved in the opposite way, showing recovery after TBI. In the follow-up assessment, the memory decline and impairments in the set-shifting domain of executive functions were associated with MRI-volumetric measures, whereas reduction in information-processing speed was not associated with the MRI measures. The presence of local contusions was only weakly associated with cognitive functions. Only few cognitive methods for attention were capable of discriminating TBI patients with and without depressive symptoms. On the other hand, most complex attentional tests were sensitive enough to discriminate TBI patients (non-depressive) from controls. This means that complex attention functions, mediated by the frontal lobes, are relatively independent of depressive symptoms post-TBI. The presence of ApoE4 was associated with different kinds of memory processes including verbal and visual episodic memory, semantic memory and verbal working memory, depending on the length of time since TBI. Many other cognitive processes were not affected by the presence of ApoE4. Age at injury and poor vocational outcome were independent risk factors for reduced survival in the multivariate analysis. Late mortality was higher among younger subjects (age < 40 years at death) compared with the general population which should be borne in mind when assessing the need for rehabilitation services and long-term follow-up after TBI.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The significance of services as business and human activities has increased dramatically throughout the world in the last three decades. Becoming a more and more competitive and efficient service provider while still being able to provide unique value opportunities for customers requires new knowledge and ideas. Part of this knowledge is created and utilized in daily activities in every service organization, but not all of it, and therefore an emerging phenomenon in the service context is information awareness. Terms like big data and Internet of things are not only modern buzz-words but they are also describing urgent requirements for a new type of competences and solutions. When the amount of information increases and the systems processing information become more efficient and intelligent, it is the human understanding and objectives that may get separated from the automated processes and technological innovations. This is an important challenge and the core driver for this dissertation: What kind of information is created, possessed and utilized in the service context, and even more importantly, what information exists but is not acknowledged or used? In this dissertation the focus is on the relationship between service design and service operations. Reframing this relationship refers to viewing the service system from the architectural perspective. The selected perspective allows analysing the relationship between design activities and operational activities as an information system while maintaining the tight connection to existing service research contributions and approaches. This type of an innovative approach is supported by research methodology that relies on design science theory. The methodological process supports the construction of a new design artifact based on existing theoretical knowledge, creation of new innovations and testing the design artifact components in real service contexts. The relationship between design and operations is analysed in the health care and social care service systems. The existing contributions in service research tend to abstract services and service systems as value creation, working or interactive systems. This dissertation adds an important information processing system perspective to the research. The main contribution focuses on the following argument: Only part of the service information system is automated and computerized, whereas a significant part of information processing is embedded in human activities, communication and ad-hoc reactions. The results indicate that the relationship between service design and service operations is more complex and dynamic than the existing scientific and managerial models tend to view it. Both activities create, utilize, mix and share information, making service information management a necessary but relatively unknown managerial task. On the architectural level, service system -specific elements seem to disappear, but access to more general information elements and processes can be found. While this dissertation focuses on conceptual-level design artifact construction, the results provide also very practical implications for service providers. Personal, visual and hidden activities of service, and more importantly all changes that take place in any service system have also an information dimension. Making this information dimension visual and prioritizing the processed information based on service dimensions is likely to provide new opportunities to increase activities and provide a new type of service potential for customers.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This monograph dissertation looks into the field of ICT-mediated health and well-being services. Through six chapters that extend the work done in the reviewed and published articles, the dissertation focuses on new and emerging technologies, and to impact of their use on the beneficiary; the individual who eventually derives advantage from the services. As the field is currently going through major changes particularly in the OECD countries, the focus is on shortterm developments in the field and the analysis on the long term developments is cursory by nature. The dissertation includes theoretical and empirical elements. Most of the empirical elements are linked to product development and conceptualization performed in the national MyWellbeing project that ended in 2010. In the project, the emphasis was on conceptualization of a personal aid for the beneficiary that could be used for managing information and services in the field of health and well-being services. This work continued the theme of developing individual-centric solutions for the field; a work that started in the InnoElli Senior program in 2006. The nature of this thesis is foremost a conceptual elaboration based on a literature review, illustrated in empirical work performed in different projects. As a theoretical contribution, this dissertation elaborates the role of a mediator, i.e. an intermediary, and it is used as an overarching theme. The role acts as a ‘lens’ through which a number of technology-related phenomena are looked at, pinned down and addressed to a degree. This includes introduction of solutions, ranging from anthropomorphic artefacts to decision support systems that may change the way individuals experience clinical encounters in the near-future. Due to the complex and multiform nature of the field, it is impractical and effectively impossible to cover all aspects that are related to mediation in a single work. Issues such as legislation, financing and privacy are all of equal importance. Consideration of all these issues is beyond the scope of this dissertation and their investigation is left to other work. It follows from this that the investigation on the role is not intended as inclusive one. The role of the mediator is also used to highlight some of the ethical issues related to personal health information management, and to mediating health and well-being related issues on behalf of another individual, such as an elderly relative or a fellow member of a small unit in the armed forces. The dissertation concludes in a summary about the use and functions of the mediator, describing some potential avenues for implementing such support mechanisms to the changing field of ICT-mediated health and well-being services. The conclusions also describe some of the limitations of this dissertation, including remarks on methodology and content.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

La théorie de l'information quantique étudie les limites fondamentales qu'imposent les lois de la physique sur les tâches de traitement de données comme la compression et la transmission de données sur un canal bruité. Cette thèse présente des techniques générales permettant de résoudre plusieurs problèmes fondamentaux de la théorie de l'information quantique dans un seul et même cadre. Le théorème central de cette thèse énonce l'existence d'un protocole permettant de transmettre des données quantiques que le receveur connaît déjà partiellement à l'aide d'une seule utilisation d'un canal quantique bruité. Ce théorème a de plus comme corollaires immédiats plusieurs théorèmes centraux de la théorie de l'information quantique. Les chapitres suivants utilisent ce théorème pour prouver l'existence de nouveaux protocoles pour deux autres types de canaux quantiques, soit les canaux de diffusion quantiques et les canaux quantiques avec information supplémentaire fournie au transmetteur. Ces protocoles traitent aussi de la transmission de données quantiques partiellement connues du receveur à l'aide d'une seule utilisation du canal, et ont comme corollaires des versions asymptotiques avec et sans intrication auxiliaire. Les versions asymptotiques avec intrication auxiliaire peuvent, dans les deux cas, être considérées comme des versions quantiques des meilleurs théorèmes de codage connus pour les versions classiques de ces problèmes. Le dernier chapitre traite d'un phénomène purement quantique appelé verrouillage: il est possible d'encoder un message classique dans un état quantique de sorte qu'en lui enlevant un sous-système de taille logarithmique par rapport à sa taille totale, on puisse s'assurer qu'aucune mesure ne puisse avoir de corrélation significative avec le message. Le message se trouve donc «verrouillé» par une clé de taille logarithmique. Cette thèse présente le premier protocole de verrouillage dont le critère de succès est que la distance trace entre la distribution jointe du message et du résultat de la mesure et le produit de leur marginales soit suffisamment petite.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Thèse numérisée par la Division de la gestion de documents et des archives de l'Université de Montréal

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Sharing of information with those in need of it has always been an idealistic goal of networked environments. With the proliferation of computer networks, information is so widely distributed among systems, that it is imperative to have well-organized schemes for retrieval and also discovery. This thesis attempts to investigate the problems associated with such schemes and suggests a software architecture, which is aimed towards achieving a meaningful discovery. Usage of information elements as a modelling base for efficient information discovery in distributed systems is demonstrated with the aid of a novel conceptual entity called infotron.The investigations are focused on distributed systems and their associated problems. The study was directed towards identifying suitable software architecture and incorporating the same in an environment where information growth is phenomenal and a proper mechanism for carrying out information discovery becomes feasible. An empirical study undertaken with the aid of an election database of constituencies distributed geographically, provided the insights required. This is manifested in the Election Counting and Reporting Software (ECRS) System. ECRS system is a software system, which is essentially distributed in nature designed to prepare reports to district administrators about the election counting process and to generate other miscellaneous statutory reports.Most of the distributed systems of the nature of ECRS normally will possess a "fragile architecture" which would make them amenable to collapse, with the occurrence of minor faults. This is resolved with the help of the penta-tier architecture proposed, that contained five different technologies at different tiers of the architecture.The results of experiment conducted and its analysis show that such an architecture would help to maintain different components of the software intact in an impermeable manner from any internal or external faults. The architecture thus evolved needed a mechanism to support information processing and discovery. This necessitated the introduction of the noveI concept of infotrons. Further, when a computing machine has to perform any meaningful extraction of information, it is guided by what is termed an infotron dictionary.The other empirical study was to find out which of the two prominent markup languages namely HTML and XML, is best suited for the incorporation of infotrons. A comparative study of 200 documents in HTML and XML was undertaken. The result was in favor ofXML.The concept of infotron and that of infotron dictionary, which were developed, was applied to implement an Information Discovery System (IDS). IDS is essentially, a system, that starts with the infotron(s) supplied as clue(s), and results in brewing the information required to satisfy the need of the information discoverer by utilizing the documents available at its disposal (as information space). The various components of the system and their interaction follows the penta-tier architectural model and therefore can be considered fault-tolerant. IDS is generic in nature and therefore the characteristics and the specifications were drawn up accordingly. Many subsystems interacted with multiple infotron dictionaries that were maintained in the system.In order to demonstrate the working of the IDS and to discover the information without modification of a typical Library Information System (LIS), an Information Discovery in Library Information System (lDLIS) application was developed. IDLIS is essentially a wrapper for the LIS, which maintains all the databases of the library. The purpose was to demonstrate that the functionality of a legacy system could be enhanced with the augmentation of IDS leading to information discovery service. IDLIS demonstrates IDS in action. IDLIS proves that any legacy system could be augmented with IDS effectively to provide the additional functionality of information discovery service.Possible applications of IDS and scope for further research in the field are covered.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In the vision of Mark Weiser on ubiquitous computing, computers are disappearing from the focus of the users and are seamlessly interacting with other computers and users in order to provide information and services. This shift of computers away from direct computer interaction requires another way of applications to interact without bothering the user. Context is the information which can be used to characterize the situation of persons, locations, or other objects relevant for the applications. Context-aware applications are capable of monitoring and exploiting knowledge about external operating conditions. These applications can adapt their behaviour based on the retrieved information and thus to replace (at least a certain amount) the missing user interactions. Context awareness can be assumed to be an important ingredient for applications in ubiquitous computing environments. However, context management in ubiquitous computing environments must reflect the specific characteristics of these environments, for example distribution, mobility, resource-constrained devices, and heterogeneity of context sources. Modern mobile devices are equipped with fast processors, sufficient memory, and with several sensors, like Global Positioning System (GPS) sensor, light sensor, or accelerometer. Since many applications in ubiquitous computing environments can exploit context information for enhancing their service to the user, these devices are highly useful for context-aware applications in ubiquitous computing environments. Additionally, context reasoners and external context providers can be incorporated. It is possible that several context sensors, reasoners and context providers offer the same type of information. However, the information providers can differ in quality levels (e.g. accuracy), representations (e.g. position represented in coordinates and as an address) of the offered information, and costs (like battery consumption) for providing the information. In order to simplify the development of context-aware applications, the developers should be able to transparently access context information without bothering with underlying context accessing techniques and distribution aspects. They should rather be able to express which kind of information they require, which quality criteria this information should fulfil, and how much the provision of this information should cost (not only monetary cost but also energy or performance usage). For this purpose, application developers as well as developers of context providers need a common language and vocabulary to specify which information they require respectively they provide. These descriptions respectively criteria have to be matched. For a matching of these descriptions, it is likely that a transformation of the provided information is needed to fulfil the criteria of the context-aware application. As it is possible that more than one provider fulfils the criteria, a selection process is required. In this process the system has to trade off the provided quality of context and required costs of the context provider against the quality of context requested by the context consumer. This selection allows to turn on context sources only if required. Explicitly selecting context services and thereby dynamically activating and deactivating the local context provider has the advantage that also the resource consumption is reduced as especially unused context sensors are deactivated. One promising solution is a middleware providing appropriate support in consideration of the principles of service-oriented computing like loose coupling, abstraction, reusability, or discoverability of context providers. This allows us to abstract context sensors, context reasoners and also external context providers as context services. In this thesis we present our solution consisting of a context model and ontology, a context offer and query language, a comprehensive matching and mediation process and a selection service. Especially the matching and mediation process and the selection service differ from the existing works. The matching and mediation process allows an autonomous establishment of mediation processes in order to transfer information from an offered representation into a requested representation. In difference to other approaches, the selection service selects not only a service for a service request, it rather selects a set of services in order to fulfil all requests which also facilitates the sharing of services. The approach is extensively reviewed regarding the different requirements and a set of demonstrators shows its usability in real-world scenarios.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This research investigates what information German Fairtrade coffee consumers search for during pre-purchase information seeking and to what extent information is retrieved. Furthermore, the sequence of the information search as well as the degree of cognitive involvement is highlighted. The role of labeling, the importance of additional ethical information and its quality in terms of concreteness as well as the importance of product price and organic origin are addressed. A set of information relevant to Fairtrade consumers was tested by means of the Information Display Matrix (IDM) method with 389 Fairtrade consumers. Results show that prior to purchase, information on product packages plays an important role and is retrieved rather extensively, but search strategies that reduce the information processing effort are applied as well. Furthermore, general information is preferred over specific information. Results of two regression analyses indicate that purchase decisions are related to search behavior variables rather than to socio-demographic variables and purchase motives. In order to match product information with consumers’ needs, marketers should offer information that is reduced to the central aspects of Fairtrade.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Integration of inputs by cortical neurons provides the basis for the complex information processing performed in the cerebral cortex. Here, we propose a new analytic framework for understanding integration within cortical neuronal receptive fields. Based on the synaptic organization of cortex, we argue that neuronal integration is a systems--level process better studied in terms of local cortical circuitry than at the level of single neurons, and we present a method for constructing self-contained modules which capture (nonlinear) local circuit interactions. In this framework, receptive field elements naturally have dual (rather than the traditional unitary influence since they drive both excitatory and inhibitory cortical neurons. This vector-based analysis, in contrast to scalarsapproaches, greatly simplifies integration by permitting linear summation of inputs from both "classical" and "extraclassical" receptive field regions. We illustrate this by explaining two complex visual cortical phenomena, which are incompatible with scalar notions of neuronal integration.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Las tecnologías de la información han empezado a ser un factor importante a tener en cuenta en cada uno de los procesos que se llevan a cabo en la cadena de suministro. Su implementación y correcto uso otorgan a las empresas ventajas que favorecen el desempeño operacional a lo largo de la cadena. El desarrollo y aplicación de software han contribuido a la integración de los diferentes miembros de la cadena, de tal forma que desde los proveedores hasta el cliente final, perciben beneficios en las variables de desempeño operacional y nivel de satisfacción respectivamente. Por otra parte es importante considerar que su implementación no siempre presenta resultados positivos, por el contrario dicho proceso de implementación puede verse afectado seriamente por barreras que impiden maximizar los beneficios que otorgan las TIC.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Monomer-sequence information in synthetic copolyimides can be recognised by tweezer-type molecules binding to adjacent triplet-sequences on the polymer chains. In the present paper different tweezer-molecules are found to have different sequence-selectivities, as demonstrated in solution by 1H NMR spectroscopy and in the solid state by single crystal X-ray analyses of tweezer-complexes with linear and macrocyclic oligo-imides. This work provides clear-cut confirmation of polyimide chain-folding and adjacent-tweezer-binding. It also reveals a new and entirely unexpected mechanism for sequence-recognition which, by analogy with a related process in biomolecular information processing, may be termed "frameshift-reading". The ability of one particular tweezer-molecule to detect, with exceptionally high sensitivity, long-range sequence-information in chain-folding aromatic copolyimides, is readily explained by this novel process.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Managing a construction project supply chain effectively and efficiently is extremely difficult due to involvement of numerous sectors that are supported by ineffective communication system. An efficient construction supply chain system ensures the delivery of materials and other services to construction site while minimising costs and rewarding all sectors based on value added to the supply chain. The advancement of information, communication and wireless technologies is driving construction companies to deploy supply chain management strategies to seek better outputs. As part of the emerging wireless technologies, contextaware computing capability represents the next generation of ICT to the construction services. Conceptually, context-awareness could be integrated with Web Services in order to ensure the delivery of pertinent information to construction site and enhance construction supply chain collaboration. An initial study has indicated that this integrated system has the potential of serving and improving the construction services delivery through access to context-specific data, information and services on as-needed basis.