1000 resultados para Sistemas de Computação


Relevância:

100.00% 100.00%

Publicador:

Resumo:

In this work will applied the technique of Differential Cryptanalysis, introduced in 1990 by Biham and Shamir, on Papílio s cryptosystem, developed by Karla Ramos, to test and most importantly, to prove its relevance to other block ciphers such as DES, Blowfish and FEAL-N (X). This technique is based on the analysis of differences between plaintext and theirs respective ciphertext, in search of patterns that will assist in the discovery of the subkeys and consequently in the discovery of master key. These differences are obtained by XOR operations. Through this analysis, in addition to obtaining patterns of Pap´ılio, it search to obtain also the main characteristics and behavior of Papilio throughout theirs 16 rounds, identifying and replacing when necessary factors that can be improved in accordance with pre-established definitions of the same, thus providing greater security in the use of his algoritm

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)

Relevância:

100.00% 100.00%

Publicador:

Resumo:

O uso de computação em nuvem tem se tornado cada vez mais intenso nos últimos anos. Devido principalmente a fatores econômicos, a avaliação de desempenho de sistemas de computação em nuvem, principalmente através de simulação, tem sido amplamente aplicada. Neste trabalho se apresenta a inclusão da funcionalidade de modelagem e simulação de computação em nuvem no simulador de ambientes distribuídos iSPD.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

Interactive systems users still face several challenge. Besides current improvements in usability and intuitiveness users have to adapt to the systems proposed to satisfy their needs. For instance, they must learn how to achieve tasks, how to interact with the system, etc. This paper proposes a methodology to improve this situation supporting the use of interactive systems by users. To achieve this goal the approach is based on enriched task models and picture-driven computing. An example based on a text editor illustrates the approach.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

Esta dissertação é fruto de pesquisas, estudos e testes sobre os modelos de representação do conhecimento e os instrumentos utilizados na construção de sistemas especialistas. Visando estabelecer critérios que permitam, a partir da definição das estruturas do conhecimento de um determinado domínio, elaborar um modelo de representação do conhecimento adequado a esse domínio e, também, escolher o instrumento de desenvolvimento e implementação (uma linguaguem ou shell) mais adequado à aplicação em consideração

Relevância:

70.00% 70.00%

Publicador:

Resumo:

Os sistemas baseados em conhecimento estão conquistando espaço entre as tecnologias de informação à medida que o hardware se desenvolve. No ambiente de acirrada competitividade presente nas empresas, esta tecnologia vem apoiá-Ias estrategicamente na agilização, principalmente, do processo decisório e na solução de diversos problemas considerados complexos. Este estudo faz uma investigação das aplicações potenciais da tecnologia de sistemas baseados em conhecimento - ou sistemas .especialistas, como comumente conhecidos -, examinando também as tendências da tecnologia e os impactos e implicações organizacionais provocados quando de sua implementação nas empresas. Um estudo exploratório em empresas brasileiras é acrescido ao perfil de uso e tendências dos sistemas baseados em conhecimento no âmbito nacional

Relevância:

70.00% 70.00%

Publicador:

Resumo:

Trata da questão das tendências da evolução da tecnologia dos Sistemas Especialistas (SE). Discute a situação atual desta tecnologia apresentando motivos para a disseminação limitada desta tecnologia nas organizações. Discute as tendências da tecnologia dos SE. Aborda a disponibilidade futura dos SE, sua integração a novos sistemas de informação e sua difusão seletiva nas organizações. Aponta a oportunidade de cetos segmentos de mercado para os fornecedores desta tecnologia

Relevância:

70.00% 70.00%

Publicador:

Resumo:

Até recentemente, pesquisas sobre a relação do usuário com o uso de tecnologia focaram sistemas utilitários. Com a popularização dos PCs para uso doméstico e o surgimento da Internet, a interação dos usuários com tecnologia deixou de ser vista apenas de forma utilitária, e passou a ser orientada também ao atendimento de necessidades hedônicas, fazendo com que as pesquisas em Tecnologia e Sistemas de Informação (SI) extrapolassem os limites da empresa. As pesquisas no campo tem dado atenção cada vez maior ao uso de sistemas hedônicos, cujo objetivo é satisfazer necessidades como entretenimento e diversão, ao invés de um valor instrumental. Entre os sistemas hedônicos, um em particular recebe cada vez mais destaque, os videogames. Jogos digitais estão presentes por toda a sociedade, existindo uma crescente adoção, institucionalização e ubiquidade dos videogames na vida diária. A nova geração foi criada com jogos interativos e espera o mesmo tipo de interação nos demais SI. Por serem divertidos e envolventes, organizações de diversos setores estão tentando integrar características dos jogos digitais em sistemas organizacionais, com a expectativa de melhorar a motivação e desempenho dos funcionários e engajar consumidores online. Para analisar a relação do usuário com jogos digitais foram desenvolvidas três pesquisas. A primeira aborda os motivos pelos quais o indivíduo se engaja com o jogo, definindo Engajamento com base em desafio, imersão e curiosidade. Já as motivações dos indivíduos são analisadas pelo divertimento e a competição. A partir do momento em que o indivíduo está engajado, é necessário medir a experiência holística que ele experimenta desta interação. Para tanto, a segunda pesquisa analisa os dois construtos mais utilizados no campo para mensurar tal experiência, Flow e Absorção Cognitiva. A última pesquisa se utiliza dos conceitos explorados nas duas anteriores para desenvolver uma rede nomológica para a predição da intenção comportamental do indivíduo para o uso de jogos digitais. Como resultados temos um esclarecimento sobre os fatores que levam a um comportamento engajado dos jogadores, com maior importância do desafio de dominar o jogo e divertimento em relação a competição com outros jogadores. Outra contribuição é a conclusão que Flow e Absorção Cognitiva, dois dos principais construtos utilizados em SI para medir a experiência hedônica do usuário, não são adequados para mensurar tal experiência para videogames. Por fim existe a proposta de uma rede nomológica em um contexto de utilização de um sistema hedônico interativo, permitindo que novos trabalhos possam se beneficiar.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

One of the current challenges of Ubiquitous Computing is the development of complex applications, those are more than simple alarms triggered by sensors or simple systems to configure the environment according to user preferences. Those applications are hard to develop since they are composed by services provided by different middleware and it is needed to know the peculiarities of each of them, mainly the communication and context models. This thesis presents OpenCOPI, a platform which integrates various services providers, including context provision middleware. It provides an unified ontology-based context model, as well as an environment that enable easy development of ubiquitous applications via the definition of semantic workflows that contains the abstract description of the application. Those semantic workflows are converted into concrete workflows, called execution plans. An execution plan consists of a workflow instance containing activities that are automated by a set of Web services. OpenCOPI supports the automatic Web service selection and composition, enabling the use of services provided by distinct middleware in an independent and transparent way. Moreover, this platform also supports execution adaptation in case of service failures, user mobility and degradation of services quality. The validation of OpenCOPI is performed through the development of case studies, specifically applications of the oil industry. In addition, this work evaluates the overhead introduced by OpenCOPI and compares it with the provided benefits, and the efficiency of OpenCOPI s selection and adaptation mechanism

Relevância:

70.00% 70.00%

Publicador:

Resumo:

It bet on the next generation of computers as architecture with multiple processors and/or multicore processors. In this sense there are challenges related to features interconnection, operating frequency, the area on chip, power dissipation, performance and programmability. The mechanism of interconnection and communication it was considered ideal for this type of architecture are the networks-on-chip, due its scalability, reusability and intrinsic parallelism. The networks-on-chip communication is accomplished by transmitting packets that carry data and instructions that represent requests and responses between the processing elements interconnected by the network. The transmission of packets is accomplished as in a pipeline between the routers in the network, from source to destination of the communication, even allowing simultaneous communications between pairs of different sources and destinations. From this fact, it is proposed to transform the entire infrastructure communication of network-on-chip, using the routing mechanisms, arbitration and storage, in a parallel processing system for high performance. In this proposal, the packages are formed by instructions and data that represent the applications, which are executed on routers as well as they are transmitted, using the pipeline and parallel communication transmissions. In contrast, traditional processors are not used, but only single cores that control the access to memory. An implementation of this idea is called IPNoSys (Integrated Processing NoC System), which has an own programming model and a routing algorithm that guarantees the execution of all instructions in the packets, preventing situations of deadlock, livelock and starvation. This architecture provides mechanisms for input and output, interruption and operating system support. As proof of concept was developed a programming environment and a simulator for this architecture in SystemC, which allows configuration of various parameters and to obtain several results to evaluate it

Relevância:

70.00% 70.00%

Publicador:

Resumo:

Although some individual techniques of supervised Machine Learning (ML), also known as classifiers, or algorithms of classification, to supply solutions that, most of the time, are considered efficient, have experimental results gotten with the use of large sets of pattern and/or that they have a expressive amount of irrelevant data or incomplete characteristic, that show a decrease in the efficiency of the precision of these techniques. In other words, such techniques can t do an recognition of patterns of an efficient form in complex problems. With the intention to get better performance and efficiency of these ML techniques, were thought about the idea to using some types of LM algorithms work jointly, thus origin to the term Multi-Classifier System (MCS). The MCS s presents, as component, different of LM algorithms, called of base classifiers, and realized a combination of results gotten for these algorithms to reach the final result. So that the MCS has a better performance that the base classifiers, the results gotten for each base classifier must present an certain diversity, in other words, a difference between the results gotten for each classifier that compose the system. It can be said that it does not make signification to have MCS s whose base classifiers have identical answers to the sames patterns. Although the MCS s present better results that the individually systems, has always the search to improve the results gotten for this type of system. Aim at this improvement and a better consistency in the results, as well as a larger diversity of the classifiers of a MCS, comes being recently searched methodologies that present as characteristic the use of weights, or confidence values. These weights can describe the importance that certain classifier supplied when associating with each pattern to a determined class. These weights still are used, in associate with the exits of the classifiers, during the process of recognition (use) of the MCS s. Exist different ways of calculating these weights and can be divided in two categories: the static weights and the dynamic weights. The first category of weights is characterizes for not having the modification of its values during the classification process, different it occurs with the second category, where the values suffers modifications during the classification process. In this work an analysis will be made to verify if the use of the weights, statics as much as dynamics, they can increase the perfomance of the MCS s in comparison with the individually systems. Moreover, will be made an analysis in the diversity gotten for the MCS s, for this mode verify if it has some relation between the use of the weights in the MCS s with different levels of diversity

Relevância:

70.00% 70.00%

Publicador:

Resumo:

In systems that combine the outputs of classification methods (combination systems), such as ensembles and multi-agent systems, one of the main constraints is that the base components (classifiers or agents) should be diverse among themselves. In other words, there is clearly no accuracy gain in a system that is composed of a set of identical base components. One way of increasing diversity is through the use of feature selection or data distribution methods in combination systems. In this work, an investigation of the impact of using data distribution methods among the components of combination systems will be performed. In this investigation, different methods of data distribution will be used and an analysis of the combination systems, using several different configurations, will be performed. As a result of this analysis, it is aimed to detect which combination systems are more suitable to use feature distribution among the components

Relevância:

70.00% 70.00%

Publicador:

Resumo:

Pervasive applications use context provision middleware support as infrastructures to provide context information. Typically, those applications use communication publish/subscribe to eliminate the direct coupling between components and to allow the selective information dissemination based in the interests of the communicating elements. The use of composite events mechanisms together with such middlewares to aggregate individual low level events, originating from of heterogeneous sources, in high level context information relevant for the application. CES (Composite Event System) is a composite events mechanism that works simultaneously in cooperation with several context provision middlewares. With that integration, applications use CES to subscribe to composite events and CES, in turn, subscribes to the primitive events in the appropriate underlying middlewares and notifies the applications when the composed events happen. Furthermore, CES offers a language with a group of operators for the definition of composite events that also allows context information sharing

Relevância:

70.00% 70.00%

Publicador:

Resumo:

The increase of capacity to integrate transistors permitted to develop completed systems, with several components, in single chip, they are called SoC (System-on-Chip). However, the interconnection subsystem cans influence the scalability of SoCs, like buses, or can be an ad hoc solution, like bus hierarchy. Thus, the ideal interconnection subsystem to SoCs is the Network-on-Chip (NoC). The NoCs permit to use simultaneous point-to-point channels between components and they can be reused in other projects. However, the NoCs can raise the complexity of project, the area in chip and the dissipated power. Thus, it is necessary or to modify the way how to use them or to change the development paradigm. Thus, a system based on NoC is proposed, where the applications are described through packages and performed in each router between source and destination, without traditional processors. To perform applications, independent of number of instructions and of the NoC dimensions, it was developed the spiral complement algorithm, which finds other destination until all instructions has been performed. Therefore, the objective is to study the viability of development that system, denominated IPNoSys system. In this study, it was developed a tool in SystemC, using accurate cycle, to simulate the system that performs applications, which was implemented in a package description language, also developed to this study. Through the simulation tool, several result were obtained that could be used to evaluate the system performance. The methodology used to describe the application corresponds to transform the high level application in data-flow graph that become one or more packages. This methodology was used in three applications: a counter, DCT-2D and float add. The counter was used to evaluate a deadlock solution and to perform parallel application. The DCT was used to compare to STORM platform. Finally, the float add aimed to evaluate the efficiency of the software routine to perform a unimplemented hardware instruction. The results from simulation confirm the viability of development of IPNoSys system. They showed that is possible to perform application described in packages, sequentially or parallelly, without interruptions caused by deadlock, and also showed that the execution time of IPNoSys is more efficient than the STORM platform