142 resultados para virtualization


Relevância:

10.00% 10.00%

Publicador:

Resumo:

Pós-graduação em Ciência da Computação - IBILCE

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Pós-graduação em Ciência da Computação - IBILCE

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Pós-graduação em Desenvolvimento Humano e Tecnologias - IBRC

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The displacement of writing emerged in urban centers due to its investment in technology and has been developed in major newspapers around the world. With the advent of the internet and technological development, the writing went from analog to digital. That made it no longer required the presence of the journalist in the physical redaction means modifying the various management structures of a writing routine and professional. This happened to capture, edit and distribute your notes, news and reports in online mode. This monograph aims to understand and define the macro change of virtualization in journalism, considering the stages of production of Le Monde Diplomatique, a newspaper whose writing is in process of deterritorialization, the people involved in this activity and the resources used and contributing to the debate on new forms of organization of an editorial environment

Relevância:

10.00% 10.00%

Publicador:

Resumo:

This article analyzes the human bonds in the context of contemporary society from the work of the polish sociologist Zygmunt Bauman. Taking as starting point the bond of friendship and relational virtualization process motivated by the development of new technologies, it was observed the occurrence in the present world an affective-relational impoverishment. As an alternative to this, it is argued the exercise of building and maintenance of dense bonds that provide an alternative to the bind model of momentary satisfaction, this time marks. Finally, it is indicated that the discussion of friendship necessarily imply on the problematization of contemporary ethics which will provide the foundation for building a new affective political.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The objective of this study is a explanation about the action possibilities of the Public Relations professional as a manager of communication between the musical celebrity and his publics. For this purpose, a bibliographical research was made about Internet question like his history, the evolution of Web 2.0 and even about different kinds of virtualization, introducing with interactivity. Other researches was been made like social culture topics and even the studies of Public Relations actions in the artistic scene, acting like a manager of different publics relationships in the virtual contacts. A study case with the virtual communication tools employed in the career of Hugo e Tiago, one of the country artists in Brazil, is also in this studies. Otherwise, the study explains about the different ways of Public Relations actions, his upgrades along with the new communication tools and his new types of making relations with more and more different publics. Working with new strategies and keeping his objectives in the organization and public relationship, Public Relations brands a new vision to this type of communication professional, who becomes even more important in the actual organizations scenario

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Network virtualization is a promising technique for building the Internet of the future since it enables the low cost introduction of new features into network elements. An open issue in such virtualization is how to effect an efficient mapping of virtual network elements onto those of the existing physical network, also called the substrate network. Mapping is an NP-hard problem and existing solutions ignore various real network characteristics in order to solve the problem in a reasonable time frame. This paper introduces new algorithms to solve this problem based on 0–1 integer linear programming, algorithms based on a whole new set of network parameters not taken into account by previous proposals. Approximative algorithms proposed here allow the mapping of virtual networks on large network substrates. Simulation experiments give evidence of the efficiency of the proposed algorithms.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

[ES] Las necesidades básicas de las empresas suelen ser las mismas, ya sea una empresa grande que pequeña, la infraestructura sobre la que montan sus procesos de negocio y las aplicaciones para gestionarlos suelen ser casi iguales. Si dividimos la infraestructura TIC de una empresa en hardware, sistema y aplicaciones, podemos ver que en la mayoría de ellas el sistema es casi idéntico. Además, gracias a la virtualización, que ha entrado de manera arrolladora en el mundo de la informática, podemos independizar totalmente el software del hardware, de forma que obtenemos una flexibilidad enorme a la hora de planificar despliegues de infraestructura. Sobre estas dos ideas, uniformidad de sistema e independencia de hardware, son sobre las que se va a desarrollar el siguiente TFG. Para el desarrollo de la primera de ellas se realizará el estudio de la infraestructura básica ( sistema) que cualquier empresa suele tener. Se intentará dar una solución que sea válida para una gran cantidad de empresas de nuestro entorno y se realizará el diseño del mismo. Con la segunda idea desarrollaremos un sistema basado en servicios, que sea lo suficientemente completa para poder dar respuesta a las necesidades vistas pero, a su vez, suficientemente flexible para que el crecimiento en capacidades o servicios se pueda realizar de forma sencilla sin que la estructura del sistema, o sus módulos deban modificarse para realizarlos. Por tanto, vamos a realizar un diseño integral y completa, de forma que será tanto de hardware como de software, haciendo énfasis en la integración de los sistemas y la interrelación entre los distintos elementos de ellos. Se dará, a su vez, la valoración económica del mismo. Por último, y como ejemplo de la flexibilidad del diseño elegido veremos dos modificaciones sobre el diseño original. El primero de ellos será una ampliación para dar mayor seguridad en cuanto a redundancia de almacenamiento y, ya en un paso definitivo, montar un CPD remoto. El segundo de ellos será un diseño de bajo coste, en el que, mantenimiento los mismos servicios, bajaremos el coste del diseño con productos con algo menos de prestaciones, pero manteniendo la solución en conjunto unos altos niveles de calidad y servicio.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

[ES] Este trabajo consiste en la creación de un sistema de gestión abierto de cita previa, surge de la necesidad del personal de las Administración Tributaria de Canarias que trabajan en las oficinas de atención tributaria de tener un sistema que facilite al contribuyente la forma de gestionar una cita previa para solicitar atención tributaria, además de facilitar al coordinador, la gestión de citas solicitadas por los contribuyentes. Para ello se ha realizado un análisis previo para cubrir todos los requisitos que eran necesarios para que se cubrieran todas las necesidades del sistema y que fuera totalmente escalable. Para el desarrollo del sistema se ha utilizado la plataforma Java EE que dota de todas las herramientas necesarias para la creación de un software de estas características, además de numerosas herramientas que ayudan a la consecución satisfactoria del sistema, como pueden ser Hibernate y Spring. Se ha utilizado un sistema de virtualización para tratar de imitar lo máximo posible la forma de desarrollo que tiene la empresa para que así la integración del sistema sea absoluto, por eso se han utilizado las mismas herramientas que utilizan y la misma estructura de base de datos. Además de las diversas herramientas se han utilizado distintos patrones de diseño para el desarrollo del software como pueden ser: Front Controller, Facade y Data Access Object además del patrón Modelo-Vista-Controlador para la arquitectura de este sistema. La utilización de estos patrones ayuda a que el mantenimiento de este sistema sea mucho más óptimo.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

[ES]Diseño e instalación de un servicio web de tienda virtual en un entorno de computación en clúster para proporcionar alta disponibilidad y balanceo de carga del mencionado servicio. La infraestructura requerida para este despliegue será virtual, utilizándose como plataforma de virtualización KVM. Las tareas se organizan en : 1.Tareas para la creación de un clúster de balanceo de carga donde una máquina será la encargada de recibir las peticiones de los clientes y de redirigirlas a los servidores web según la carga de éstos. 2. Tareas para la creación de un clúster de alta disponibilidad para que en caso de fallo siempre haya un servidor de base de datos que pueda atender a las peticiones de los servidores web.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Le nuove teorie di rete come Software Defined Networking Network Function Virtualization, insieme alle teorie Cognitive/Autonomics consentono di abilitare scenari futuri “disruptive” di rete. Lo scopo di questa tesi è quello di esplorare questi scenari futuri e di capire il ruolo della migrazione di funzioni di rete, sotto forma di Virtual Machine. Si vuole affrontare la migrazione di Virtual Machine dal punto di vista delle performance, ma anche come strumento di gestione delle risorse in uno scenario di rete d'accesso autonomica.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

During the last few decades an unprecedented technological growth has been at the center of the embedded systems design paramount, with Moore’s Law being the leading factor of this trend. Today in fact an ever increasing number of cores can be integrated on the same die, marking the transition from state-of-the-art multi-core chips to the new many-core design paradigm. Despite the extraordinarily high computing power, the complexity of many-core chips opens the door to several challenges. As a result of the increased silicon density of modern Systems-on-a-Chip (SoC), the design space exploration needed to find the best design has exploded and hardware designers are in fact facing the problem of a huge design space. Virtual Platforms have always been used to enable hardware-software co-design, but today they are facing with the huge complexity of both hardware and software systems. In this thesis two different research works on Virtual Platforms are presented: the first one is intended for the hardware developer, to easily allow complex cycle accurate simulations of many-core SoCs. The second work exploits the parallel computing power of off-the-shelf General Purpose Graphics Processing Units (GPGPUs), with the goal of an increased simulation speed. The term Virtualization can be used in the context of many-core systems not only to refer to the aforementioned hardware emulation tools (Virtual Platforms), but also for two other main purposes: 1) to help the programmer to achieve the maximum possible performance of an application, by hiding the complexity of the underlying hardware. 2) to efficiently exploit the high parallel hardware of many-core chips in environments with multiple active Virtual Machines. This thesis is focused on virtualization techniques with the goal to mitigate, and overtake when possible, some of the challenges introduced by the many-core design paradigm.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In questo elaborato andremo a descrivere, sia a livello strutturale che implementativo, il progetto sperimentale da noi ideato volto alla creazione di un laboratorio virtuale di testing. Lo scopo principale del laboratorio è appunto quello di delocalizzare la fase preliminare di testing di un'applicazione; nel nostro caso specifico siamo partiti dallo scenario riguardante la rete ferroviaria ed abbiamo preso in esame il software utilizzato per la gestione del traffico. Il vincolo principale che grava su questo progetto è che l'applicazione da collaudare non deve minimamente essere modificata, e ciò significa che occorre creare intorno ad essa un ambiente di testing conforme alle caratteristiche in cui il software si aspetta di operare. Ed è qui che entra in gioco la virtualizzazione.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Questo documento si interroga sulle nuove possibilità offerte agli operatori del mondo delle Reti di Telecomunicazioni dai paradigmi di Network Functions Virtualization, Cloud Computing e Software Defined Networking: questi sono nuovi approcci che permettono la creazione di reti dinamiche e altamente programmabili, senza disdegnare troppo il lato prestazionale. L'intento finale è valutare se con un approccio di questo genere si possano implementare dinamicamente delle concatenazioni di servizi di rete e se le prestazioni finali rispecchiano ciò che viene teorizzato dai suddetti paradigmi. Tutto ciò viene valutato per cercare una soluzione efficace al problema dell'ossificazione di Internet: infatti le applicazioni di rete, dette middle-boxes, comportano costi elevati, situazioni di dipendenza dal vendor e staticità delle reti stesse, portando all'impossibilità per i providers di sviluppare nuovi servizi. Il caso di studio si basa proprio su una rete che implementa questi nuovi paradigmi: si farà infatti riferimento a due diverse topologie, una relativa al Livello L2 del modello OSI (cioè lo strato di collegamento) e una al Livello L3 (strato di rete). Le misure effettuate infine mostrano come le potenzialità teorizzate siano decisamente interessanti e innovative, aprendo un ventaglio di infinite possibilità per il futuro sviluppo di questo settore.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Nowadays, data handling and data analysis in High Energy Physics requires a vast amount of computational power and storage. In particular, the world-wide LHC Com- puting Grid (LCG), an infrastructure and pool of services developed and deployed by a ample community of physicists and computer scientists, has demonstrated to be a game changer in the efficiency of data analyses during Run-I at the LHC, playing a crucial role in the Higgs boson discovery. Recently, the Cloud computing paradigm is emerging and reaching a considerable adoption level by many different scientific organizations and not only. Cloud allows to access and utilize not-owned large computing resources shared among many scientific communities. Considering the challenging requirements of LHC physics in Run-II and beyond, the LHC computing community is interested in exploring Clouds and see whether they can provide a complementary approach - or even a valid alternative - to the existing technological solutions based on Grid. In the LHC community, several experiments have been adopting Cloud approaches, and in particular the experience of the CMS experiment is of relevance to this thesis. The LHC Run-II has just started, and Cloud-based solutions are already in production for CMS. However, other approaches of Cloud usage are being thought of and are at the prototype level, as the work done in this thesis. This effort is of paramount importance to be able to equip CMS with the capability to elastically and flexibly access and utilize the computing resources needed to face the challenges of Run-III and Run-IV. The main purpose of this thesis is to present forefront Cloud approaches that allow the CMS experiment to extend to on-demand resources dynamically allocated as needed. Moreover, a direct access to Cloud resources is presented as suitable use case to face up with the CMS experiment needs. Chapter 1 presents an overview of High Energy Physics at the LHC and of the CMS experience in Run-I, as well as preparation for Run-II. Chapter 2 describes the current CMS Computing Model, and Chapter 3 provides Cloud approaches pursued and used within the CMS Collaboration. Chapter 4 and Chapter 5 discuss the original and forefront work done in this thesis to develop and test working prototypes of elastic extensions of CMS computing resources on Clouds, and HEP Computing “as a Service”. The impact of such work on a benchmark CMS physics use-cases is also demonstrated.