863 resultados para Business enterprises - Computer networks
Resumo:
Using panel data of 57 countries during the period of 1995-2012, this study investigates the impact of intellectual property rights (IPR) processes on productivity growth. The IPR processes are decomposed into three stages, innovation process, commercialization process, and IPR protection process. Our results suggest that better IPR protection is directly associated with productivity improvement only in developed economies. In addition, the contribution of IPR processes on growth through foreign direct investment (FDI) appears to be very limited. Only FDI inflows in developed countries which help to create a better innovative capability lead to a higher growth. And in connection with FDI outflows, only IPR protection and commercialization processes are proven to improve productivity in the case of developing countries, particularly when the country acts as the investing country.
Resumo:
We analyze competition in emerging markets between firms in developing and developed countries from the viewpoint of the boundaries of the firm. Although indigenous firms generally face a disadvantage in technology compared with foreign firms, they have an advantage in marketing as local firms. Moreover, they have opportunities to leave weaker fields to independent specialized firms and use lower wages. On the other hand, foreign firms also have their own advantages and disadvantages for growth. Therefore, entry conditions for indigenous firms can vary greatly depending on the situation. We classify these conditions into eight cases by developing a model and showing each boundary choice for indigenous firms.
Resumo:
This study compares the innovation process of a privately-owned enterprise and a state-owned enterprise in China using their patent data. Huawei and ZTE were selected for this study because they experienced the same historical environment in the same industry from the same region in China leaving their owner types as their critical difference. This study investigates the difference in the innovation process in R&D between a privately-owned and a state-owned enterprise by analyzing (1) domestic and international patent application pattern, (2) co-application and co-applicants, (3) knowledge accumulation inside Huawei and ZTE, and (4) knowledge spillover to domestic and foreign firms.
Resumo:
The objective of the present study is to examine the determinants of ISO 9001 certification, focusing on the effect of Product-related Environmental Regulations on Chemicals (PRERCs) and FDI using the answers to several questions in our Vietnam survey conducted from December 2011 to January 2012. Our findings suggest that PRERCs may help with the improvement in quality control of Vietnamese firms. If Vietnamese manufacturing firms with ISO 9001 certification are more likely to adopt ISO 14001, as well as firms in developed countries, our results indicate that the European chemical regulations may assist in the reduction of various environmental impacts in Vietnam. In addition, we found that FDI promotes the adoption of ISO 9001. If FDI firms in Vietnam certify ISO 14001 after the adoption of ISO 9001, as in the case of Malaysia and the developed economies, FDI firms may also be able to improve environmental performance as a result of ISO 14001.
Resumo:
Research on multinational firms’ activity has been conducted widely since late 1980s. The literature is differentiated into three types: horizontal FDI, vertical FDI, and three-country FDI, represented by export platform FDI. There are other methods of differentiation of the literature by approach, for example, the pure theory approach represented by Krugman and Melitz and the numerical simulation approach represented by Markusen. This paper surveys Markusen type literature by firm type. There is little literature focused on intermediate goods trade, although intermediate goods trade is considered to be strongly related to the production patterns of MNEs. In this paper, we introduce a model to explicitly treat intermediate goods trade and present simulation analysis for empirical estimation.
Resumo:
This paper develops a quantitative measure of allocation efficiency, which is an extension of the dynamic Olley-Pakes productivity decomposition proposed by Melitz and Polanec (2015). The extended measure enables the simultaneous capture of the degree of misallocation within a group and between groups and parallel to capturing the contribution of entering and exiting firms to aggregate productivity growth. This measure empirically assesses the degree of misallocation in China using manufacturing firm-level data from 2004 to 2007. Misallocation among industrial sectors has been found to increase over time, and allocation efficiency within an industry has been found to worsen in industries that use more capital and have firms with relatively higher state-owned market shares. Allocation efficiency among three ownership sectors (state-owned, domestic private, and foreign sectors) tends to improve in industries wherein the market share moves from a less-productive state-owned sector to a more productive private sector.
Resumo:
This study extends Melitz's model with heterogeneous firms by introducing shared fixed costs in a marketplace. It aims to explain heterogeneous firms' choice between traditional marketplaces and modern distribution channels on the basis of their productivities. The results reveal that the co-existence of a traditional marketplace and modern distribution channels improves social welfare. In addition, a deregulation policy for firm entry outside a marketplace and accumulation of human capital are factors that contribute to improve the social welfare.
Resumo:
The emerging use of real-time 3D-based multimedia applications imposes strict quality of service (QoS) requirements on both access and core networks. These requirements and their impact to provide end-to-end 3D videoconferencing services have been studied within the Spanish-funded VISION project, where different scenarios were implemented showing an agile stereoscopic video call that might be offered to the general public in the near future. In view of the requirements, we designed an integrated access and core converged network architecture which provides the requested QoS to end-to-end IP sessions. Novel functional blocks are proposed to control core optical networks, the functionality of the standard ones is redefined, and the signaling improved to better meet the requirements of future multimedia services. An experimental test-bed to assess the feasibility of the solution was also deployed. In such test-bed, set-up and release of end-to-end sessions meeting specific QoS requirements are shown and the impact of QoS degradation in terms of the user perceived quality degradation is quantified. In addition, scalability results show that the proposed signaling architecture is able to cope with large number of requests introducing almost negligible delay.
Resumo:
Cloud computing is one the most relevant computing paradigms available nowadays. Its adoption has increased during last years due to the large investment and research from business enterprises and academia institutions. Among all the services cloud providers usually offer, Infrastructure as a Service has reached its momentum for solving HPC problems in a more dynamic way without the need of expensive investments. The integration of a large number of providers is a major goal as it enables the improvement of the quality of the selected resources in terms of pricing, speed, redundancy, etc. In this paper, we propose a system architecture, based on semantic solutions, to build an interoperable scheduler for federated clouds that works with several IaaS (Infrastructure as a Service) providers in a uniform way. Based on this architecture we implement a proof-of-concept prototype and test it with two different cloud solutions to provide some experimental results about the viability of our approach.
Resumo:
This paper hallmarks the most relevant contributions carried out by the authors in the VOTESCRIPT project (TIC2000-1630-C02). The main goal of this project was the analysis, definition and implementation of a system which copes with every phases and elements existing in a process of electronic voting using computer networks. A summary of the main criticisms of electronic voting is presented to disclose that the most relevant voting schemes only take into account a technological perspective, just trying to imitate the conventional voting schemes. Nevertheless in these proposals important aspects such individual and global verification are not properly undertaken. The paper includes the proposed solutions of the project to solve these mentioned problems.
Resumo:
Este artículo presenta los aspectos más relevantes del trabajo realizado por los autores dentro del proyecto VOTESCRIPT (TIC2000-1630-C02). El objetivo principal de este proyecto fue el análisis, definición e implementación de un sistema que abarcara todas las fases y elementos existentes en un proceso de votación electrónica sobre redes de ordenadores. El artículo incluye las soluciones propuestas dentro del proyecto. This paper hallmarks the most relevant contributions carried out by the authors in the VOTESCRIPT project (TIC2000-1630-C02). The main goal of this project was the analysis, definition and implementation of a system, which copes with every phases and elements existing in a process of electronic voting using computer networks. The paper includes the proposed solutions of the project to solve these problems.
Resumo:
A system for simultaneous 2D estimation of rectangular room and transceiver localization is proposed. The system is based on two radio transceivers, both capable of full duplex operations (simultaneous transmission and reception). This property enables measurements of channel impulse response (CIR) at the same place the signal is transmitted (generated), commonly known as self-to-self CIR. Another novelty of the proposed system is the spatial CIR discrimination that is possible with the receiver antenna design which consists of eight sectorized antennas with 45° aperture in the horizontal plane and total coverage equal to the isotropic one. The dimensions of a rectangular room are reconstructed directly from spatial radio impulse responses by extracting the information regarding round trip time (RTT). Using radar approach estimation of walls and corners positions is derived. Tests using measured data were performed, and the simulation results confirm the feasibility of the approach.
Resumo:
Desde la aparición de Internet, hace ya más de 20 años ha existido por parte de diversos sectores de la sociedad, científicos, empresas, usuarios, etc. la inquietud por la aplicación de esta tecnología a lo que se ha dado en llamar “El Internet de las Cosas”, que no es más que el control a distancia de cualquier elemento útil o necesario para la vida cotidiana y la industria. Sin embargo el desarrollo masivo de aplicaciones orientadas a esto, no ha evolucionado hasta que no se han producido avances importantes en dos campos: por un lado, en las Redes Inalámbricas de Sensores (WSN), redes compuestas por un conjunto de pequeños dispositivos capaces de transmitir la información que recogen, haciéndola llegar desde su propia red inalámbrica, a otras de amplia cobertura y por otro con la miniaturización cada vez mayor de dispositivos capaces de tener una autonomía suficiente como para procesar datos e interconectarse entre sí. Al igual que en las redes de ordenadores convencionales, las WSN se pueden ver comprometidas en lo que a seguridad se refiere, ya que la masiva implementación de estas redes hará que millones de Terabytes de datos, muchas veces comprometidos o sometidos a estrictas Leyes de protección de los mismos, circulen en la sociedad de la información, de forma que lo que nace como una ventaja muy interesante para sus usuarios, puede convertirse en una pesadilla debido a la amenaza constante hacia los servicios mínimos de seguridad que las compañías desarrolladoras han de garantizar a los usuarios de sus aplicaciones. Éstas, y con el objetivo de proveer un ámbito de seguridad mínimo, deben de realizar un minucioso estudio de la aplicación en particular que se quiere ofrecer con una WSN y también de las características específicas de la red ya que, al estar formadas por dispositivos prácticamente diminutos, pueden tener ciertas limitaciones en cuanto al tamaño de la batería, capacidad de procesamiento, memoria, etc. El presente proyecto desarrolla una aplicación, única, ya que en la actualidad no existe un software con similares características y que aporta un avance importante en dos campos principalmente: por un lado ayudará a los usuarios que deseen desplegar una aplicación en una red WSN a determinar de forma automática cuales son los mecanismos y servicios específicos de seguridad que se han de implementar en dicha red para esa aplicación concreta y, por otro lado proporcionará un apoyo extra a expertos de seguridad que estén investigando en la materia ya que, servirá de plataforma de pruebas para centralizar la información sobre seguridad que se tengan en ese momento en una base de conocimientos única, proporcionando también un método útil de prueba para posibles escenarios virtuales. ABSTRACT. It has been more than 20 years since the Internet appeared and with it, scientists, companies, users, etc. have been wanted to apply this technology to their environment which means to control remotely devices, which are useful for the industry or aspects of the daily life. However, the huge development of these applications oriented to that use, has not evolve till some important researches has been occurred in two fields: on one hand, the field of the Wireless Sensor Networks (WSN) which are networks composed of little devices that are able to transmit the information that they gather making it to pass through from their wireless network to other wider networks and on the other hand with the increase of the miniaturization of the devices which are able to work in autonomous mode so that to process data and connect to each other. WSN could be compromised in the matter of security as well as the conventional computer networks, due to the massive implementation of this kind of networks will cause that millions of Terabytes of data will be going around in the information society, thus what it is thought at first as an interesting advantage for people, could turn to be a nightmare because of the continuous threat to the minimal security services that developing companies must guarantee their applications users. These companies, and with the aim to provide a minimal security realm, they have to do a strict research about the application that they want to implement in one WSN and the specific characteristics of the network as they are made by tiny devices so that they could have certain limitations related to the battery, throughput, memory, etc. This project develops a unique application since, nowadays, there is not any software with similar characteristics and it will be really helpful in mainly two areas: on one side, it will help users who want to deploy an application in one WSN to determine in an automatically way, which ones security services and mechanisms are those which is necessary to implement in that network for the concrete application and, on the other side, it will provide an extra help for the security experts who are researching in wireless sensor network security so that ti will an exceptional platform in order to centralize information about security in the Wireless Sensor Networks in an exclusive knowledge base, providing at the same time a useful method to test virtual scenarios.
Resumo:
El presente documento pretende ofrecer una visión general del estado del conjunto de herramientas disponibles para el análisis y explotación de vulnerabilidades en sistemas informáticos y más concretamente en redes de ordenadores. Por un lado se ha procedido a describir analíticamente el conjunto de herramientas de software libre que se ofrecen en la actualidad para analizar y detectar vulnerabilidades en sistemas informáticos. Se ha descrito el funcionamiento, las opciones, y la motivación de uso para dichas herramientas, comparándolas con otras en algunos casos, describiendo sus diferencias en otros, y justificando su elección en todos ellos. Por otro lado se ha procedido a utilizar dichas herramientas analizadas con el objetivo de desarrollar ejemplos concretos de uso con sus diferentes parámetros seleccionados observando su comportamiento y tratando de discernir qué datos son útiles para obtener información acerca de las vulnerabilidades existentes en el sistema. Además, se ha desarrollado un caso práctico en el que se pone en práctica el conocimiento teórico presentado de forma que el lector sea capaz de asentar lo aprendido comprobando mediante un caso real la utilidad de las herramientas descritas. Los resultados obtenidos han demostrado que el análisis y detección de vulnerabilidades por parte de un administrador de sistemas competente permite ofrecer a la organización en cuestión un conjunto de técnicas para mejorar su seguridad informática y así evitar problemas con potenciales atacantes. ABSTRACT. This paper tries to provide an overview of the features of the set of tools available for the analysis and exploitation of vulnerabilities in computer systems and more specifically in computer networks. On the one hand we pretend analytically describe the set of free software tools that are offered today to analyze and detect vulnerabilities in computer systems. We have described the operation, options, and motivation to use these tools in comparison with other in some case, describing their differences in others, and justifying them in all cases. On the other hand we proceeded to use these analyzed tools in order to develop concrete examples of use with different parameters selected by observing their behavior and trying to discern what data are useful for obtaining information on existing vulnerabilities in the system. In addition, we have developed a practical case in which we put in practice the theoretical knowledge presented so that the reader is able to settle what has been learned through a real case verifying the usefulness of the tools previously described. The results have shown that vulnerabilities analysis and detection made by a competent system administrator can provide to an organization a set of techniques to improve its systems and avoid any potential attacker.
Resumo:
En este proyecto se hace un análisis en profundidad de las técnicas de ataque a las redes de ordenadores conocidas como APTs (Advanced Persistent Threats), viendo cuál es el impacto que pueden llegar a tener en los equipos de una empresa y el posible robo de información y pérdida monetaria que puede llevar asociada. Para hacer esta introspección veremos qué técnicas utilizan los atacantes para introducir el malware en la red y también cómo dicho malware escala privilegios, obtiene información privilegiada y se mantiene oculto. Además, y cómo parte experimental de este proyecto se ha desarrollado una plataforma para la detección de malware de una red en base a las webs, URLs e IPs que visitan los nodos que la componen. Obtendremos esta visión gracias a la extracción de los logs y registros de DNS de consulta de la compañía, sobre los que realizaremos un análisis exhaustivo. Para poder inferir correctamente qué equipos están infectados o no se ha utilizado un algoritmo de desarrollo propio inspirado en la técnica Belief Propagation (“Propagación basada en creencia”) que ya ha sido usada antes por desarrolladores cómo los de los Álamos en Nuevo México (Estados Unidos) para fines similares a los que aquí se muestran. Además, para mejorar la velocidad de inferencia y el rendimiento del sistema se propone un algoritmo adaptado a la plataforma Hadoop de Apache, por lo que se modifica el paradigma de programación habitual y se busca un nuevo paradigma conocido como MapReduce que consiste en la división de la información en conceptos clave-valor. Por una parte, los algoritmos que existen basados en Belief Propagation para el descubrimiento de malware son propietarios y no han sido publicados completamente hasta la fecha, por otra parte, estos algoritmos aún no han sido adaptados a Hadoop ni a ningún modelo de programación distribuida aspecto que se abordará en este proyecto. No es propósito de este proyecto desarrollar una plataforma comercial o funcionalmente completa, sino estudiar el problema de las APTs y una implementación que demuestre que la plataforma mencionada es factible de implementar. Este proyecto abre, a su vez, un horizonte nuevo de investigación en el campo de la adaptación al modelo MapReduce de algoritmos del tipo Belief Propagation basados en la detección del malware mediante registros DNS. ABSTRACT. This project makes an in-depth investigation about problems related to APT in computer networks nowadays, seeing how much damage could they inflict on the hosts of a Company and how much monetary and information loss may they cause. In our investigation we will find what techniques are generally applied by attackers to inject malware into networks and how this malware escalates its privileges, extracts privileged information and stays hidden. As the main part of this Project, this paper shows how to develop and configure a platform that could detect malware from URLs and IPs visited by the hosts of the network. This information can be extracted from the logs and DNS query records of the Company, on which we will make an analysis in depth. A self-developed algorithm inspired on Belief Propagation technique has been used to infer which hosts are infected and which are not. This technique has been used before by developers of Los Alamos Lab (New Mexico, USA) for similar purposes. Moreover, this project proposes an algorithm adapted to Apache Hadoop Platform in order to improve the inference speed and system performance. This platform replaces the traditional coding paradigm by a new paradigm called MapReduce which splits and shares information among hosts and uses key-value tokens. On the one hand, existing algorithms based on Belief Propagation are part of owner software and they have not been published yet because they have been patented due to the huge economic benefits they could give. On the other hand these algorithms have neither been adapted to Hadoop nor to other distributed coding paradigms. This situation turn the challenge into a complicated problem and could lead to a dramatic increase of its installation difficulty on a client corporation. The purpose of this Project is to develop a complete and 100% functional brand platform. Herein, show a short summary of the APT problem will be presented and make an effort will be made to demonstrate the viability of an APT discovering platform. At the same time, this project opens up new horizons of investigation about adapting Belief Propagation algorithms to the MapReduce model and about malware detection with DNS records.