681 resultados para cluster computing
Resumo:
Hoy en día el uso de las TIC en una empresa se ha convertido en indispensable sin importar su tamaño o industria a tal punto que ya no es una ventaja competitiva el simple hecho de usarla sino que la clave es determinar cuál es la mejor para la empresa. Existen en el mercado servicios en-línea que igualan o sobrepasan los estándares de tecnologías pagadas que incluso deben ser instaladas dentro de la empresa consumiendo recursos y en algunos casos subutilizándolos, los servicios del cloud computing poseen la flexibilidad y capacidad que la empresa requiere de acuerdo a sus necesidades de crecimiento y de cambio continuo que el mercado exige en la actualidad. Algunas empresas no invierten en dichas tecnologías porque no saben que pueden accederlas a costos bajos e incluso en algunos casos sin pagar un solo centavo y pierden competitividad y productividad por simple desconocimiento. El uso de servicios SaaS e IaaS afecta al flujo de caja de manera positiva comparada con una inversión en infraestructura propia, sobre todo en empresas que están empezando a funcionar o en aquellas que están en proceso de renovación tecnológica. Después de realizada esta investigación se recomienda el uso de los servicios de cloud computing, pero es evidente que no son útiles para todas las empresas y la utilidad depende de la criticidad de las aplicaciones o infraestructuras, el momento por el que pasa la compañía, tamaño de la empresa, ¿Cloud pública o Cloud privada?, en ocasiones, virtualizar también es racionalizar y la visión integral de la seguridad. Por otro lado, después de tabular los datos obtenidos en la encuesta número 2 se evidenció que el lanzamiento del Servicio en la nube de EcuFlow incrementó la demanda de esta herramienta en el mercado.
Resumo:
El comercio mundial tiene múltiples actores que está sumamente bien posicionados y otros que buscan nuevas estrategias para mejorar su posicionamiento. Asimismo, las diferencias entre los mercados internacionales, nacionales y/o locales son notables, mientras los unos se expanden a pasos agigantados, los otros lo hacen paso a paso. Por lo tanto, el presente trabajo tiene como objetivo fundamental determinar las estrategias que las empresas del sector de las confecciones del cluster textil de Atuntaqui pueden implementar para hacer frente a la competencia internacional. Con esta referencia, el trabajo se ha dividido en tres capítulos. El primer capítulo, está conformado por un análisis de los principales elementos conceptuales (cadena global de valor, managment dentro de la CGV y cluster industriales); seguido, se presenta un breve recuento de la industria, el comercio mundial y los principales importadores y exportadores. Luego, se analiza las tendencias globales utilizadas por los países de América Latina y el Caribe, entre los que están: plataforma de exportación, clusters y logística internacional. En el segundo capítulo, se analizan las principales estadísticas sobre el comportamiento histórico de las importaciones y exportaciones, la caracterización de las empresas, el ámbito tecnológico, los costos y gastos en los que se ha incurrido en algunos cantones y sobre el mercado laboral. Para finalizar, el tercer capítulo, cuenta con una perspectiva local de la industria en Atuntaqui, con temas, como: el “cluster” de la industria, la caracterización de las empresas, el clima de negocios, el diagnostico FODA y las principales estrategias.
Resumo:
Medical universities and teaching hospitals in Iraq are facing a lack of professional staff due to the ongoing violence that forces them to flee the country. The professionals are now distributed outside the country which reduces the chances for the staff and students to be physically in one place to continue the teaching and limits the efficiency of the consultations in hospitals. A survey was done among students and professional staff in Iraq to find the problems in the learning and clinical systems and how Information and Communication Technology could improve it. The survey has shown that 86% of the participants use the Internet as a learning resource and 25% for clinical purposes while less than 11% of them uses it for collaboration between different institutions. A web-based collaborative tool is proposed to improve the teaching and clinical system. The tool helps the users to collaborate remotely to increase the quality of the learning system as well as it can be used for remote medical consultation in hospitals.
Computing the continuous-spectrum linearised bounded standing wave on a plane bed of arbitrary slope
Resumo:
One among the most influential and popular data mining methods is the k-Means algorithm for cluster analysis. Techniques for improving the efficiency of k-Means have been largely explored in two main directions. The amount of computation can be significantly reduced by adopting geometrical constraints and an efficient data structure, notably a multidimensional binary search tree (KD-Tree). These techniques allow to reduce the number of distance computations the algorithm performs at each iteration. A second direction is parallel processing, where data and computation loads are distributed over many processing nodes. However, little work has been done to provide a parallel formulation of the efficient sequential techniques based on KD-Trees. Such approaches are expected to have an irregular distribution of computation load and can suffer from load imbalance. This issue has so far limited the adoption of these efficient k-Means variants in parallel computing environments. In this work, we provide a parallel formulation of the KD-Tree based k-Means algorithm for distributed memory systems and address its load balancing issue. Three solutions have been developed and tested. Two approaches are based on a static partitioning of the data set and a third solution incorporates a dynamic load balancing policy.
Resumo:
While the Cluster spacecraft were located near the high-latitude magnetopause, between 10:10 and 10:40 UT on 16 January 2004, three typical flux transfer event (FTE) signatures were observed. During this interval, simultaneous and conjugated all-sky camera measurements, recorded at Yellow River Station, Svalbard, are available at 630.0 and 557.7nm that show poleward-moving auroral forms (PMAFs), consistent with magnetic reconnection at dayside magnetopause. Simultaneous FTEs seen at the magnetopause mainly move northward, but having duskward (eastward) and tailward velocity components, roughly consistent with the observed direction of motion of the PMAFs in all-sky images. Between the PMAFs meridional keograms, extracted from the all-sky images, show intervals of lower intensity aurora which migrate equatorward just before the PMAFs intensify. This is strong evidence for an equatorward eroding and poleward moving open-closed boundary (OCB) associated with a variable magnetopause reconnection rate under variable IMF conditions. From the durations of the PMAFs we infer that the evolution time of FTEs is 5-11 minutes from its origin on magnetopause to its addition to the polar cap.
Resumo:
It is known that germin, which is a marker of the onset of growth in germinating wheat, is an oxalate oxidase, and also that germins possess sequence similarity with legumin and vicilin seed storage proteins. These two pieces of information have been combined in order to generate a 3D model of germin based on the structure of vicilin and to examine the model with regard to a potential oxalate oxidase active site. A cluster of three histidine residues has been located within the conserved beta-barrel structure. While there is a relatively low level of overall sequence similarity between the model and the vicilin structures, the conservation of amino acids important in maintaining the scaffold of the beta-barrel lends confidence to the juxtaposition of the histidine residues. The cluster is similar structurally to those found in copper amine oxidase and other proteins, leading to the suggestion that it defines a metal-binding location within the oxalate oxidase active site. It is also proposed that the structural elements involved in intermolecular interactions in vicilins may play a role in oligomer formation in germin/oxalate oxidase.
Resumo:
A full assessment of para-virtualization is important, because without knowledge about the various overheads, users can not understand whether using virtualization is a good idea or not. In this paper we are very interested in assessing the overheads of running various benchmarks on bare-‐metal, as well as on para-‐virtualization. The idea is to see what the overheads of para-‐ virtualization are, as well as looking at the overheads of turning on monitoring and logging. The knowledge from assessing various benchmarks on these different systems will help a range of users understand the use of virtualization systems. In this paper we assess the overheads of using Xen, VMware, KVM and Citrix, see Table 1. These different virtualization systems are used extensively by cloud-‐users. We are using various Netlib1 benchmarks, which have been developed by the University of Tennessee at Knoxville (UTK), and Oak Ridge National Laboratory (ORNL). In order to assess these virtualization systems, we run the benchmarks on bare-‐metal, then on the para-‐virtualization, and finally we turn on monitoring and logging. The later is important as users are interested in Service Level Agreements (SLAs) used by the Cloud providers, and the use of logging is a means of assessing the services bought and used from commercial providers. In this paper we assess the virtualization systems on three different systems. We use the Thamesblue supercomputer, the Hactar cluster and IBM JS20 blade server (see Table 2), which are all servers available at the University of Reading. A functional virtualization system is multi-‐layered and is driven by the privileged components. Virtualization systems can host multiple guest operating systems, which run on its own domain, and the system schedules virtual CPUs and memory within each Virtual Machines (VM) to make the best use of the available resources. The guest-‐operating system schedules each application accordingly. You can deploy virtualization as full virtualization or para-‐virtualization. Full virtualization provides a total abstraction of the underlying physical system and creates a new virtual system, where the guest operating systems can run. No modifications are needed in the guest OS or application, e.g. the guest OS or application is not aware of the virtualized environment and runs normally. Para-‐virualization requires user modification of the guest operating systems, which runs on the virtual machines, e.g. these guest operating systems are aware that they are running on a virtual machine, and provide near-‐native performance. You can deploy both para-‐virtualization and full virtualization across various virtualized systems. Para-‐virtualization is an OS-‐assisted virtualization; where some modifications are made in the guest operating system to enable better performance. In this kind of virtualization, the guest operating system is aware of the fact that it is running on the virtualized hardware and not on the bare hardware. In para-‐virtualization, the device drivers in the guest operating system coordinate the device drivers of host operating system and reduce the performance overheads. The use of para-‐virtualization [0] is intended to avoid the bottleneck associated with slow hardware interrupts that exist when full virtualization is employed. It has revealed [0] that para-‐ virtualization does not impose significant performance overhead in high performance computing, and this in turn this has implications for the use of cloud computing for hosting HPC applications. The “apparent” improvement in virtualization has led us to formulate the hypothesis that certain classes of HPC applications should be able to execute in a cloud environment, with minimal performance degradation. In order to support this hypothesis, first it is necessary to define exactly what is meant by a “class” of application, and secondly it will be necessary to observe application performance, both within a virtual machine and when executing on bare hardware. A further potential complication is associated with the need for Cloud service providers to support Service Level Agreements (SLA), so that system utilisation can be audited.