939 resultados para mobile telefonia back-end
Resumo:
Questo elaborato di tesi illustra la realizzazione del back-end di LogicPlayer, un'app Android per la didattica della Deduzione Naturale
Resumo:
L’argomento centrale della tesi sono i centri sportivi, l’applicazione permette quindi all’utente di cercare un centro sportivo per nome, per città o per provincia. Consente inoltre di visualizzare la disponibilità per ogni campo offerto dalle strutture ed eventualmente di effettuare una prenotazione. Il centro sportivo renderà disponibili informazioni altrimenti difficilmente reperibili come gli orari, il numero telefonico, l’indirizzo, ecc.. Il progetto si compone di una parte front end e una parte back end. Il front consiste in un’applicazione android nativo (sviluppata in java). Il back-end invece vede un applicativo basato su ASP.NET Web API 2, con db Entity Framework Code First. Per la gestione degli user è stato scelto il framework ASP.NET Identity 2.1.
Resumo:
Lo scopo della tesi è quello di descrivere e mettere a confronto tre diversi linguaggi, e quindi approcci, alla programmazione server-side e di back-end, ovvero il linguaggio PHP, il linguaggio Python ed il linguaggio Javascript, utilizzato però per una programmazione “Server Side”, e quindi associato al framework NodeJS. Questo confronto si pone l’obiettivo di sottolineare le differenti caratteristiche di ogni linguaggio, gli scopi a cui esso maggiormente si addice e di fornire una sorta di guida per far in modo che si possa comprendere al meglio quale dei tre linguaggi maggiormente usati per la programmazione backend si conformi meglio all’obiettivo prepostosi.
Resumo:
Spreading the PSF over a quite large amount of pixels is an increasingly used observing technique in order to reach extremely precise photometry, such as in the case of exoplanets searching and characterization via transits observations. A PSF top-hat profile helps to minimize the errors contribution due to the uncertainty on the knowledge of the detector flat field. This work has been carried out during the recent design study in the framework of the ESA small mission CHEOPS. Because of lack of perfect flat-fielding information, in the CHEOPS optics it is required to spread the light of a source into a well defined angular area, in a manner as uniform as possible. Furthermore this should be accomplished still retaining the features of a true focal plane onto the detector. In this way, for instance, the angular displacement on the focal plane is fully retained and in case of several stars in a field these look as separated as their distance is larger than the spreading size. An obvious way is to apply a defocus, while the presence of an intermediate pupil plane in the Back End Optics makes attractive to introduce here an optical device that is able to spread the light in a well defined manner, still retaining the direction of the chief ray hitting it. This can be accomplished through an holographic diffuser or through a lenslet array. Both techniques implement the concept of segmenting the pupil into several sub-zones where light is spread to a well defined angle. We present experimental results on how to deliver such PSF profile by mean of holographic diffuser and lenslet array. Both the devices are located in an intermediate pupil plane of a properly scaled laboratory setup mimicking the CHEOPS optical design configuration. © (2014) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Resumo:
We present and evaluate a compiler from Prolog (and extensions) to JavaScript which makes it possible to use (constraint) logic programming to develop the client side of web applications while being compliant with current industry standards. Targeting JavaScript makes (C)LP programs executable in virtually every modern computing device with no additional software requirements from the point of view of the user. In turn, the use of a very high-level language facilitates the development of high-quality, complex software. The compiler is a back end of the Ciao system and supports most of its features, including its module system and its rich language extension mechanism based on packages. We present an overview of the compilation process and a detailed description of the run-time system, including the support for modular compilation into separate JavaScript code. We demonstrate the maturity of the compiler by testing it with complex code such as a CLP(FD) library written in Prolog with attributed variables. Finally, we validate our proposal by measuring the performance of some LP and CLP(FD) benchmarks running on top of major JavaScript engines.
Resumo:
Spanish Young Generation in Nuclear (Jóvenes Nucleares, JJNN) is a non-profrt organization that depends on the Spanish Nuclear Society (Sociedad Nuclear Española, SNE).Since one of rts main goals is to spread the knowledge about nuclear power,severa! technical tours to facilities wrth an importan!role in the nuclear fuel cycle have been organized for the purpose ofleaming about the different stages of the Spanish tuel cycle. Spanish Young Generation in Nuclear had the opportunity to visit ENUSA Fuel Assembly Factory in Juzbado (Salamanca, Spain), Where it could be understood the front-end cycle which involves the uranium supply and storage, design and manufacturing of fuel bundles for European nuclear power plants. Alterwards, due to the tour of Almaraz NPP (PWR) and Santa María de Garoña NPP (BWR), rt could be comprehended how to obtain energy from this fuel in two different types of reactors.Furthermore,in these two plants, the facilities related to the back-end cycle could be toured. lt was possible to watch the Spent FuelPools, where the fuel bundles are stored under water until their activity is reduced enough to transport them to an Individual Temporary Storage Facility orto the Centralized Temporary Storage. Finally, a technical tour to ENSA Heavy Components Factory (ENSA) was accomplished, Where it could be experienced at first hand how different Nuclear Steam Supply System (NSSS) components and other nuclear elements, such as racks or shipping and storage casks for spent nuclear fuel, are manulactured.
All these perlonned technical tours were a complete success thanks to a generous care and know-how of the wor1
Resumo:
El estudio de los ciclos del combustible nuclear requieren de herramientas computacionales o "códigos" versátiles para dar respuestas al problema multicriterio de evaluar los actuales ciclos o las capacidades de las diferentes estrategias y escenarios con potencial de desarrollo en a nivel nacional, regional o mundial. Por otra parte, la introducción de nuevas tecnologías para reactores y procesos industriales hace que los códigos existentes requieran nuevas capacidades para evaluar la transición del estado actual del ciclo del combustible hacia otros más avanzados y sostenibles. Brevemente, esta tesis se centra en dar respuesta a las principales preguntas, en términos económicos y de recursos, al análisis de escenarios de ciclos de combustible, en particular, para el análisis de los diferentes escenarios del ciclo del combustible de relativa importancia para España y Europa. Para alcanzar este objetivo ha sido necesaria la actualización y el desarrollo de nuevas capacidades del código TR_EVOL (Transition Evolution code). Este trabajo ha sido desarrollado en el Programa de Innovación Nuclear del CIEMAT desde el año 2010. Esta tesis se divide en 6 capítulos. El primer capítulo ofrece una visión general del ciclo de combustible nuclear, sus principales etapas y los diferentes tipos utilizados en la actualidad o en desarrollo para el futuro. Además, se describen las fuentes de material nuclear que podrían ser utilizadas como combustible (uranio y otros). También se puntualizan brevemente una serie de herramientas desarrolladas para el estudio de estos ciclos de combustible nuclear. El capítulo 2 está dirigido a dar una idea básica acerca de los costes involucrados en la generación de electricidad mediante energía nuclear. Aquí se presentan una clasificación de estos costos y sus estimaciones, obtenidas en la bibliografía, y que han sido evaluadas y utilizadas en esta tesis. Se ha incluido también una breve descripción del principal indicador económico utilizado en esta tesis, el “coste nivelado de la electricidad”. El capítulo 3 se centra en la descripción del código de simulación desarrollado para el estudio del ciclo del combustible nuclear, TR_EVOL, que ha sido diseñado para evaluar diferentes opciones de ciclos de combustibles. En particular, pueden ser evaluados las diversos reactores con, posiblemente, diferentes tipos de combustibles y sus instalaciones del ciclo asociadas. El módulo de evaluaciones económica de TR_EVOL ofrece el coste nivelado de la electricidad haciendo uso de las cuatro fuentes principales de información económica y de la salida del balance de masas obtenido de la simulación del ciclo en TR_EVOL. Por otra parte, la estimación de las incertidumbres en los costes también puede ser efectuada por el código. Se ha efectuado un proceso de comprobación cruzada de las funcionalidades del código y se descrine en el Capítulo 4. El proceso se ha aplicado en cuatro etapas de acuerdo con las características más importantes de TR_EVOL, balance de masas, composición isotópica y análisis económico. Así, la primera etapa ha consistido en el balance masas del ciclo de combustible nuclear actual de España. La segunda etapa se ha centrado en la comprobación de la composición isotópica del flujo de masas mediante el la simulación del ciclo del combustible definido en el proyecto CP-ESFR UE. Las dos últimas etapas han tenido como objetivo validar el módulo económico. De este modo, en la tercera etapa han sido evaluados los tres principales costes (financieros, operación y mantenimiento y de combustible) y comparados con los obtenidos por el proyecto ARCAS, omitiendo los costes del fin del ciclo o Back-end, los que han sido evaluado solo en la cuarta etapa, haciendo uso de costes unitarios y parámetros obtenidos a partir de la bibliografía. En el capítulo 5 se analizan dos grupos de opciones del ciclo del combustible nuclear de relevante importancia, en términos económicos y de recursos, para España y Europa. Para el caso español, se han simulado dos grupos de escenarios del ciclo del combustible, incluyendo estrategias de reproceso y extensión de vida de los reactores. Este análisis se ha centrado en explorar las ventajas y desventajas de reprocesado de combustible irradiado en un país con una “relativa” pequeña cantidad de reactores nucleares. Para el grupo de Europa se han tratado cuatro escenarios, incluyendo opciones de transmutación. Los escenarios incluyen los reactores actuales utilizando la tecnología reactor de agua ligera y ciclo abierto, un reemplazo total de los reactores actuales con reactores rápidos que queman combustible U-Pu MOX y dos escenarios del ciclo del combustible con transmutación de actínidos minoritarios en una parte de los reactores rápidos o en sistemas impulsados por aceleradores dedicados a transmutación. Finalmente, el capítulo 6 da las principales conclusiones obtenidas de esta tesis y los trabajos futuros previstos en el campo del análisis de ciclos de combustible nuclear. ABSTRACT The study of the nuclear fuel cycle requires versatile computational tools or “codes” to provide answers to the multicriteria problem of assessing current nuclear fuel cycles or the capabilities of different strategies and scenarios with potential development in a country, region or at world level. Moreover, the introduction of new technologies for reactors and industrial processes makes the existing codes to require new capabilities to assess the transition from current status of the fuel cycle to the more advanced and sustainable ones. Briefly, this thesis is focused in providing answers to the main questions about resources and economics in fuel cycle scenario analyses, in particular for the analysis of different fuel cycle scenarios with relative importance for Spain and Europe. The upgrade and development of new capabilities of the TR_EVOL code (Transition Evolution code) has been necessary to achieve this goal. This work has been developed in the Nuclear Innovation Program at CIEMAT since year 2010. This thesis is divided in 6 chapters. The first one gives an overview of the nuclear fuel cycle, its main stages and types currently used or in development for the future. Besides the sources of nuclear material that could be used as fuel (uranium and others) are also viewed here. A number of tools developed for the study of these nuclear fuel cycles are also briefly described in this chapter. Chapter 2 is aimed to give a basic idea about the cost involved in the electricity generation by means of the nuclear energy. The main classification of these costs and their estimations given by bibliography, which have been evaluated in this thesis, are presented. A brief description of the Levelized Cost of Electricity, the principal economic indicator used in this thesis, has been also included. Chapter 3 is focused on the description of the simulation tool TR_EVOL developed for the study of the nuclear fuel cycle. TR_EVOL has been designed to evaluate different options for the fuel cycle scenario. In particular, diverse nuclear power plants, having possibly different types of fuels and the associated fuel cycle facilities can be assessed. The TR_EVOL module for economic assessments provides the Levelized Cost of Electricity making use of the TR_EVOL mass balance output and four main sources of economic information. Furthermore, uncertainties assessment can be also carried out by the code. A cross checking process of the performance of the code has been accomplished and it is shown in Chapter 4. The process has been applied in four stages according to the most important features of TR_EVOL. Thus, the first stage has involved the mass balance of the current Spanish nuclear fuel cycle. The second stage has been focused in the isotopic composition of the mass flow using the fuel cycle defined in the EU project CP-ESFR. The last two stages have been aimed to validate the economic module. In the third stage, the main three generation costs (financial cost, O&M and fuel cost) have been assessed and compared to those obtained by ARCAS project, omitting the back-end costs. This last cost has been evaluated alone in the fourth stage, making use of some unit cost and parameters obtained from the bibliography. In Chapter 5 two groups of nuclear fuel cycle options with relevant importance for Spain and Europe are analyzed in economic and resources terms. For the Spanish case, two groups of fuel cycle scenarios have been simulated including reprocessing strategies and life extension of the current reactor fleet. This analysis has been focused on exploring the advantages and disadvantages of spent fuel reprocessing in a country with relatively small amount of nuclear power plants. For the European group, four fuel cycle scenarios involving transmutation options have been addressed. Scenarios include the current fleet using Light Water Reactor technology and open fuel cycle, a full replacement of the initial fleet with Fast Reactors burning U-Pu MOX fuel and two fuel cycle scenarios with Minor Actinide transmutation in a fraction of the FR fleet or in dedicated Accelerator Driven Systems. Finally, Chapter 6 gives the main conclusions obtained from this thesis and the future work foreseen in the field of nuclear fuel cycle analysis.
Resumo:
El software se ha convertido en el eje central del mundo actual, una compleja creación humana que influye en la vida, negocios y comunicación de todas las personas pertenecientes a la Sociedad de la Información. El rápido crecimiento experimentado en el ámbito del desarrollo software ha permitido la creación de avanzadas estructuras tecnológicas, denominadas “Sistemas Intensivos Software”, capaces de comunicarse con otros sistemas, dispositivos, sensores y personas. A lo largo de los próximos años los sistemas se enfrentarán a una mayor complejidad, surgida de la necesidad de operar en entornos de grandes dimensiones y de comportamientos no deterministas. Los métodos y herramientas actuales no son lo suficientemente potentes para diseñar, construir,implementar y mantener sistemas intensivos software con estas características, y detener la construcción de sistemas intensivos software o construir sistemas poco flexibles o fiables no es una alternativa real. En el desarrollo de “Sistemas Intensivos Software” pueden llegar a intervenir distintas entidades o compañías software que suelen estar en ubicaciones geográficas distintas y constituidas por grandes equipos de desarrollo, multidisciplinares e incluso multilingües. Debido a la criticidad del resultado de las actividades realizadas de forma independiente en el sistema resultante, éstas se han de controlar y monitorizar para asegurar la correcta integración de todos los elementos del sistema completo. El objetivo de este proyecto es la creación de una herramienta software para dar soporte a la gestión y monitorización de la construcción e integración de sistemas intensivos software, siendo extensible también a proyectos de otra índole. La herramienta resultante se denomina Positioning System, una aplicación web del tipo SPA (Single Page Application) creada con tecnología de última generación como el framework JavaScript AngularJS y tecnología de back-end como SlimPHP. Positioning System provee la funcionalidad necesaria para la creación de proyectos, familias y subfamilias de productos que constituyen los productos software de los proyectos creados, así como la gestión de socios comerciales y gestión de contactos de dichos proyectos. Todas estas funcionalidades son fácilmente monitorizadas y controladas por gráficos estadísticos generados para cada proyecto. ABSTRACT Software has become the backbone of today’s world, a complex human creation that has an important impact in the life, business and communication of all people involved with the Information Society. The quick growth that software development has undergone for last years has enabled the creation of advanced technological structures called “Software Intensive Systems”. They are able to communicate with other systems, devices, sensors and people. Next years, systems will face more complexity. It arises from the need of operating systems of large dimensions with non-deterministic behaviors. Current methods and tools are not powerful enough to design, build, implement and maintain software intensive systems; however stopping the development or developing unreliable and non-flexible systems is not a real alternative. Software Intensive Systems” development may involve different entities or software companies which may be in different geographical locations and may be constituted by large, multidisciplinary and even multilingual development teams. Due to the criticality of the result of each conducted activity, independently in the resulting system, these activities must be controlled and monitored to ensure the proper integration of all the elements within the complete system. The goal of this project is the creation of a software tool to support the management and monitoring of the construction and integration of software intensive systems, being possible to be extended to other kind of projects. The resultant tool is called Positioning System, a web application that follows the SPA (Single Page Application) style. It was created with the latest technologies, such as, the AngularJS framework and SlimPHP. The Positioning System provides the necessary features for the creation of projects, families and subfamilies of products that constitute the software products of the created projects, as well as the management of business partners and contacts of these projects. All these features are easily monitored and controlled by statistical graphs generated for each project.
Resumo:
Este proyecto fin de grado presenta dos herramientas, Papify y Papify-Viewer, para medir y visualizar, respectivamente, las prestaciones a bajo nivel de especificaciones RVC-CAL basándose en eventos hardware. RVC-CAL es un lenguaje de flujo de datos estandarizado por MPEG y utilizado para definir herramientas relacionadas con la codificación de vídeo. La estructura de los programas descritos en RVC-CAL se basa en unidades funcionales llamadas actores, que a su vez se subdividen en funciones o procedimientos llamados acciones. ORCC (Open RVC-CAL Compiler) es un compilador de código abierto que utiliza como entrada descripciones RVC-CAL y genera a partir de ellas código fuente en un lenguaje dado, como por ejemplo C. Internamente, el compilador ORCC se divide en tres etapas distinguibles: front-end, middle-end y back-end. La implementación de Papify consiste en modificar la etapa del back-end del compilador, encargada de la generación de código, de modo tal que los actores, al ser traducidos a lenguaje C, queden instrumentados con PAPI (Performance Application Programing Interface), una herramienta utilizada como interfaz a los registros contadores de rendimiento (PMC) de los procesadores. Además, también se modifica el front-end para permitir identificar cierto tipo de anotaciones en las descripciones RVC-CAL, utilizadas para que el diseñador pueda indicar qué actores o acciones en particular se desean analizar. Los actores instrumentados, además de conservar su funcionalidad original, generan una serie de ficheros que contienen datos sobre los distintos eventos hardware que suceden a lo largo de su ejecución. Los eventos incluidos en estos ficheros son configurables dentro de las anotaciones previamente mencionadas. La segunda herramienta, Papify-Viewer, utiliza los datos generados por Papify y los procesa, obteniendo una representación visual de la información a dos niveles: por un lado, representa cronológicamente la ejecución de la aplicación, distinguiendo cada uno de los actores a lo largo de la misma. Por otro lado, genera estadísticas sobre la cantidad de eventos disparados por acción, actor o núcleo de ejecución y las representa mediante gráficos de barra. Ambas herramientas pueden ser utilizadas en conjunto para verificar el funcionamiento del programa, balancear la carga de los actores o la distribución por núcleos de los mismos, mejorar el rendimiento y diagnosticar problemas. ABSTRACT. This diploma project presents two tools, Papify and Papify-Viewer, used to measure and visualize the low level performance of RVC-CAL specifications based on hardware events. RVC-CAL is a dataflow language standardized by MPEG which is used to define video codec tools. The structure of the applications described in RVC-CAL is based on functional units called actors, which are in turn divided into smaller procedures called actions. ORCC (Open RVC-CAL Compiler) is an open-source compiler capable of transforming RVC-CAL descriptions into source code in a given language, such as C. Internally, the compiler is divided into three distinguishable stages: front-end, middle-end and back-end. Papify’s implementation consists of modifying the compiler’s back-end stage, which is responsible for generating the final source code, so that translated actors in C code are now instrumented with PAPI (Performance Application Programming Interface), a tool that provides an interface to the microprocessor’s performance monitoring counters (PMC). In addition, the front-end is also modified in such a way that allows identification of a certain type of annotations in the RVC-CAL descriptions, allowing the designer to set the actors or actions to be included in the measurement. Besides preserving their initial behavior, the instrumented actors will also generate a set of files containing data about the different events triggered throughout the program’s execution. The events included in these files can be configured inside the previously mentioned annotations. The second tool, Papify-Viewer, makes use of the files generated by Papify to process them and provide a visual representation of the information in two different ways: on one hand, a chronological representation of the application’s execution where each actor has its own timeline. On the other hand, statistical information is generated about the amount of triggered events per action, actor or core. Both tools can be used together to assert the normal functioning of the program, balance the load between actors or cores, improve performance and identify problems.
Resumo:
En esta memoria se presenta el diseño y desarrollo de una aplicación en la nube destinada a la compartición de objetos y servicios. El desarrollo de esta aplicación surge dentro del proyecto de I+D+i, SITAC: Social Internet of Things – Apps by and for the Crowd ITEA 2 11020, que trata de crear una arquitectura integradora y un “ecosistema” que incluya plataformas, herramientas y metodologías para facilitar la conexión y cooperación de entidades de distinto tipo conectadas a la red bien sean sistemas, máquinas, dispositivos o personas con dispositivos móviles personales como tabletas o teléfonos móviles. El proyecto innovará mediante la utilización de un modelo inspirado en las redes sociales para facilitar y unificar las interacciones tanto entre personas como entre personas y dispositivos. En este contexto surge la necesidad de desarrollar una aplicación destinada a la compartición de recursos en la nube que pueden ser tanto lógicos como físicos, y que esté orientada al big data. Ésta será la aplicación presentada en este trabajo, el “Resource Sharing Center”, que ofrece un servicio web para el intercambio y compartición de contenido, y un motor de recomendaciones basado en las preferencias de los usuarios. Con este objetivo, se han usado tecnologías de despliegue en la nube, como Elastic Beanstalk (el PaaS de Amazon Web Services), S3 (el sistema de almacenamiento de Amazon Web Services), SimpleDB (base de datos NoSQL) y HTML5 con JavaScript y Twitter Bootstrap para el desarrollo del front-end, siendo Python y Node.js las tecnologías usadas en el back end, y habiendo contribuido a la mejora de herramientas de clustering sobre big data. Por último, y de cara a realizar el estudio sobre las pruebas de carga de la aplicación se ha usado la herramienta ApacheJMeter.
Resumo:
Translation of La vie simple.
Resumo:
INTAMAP is a web processing service for the automatic interpolation of measured point data. Requirements were (i) using open standards for spatial data such as developed in the context of the open geospatial consortium (OGC), (ii) using a suitable environment for statistical modelling and computation, and (iii) producing an open source solution. The system couples the 52-North web processing service, accepting data in the form of an observations and measurements (O&M) document with a computing back-end realized in the R statistical environment. The probability distribution of interpolation errors is encoded with UncertML, a new markup language to encode uncertain data. Automatic interpolation needs to be useful for a wide range of applications and the algorithms have been designed to cope with anisotropies and extreme values. In the light of the INTAMAP experience, we discuss the lessons learnt.
Resumo:
INTAMAP is a Web Processing Service for the automatic spatial interpolation of measured point data. Requirements were (i) using open standards for spatial data such as developed in the context of the Open Geospatial Consortium (OGC), (ii) using a suitable environment for statistical modelling and computation, and (iii) producing an integrated, open source solution. The system couples an open-source Web Processing Service (developed by 52°North), accepting data in the form of standardised XML documents (conforming to the OGC Observations and Measurements standard) with a computing back-end realised in the R statistical environment. The probability distribution of interpolation errors is encoded with UncertML, a markup language designed to encode uncertain data. Automatic interpolation needs to be useful for a wide range of applications and the algorithms have been designed to cope with anisotropy, extreme values, and data with known error distributions. Besides a fully automatic mode, the system can be used with different levels of user control over the interpolation process.
Resumo:
The research presented in this dissertation is comprised of several parts which jointly attain the goal of Semantic Distributed Database Management with Applications to Internet Dissemination of Environmental Data. ^ Part of the research into more effective and efficient data management has been pursued through enhancements to the Semantic Binary Object-Oriented database (Sem-ODB) such as more effective load balancing techniques for the database engine, and the use of Sem-ODB as a tool for integrating structured and unstructured heterogeneous data sources. Another part of the research in data management has pursued methods for optimizing queries in distributed databases through the intelligent use of network bandwidth; this has applications in networks that provide varying levels of Quality of Service or throughput. ^ The application of the Semantic Binary database model as a tool for relational database modeling has also been pursued. This has resulted in database applications that are used by researchers at the Everglades National Park to store environmental data and to remotely-sensed imagery. ^ The areas of research described above have contributed to the creation TerraFly, which provides for the dissemination of geospatial data via the Internet. TerraFly research presented herein ranges from the development of TerraFly's back-end database and interfaces, through the features that are presented to the public (such as the ability to provide autopilot scripts and on-demand data about a point), to applications of TerraFly in the areas of hazard mitigation, recreation, and aviation. ^
Resumo:
OBJECTIVES: In natural hearing, cochlear mechanical compression is dynamically adjusted via the efferent medial olivocochlear reflex (MOCR). These adjustments probably help understanding speech in noisy environments and are not available to the users of current cochlear implants (CIs). The aims of the present study are to: (1) present a binaural CI sound processing strategy inspired by the control of cochlear compression provided by the contralateral MOCR in natural hearing; and (2) assess the benefits of the new strategy for understanding speech presented in competition with steady noise with a speech-like spectrum in various spatial configurations of the speech and noise sources. DESIGN: Pairs of CI sound processors (one per ear) were constructed to mimic or not mimic the effects of the contralateral MOCR on compression. For the nonmimicking condition (standard strategy or STD), the two processors in a pair functioned similarly to standard clinical processors (i.e., with fixed back-end compression and independently of each other). When configured to mimic the effects of the MOCR (MOC strategy), the two processors communicated with each other and the amount of back-end compression in a given frequency channel of each processor in the pair decreased/increased dynamically (so that output levels dropped/increased) with increases/decreases in the output energy from the corresponding frequency channel in the contralateral processor. Speech reception thresholds in speech-shaped noise were measured for 3 bilateral CI users and 2 single-sided deaf unilateral CI users. Thresholds were compared for the STD and MOC strategies in unilateral and bilateral listening conditions and for three spatial configurations of the speech and noise sources in simulated free-field conditions: speech and noise sources colocated in front of the listener, speech on the left ear with noise in front of the listener, and speech on the left ear with noise on the right ear. In both bilateral and unilateral listening, the electrical stimulus delivered to the test ear(s) was always calculated as if the listeners were wearing bilateral processors. RESULTS: In both unilateral and bilateral listening conditions, mean speech reception thresholds were comparable with the two strategies for colocated speech and noise sources, but were at least 2 dB lower (better) with the MOC than with the STD strategy for spatially separated speech and noise sources. In unilateral listening conditions, mean thresholds improved with increasing the spatial separation between the speech and noise sources regardless of the strategy but the improvement was significantly greater with the MOC strategy. In bilateral listening conditions, thresholds improved significantly with increasing the speech-noise spatial separation only with the MOC strategy. CONCLUSIONS: The MOC strategy (1) significantly improved the intelligibility of speech presented in competition with a spatially separated noise source, both in unilateral and bilateral listening conditions; (2) produced significant spatial release from masking in bilateral listening conditions, something that did not occur with fixed compression; and (3) enhanced spatial release from masking in unilateral listening conditions. The MOC strategy as implemented here, or a modified version of it, may be usefully applied in CIs and in hearing aids.