874 resultados para Distributed data access
Resumo:
The development of large discount retailers, or big-boxes as they are sometimes referred to, are often subject to heated debate and their entry on a market is greeted with either great enthusiasm or dread. For instance, the world’s largest retailer Wal-Mart (Forbes 2014) has a number of anti- and pro-groups dedicated to its being and the event of a Wal-Mart entry tends to be met with protests and campaigns (Decamme 2013) but also welcomed by, for instance, consumers (Davis & DeBonis 2013). Also in Sweden, the entry of a big box is a hot topic and before IKEA’s opening i Borlänge 2013, the first in Sweden in more than five years, great expectations were mixed with worry (Västerbottens-Kuriren 2011).The presence of large scale discount retailers is not, however, a novel phenomenon but a part of a long-term change in retailing that has taken place globally over the past couple of decades (Taylor & Smalling, 2005). As noted by Dawson (2006), the trend in Europe has over the past few decades gone towards an increasing concentration of large firms along with a decrease of smaller firms.This trend is also detectable in the Swedish retail industry. Over the past decade, the retailing industry in Sweden has increased by around 190 Billion SEK, and its share of GDP has risen from 2,7% to 2,9%, while the number of employees have increased from 200 000 to 250 000 (HUI 2013). This growth, however, has not been distributed evenly but rather it has been oriented mainly towards out-of-town retail clusters. Parallel to this development, the number of large retailers has risen at the expense of market shares of smaller independent firms (Rämme et al 2010). Thereby, the presence of large scale retailers is simply part of a changing retail landscape.The effects of this development, where large scale retailing agents relocate shopping to out-of-town shopping areas, have been heavily debated. On the one hand, the big-boxes are accused of displacing independent small retail businesses in the city-centers and the residential areas, resulting in, to some extent, reduced employment opportunities and less availability for the consumers - especially the elderly (Ljungberg et al 2006). In addition, as access to shopping now tends to require some sort of a motorized vehicle, environmental aspects to the discussion have emerged. Ultimately these types of concerns have resulted in calls for regulations against this development (Olsson 2010). On the other hand, the proponents of the new shopping landscape argue that this evolution implies productivity gains, the benefits of lower prices and an increased variety of products (Maican & Orth 2012). Moreover it is argued that it leads to, for instance, better services (such as longer opening hours) and a creative destruction transformation pressure on retailers, which brings about a renewal of city-centerIIretail and services, increasing their attractivity (Bergström 2010). The belief in benefits of a big box entry can be exemplified by the attractivity of IKEA, and the fact that municipalities are prepared to commit to expenses amounting up to hundreds of millions in order to attract the entry of this big-box. Borlänge municipality, for instance, agreed to expenses of about 350 million SEK in order to secure the entry of IKEA, which opened in 2013 (Blomgren 2009).Against this backdrop, the overall effects of large discount retailers become important: Are the economic benefits enough to warrant subsidies or are there, on the contrary, some very compelling grounds for regulations against these types of establishments? In other words; how is overall retail in a region where a store like IKEA enters affected? And how are local retail firms affected?In order to answer these questions, the purpose of this thesis is to study how entry of a big-box retailer affects the entry region. The object of this study is IKEA - one of the world’s largest retailers, with 345 stores, active in over 40 countries and with profits of about 3.3 billion (IKEA 2013; IKEA 2014). By studying the effects of IKEA-entry, both on an aggregated level and on firm level, this thesis intends to find indications of how large discount retail establishments in general can be expected to affect the economic development both in a region overall, but also on the local firm level, something which is of interest to both policymakers as well as the retailing industry in general.The first paper examines the effects of IKEA on retail revenues and employment in the municipalities that IKEA chose to enter between 2000 and 2011; Gothenburg, Haparanda, Kalmar and Karlstad. By means of a matching method we first identify non-entry municipalities that have a similar probability of IKEA entry as the true entry municipalities. Then, using these non-entry municipalities as a control group, the causal effects of IKEA entry can be estimated using a treatment-control approach. We also extend the analysis to examine the spatial impact of IKEA by estimating the effects on retail in neighboring municipalities. It is found that a new IKEA store increases revenues in durable goods trade with 20% in the entry municipality and the number of employees with 17%. Only small, and in most cases statistically insignificant, negative effects were found in neighboring municipalities.It appears that there is a positive net effect on durables retail sales and employment in the entry municipality. However, the analysis is based on data on an aggregated municipality level and thereby it remains unclear if and how the effects vary within the entry municipalities. In addition, the data used in the first study includes the sales and employment of IKEA itself, which could account for the majority of the increases in employment and retail. Thereby the potential spillover effects on incumbent retailers in the entry municipalities cannot be discerned in the first study.IIITo examine effects of IKEA entry on incumbent retail firms, the second paper in this thesis analyses how IKEA entry affects the revenues and employment of local retail firms in three municipalities; Haparanda, Kalmar and Karlstad, which experienced entry by IKEA between 2000 and 2010. In this second study, we exclude Gothenburg due to the fact that big-box entry appears to have weaker effects in metropolitan areas (as indicated by Artz & Stone 2006). By excluding Gothenburg we aim to reduce the geographical heterogeneity in our study. We obtain control municipalities that are as similar as possible to the three entry municipalities using the same method as in the previous study, but including a slightly different set of variables in the selection equation. Using similar retail firms in the control municipalities as our comparison group, we estimate the impact of IKEA entry on revenues and employment for retail firms located at varying distances from the IKEA entry site.The results generated in this study imply that entry by IKEA increases revenues in incumbent retail firms by, on average, 11% in the entry municipalities. In addition, we do not find any significant impact on retail revenues in the city centers of the entry municipalities. However, we do find that retail firms within 1 km of the IKEA experience increases in revenues of about 26%, which indicates large spillover effects in the area nearby the entry site. As expected, this impact decreases as we expand the buffer zone: firms located between 0-2 km experiences a 14% increase and firms in 2-5 km experiences an increase of 10%. We do not find any significant impacts on retail employment.
Resumo:
Background qtl.outbred is an extendible interface in the statistical environment, R, for combining quantitative trait loci (QTL) mapping tools. It is built as an umbrella package that enables outbred genotype probabilities to be calculated and/or imported into the software package R/qtl. Findings Using qtl.outbred, the genotype probabilities from outbred line cross data can be calculated by interfacing with a new and efficient algorithm developed for analyzing arbitrarily large datasets (included in the package) or imported from other sources such as the web-based tool, GridQTL. Conclusion qtl.outbred will improve the speed for calculating probabilities and the ability to analyse large future datasets. This package enables the user to analyse outbred line cross data accurately, but with similar effort than inbred line cross data.
Resumo:
To have good data quality with high complexity is often seen to be important. Intuition says that the higher accuracy and complexity the data have the better the analytic solutions becomes if it is possible to handle the increasing computing time. However, for most of the practical computational problems, high complexity data means that computational times become too long or that heuristics used to solve the problem have difficulties to reach good solutions. This is even further stressed when the size of the combinatorial problem increases. Consequently, we often need a simplified data to deal with complex combinatorial problems. In this study we stress the question of how the complexity and accuracy in a network affect the quality of the heuristic solutions for different sizes of the combinatorial problem. We evaluate this question by applying the commonly used p-median model, which is used to find optimal locations in a network of p supply points that serve n demand points. To evaluate this, we vary both the accuracy (the number of nodes) of the network and the size of the combinatorial problem (p). The investigation is conducted by the means of a case study in a region in Sweden with an asymmetrically distributed population (15,000 weighted demand points), Dalecarlia. To locate 5 to 50 supply points we use the national transport administrations official road network (NVDB). The road network consists of 1.5 million nodes. To find the optimal location we start with 500 candidate nodes in the network and increase the number of candidate nodes in steps up to 67,000 (which is aggregated from the 1.5 million nodes). To find the optimal solution we use a simulated annealing algorithm with adaptive tuning of the temperature. The results show that there is a limited improvement in the optimal solutions when the accuracy in the road network increase and the combinatorial problem (low p) is simple. When the combinatorial problem is complex (large p) the improvements of increasing the accuracy in the road network are much larger. The results also show that choice of the best accuracy of the network depends on the complexity of the combinatorial (varying p) problem.
Resumo:
Background. Through a national policy agreement, over 167 million Euros will be invested in the Swedish National Quality Registries (NQRs) between 2012 and 2016. One of the policy agreement¿s intentions is to increase the use of NQR data for quality improvement (QI). However, the evidence is fragmented as to how the use of medical registries and the like lead to quality improvement, and little is known about non-clinical use. The aim was therefore to investigate the perspectives of Swedish politicians and administrators on quality improvement based on national registry data. Methods. Politicians and administrators from four county councils were interviewed. A qualitative content analysis guided by the Consolidated Framework for Implementation Research (CFIR) was performed. Results. The politicians and administrators perspectives on the use of NQR data for quality improvement were mainly assigned to three of the five CFIR domains. In the domain of intervention characteristics, data reliability and access in reasonable time were not considered entirely satisfactory, making it difficult for the politico-administrative leaderships to initiate, monitor, and support timely QI efforts. Still, politicians and administrators trusted the idea of using the NQRs as a base for quality improvement. In the domain of inner setting, the organizational structures were not sufficiently developed to utilize the advantages of the NQRs, and readiness for implementation appeared to be inadequate for two reasons. Firstly, the resources for data analysis and quality improvement were not considered sufficient at politico-administrative or clinical level. Secondly, deficiencies in leadership engagement at multiple levels were described and there was a lack of consensus on the politicians¿ role and level of involvement. Regarding the domain of outer setting, there was a lack of communication and cooperation between the county councils and the national NQR organizations. Conclusions. The Swedish experiences show that a government-supported national system of well-funded, well-managed, and reputable national quality registries needs favorable local politico-administrative conditions to be used for quality improvement; such conditions are not yet in place according to local politicians and administrators.
Resumo:
Delineation of commuting regions has always been based on statistical units, often municipalities or wards. However, using these units has certain disadvantages as their land areas differ considerably. Much information is lost in the larger spatial base units and distortions in self-containment values, the main criterion in rule-based delineation procedures, occur. Alternatively, one can start from relatively small standard size units such as hexagons. In this way, much greater detail in spatial patterns is obtained. In this paper, regions are built by means of intrazonal maximization (Intramax) on the basis of hexagons. The use of geoprocessing tools, specifically developed for the processing ofcommuting data, speeds up processing time considerably. The results of the Intramax analysis are evaluated with travel-to-work area constraints, and comparisons are made with commuting fields, accessibility to employment, commuting flow density and network commuting flow size. From selected steps in the regionalization process, a hierarchy of nested commuting regions emerges, revealing the complexity of commuting patterns.
Resumo:
As scientific workflows and the data they operate on, grow in size and complexity, the task of defining how those workflows should execute (which resources to use, where the resources must be in readiness for processing etc.) becomes proportionally more difficult. While "workflow compilers", such as Pegasus, reduce this burden, a further problem arises: since specifying details of execution is now automatic, a workflow's results are harder to interpret, as they are partly due to specifics of execution. By automating steps between the experiment design and its results, we lose the connection between them, hindering interpretation of results. To reconnect the scientific data with the original experiment, we argue that scientists should have access to the full provenance of their data, including not only parameters, inputs and intermediary data, but also the abstract experiment, refined into a concrete execution by the "workflow compiler". In this paper, we describe preliminary work on adapting Pegasus to capture the process of workflow refinement in the PASOA provenance system.
Resumo:
The reliable evaluation of the flood forecasting is a crucial problem for assessing flood risk and consequent damages. Different hydrological models (distributed, semi-distributed or lumped) have been proposed in order to deal with this issue. The choice of the proper model structure has been investigated by many authors and it is one of the main sources of uncertainty for a correct evaluation of the outflow hydrograph. In addition, the recent increasing of data availability makes possible to update hydrological models as response of real-time observations. For these reasons, the aim of this work it is to evaluate the effect of different structure of a semi-distributed hydrological model in the assimilation of distributed uncertain discharge observations. The study was applied to the Bacchiglione catchment, located in Italy. The first methodological step was to divide the basin in different sub-basins according to topographic characteristics. Secondly, two different structures of the semi-distributed hydrological model were implemented in order to estimate the outflow hydrograph. Then, synthetic observations of uncertain value of discharge were generated, as a function of the observed and simulated value of flow at the basin outlet, and assimilated in the semi-distributed models using a Kalman Filter. Finally, different spatial patterns of sensors location were assumed to update the model state as response of the uncertain discharge observations. The results of this work pointed out that, overall, the assimilation of uncertain observations can improve the hydrologic model performance. In particular, it was found that the model structure is an important factor, of difficult characterization, since can induce different forecasts in terms of outflow discharge. This study is partly supported by the FP7 EU Project WeSenseIt.
Resumo:
Instrumentation and automation plays a vital role to managing the water industry. These systems generate vast amounts of data that must be effectively managed in order to enable intelligent decision making. Time series data management software, commonly known as data historians are used for collecting and managing real-time (time series) information. More advanced software solutions provide a data infrastructure or utility wide Operations Data Management System (ODMS) that stores, manages, calculates, displays, shares, and integrates data from multiple disparate automation and business systems that are used daily in water utilities. These ODMS solutions are proven and have the ability to manage data from smart water meters to the collaboration of data across third party corporations. This paper focuses on practical, utility successes in the water industry where utility managers are leveraging instantaneous access to data from proven, commercial off-the-shelf ODMS solutions to enable better real-time decision making. Successes include saving $650,000 / year in water loss control, safeguarding water quality, saving millions of dollars in energy management and asset management. Immediate opportunities exist to integrate the research being done in academia with these ODMS solutions in the field and to leverage these successes to utilities around the world.
Resumo:
The Short-term Water Information and Forecasting Tools (SWIFT) is a suite of tools for flood and short-term streamflow forecasting, consisting of a collection of hydrologic model components and utilities. Catchments are modeled using conceptual subareas and a node-link structure for channel routing. The tools comprise modules for calibration, model state updating, output error correction, ensemble runs and data assimilation. Given the combinatorial nature of the modelling experiments and the sub-daily time steps typically used for simulations, the volume of model configurations and time series data is substantial and its management is not trivial. SWIFT is currently used mostly for research purposes but has also been used operationally, with intersecting but significantly different requirements. Early versions of SWIFT used mostly ad-hoc text files handled via Fortran code, with limited use of netCDF for time series data. The configuration and data handling modules have since been redesigned. The model configuration now follows a design where the data model is decoupled from the on-disk persistence mechanism. For research purposes the preferred on-disk format is JSON, to leverage numerous software libraries in a variety of languages, while retaining the legacy option of custom tab-separated text formats when it is a preferred access arrangement for the researcher. By decoupling data model and data persistence, it is much easier to interchangeably use for instance relational databases to provide stricter provenance and audit trail capabilities in an operational flood forecasting context. For the time series data, given the volume and required throughput, text based formats are usually inadequate. A schema derived from CF conventions has been designed to efficiently handle time series for SWIFT.
Resumo:
Distributed energy and water balance models require time-series surfaces of the meteorological variables involved in hydrological processes. Most of the hydrological GIS-based models apply simple interpolation techniques to extrapolate the point scale values registered at weather stations at a watershed scale. In mountainous areas, where the monitoring network ineffectively covers the complex terrain heterogeneity, simple geostatistical methods for spatial interpolation are not always representative enough, and algorithms that explicitly or implicitly account for the features creating strong local gradients in the meteorological variables must be applied. Originally developed as a meteorological pre-processing tool for a complete hydrological model (WiMMed), MeteoMap has become an independent software. The individual interpolation algorithms used to approximate the spatial distribution of each meteorological variable were carefully selected taking into account both, the specific variable being mapped, and the common lack of input data from Mediterranean mountainous areas. They include corrections with height for both rainfall and temperature (Herrero et al., 2007), and topographic corrections for solar radiation (Aguilar et al., 2010). MeteoMap is a GIS-based freeware upon registration. Input data include weather station records and topographic data and the output consists of tables and maps of the meteorological variables at hourly, daily, predefined rainfall event duration or annual scales. It offers its own pre and post-processing tools, including video outlook, map printing and the possibility of exporting the maps to images or ASCII ArcGIS formats. This study presents the friendly user interface of the software and shows some case studies with applications to hydrological modeling.
Resumo:
o ambiente econômico atual tem exigido empenho das empresas em conhecer, interagir, diferenciar e personalizar cada vez mais produtos e serviços para os clientes. Este cenário requer ferramentas e modelos de gestão para gerenciar as relações com os clientes, com o objetivo de permitir que a empresa consiga perceber e responder rapidamente a exigências dos consumidores. Este trabalho revisa conceitos de CRM (Customer Relationschip Management ou Gerenciamento das Relações com os Clientes) e descreve a implementação de ferramenta de gestão de relacionamento com clientes em empresa de consórcio. O desenvolvimento do trabalho reflete uma necessidade apontada no planejamento estratégico da empresa, sendo que ferramentas de tecnologia de informação e software de banco de dados foram usadas como suporte aos propósitos da gestão empresarial. Como resultado do trabalho, a empresa está hoje atuando com um sistema de Data Base Marketing, o qual foi criado para auxiliar os profissionais envolvidos no processo de atendimento e gestão de relacionamento com clientes. O Data Base Marketing esta sendo utilizado para coletar dados de atendimento a clientes, tais como históricos de atendimento, dados cadastrais, perfil demográfico, perfil psicográfico e categoria de valor dos clientes. Durante o processo de interação com clientes, o sistema facilita o trabalho dos especialistas e permite melhorar a qualidade do atendimento aos clientes, contemplando necessidades dos diversos especialistas da empresa em assuntos como vendas, qualidade em serviços, finanças e gestão empresarial.O processo começou pela constituição de um grupo de trabalho interno para discutir estratégias e cronograma de implantação. A primeira decisão do grupo foi pelo desenvolvimento interno do software visando atender plenamente o "core business" da empresa. O processo começou pela constituição de um grupo de trabalho interno para discutir estratégias e cronograma de implantação. A primeira decisão do grupo foi pelo desenvolvimento interno do software visando atender plenamente o "core business" da empresa. O projeto contou com o conhecimento do negócio dos profissionais da empresa e auxilio de especialistas e consultores externos. O detalhamento do projeto, bem como os passos da pesquisa-ação, está descrito no corpo da dissertação.
Resumo:
A quantificação da precipitação é dificultada pela extrema aleatoriedade do fenômeno na natureza. Os métodos convencionais para mensuração da precipitação atuam no sentido de espacializar a precipitação mensurada pontualmente em postos pluviométricos para toda a área de interesse e, desta forma, uma rede com elevado número de postos bem distribuídos em toda a área de interesse é necessária para um resultado satisfatório. No entanto, é notória a escassez de postos pluviométricos e a má distribuição espacial dos poucos existentes, não somente no Brasil, mas em vastas áreas do globo. Neste contexto, as estimativas da precipitação com técnicas de sensoriamento remoto e geoprocessamento pretendem potencializar a utilização dos postos pluviométricos existentes através de uma espacialização baseada em critérios físicos. Além disto, o sensoriamento remoto é a ferramenta mais capaz para gerar estimativas de precipitação nos oceanos e nas vastas áreas continentais desprovidas de qualquer tipo de informação pluviométrica. Neste trabalho investigou-se o emprego de técnicas de sensoriamento remoto e geoprocessamento para estimativas de precipitação no sul do Brasil. Três algoritmos computadorizados foram testados, sendo utilizadas as imagens dos canais 1, 3 e 4 (visível, vapor d’água e infravermelho) do satélite GOES 8 (Geostacionary Operational Environmental Satellite – 8) fornecidas pelo Centro de Previsão de Tempo e Estudos Climáticos do Instituto Nacional de Pesquisas Espaciais. A área de estudo compreendeu todo o estado do Rio Grande do Sul, onde se utilizaram os dados pluviométricos diários derivados de 142 postos no ano de 1998. Os algoritmos citados buscam identificar as nuvens precipitáveis para construir modelos estatísticos que correlacionem as precipitações diária e decendial observadas em solo com determinadas características físicas das nuvens acumuladas durante o mesmo período de tempo e na mesma posição geográfica de cada pluviômetro considerado. Os critérios de decisão que norteiam os algoritmos foram baseados na temperatura do topo das nuvens (através do infravermelho termal), reflectância no canal visível, características de vizinhança e no plano de temperatura x gradiente de temperatura Os resultados obtidos pelos modelos estatísticos são expressos na forma de mapas de precipitação por intervalo de tempo que podem ser comparados com mapas de precipitação obtidas por meios convencionais.
Resumo:
Este trabalho apresenta um estudo de caso de mineração de dados no varejo. O negócio em questão é a comercialização de móveis e materiais de construção. A mineração foi realizada sobre informações geradas das transações de vendas por um período de 8 meses. Informações cadastrais de clientes também foram usadas e cruzadas com informações de venda, visando obter resultados que possam ser convertidos em ações que, por conseqüência, gerem lucro para a empresa. Toda a modelagem, preparação e transformação dos dados, foi feita visando facilitar a aplicação das técnicas de mineração que as ferramentas de mineração de dados proporcionam para a descoberta de conhecimento. O processo foi detalhado para uma melhor compreensão dos resultados obtidos. A metodologia CRISP usada no trabalho também é discutida, levando-se em conta as dificuldades e facilidades que se apresentaram durante as fases do processo de obtenção dos resultados. Também são analisados os pontos positivos e negativos das ferramentas de mineração utilizadas, o IBM Intelligent Miner e o WEKA - Waikato Environment for Knowledge Analysis, bem como de todos os outros softwares necessários para a realização do trabalho. Ao final, os resultados obtidos são apresentados e discutidos, sendo também apresentada a opinião dos proprietários da empresa sobre tais resultados e qual valor cada um deles poderá agregar ao negócio.