953 resultados para Interval discrete log problem
Resumo:
In the last decades, the globalized competition among cities and regions made them develop new strategies for branding and promoting their territory to attract tourists, investors, companies and residents. Major sports events - such as the Olympic Games, the FIFA World Cup or World and Continental Championships - have played an integral part in these strategies. Believing, with or without evidence, in the capacity of those events to improve the visibility and the economy of the host destination, many cities, regions and even countries have engaged in establishing sports events hosting strategies. The problem of the globalized competition in the sports events "market" is that many cities and regions do not have the resources - either financial, human or in terms of infrastructure - to compete in hosting major sports events. Consequently, many cities or regions have to turn to second-tier sports events. To organise those smaller events means less media coverage and more difficulty in finding sponsors, while the costs - both financial and in terms of services - stay high for the community. This paper analyses how Heritage Sporting Events (HSE) might be an opportunity for cities and regions engaged in sports events hosting strategies. HSE is an emerging concept that to date has been under-researched in the academic literature. Therefore, this paper aims to define the concept of HSE through an exploratory research study. A multidisciplinary literature review reveals two major characteristics of HSEs: the sustainability in the territory and the authenticity of the event constructed through a differentiation process. These characteristics, defined through multiple variables, give us the opportunity to observe the construction process of a sports event into a heritage object. This paper argues that HSEs can be seen as territorial resources that can represent a competitive advantage for host destinations. In conclusion, academics are invited to further research HSEs to better understand their construction process and their impacts on the territory, while local authorities are invited to consider HSEs for the branding and the promotion of their territory.
Resumo:
From a managerial point of view, the more effcient, simple, and parameter-free (ESP) an algorithm is, the more likely it will be used in practice for solving real-life problems. Following this principle, an ESP algorithm for solving the Permutation Flowshop Sequencing Problem (PFSP) is proposed in this article. Using an Iterated Local Search (ILS) framework, the so-called ILS-ESP algorithm is able to compete in performance with other well-known ILS-based approaches, which are considered among the most effcient algorithms for the PFSP. However, while other similar approaches still employ several parameters that can affect their performance if not properly chosen, our algorithm does not require any particular fine-tuning process since it uses basic "common sense" rules for the local search, perturbation, and acceptance criterion stages of the ILS metaheuristic. Our approach defines a new operator for the ILS perturbation process, a new acceptance criterion based on extremely simple and transparent rules, and a biased randomization process of the initial solution to randomly generate different alternative initial solutions of similar quality -which is attained by applying a biased randomization to a classical PFSP heuristic. This diversification of the initial solution aims at avoiding poorly designed starting points and, thus, allows the methodology to take advantage of current trends in parallel and distributed computing. A set of extensive tests, based on literature benchmarks, has been carried out in order to validate our algorithm and compare it against other approaches. These tests show that our parameter-free algorithm is able to compete with state-of-the-art metaheuristics for the PFSP. Also, the experiments show that, when using parallel computing, it is possible to improve the top ILS-based metaheuristic by just incorporating to it our biased randomization process with a high-quality pseudo-random number generator.
Resumo:
Biplots are graphical displays of data matrices based on the decomposition of a matrix as the product of two matrices. Elements of these two matrices are used as coordinates for the rows and columns of the data matrix, with an interpretation of the joint presentation that relies on the properties of the scalar product. Because the decomposition is not unique, there are several alternative ways to scale the row and column points of the biplot, which can cause confusion amongst users, especially when software packages are not united in their approach to this issue. We propose a new scaling of the solution, called the standard biplot, which applies equally well to a wide variety of analyses such as correspondence analysis, principal component analysis, log-ratio analysis and the graphical results of a discriminant analysis/MANOVA, in fact to any method based on the singular-value decomposition. The standard biplot also handles data matrices with widely different levels of inherent variance. Two concepts taken from correspondence analysis are important to this idea: the weighting of row and column points, and the contributions made by the points to the solution. In the standard biplot one set of points, usually the rows of the data matrix, optimally represent the positions of the cases or sample units, which are weighted and usually standardized in some way unless the matrix contains values that are comparable in their raw form. The other set of points, usually the columns, is represented in accordance with their contributions to the low-dimensional solution. As for any biplot, the projections of the row points onto vectors defined by the column points approximate the centred and (optionally) standardized data. The method is illustrated with several examples to demonstrate how the standard biplot copes in different situations to give a joint map which needs only one common scale on the principal axes, thus avoiding the problem of enlarging or contracting the scale of one set of points to make the biplot readable. The proposal also solves the problem in correspondence analysis of low-frequency categories that are located on the periphery of the map, giving the false impression that they are important.
Resumo:
Estudo descritivo e analítico, de base populacional, realizado com uma amostra aleatória e probabilística de 340 hipertensos, representativa da Estratégia Saúde da Família (ESF) de João Pessoa, PB. O estudo compõe a primeira parte de uma coorte iniciada em 2008. O instrumento utilizado foi adaptado do Primary Care Assessment Tool, revalidado no Brasil. A regressão logística avaliou a associação entre o controle pressórico, as variáveis sócio-demográficas e o indicador de adesão/vínculo. Dentre os 340 hipertensos, 32,6% era acompanhado pela ESF e 89,1% apresentou adesão/vínculo satisfatória. Os idosos apresentaram maior chance de controlar a pressão, o que sugere uma percepção melhor do autocuidado e maior adesão ao tratamento. O estudo permitiu dar visibilidade à problemática do controle da hipertensão por meio da avaliação do serviço. Espera-se que esse modelo possa ser adotado em outras localidades, gerando parâmetros para comparações entre municípios distintos.
Resumo:
The past four decades have witnessed an explosive growth in the field of networkbased facility location modeling. This is not at all surprising since location policy is one of the most profitable areas of applied systems analysis in regional science and ample theoretical and applied challenges are offered. Location-allocation models seek the location of facilities and/or services (e.g., schools, hospitals, and warehouses) so as to optimize one or several objectives generally related to the efficiency of the system or to the allocation of resources. This paper concerns the location of facilities or services in discrete space or networks, that are related to the public sector, such as emergency services (ambulances, fire stations, and police units), school systems and postal facilities. The paper is structured as follows: first, we will focus on public facility location models that use some type of coverage criterion, with special emphasis in emergency services. The second section will examine models based on the P-Median problem and some of the issues faced by planners when implementing this formulation in real world locational decisions. Finally, the last section will examine new trends in public sector facility location modeling.
Resumo:
When continuous data are coded to categorical variables, two types of coding are possible: crisp coding in the form of indicator, or dummy, variables with values either 0 or 1; or fuzzy coding where each observation is transformed to a set of "degrees of membership" between 0 and 1, using co-called membership functions. It is well known that the correspondence analysis of crisp coded data, namely multiple correspondence analysis, yields principal inertias (eigenvalues) that considerably underestimate the quality of the solution in a low-dimensional space. Since the crisp data only code the categories to which each individual case belongs, an alternative measure of fit is simply to count how well these categories are predicted by the solution. Another approach is to consider multiple correspondence analysis equivalently as the analysis of the Burt matrix (i.e., the matrix of all two-way cross-tabulations of the categorical variables), and then perform a joint correspondence analysis to fit just the off-diagonal tables of the Burt matrix - the measure of fit is then computed as the quality of explaining these tables only. The correspondence analysis of fuzzy coded data, called "fuzzy multiple correspondence analysis", suffers from the same problem, albeit attenuated. Again, one can count how many correct predictions are made of the categories which have highest degree of membership. But here one can also defuzzify the results of the analysis to obtain estimated values of the original data, and then calculate a measure of fit in the familiar percentage form, thanks to the resultant orthogonal decomposition of variance. Furthermore, if one thinks of fuzzy multiple correspondence analysis as explaining the two-way associations between variables, a fuzzy Burt matrix can be computed and the same strategy as in the crisp case can be applied to analyse the off-diagonal part of this matrix. In this paper these alternative measures of fit are defined and applied to a data set of continuous meteorological variables, which are coded crisply and fuzzily into three categories. Measuring the fit is further discussed when the data set consists of a mixture of discrete and continuous variables.
Resumo:
This study aimed at analyzing nipple trauma resulted from breastfeeding based on dermatological approach. Two integrative reviews of literature were conducted, the first related to definitions, classification and evaluation methods of nipple trauma and another about validation studies related to this theme. In the first part were included 20 studies and only one third defined nipple trauma, more than half did not defined the nipple’s injuries reported, and each author showed a particular way to assess the injuries, without consensus. In the second integrative review, no validation study or algorithm related to nipple trauma resulted from breastfeeding was found. This fact demonstrated that the nipple’s injuries mentioned in the first review did not go through validation studies, justifying the lack of consensus identified as far as definition, classification and assessment methods of nipple trauma.
Resumo:
This research paper has been written with the intention to discuss the problem of discipline in Cape Verdean secondary schools. While many of us discuss the effects that student misbehavior has on the student misbehavior has on the student, school and society as a whole, very few of us seek solutions which would impact on the prevention and management of this problem that each day becomes more complicated and harder to handle. This paper will discuss the need to better define discipline at the school level; identify the causes and factors that aggravate the problem, in addition, to provide what I hope to be useful strategies to better manage the problem as we make the effort to reclaim our schools and better educate our students. My research included surveys completed by teachers and student alike as they baffled over the question: what is discipline and how can we better manage discipline problems at our schools?
Resumo:
We start with a generalization of the well-known three-door problem:the n-door problem. The solution of this new problem leads us toa beautiful representation system for real numbers in (0,1] as alternated series, known in the literature as Pierce expansions. A closer look to Pierce expansions will take us to some metrical properties of sets defined through the Pierce expansions of its elements. Finally, these metrical properties will enable us to present 'strange' sets, similar to the classical Cantor set.
Resumo:
One of the assumptions of the Capacitated Facility Location Problem (CFLP) is thatdemand is known and fixed. Most often, this is not the case when managers take somestrategic decisions such as locating facilities and assigning demand points to thosefacilities. In this paper we consider demand as stochastic and we model each of thefacilities as an independent queue. Stochastic models of manufacturing systems anddeterministic location models are put together in order to obtain a formula for thebacklogging probability at a potential facility location.Several solution techniques have been proposed to solve the CFLP. One of the mostrecently proposed heuristics, a Reactive Greedy Adaptive Search Procedure, isimplemented in order to solve the model formulated. We present some computationalexperiments in order to evaluate the heuristics performance and to illustrate the use ofthis new formulation for the CFLP. The paper finishes with a simple simulationexercise.
Resumo:
The problems arising in commercial distribution are complex and involve several players and decision levels. One important decision is relatedwith the design of the routes to distribute the products, in an efficient and inexpensive way.This article deals with a complex vehicle routing problem that can beseen as a new extension of the basic vehicle routing problem. The proposed model is a multi-objective combinatorial optimization problemthat considers three objectives and multiple periods, which models in a closer way the real distribution problems. The first objective is costminimization, the second is balancing work levels and the third is amarketing objective. An application of the model on a small example, with5 clients and 3 days, is presented. The results of the model show the complexity of solving multi-objective combinatorial optimization problems and the contradiction between the several distribution management objective.
Resumo:
Scoring rules that elicit an entire belief distribution through the elicitation of point beliefsare time-consuming and demand considerable cognitive e¤ort. Moreover, the results are validonly when agents are risk-neutral or when one uses probabilistic rules. We investigate a classof rules in which the agent has to choose an interval and is rewarded (deterministically) onthe basis of the chosen interval and the realization of the random variable. We formulatean e¢ ciency criterion for such rules and present a speci.c interval scoring rule. For single-peaked beliefs, our rule gives information about both the location and the dispersion of thebelief distribution. These results hold for all concave utility functions.
Resumo:
Blowflies utilize discrete and ephemeral sites for breeding and larval nutrition. After the exhaustion of food, the larvae begin dispersing to search for sites to pupate or to additional food source, process referred as postfeeding larval dispersal. Some of the most important aspects of this process were investigated in Chrysomya megacephala, utilizing a circular arena to permit the radial dispersion of larvae from the center. To determinate the localization of each pupa, the arena was split in 72 equal sectors from the center. For each pupa, distance from the center of arena, weight and depth were determined. Statistical tests were performed to verify the relation among weight, depth and distance of burying for pupation. It was verified that the larvae that disperse farther are those with higher weights. The majority of individuals reached the depth of burying for pupation between 7 and 18 cm. The study of this process of dispersion can be utilized in the estimation of postmortem interval (PMI) for human corpses in medico-criminal investigations.