272 resultados para cocktail party problem
Resumo:
A number of online algorithms have been developed that have small additional loss (regret) compared to the best “shifting expert”. In this model, there is a set of experts and the comparator is the best partition of the trial sequence into a small number of segments, where the expert of smallest loss is chosen in each segment. The regret is typically defined for worst-case data / loss sequences. There has been a recent surge of interest in online algorithms that combine good worst-case guarantees with much improved performance on easy data. A practically relevant class of easy data is the case when the loss of each expert is iid and the best and second best experts have a gap between their mean loss. In the full information setting, the FlipFlop algorithm by De Rooij et al. (2014) combines the best of the iid optimal Follow-The-Leader (FL) and the worst-case-safe Hedge algorithms, whereas in the bandit information case SAO by Bubeck and Slivkins (2012) competes with the iid optimal UCB and the worst-case-safe EXP3. We ask the same question for the shifting expert problem. First, we ask what are the simple and efficient algorithms for the shifting experts problem when the loss sequence in each segment is iid with respect to a fixed but unknown distribution. Second, we ask how to efficiently unite the performance of such algorithms on easy data with worst-case robustness. A particular intriguing open problem is the case when the comparator shifts within a small subset of experts from a large set under the assumption that the losses in each segment are iid.
Resumo:
Staffing rural and remote schools is an important policy issue for the public good. This paper examines the private issues it also poses for teachers with families working in these communities, as they seek to reconcile careers with educational choices for children. The paper first considers historical responses to staffing rural and remote schools in Australia, and the emergence of neoliberal policy encouraging marketisation of the education sector. We report on interviews about considerations motivating household mobility with 11 teachers across regional, rural and remote communities in Queensland. Like other middle-class parents, these teachers prioritised their children’s educational opportunities over career opportunities. The analysis demonstrates how teachers in rural and remote communities constitute a special group of educational consumers with insider knowledge and unique dilemmas around school choice. Their heightened anxieties around school choice under neoliberal policy are shown to contribute to the public issue of staffing rural and remote schools.
Resumo:
Guaranteeing Quality of Service (QoS) with minimum computation cost is the most important objective of cloud-based MapReduce computations. Minimizing the total computation cost of cloud-based MapReduce computations is done through MapReduce placement optimization. MapReduce placement optimization approaches can be classified into two categories: homogeneous MapReduce placement optimization and heterogeneous MapReduce placement optimization. It is generally believed that heterogeneous MapReduce placement optimization is more effective than homogeneous MapReduce placement optimization in reducing the total running cost of cloud-based MapReduce computations. This paper proposes a new approach to the heterogeneous MapReduce placement optimization problem. In this new approach, the heterogeneous MapReduce placement optimization problem is transformed into a constrained combinatorial optimization problem and is solved by an innovative constructive algorithm. Experimental results show that the running cost of the cloud-based MapReduce computation platform using this new approach is 24:3%-44:0% lower than that using the most popular homogeneous MapReduce placement approach, and 2:0%-36:2% lower than that using the heterogeneous MapReduce placement approach not considering the spare resources from the existing MapReduce computations. The experimental results have also demonstrated the good scalability of this new approach.
Resumo:
Trade union membership, both in aggregate numbers and in density, has declined in the majority of advanced economies globally over recent decades (Blanchflower, 2007). In Australia, the decline in the 1990s was somewhat more precipitate than in most countries (Peetz, 1998). As discussed in Chapter 1, reasons for the decline are multifactorial, including a more hostile environment to unionism created by employers and the state, difficulties ·with workplace union organisation, and structural change in the economy (Bryson and Gomez, 2005; Bryson et a!., 2011; Ebbinghaus et al., 2011; Payne, 1989; Waddington and Kerr, 2002; Waddington and Whitson, 1997). Our purpose in this chapter is to look beyond aggregate Australian union density data, to examine how age relates to membership decline, and how different age groups, particularly younger workers, are located in the story of union decline. The practical implications of this research are that understanding how unions relate to workers of different age groups, and to workers of different genders amongst those age groups, may lead to improved recruitment and better union organisation.
Resumo:
Particle Swarm Optimization (PSO) is a biologically inspired computational search and optimization method based on the social behaviors of birds flocking or fish schooling. Although, PSO is represented in solving many well-known numerical test problems, but it suffers from the premature convergence. A number of basic variations have been developed due to solve the premature convergence problem and improve quality of solution founded by the PSO. This study presents a comprehensive survey of the various PSO-based algorithms. As part of this survey, the authors have included a classification of the approaches and they have identify the main features of each proposal. In the last part of the study, some of the topics within this field that are considered as promising areas of future research are listed.
Resumo:
Index tracking is an investment approach where the primary objective is to keep portfolio return as close as possible to a target index without purchasing all index components. The main purpose is to minimize the tracking error between the returns of the selected portfolio and a benchmark. In this paper, quadratic as well as linear models are presented for minimizing the tracking error. The uncertainty is considered in the input data using a tractable robust framework that controls the level of conservatism while maintaining linearity. The linearity of the proposed robust optimization models allows a simple implementation of an ordinary optimization software package to find the optimal robust solution. The proposed model of this paper employs Morgan Stanley Capital International Index as the target index and the results are reported for six national indices including Japan, the USA, the UK, Germany, Switzerland and France. The performance of the proposed models is evaluated using several financial criteria e.g. information ratio, market ratio, Sharpe ratio and Treynor ratio. The preliminary results demonstrate that the proposed model lowers the amount of tracking error while raising values of portfolio performance measures.
Resumo:
Lattice-based cryptographic primitives are believed to offer resilience against attacks by quantum computers. We demonstrate the practicality of post-quantum key exchange by constructing cipher suites for the Transport Layer Security (TLS) protocol that provide key exchange based on the ring learning with errors (R-LWE) problem, we accompany these cipher suites with a rigorous proof of security. Our approach ties lattice-based key exchange together with traditional authentication using RSA or elliptic curve digital signatures: the post-quantum key exchange provides forward secrecy against future quantum attackers, while authentication can be provided using RSA keys that are issued by today's commercial certificate authorities, smoothing the path to adoption. Our cryptographically secure implementation, aimed at the 128-bit security level, reveals that the performance price when switching from non-quantum-safe key exchange is not too high. With our R-LWE cipher suites integrated into the Open SSL library and using the Apache web server on a 2-core desktop computer, we could serve 506 RLWE-ECDSA-AES128-GCM-SHA256 HTTPS connections per second for a 10 KiB payload. Compared to elliptic curve Diffie-Hellman, this means an 8 KiB increased handshake size and a reduction in throughput of only 21%. This demonstrates that provably secure post-quantum key-exchange can already be considered practical.
Resumo:
It is the Journal of Business Venturing's (JBV) 30th birthday. Although the community of entrepreneurship scholars deserves to celebrate JBV's achievements over the last 30 years (and congratulate the journal's parents—Ian Macmillan and S. Venkataraman), my focus is more on the future of entrepreneurship (and by extension JBV). A focus on entrepreneurship is both timeless and timely. On the one hand, entrepreneurship is timeless given the long-recognized importance of entrepreneurs to economies and societies (e.g., Jean Baptiste who supposedly coined the term in about 1800). On the other hand, a discussion of entrepreneurship is timely because now that the field of entrepreneurship has achieved legitimacy, it faces both opportunities and threats. It is thus timely to acknowledge the threats and think about opportunities to advance the field. A discussion of entrepreneurship is also timely because society faces a number of grand challenges (including the durability of poverty, environmental degradation [ Dorado and Ventresca, 2013]), challenges well suited to entrepreneurial responses...
Resumo:
In the mining optimisation literature, most researchers focused on two strategic-level and tactical-level open-pit mine optimisation problems, which are respectively termed ultimate pit limit (UPIT) or constrained pit limit (CPIT). However, many researchers indicate that the substantial numbers of variables and constraints in real-world instances (e.g., with 50-1000 thousand blocks) make the CPIT’s mixed integer programming (MIP) model intractable for use. Thus, it becomes a considerable challenge to solve the large scale CPIT instances without relying on exact MIP optimiser as well as the complicated MIP relaxation/decomposition methods. To take this challenge, two new graph-based algorithms based on network flow graph and conjunctive graph theory are developed by taking advantage of problem properties. The performance of our proposed algorithms is validated by testing recent large scale benchmark UPIT and CPIT instances’ datasets of MineLib in 2013. In comparison to best known results from MineLib, it is shown that the proposed algorithms outperform other CPIT solution approaches existing in the literature. The proposed graph-based algorithms leads to a more competent mine scheduling optimisation expert system because the third-party MIP optimiser is no longer indispensable and random neighbourhood search is not necessary.
Resumo:
In the past few years, the virtual machine (VM) placement problem has been studied intensively and many algorithms for the VM placement problem have been proposed. However, those proposed VM placement algorithms have not been widely used in today's cloud data centers as they do not consider the migration cost from current VM placement to the new optimal VM placement. As a result, the gain from optimizing VM placement may be less than the loss of the migration cost from current VM placement to the new VM placement. To address this issue, this paper presents a penalty-based genetic algorithm (GA) for the VM placement problem that considers the migration cost in addition to the energy-consumption of the new VM placement and the total inter-VM traffic flow in the new VM placement. The GA has been implemented and evaluated by experiments, and the experimental results show that the GA outperforms two well known algorithms for the VM placement problem.
Resumo:
This chapter addresses opportunities for problem posing in developing young children’s statistical literacy, with a focus on student-directed investigations. Although the notion of problem posing has broadened in recent years, there nevertheless remains limited research on how problem posing can be integrated within the regular mathematics curriculum, especially in the areas of statistics and probability. The chapter first reviews briefly aspects of problem posing that have featured in the literature over the years. Consideration is next given to the importance of developing children’s statistical literacy in which problem posing is an inherent feature. Some findings from a school playground investigation conducted in four, fourth-grade classes illustrate the different ways in which children posed investigative questions, how they made predictions about their outcomes and compared these with their findings, and the ways in which they chose to represent their findings.
Resumo:
Engineering-based modeling activities provide a rich source of meaningful situations that capitalize on and extend students’ routine learning. By integrating such activities within existing curricula, students better appreciate how their school learning in mathematics and science applies to problems in the outside world...
Resumo:
Research on problem solving in the mathematics curriculum has spanned many decades, yielding pendulum-like swings in recommendations on various issues. Ongoing debates concern the effectiveness of teaching general strategies and heuristics, the role of mathematical content (as the means versus the learning goal of problem solving), the role of context, and the proper emphasis on the social and affective dimensions of problem solving (e.g., Lesh & Zawojewski, 2007; Lester, 2013; Lester & Kehle, 2003; Schoenfeld, 1985, 2008; Silver, 1985). Various scholarly perspectives—including cognitive and behavioral science, neuroscience, the discipline of mathematics, educational philosophy, and sociocultural stances—have informed these debates, often generating divergent resolutions. Perhaps due to this uncertainty, educators’ efforts over the years to improve students’ mathematical problem-solving skills have had disappointing results. Qualitative and quantitative studies consistently reveal mathematics students’ struggles to solve problems more significant than routine exercises (OECD, 2014; Boaler, 2009)...
What triggers problem recognition? An exploration on young Australian male problematic online gamers
Resumo:
Help-seeking is a complex decision-making process that first begins with problem recognition. However, little is understood about the conceptualisation of the helpseeking process and the triggers of problem recognition. This research proposes the use of the Critical Incident Technique (CIT) to examine and classify incidents that serve as key triggers of problem recognition among young Australian male problematic online gamers. The research provides a classification of five different types of triggers that will aid social marketers into developing effective early detection, prevention and treatment focused social marketing interventions.
Resumo:
Concept mapping involves determining relevant concepts from a free-text input, where concepts are defined in an external reference ontology. This is an important process that underpins many applications for clinical information reporting, derivation of phenotypic descriptions, and a number of state-of-the-art medical information retrieval methods. Concept mapping can be cast into an information retrieval (IR) problem: free-text mentions are treated as queries and concepts from a reference ontology as the documents to be indexed and retrieved. This paper presents an empirical investigation applying general-purpose IR techniques for concept mapping in the medical domain. A dataset used for evaluating medical information extraction is adapted to measure the effectiveness of the considered IR approaches. Standard IR approaches used here are contrasted with the effectiveness of two established benchmark methods specifically developed for medical concept mapping. The empirical findings show that the IR approaches are comparable with one benchmark method but well below the best benchmark.