5 resultados para Selection process


Relevância:

60.00% 60.00%

Publicador:

Resumo:

Estimates of HIV prevalence are important for policy in order to establish the health status of a country's population and to evaluate the effectiveness of population-based interventions and campaigns. However, participation rates in testing for surveillance conducted as part of household surveys, on which many of these estimates are based, can be low. HIV positive individuals may be less likely to participate because they fear disclosure, in which case estimates obtained using conventional approaches to deal with missing data, such as imputation-based methods, will be biased. We develop a Heckman-type simultaneous equation approach which accounts for non-ignorable selection, but unlike previous implementations, allows for spatial dependence and does not impose a homogeneous selection process on all respondents. In addition, our framework addresses the issue of separation, where for instance some factors are severely unbalanced and highly predictive of the response, which would ordinarily prevent model convergence. Estimation is carried out within a penalized likelihood framework where smoothing is achieved using a parametrization of the smoothing criterion which makes estimation more stable and efficient. We provide the software for straightforward implementation of the proposed approach, and apply our methodology to estimating national and sub-national HIV prevalence in Swaziland, Zimbabwe and Zambia.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Variations are inherent in all manufacturing processes and can significantly affect the quality of a final assembly, particularly in multistage assembly systems. Existing research in variation management has primarily focused on incorporating GD&T factors into variation propagation models in order to predict product quality and allocate tolerances. However, process induced variation, which has a key influence on process planning, has not been fully studied. Furthermore, the link between variation and cost has not been well established, in particular the effect that assembly process selection has on the final quality and cost of a product. To overcome these barriers, this paper proposes a novel method utilizing process capabilities to establish the relationship between variation and cost. The methodology is discussed using a real industrial case study. The benefits include determining the optimum configuration of an assembly system and facilitating rapid introduction of novel assembly techniques to achieve a competitive edge.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this paper, we investigate the secrecy outage performance of spectrum sharing multiple-input multiple-output networks using generalized transmit antenna selection with maximal ratio combining over Nakagami-m channels. In particular, the outdated channel state information is considered at the process of antenna selection due to feedback delay. Considering a practical passive eavesdropper scenario, we derive the exact and asymptotic closed-form expressions of secrecy outage probability, which enable us to evaluate the secrecy performance with high efficiency and present a new design insight into the impact of key parameters on the secrecy performance. In addition, the analytical results demonstrate that the achievable secrecy diversity order is only determined by the parameters of the secondary network, while other parameters related to primary or eavesdropper’s channels have a significantly impact on the secrecy coding gain. 

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Traditional heuristic approaches to the Examination Timetabling Problem normally utilize a stochastic method during Optimization for the selection of the next examination to be considered for timetabling within the neighbourhood search process. This paper presents a technique whereby the stochastic method has been augmented with information from a weighted list gathered during the initial adaptive construction phase, with the purpose of intelligently directing examination selection. In addition, a Reinforcement Learning technique has been adapted to identify the most effective portions of the weighted list in terms of facilitating the greatest potential for overall solution improvement. The technique is tested against the 2007 International Timetabling Competition datasets with solutions generated within a time frame specified by the competition organizers. The results generated are better than those of the competition winner in seven of the twelve examinations, while being competitive for the remaining five examinations. This paper also shows experimentally how using reinforcement learning has improved upon our previous technique.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

To maintain the pace of development set by Moore's law, production processes in semiconductor manufacturing are becoming more and more complex. The development of efficient and interpretable anomaly detection systems is fundamental to keeping production costs low. As the dimension of process monitoring data can become extremely high anomaly detection systems are impacted by the curse of dimensionality, hence dimensionality reduction plays an important role. Classical dimensionality reduction approaches, such as Principal Component Analysis, generally involve transformations that seek to maximize the explained variance. In datasets with several clusters of correlated variables the contributions of isolated variables to explained variance may be insignificant, with the result that they may not be included in the reduced data representation. It is then not possible to detect an anomaly if it is only reflected in such isolated variables. In this paper we present a new dimensionality reduction technique that takes account of such isolated variables and demonstrate how it can be used to build an interpretable and robust anomaly detection system for Optical Emission Spectroscopy data.