11 resultados para cluster algorithms
em Dalarna University College Electronic Archive
Resumo:
To connect different electrical, network and data devices with the minimum cost and shortest path, is a complex job. In huge buildings, where the devices are placed at different locations on different floors and only some specific routes are available to pass the cables and buses, the shortest path search becomes more complex. The aim of this thesis project is, to develop an application which indentifies the best path to connect all objects or devices by following the specific routes.To address the above issue we adopted three algorithms Greedy Algorithm, Simulated Annealing and Exhaustive search and analyzed their results. The given problem is similar to Travelling Salesman Problem. Exhaustive search is a best algorithm to solve this problem as it checks each and every possibility and give the accurate result but it is an impractical solution because of huge time consumption. If no. of objects increased from 12 it takes hours to search the shortest path. Simulated annealing is emerged with some promising results with lower time cost. As of probabilistic nature, Simulated annealing could be non optimal but it gives a near optimal solution in a reasonable duration. Greedy algorithm is not a good choice for this problem. So, simulated annealing is proved best algorithm for this problem. The project has been implemented in C-language which takes input and store output in an Excel Workbook
Resumo:
Parkinson's disease (PD) is the second most common neurodegenerative disorder (after Alzheimer's disease) and directly affects upto 5 million people worldwide. The stages (Hoehn and Yaar) of disease has been predicted by many methods which will be helpful for the doctors to give the dosage according to it. So these methods were brought up based on the data set which includes about seventy patients at nine clinics in Sweden. The purpose of the work is to analyze unsupervised technique with supervised neural network techniques in order to make sure the collected data sets are reliable to make decisions. The data which is available was preprocessed before calculating the features of it. One of the complex and efficient feature called wavelets has been calculated to present the data set to the network. The dimension of the final feature set has been reduced using principle component analysis. For unsupervised learning k-means gives the closer result around 76% while comparing with supervised techniques. Back propagation and J4 has been used as supervised model to classify the stages of Parkinson's disease where back propagation gives the variance percentage of 76-82%. The results of both these models have been analyzed. This proves that the data which are collected are reliable to predict the disease stages in Parkinson's disease.
Resumo:
Parkinson's disease (PD) is a degenerative illness whose cardinal symptoms include rigidity, tremor, and slowness of movement. In addition to its widely recognized effects PD can have a profound effect on speech and voice.The speech symptoms most commonly demonstrated by patients with PD are reduced vocal loudness, monopitch, disruptions of voice quality, and abnormally fast rate of speech. This cluster of speech symptoms is often termed Hypokinetic Dysarthria.The disease can be difficult to diagnose accurately, especially in its early stages, due to this reason, automatic techniques based on Artificial Intelligence should increase the diagnosing accuracy and to help the doctors make better decisions. The aim of the thesis work is to predict the PD based on the audio files collected from various patients.Audio files are preprocessed in order to attain the features.The preprocessed data contains 23 attributes and 195 instances. On an average there are six voice recordings per person, By using data compression technique such as Discrete Cosine Transform (DCT) number of instances can be minimized, after data compression, attribute selection is done using several WEKA build in methods such as ChiSquared, GainRatio, Infogain after identifying the important attributes, we evaluate attributes one by one by using stepwise regression.Based on the selected attributes we process in WEKA by using cost sensitive classifier with various algorithms like MultiPass LVQ, Logistic Model Tree(LMT), K-Star.The classified results shows on an average 80%.By using this features 95% approximate classification of PD is acheived.This shows that using the audio dataset, PD could be predicted with a higher level of accuracy.
Resumo:
Colour segmentation is the most commonly used method in road signs detection. Road sign contains several basic colours such as red, yellow, blue and white which depends on countries.The objective of this thesis is to do an evaluation of the four colour segmentation algorithms. Dynamic Threshold Algorithm, A Modification of de la Escalera’s Algorithm, the Fuzzy Colour Segmentation Algorithm and Shadow and Highlight Invariant Algorithm. The processing time and segmentation success rate as criteria are used to compare the performance of the four algorithms. And red colour is selected as the target colour to complete the comparison. All the testing images are selected from the Traffic Signs Database of Dalarna University [1] randomly according to the category. These road sign images are taken from a digital camera mounted in a moving car in Sweden.Experiments show that the Fuzzy Colour Segmentation Algorithm and Shadow and Highlight Invariant Algorithm are more accurate and stable to detect red colour of road signs. And the method could also be used in other colours analysis research. The yellow colour which is chosen to evaluate the performance of the four algorithms can reference Master Thesis of Yumei Liu.
Resumo:
The Intelligent Algorithm is designed for theusing a Battery source. The main function is to automate the Hybrid System through anintelligent Algorithm so that it takes the decision according to the environmental conditionsfor utilizing the Photovoltaic/Solar Energy and in the absence of this, Fuel Cell energy isused. To enhance the performance of the Fuel Cell and Photovoltaic Cell we used batterybank which acts like a buffer and supply the current continuous to the load. To develop the main System whlogic based controller was used. Fuzzy Logic based controller used to develop this system,because they are chosen to be feasible for both controlling the decision process and predictingthe availability of the available energy on the basis of current Photovoltaic and Battery conditions. The Intelligent Algorithm is designed to optimize the performance of the system and to selectthe best available energy source(s) in regard of the input parameters. The enhance function of these Intelligent Controller is to predict the use of available energy resources and turn on thatparticular source for efficient energy utilization. A fuzzy controller was chosen to take thedecisions for the efficient energy utilization from the given resources. The fuzzy logic basedcontroller is designed in the Matlab-Simulink environment. Initially, the fuzzy based ruleswere built. Then MATLAB based simulation system was designed and implemented. Thenthis whole proposed model is simulated and tested for the accuracy of design and performanceof the system.
Resumo:
This paper presents a two-step pseudo likelihood estimation technique for generalized linear mixed models with the random effects being correlated between groups. The core idea is to deal with the intractable integrals in the likelihood function by multivariate Taylor's approximation. The accuracy of the estimation technique is assessed in a Monte-Carlo study. An application of it with a binary response variable is presented using a real data set on credit defaults from two Swedish banks. Thanks to the use of two-step estimation technique, the proposed algorithm outperforms conventional pseudo likelihood algorithms in terms of computational time.
Resumo:
In this paper, we propose a new method for solving large scale p-median problem instances based on real data. We compare different approaches in terms of runtime, memory footprint and quality of solutions obtained. In order to test the different methods on real data, we introduce a new benchmark for the p-median problem based on real Swedish data. Because of the size of the problem addressed, up to 1938 candidate nodes, a number of algorithms, both exact and heuristic, are considered. We also propose an improved hybrid version of a genetic algorithm called impGA. Experiments show that impGA behaves as well as other methods for the standard set of medium-size problems taken from Beasley’s benchmark, but produces comparatively good results in terms of quality, runtime and memory footprint on our specific benchmark based on real Swedish data.
Resumo:
In a global economy, manufacturers mainly compete with cost efficiency of production, as the price of raw materials are similar worldwide. Heavy industry has two big issues to deal with. On the one hand there is lots of data which needs to be analyzed in an effective manner, and on the other hand making big improvements via investments in cooperate structure or new machinery is neither economically nor physically viable. Machine learning offers a promising way for manufacturers to address both these problems as they are in an excellent position to employ learning techniques with their massive resource of historical production data. However, choosing modelling a strategy in this setting is far from trivial and this is the objective of this article. The article investigates characteristics of the most popular classifiers used in industry today. Support Vector Machines, Multilayer Perceptron, Decision Trees, Random Forests, and the meta-algorithms Bagging and Boosting are mainly investigated in this work. Lessons from real-world implementations of these learners are also provided together with future directions when different learners are expected to perform well. The importance of feature selection and relevant selection methods in an industrial setting are further investigated. Performance metrics have also been discussed for the sake of completion.
Resumo:
BACKGROUND: Facilitation of local women's groups may reportedly reduce neonatal mortality. It is not known whether facilitation of groups composed of local health care staff and politicians can improve perinatal outcomes. We hypothesised that facilitation of local stakeholder groups would reduce neonatal mortality (primary outcome) and improve maternal, delivery, and newborn care indicators (secondary outcomes) in Quang Ninh province, Vietnam. METHODS AND FINDINGS: In a cluster-randomized design 44 communes were allocated to intervention and 46 to control. Laywomen facilitated monthly meetings during 3 years in groups composed of health care staff and key persons in the communes. A problem-solving approach was employed. Births and neonatal deaths were monitored, and interviews were performed in households of neonatal deaths and of randomly selected surviving infants. A latent period before effect is expected in this type of intervention, but this timeframe was not pre-specified. Neonatal mortality rate (NMR) from July 2008 to June 2011 was 16.5/1,000 (195 deaths per 11,818 live births) in the intervention communes and 18.4/1,000 (194 per 10,559 live births) in control communes (adjusted odds ratio [OR] 0.96 [95% CI 0.73-1.25]). There was a significant downward time trend of NMR in intervention communes (p = 0.003) but not in control communes (p = 0.184). No significant difference in NMR was observed during the first two years (July 2008 to June 2010) while the third year (July 2010 to June 2011) had significantly lower NMR in intervention arm: adjusted OR 0.51 (95% CI 0.30-0.89). Women in intervention communes more frequently attended antenatal care (adjusted OR 2.27 [95% CI 1.07-4.8]). CONCLUSIONS: A randomized facilitation intervention with local stakeholder groups composed of primary care staff and local politicians working for three years with a perinatal problem-solving approach resulted in increased attendance to antenatal care and reduced neonatal mortality after a latent period. TRIAL REGISTRATION: Current Controlled Trials ISRCTN44599712. Please see later in the article for the Editors' Summary.
Resumo:
BACKGROUND: Reminder systems in electronic patient records (EPR) have proven to affect both health care professionals' behaviour and patient outcomes. The aim of this cluster randomised trial was to investigate the effects of implementing a clinical practice guideline (CPG) for peripheral venous catheters (PVCs) in paediatric care in the format of reminders integrated in the EPRs, on PVC-related complications, and on registered nurses' (RNs') self-reported adherence to the guideline. An additional aim was to study the relationship between contextual factors and the outcomes of the intervention. METHODS: The study involved 12 inpatient units at a paediatric university hospital. The reminders included choice of PVC, hygiene, maintenance, and daily inspection of PVC site. Primary outcome was documented signs and symptoms of PVC-related complications at removal, retrieved from the EPR. Secondary outcome was RNs' adherence to a PVC guideline, collected through a questionnaire that also included RNs' perceived work context, as measured by the Alberta Context Tool. Units were allocated into two strata, based on occurrence of PVCs. A blinded simple draw of lots from each stratum randomised six units to the control and intervention groups, respectively. Units were not blinded. The intervention group included 626 PVCs at baseline and 618 post-intervention and the control group 724 PVCs at baseline and 674 post-intervention. RNs included at baseline were 212 (65.4 %) and 208 (71.5 %) post-intervention. RESULTS: No significant effect was found for the computer reminders on PVC-related complications nor on RNs' adherence to the guideline recommendations. The complication rate at baseline and post-intervention was 40.6 % (95 % confidence interval (CI) 36.7-44.5) and 41.9 % (95 % CI 38.0-45.8), for the intervention group and 40.3 % (95 % CI 36.8-44.0) and 46.9 % (95 % CI 43.1-50.7) for the control. In general, RNs' self-rated work context varied from moderately low to moderately high, indicating that conditions for a successful implementation to occur were less optimal. CONCLUSIONS: The reminders might have benefitted from being accompanied by a tailored intervention that targeted specific barriers, such as the low frequency of recorded reasons for removal, the low adherence to daily inspection of PVC sites, and the lack of regular feedback to the RNs. TRIAL REGISTRATION: Current Controlled Trials ISRCTN44819426.
Resumo:
Bin planning (arrangements) is a key factor in the timber industry. Improper planning of the storage bins may lead to inefficient transportation of resources, which threaten the overall efficiency and thereby limit the profit margins of sawmills. To address this challenge, a simulation model has been developed. However, as numerous alternatives are available for arranging bins, simulating all possibilities will take an enormous amount of time and it is computationally infeasible. A discrete-event simulation model incorporating meta-heuristic algorithms has therefore been investigated in this study. Preliminary investigations indicate that the results achieved by GA based simulation model are promising and better than the other meta-heuristic algorithm. Further, a sensitivity analysis has been done on the GA based optimal arrangement which contributes to gaining insights and knowledge about the real system that ultimately leads to improved and enhanced efficiency in sawmill yards. It is expected that the results achieved in the work will support timber industries in making optimal decisions with respect to arrangement of storage bins in a sawmill yard.