979 resultados para RM(rate monotonic)algorithm
Resumo:
A systolic array to implement lattice-reduction-aided lineardetection is proposed for a MIMO receiver. The lattice reductionalgorithm and the ensuing linear detections are operated in the same array, which can be hardware-efficient. All-swap lattice reduction algorithm (ASLR) is considered for the systolic design.ASLR is a variant of the LLL algorithm, which processes all lattice basis vectors within one iteration. Lattice-reduction-aided linear detection based on ASLR and LLL algorithms have very similarbit-error-rate performance, while ASLR is more time efficient inthe systolic array, especially for systems with a large number ofantennas.
Resumo:
BACKGROUND: The hospital readmission rate has been proposed as an important outcome indicator computable from routine statistics. However, most commonly used measures raise conceptual issues. OBJECTIVES: We sought to evaluate the usefulness of the computerized algorithm for identifying avoidable readmissions on the basis of minimum bias, criterion validity, and measurement precision. RESEARCH DESIGN AND SUBJECTS: A total of 131,809 hospitalizations of patients discharged alive from 49 hospitals were used to compare the predictive performance of risk adjustment methods. A subset of a random sample of 570 medical records of discharge/readmission pairs in 12 hospitals were reviewed to estimate the predictive value of the screening of potentially avoidable readmissions. MEASURES: Potentially avoidable readmissions, defined as readmissions related to a condition of the previous hospitalization and not expected as part of a program of care and occurring within 30 days after the previous discharge, were identified by a computerized algorithm. Unavoidable readmissions were considered as censored events. RESULTS: A total of 5.2% of hospitalizations were followed by a potentially avoidable readmission, 17% of them in a different hospital. The predictive value of the screen was 78%; 27% of screened readmissions were judged clearly avoidable. The correlation between the hospital rate of clearly avoidable readmission and all readmissions rate, potentially avoidable readmissions rate or the ratio of observed to expected readmissions were respectively 0.42, 0.56 and 0.66. Adjustment models using clinical information performed better. CONCLUSION: Adjusted rates of potentially avoidable readmissions are scientifically sound enough to warrant their inclusion in hospital quality surveillance.
Resumo:
Summary Background: We previously derived a clinical prognostic algorithm to identify patients with pulmonary embolism (PE) who are at low-risk of short-term mortality who could be safely discharged early or treated entirely in an outpatient setting. Objectives: To externally validate the clinical prognostic algorithm in an independent patient sample. Methods: We validated the algorithm in 983 consecutive patients prospectively diagnosed with PE at an emergency department of a university hospital. Patients with none of the algorithm's 10 prognostic variables (age >/= 70 years, cancer, heart failure, chronic lung disease, chronic renal disease, cerebrovascular disease, pulse >/= 110/min., systolic blood pressure < 100 mm Hg, oxygen saturation < 90%, and altered mental status) at baseline were defined as low-risk. We compared 30-day overall mortality among low-risk patients based on the algorithm between the validation and the original derivation sample. We also assessed the rate of PE-related and bleeding-related mortality among low-risk patients. Results: Overall, the algorithm classified 16.3% of patients with PE as low-risk. Mortality at 30 days was 1.9% among low-risk patients and did not differ between the validation and the original derivation sample. Among low-risk patients, only 0.6% died from definite or possible PE, and 0% died from bleeding. Conclusions: This study validates an easy-to-use, clinical prognostic algorithm for PE that accurately identifies patients with PE who are at low-risk of short-term mortality. Low-risk patients based on our algorithm are potential candidates for less costly outpatient treatment.
Resumo:
Background: Attention to patients with acute minor-illnesses requesting same-day consultation represents a major burden in primary care. The workload is assumed by general practitioners in many countries. A number of reports suggest that care to these patients may be provided, at in least in part, by nurses. However, there is scarce information with respect to the applicability of a program of nurse management for adult patients with acute minor-illnesses in large areas. The aim of this study is to assess the effectiveness of a program of nurse algorithm-guided care for adult patients with acute minor illnesses requesting same-day consultation in primary care in a largely populated area. Methods: A cross-sectional study of all adult patients seeking same day consultation for 16 common acute minor illnesses in a large geographical area with 284 primary care practices. Patients were included in a program of nurse case management using management algorithms. The main outcome measure was case resolution, defined as completion of the algorithm by the nurse without need of referral of the patient to the general practitioner. The secondary outcome measure was return to consultation, defined as requirement of new consultation for the same reason as the first one, in primary care within a 7-day period. Results: During a two year period (April 2009-April 2011), a total of 1,209,669 consultations were performed in the program. Case resolution was achieved by nurses in 62.5% of consultations. The remaining cases were referred to a general practitioner. Resolution rates ranged from 94.2% in patients with burns to 42% in patients with upper respiratory symptoms. None of the 16 minor illnesses had a resolution rate below 40%. Return to consultation during a 7-day period was low, only 4.6%. Conclusions: A program of algorithms-guided care is effective for nurse case management of patients requesting same day consultation for minor illnesses in primary care.
Resumo:
Background: Attention to patients with acute minor-illnesses requesting same-day consultation represents a major burden in primary care. The workload is assumed by general practitioners in many countries. A number of reports suggest that care to these patients may be provided, at in least in part, by nurses. However, there is scarce information with respect to the applicability of a program of nurse management for adult patients with acute minor-illnesses in large areas. The aim of this study is to assess the effectiveness of a program of nurse algorithm-guided care for adult patients with acute minor illnesses requesting same-day consultation in primary care in a largely populated area. Methods: A cross-sectional study of all adult patients seeking same day consultation for 16 common acute minor illnesses in a large geographical area with 284 primary care practices. Patients were included in a program of nurse case management using management algorithms. The main outcome measure was case resolution, defined as completion of the algorithm by the nurse without need of referral of the patient to the general practitioner. The secondary outcome measure was return to consultation, defined as requirement of new consultation for the same reason as the first one, in primary care within a 7-day period. Results: During a two year period (April 2009-April 2011), a total of 1,209,669 consultations were performed in the program. Case resolution was achieved by nurses in 62.5% of consultations. The remaining cases were referred to a general practitioner. Resolution rates ranged from 94.2% in patients with burns to 42% in patients with upper respiratory symptoms. None of the 16 minor illnesses had a resolution rate below 40%. Return to consultation during a 7-day period was low, only 4.6%. Conclusions: A program of algorithms-guided care is effective for nurse case management of patients requesting same day consultation for minor illnesses in primary care.
Resumo:
Background: Attention to patients with acute minor-illnesses requesting same-day consultation represents a major burden in primary care. The workload is assumed by general practitioners in many countries. A number of reports suggest that care to these patients may be provided, at in least in part, by nurses. However, there is scarce information with respect to the applicability of a program of nurse management for adult patients with acute minor-illnesses in large areas. The aim of this study is to assess the effectiveness of a program of nurse algorithm-guided care for adult patients with acute minor illnesses requesting same-day consultation in primary care in a largely populated area. Methods: A cross-sectional study of all adult patients seeking same day consultation for 16 common acute minor illnesses in a large geographical area with 284 primary care practices. Patients were included in a program of nurse case management using management algorithms. The main outcome measure was case resolution, defined as completion of the algorithm by the nurse without need of referral of the patient to the general practitioner. The secondary outcome measure was return to consultation, defined as requirement of new consultation for the same reason as the first one, in primary care within a 7-day period. Results: During a two year period (April 2009-April 2011), a total of 1,209,669 consultations were performed in the program. Case resolution was achieved by nurses in 62.5% of consultations. The remaining cases were referred to a general practitioner. Resolution rates ranged from 94.2% in patients with burns to 42% in patients with upper respiratory symptoms. None of the 16 minor illnesses had a resolution rate below 40%. Return to consultation during a 7-day period was low, only 4.6%. Conclusions: A program of algorithms-guided care is effective for nurse case management of patients requesting same day consultation for minor illnesses in primary care.
Resumo:
Background: Research in epistasis or gene-gene interaction detection for human complex traits has grown over the last few years. It has been marked by promising methodological developments, improved translation efforts of statistical epistasis to biological epistasis and attempts to integrate different omics information sources into the epistasis screening to enhance power. The quest for gene-gene interactions poses severe multiple-testing problems. In this context, the maxT algorithm is one technique to control the false-positive rate. However, the memory needed by this algorithm rises linearly with the amount of hypothesis tests. Gene-gene interaction studies will require a memory proportional to the squared number of SNPs. A genome-wide epistasis search would therefore require terabytes of memory. Hence, cache problems are likely to occur, increasing the computation time. In this work we present a new version of maxT, requiring an amount of memory independent from the number of genetic effects to be investigated. This algorithm was implemented in C++ in our epistasis screening software MBMDR-3.0.3. We evaluate the new implementation in terms of memory efficiency and speed using simulated data. The software is illustrated on real-life data for Crohn’s disease. Results: In the case of a binary (affected/unaffected) trait, the parallel workflow of MBMDR-3.0.3 analyzes all gene-gene interactions with a dataset of 100,000 SNPs typed on 1000 individuals within 4 days and 9 hours, using 999 permutations of the trait to assess statistical significance, on a cluster composed of 10 blades, containing each four Quad-Core AMD Opteron(tm) Processor 2352 2.1 GHz. In the case of a continuous trait, a similar run takes 9 days. Our program found 14 SNP-SNP interactions with a multiple-testing corrected p-value of less than 0.05 on real-life Crohn’s disease (CD) data. Conclusions: Our software is the first implementation of the MB-MDR methodology able to solve large-scale SNP-SNP interactions problems within a few days, without using much memory, while adequately controlling the type I error rates. A new implementation to reach genome-wide epistasis screening is under construction. In the context of Crohn’s disease, MBMDR-3.0.3 could identify epistasis involving regions that are well known in the field and could be explained from a biological point of view. This demonstrates the power of our software to find relevant phenotype-genotype higher-order associations.
Resumo:
3 Summary 3. 1 English The pharmaceutical industry has been facing several challenges during the last years, and the optimization of their drug discovery pipeline is believed to be the only viable solution. High-throughput techniques do participate actively to this optimization, especially when complemented by computational approaches aiming at rationalizing the enormous amount of information that they can produce. In siiico techniques, such as virtual screening or rational drug design, are now routinely used to guide drug discovery. Both heavily rely on the prediction of the molecular interaction (docking) occurring between drug-like molecules and a therapeutically relevant target. Several softwares are available to this end, but despite the very promising picture drawn in most benchmarks, they still hold several hidden weaknesses. As pointed out in several recent reviews, the docking problem is far from being solved, and there is now a need for methods able to identify binding modes with a high accuracy, which is essential to reliably compute the binding free energy of the ligand. This quantity is directly linked to its affinity and can be related to its biological activity. Accurate docking algorithms are thus critical for both the discovery and the rational optimization of new drugs. In this thesis, a new docking software aiming at this goal is presented, EADock. It uses a hybrid evolutionary algorithm with two fitness functions, in combination with a sophisticated management of the diversity. EADock is interfaced with .the CHARMM package for energy calculations and coordinate handling. A validation was carried out on 37 crystallized protein-ligand complexes featuring 11 different proteins. The search space was defined as a sphere of 15 R around the center of mass of the ligand position in the crystal structure, and conversely to other benchmarks, our algorithms was fed with optimized ligand positions up to 10 A root mean square deviation 2MSD) from the crystal structure. This validation illustrates the efficiency of our sampling heuristic, as correct binding modes, defined by a RMSD to the crystal structure lower than 2 A, were identified and ranked first for 68% of the complexes. The success rate increases to 78% when considering the five best-ranked clusters, and 92% when all clusters present in the last generation are taken into account. Most failures in this benchmark could be explained by the presence of crystal contacts in the experimental structure. EADock has been used to understand molecular interactions involved in the regulation of the Na,K ATPase, and in the activation of the nuclear hormone peroxisome proliferatoractivated receptors a (PPARa). It also helped to understand the action of common pollutants (phthalates) on PPARy, and the impact of biotransformations of the anticancer drug Imatinib (Gleevec®) on its binding mode to the Bcr-Abl tyrosine kinase. Finally, a fragment-based rational drug design approach using EADock was developed, and led to the successful design of new peptidic ligands for the a5ß1 integrin, and for the human PPARa. In both cases, the designed peptides presented activities comparable to that of well-established ligands such as the anticancer drug Cilengitide and Wy14,643, respectively. 3.2 French Les récentes difficultés de l'industrie pharmaceutique ne semblent pouvoir se résoudre que par l'optimisation de leur processus de développement de médicaments. Cette dernière implique de plus en plus. de techniques dites "haut-débit", particulièrement efficaces lorsqu'elles sont couplées aux outils informatiques permettant de gérer la masse de données produite. Désormais, les approches in silico telles que le criblage virtuel ou la conception rationnelle de nouvelles molécules sont utilisées couramment. Toutes deux reposent sur la capacité à prédire les détails de l'interaction moléculaire entre une molécule ressemblant à un principe actif (PA) et une protéine cible ayant un intérêt thérapeutique. Les comparatifs de logiciels s'attaquant à cette prédiction sont flatteurs, mais plusieurs problèmes subsistent. La littérature récente tend à remettre en cause leur fiabilité, affirmant l'émergence .d'un besoin pour des approches plus précises du mode d'interaction. Cette précision est essentielle au calcul de l'énergie libre de liaison, qui est directement liée à l'affinité du PA potentiel pour la protéine cible, et indirectement liée à son activité biologique. Une prédiction précise est d'une importance toute particulière pour la découverte et l'optimisation de nouvelles molécules actives. Cette thèse présente un nouveau logiciel, EADock, mettant en avant une telle précision. Cet algorithme évolutionnaire hybride utilise deux pressions de sélections, combinées à une gestion de la diversité sophistiquée. EADock repose sur CHARMM pour les calculs d'énergie et la gestion des coordonnées atomiques. Sa validation a été effectuée sur 37 complexes protéine-ligand cristallisés, incluant 11 protéines différentes. L'espace de recherche a été étendu à une sphère de 151 de rayon autour du centre de masse du ligand cristallisé, et contrairement aux comparatifs habituels, l'algorithme est parti de solutions optimisées présentant un RMSD jusqu'à 10 R par rapport à la structure cristalline. Cette validation a permis de mettre en évidence l'efficacité de notre heuristique de recherche car des modes d'interactions présentant un RMSD inférieur à 2 R par rapport à la structure cristalline ont été classés premier pour 68% des complexes. Lorsque les cinq meilleures solutions sont prises en compte, le taux de succès grimpe à 78%, et 92% lorsque la totalité de la dernière génération est prise en compte. La plupart des erreurs de prédiction sont imputables à la présence de contacts cristallins. Depuis, EADock a été utilisé pour comprendre les mécanismes moléculaires impliqués dans la régulation de la Na,K ATPase et dans l'activation du peroxisome proliferatoractivated receptor a (PPARa). Il a également permis de décrire l'interaction de polluants couramment rencontrés sur PPARy, ainsi que l'influence de la métabolisation de l'Imatinib (PA anticancéreux) sur la fixation à la kinase Bcr-Abl. Une approche basée sur la prédiction des interactions de fragments moléculaires avec protéine cible est également proposée. Elle a permis la découverte de nouveaux ligands peptidiques de PPARa et de l'intégrine a5ß1. Dans les deux cas, l'activité de ces nouveaux peptides est comparable à celles de ligands bien établis, comme le Wy14,643 pour le premier, et le Cilengitide (PA anticancéreux) pour la seconde.
Resumo:
Introduction New evidence from randomized controlled and etiology of fever studies, the availability of reliable RDT for malaria, and novel technologies call for revision of the IMCI strategy. We developed a new algorithm based on (i) a systematic review of published studies assessing the safety and appropriateness of RDT and antibiotic prescription, (ii) results from a clinical and microbiological investigation of febrile children aged <5 years, (iii) international expert IMCI opinions. The aim of this study was to assess the safety of the new algorithm among patients in urban and rural areas of Tanzania.Materials and Methods The design was a controlled noninferiority study. Enrolled children aged 2-59 months with any illness were managed either by a study clinician using the new Almanach algorithm (two intervention health facilities), or clinicians using standard practice, including RDT (two control HF). At day 7 and day 14, all patients were reassessed. Patients who were ill in between or not cured at day 14 were followed until recovery or death. Primary outcome was rate of complications, secondary outcome rate of antibiotic prescriptions.Results 1062 children were recruited. Main diagnoses were URTI 26%, pneumonia 19% and gastroenteritis (9.4%). 98% (531/541) were cured at D14 in the Almanach arm and 99.6% (519/521) in controls. Rate of secondary hospitalization was 0.2% in each. One death occurred in controls. None of the complications was due to withdrawal of antibiotics or antimalarials at day 0. Rate of antibiotic use was 19% in the Almanach arm and 84% in controls.Conclusion Evidence suggests that the new algorithm, primarily aimed at the rational use of drugs, is as safe as standard practice and leads to a drastic reduction of antibiotic use. The Almanach is currently being tested for clinician adherence to proposed procedures when used on paper or a mobile phone
Resumo:
INTRODUCTION: The decline of malaria and scale-up of rapid diagnostic tests calls for a revision of IMCI. A new algorithm (ALMANACH) running on mobile technology was developed based on the latest evidence. The objective was to ensure that ALMANACH was safe, while keeping a low rate of antibiotic prescription. METHODS: Consecutive children aged 2-59 months with acute illness were managed using ALMANACH (2 intervention facilities), or standard practice (2 control facilities) in Tanzania. Primary outcomes were proportion of children cured at day 7 and who received antibiotics on day 0. RESULTS: 130/842 (15∙4%) in ALMANACH and 241/623 (38∙7%) in control arm were diagnosed with an infection in need for antibiotic, while 3∙8% and 9∙6% had malaria. 815/838 (97∙3%;96∙1-98.4%) were cured at D7 using ALMANACH versus 573/623 (92∙0%;89∙8-94∙1%) using standard practice (p<0∙001). Of 23 children not cured at D7 using ALMANACH, 44% had skin problems, 30% pneumonia, 26% upper respiratory infection and 13% likely viral infection at D0. Secondary hospitalization occurred for one child using ALMANACH and one who eventually died using standard practice. At D0, antibiotics were prescribed to 15∙4% (12∙9-17∙9%) using ALMANACH versus 84∙3% (81∙4-87∙1%) using standard practice (p<0∙001). 2∙3% (1∙3-3.3) versus 3∙2% (1∙8-4∙6%) received an antibiotic secondarily. CONCLUSION: Management of children using ALMANACH improve clinical outcome and reduce antibiotic prescription by 80%. This was achieved through more accurate diagnoses and hence better identification of children in need of antibiotic treatment or not. The building on mobile technology allows easy access and rapid update of the decision chart. TRIAL REGISTRATION: Pan African Clinical Trials Registry PACTR201011000262218.
Resumo:
OBJECTIVE: To review the available knowledge on epidemiology and diagnoses of acute infections in children aged 2 to 59 months in primary care setting and develop an electronic algorithm for the Integrated Management of Childhood Illness to reach optimal clinical outcome and rational use of medicines. METHODS: A structured literature review in Medline, Embase and the Cochrane Database of Systematic Review (CDRS) looked for available estimations of diseases prevalence in outpatients aged 2-59 months, and for available evidence on i) accuracy of clinical predictors, and ii) performance of point-of-care tests for targeted diseases. A new algorithm for the management of childhood illness (ALMANACH) was designed based on evidence retrieved and results of a study on etiologies of fever in Tanzanian children outpatients. FINDINGS: The major changes in ALMANACH compared to IMCI (2008 version) are the following: i) assessment of 10 danger signs, ii) classification of non-severe children into febrile and non-febrile illness, the latter receiving no antibiotics, iii) classification of pneumonia based on a respiratory rate threshold of 50 assessed twice for febrile children 12-59 months; iv) malaria rapid diagnostic test performed for all febrile children. In the absence of identified source of fever at the end of the assessment, v) urine dipstick performed for febrile children <2 years to consider urinary tract infection, vi) classification of 'possible typhoid' for febrile children >2 years with abdominal tenderness; and lastly vii) classification of 'likely viral infection' in case of negative results. CONCLUSION: This smartphone-run algorithm based on new evidence and two point-of-care tests should improve the quality of care of <5 year children and lead to more rational use of antimicrobials.
Resumo:
La grossesse induit de profonds changements hémodynamiques et métaboliques de l’organisme maternel qui ont des conséquences sur le cœur. L’adaptation du cœur à cette condition physiologique nécessite un remodelage de sa structure et par conséquent des ajustements de sa fonction. Les mécanismes responsables de ces adaptations sont en grande partie inconnus. Cependant, ces connaissances sont essentielles pour la compréhension des complications cardiovasculaires, telle que l’hypertension gestationnelle (HG), qui constituent un risque pour la santé de la mère et du fœtus. Afin de caractériser les adaptations du cœur lors de la grossesse, l’originalité de notre approche expérimentale consistait à étudier le remodelage à l’échelle des cardiomyocytes du ventricule gauche. Ainsi, notre premier objectif était de déterminer les modifications structurales et fonctionnelles des cardiomyocytes chez la rate en vue d’identifier les altérations lors de l’HG. Chez les rates gestantes, le remodelage structural des cardiomyocytes se caractérise par une hypertrophie cellulaire avec une augmentation proportionnelle des dimensions. L’HG a été induite par un supplément sodique (0.9% NaCl) dans la diète. L’inadaptation structurale lors de l’HG se traduit par une diminution du volume cellulaire. L’étude des modifications fonctionnelles a révélé que lors de la gestation le fonctionnement contractile des cellules est dépendant de l’adaptation du métabolisme maternel. En effet, les substrats énergétiques, lactate et pyruvate, induisent une augmentation de la contractilité des cardiomyocytes. Cet effet est plus faible dans les cellules des rates hypertendues, ce qui suggère des anomalies du couplage excitation-contraction, dans lequel les courants calciques de type L (ICa-L) jouent un rôle important. Paradoxalement, le lactate et le pyruvate ont induit une augmentation de la densité des courants ICa-L seulement chez les rates hypertendues. Le récepteur aux minéralocorticoïdes (RM) est connu pour son implication dans le remodelage structuro-fonctionnel du cœur dans les conditions pathologiques mais pas dans celui induit par la grossesse. Notre deuxième objectif était donc de déterminer le rôle du RM dans l’adaptation de la morphologie et de la contractilité des cardiomyocytes. Des rates gestantes ont été traitées avec le canrénoate de potassium (20 mg/kg/jr), un antagoniste des RM. L’inhibition des RM pendant la gestation empêche l’hypertrophie cellulaire. De plus, l’inhibition des RM bloque l’effet du lactate et du pyruvate sur la contractilité. Chez la femme, la grossesse est associée à des changements des propriétés électriques du cœur. Sur l’électrocardiogramme, l’intervalle QTc est plus long, témoignant de la prolongation de la repolarisation. Les mécanismes régulant cette adaptation restent encore inconnus. Ainsi, notre troisième objectif était de déterminer le rôle du RM dans l’adaptation de la repolarisation. Chez la rate gestante, l’intervalle QTc est prolongé ce qui est corroboré par la diminution des courants potassiques Ito et IK1. L’inhibition des RM pendant la gestation empêche la prolongation de l’intervalle QTc et la diminution des courants Ito. Les travaux exposés dans cette thèse apportent une vision plus précise du remodelage cardiaque induit par la grossesse, qui est permise par l’étude à l’échelle cellulaire. Nos résultats montrent que lors de la gestation et de l’HG les cardiomyocytes subissent des remodelages morphologiques contrastés. Notre étude a aussi révélé que lors de la gestation, la fonction contractile est tributaire des adaptations métaboliques et que cette relation est altérée lors de l’HG. Nos travaux montrent que la régulation de ces adaptations gestationnelles fait intervenir le RM au niveau de la morphologie, de la relation métabolisme/fonctionnement contractile et de la repolarisation. En faisant avancer les connaissances sur l’hypertrophie de la grossesse, ces travaux vont permettre d’améliorer la compréhension des complications cardiovasculaires gestationnelles.
Resumo:
This paper presents a novel two-pass algorithm constituted by Linear Hashtable Motion Estimation Algorithm (LHMEA) and Hexagonal Search (HEXBS). compensation. for block base motion On the basis of research from previous algorithms, especially an on-the-edge motion estimation algorithm called hexagonal search (HEXBS), we propose the LHMEA and the Two-Pass Algorithm (TPA). We introduce hashtable into video compression. In this paper we employ LHMEA for the first-pass search in all the Macroblocks (MB) in the picture. Motion Vectors (MV) are then generated from the first-pass and are used as predictors for second-pass HEXBS motion estimation, which only searches a small number of MBs. The evaluation of the algorithm considers the three important metrics being time, compression rate and PSNR. The performance of the algorithm is evaluated by using standard video sequences and the results are compared to current algorithms. Experimental results show that the proposed algorithm can offer the same compression rate as the Full Search. LHMEA with TPA has significant improvement on HEXBS and shows a direction for improving other fast motion estimation algorithms, for example Diamond Search.
Resumo:
This paper presents an improved Two-Pass Hexagonal (TPA) algorithm constituted by Linear Hashtable Motion Estimation Algorithm (LHMEA) and Hexagonal Search (HEXBS) for motion estimation. In the TPA, Motion Vectors (MV) are generated from the first-pass LHMEA and are used as predictors for second-pass HEXBS motion estimation, which only searches a small number of Macroblocks (MBs). The hashtable structure of LHMEA is improved compared to the original TPA and LHMEA. The evaluation of the algorithm considers the three important metrics being processing time, compression rate and PSNR. The performance of the algorithm is evaluated by using standard video sequences and the results are compared to current algorithms.
Resumo:
An extensive set of machine learning and pattern classification techniques trained and tested on KDD dataset failed in detecting most of the user-to-root attacks. This paper aims to provide an approach for mitigating negative aspects of the mentioned dataset, which led to low detection rates. Genetic algorithm is employed to implement rules for detecting various types of attacks. Rules are formed of the features of the dataset identified as the most important ones for each attack type. In this way we introduce high level of generality and thus achieve high detection rates, but also gain high reduction of the system training time. Thenceforth we re-check the decision of the user-to- root rules with the rules that detect other types of attacks. In this way we decrease the false-positive rate. The model was verified on KDD 99, demonstrating higher detection rates than those reported by the state- of-the-art while maintaining low false-positive rate.