44 resultados para Quasi-analytical algorithms

em Instituto Politécnico do Porto, Portugal


Relevância:

20.00% 20.00%

Publicador:

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Introduction: Image resizing is a normal feature incorporated into the Nuclear Medicine digital imaging. Upsampling is done by manufacturers to adequately fit more the acquired images on the display screen and it is applied when there is a need to increase - or decrease - the total number of pixels. This paper pretends to compare the “hqnx” and the “nxSaI” magnification algorithms with two interpolation algorithms – “nearest neighbor” and “bicubic interpolation” – in the image upsampling operations. Material and Methods: Three distinct Nuclear Medicine images were enlarged 2 and 4 times with the different digital image resizing algorithms (nearest neighbor, bicubic interpolation nxSaI and hqnx). To evaluate the pixel’s changes between the different output images, 3D whole image plot profiles and surface plots were used as an addition to the visual approach in the 4x upsampled images. Results: In the 2x enlarged images the visual differences were not so noteworthy. Although, it was clearly noticed that bicubic interpolation presented the best results. In the 4x enlarged images the differences were significant, with the bicubic interpolated images presenting the best results. Hqnx resized images presented better quality than 4xSaI and nearest neighbor interpolated images, however, its intense “halo effect” affects greatly the definition and boundaries of the image contents. Conclusion: The hqnx and the nxSaI algorithms were designed for images with clear edges and so its use in Nuclear Medicine images is obviously inadequate. Bicubic interpolation seems, from the algorithms studied, the most suitable and its each day wider applications seem to show it, being assumed as a multi-image type efficient algorithm.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Introduction: A major focus of data mining process - especially machine learning researches - is to automatically learn to recognize complex patterns and help to take the adequate decisions strictly based on the acquired data. Since imaging techniques like MPI – Myocardial Perfusion Imaging on Nuclear Cardiology, can implicate a huge part of the daily workflow and generate gigabytes of data, there could be advantages on Computerized Analysis of data over Human Analysis: shorter time, homogeneity and consistency, automatic recording of analysis results, relatively inexpensive, etc.Objectives: The aim of this study relates with the evaluation of the efficacy of this methodology on the evaluation of MPI Stress studies and the process of decision taking concerning the continuation – or not – of the evaluation of each patient. It has been pursued has an objective to automatically classify a patient test in one of three groups: “Positive”, “Negative” and “Indeterminate”. “Positive” would directly follow to the Rest test part of the exam, the “Negative” would be directly exempted from continuation and only the “Indeterminate” group would deserve the clinician analysis, so allowing economy of clinician’s effort, increasing workflow fluidity at the technologist’s level and probably sparing time to patients. Methods: WEKA v3.6.2 open source software was used to make a comparative analysis of three WEKA algorithms (“OneR”, “J48” and “Naïve Bayes”) - on a retrospective study using the comparison with correspondent clinical results as reference, signed by nuclear cardiologist experts - on “SPECT Heart Dataset”, available on University of California – Irvine, at the Machine Learning Repository. For evaluation purposes, criteria as “Precision”, “Incorrectly Classified Instances” and “Receiver Operating Characteristics (ROC) Areas” were considered. Results: The interpretation of the data suggests that the Naïve Bayes algorithm has the best performance among the three previously selected algorithms. Conclusions: It is believed - and apparently supported by the findings - that machine learning algorithms could significantly assist, at an intermediary level, on the analysis of scintigraphic data obtained on MPI, namely after Stress acquisition, so eventually increasing efficiency of the entire system and potentially easing both roles of Technologists and Nuclear Cardiologists. In the actual continuation of this study, it is planned to use more patient information and significantly increase the population under study, in order to allow improving system accuracy.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A square-wave voltammetric (SWV) method and a flow injection analysis system with amperometric detection were developed for the determination of tramadol hydrochloride. The SWV method enables the determination of tramadol over the concentration range of 15-75 µM with a detection limit of 2.2 µM. Tramadol could be determined in concentrations between 9 and 50 µM at a sampling rate of 90 h-1, with a detection limit of 1.7 µM using the flow injection system. The electrochemical methods developed were successfully applied to the determination of tramadol in pharmaceutical dosage forms, without any pre-treatment of the samples. Recovery trials were performed to assess the accuracy of the results; the values were between 97 and 102% for both methods.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Celiac disease (CD) is an autoimmune enteropathy, characterized by an inappropriate T-cell-mediated immune response to the ingestion of certain dietary cereal proteins in genetically susceptible individuals. This disorder presents environmental, genetic, and immunological components. CD presents a prevalence of up to 1% in populations of European ancestry, yet a high percentage of cases remain underdiagnosed. The diagnosis and treatment should be made early since untreated disease causes growth retardation and atypical symptoms, like infertility or neurological disorders. The diagnostic criteria for CD, which requires endoscopy with small bowel biopsy, have been changing over the last few decades, especially due to the advent of serological tests with higher sensitivity and specificity. The use of serological markers can be very useful to rule out clinical suspicious cases and also to help monitor the patients, after adherence to a gluten-free diet. Since the current treatment consists of a life-long glutenfree diet, which leads to significant clinical and histological improvement, the standardization of an assay to assess in an unequivocal way gluten in gluten-free foodstuff is of major importance.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A Distribuição Individual Diária em Dose Unitária deve assegurar o maior rigor no Processo de Distribuição de Medicamentos, o que nem sempre se verificou no Hospital Pedro Hispano, dado que as Doses Unitárias devolvidas pelos Serviços de Internamento, nem sempre seguiam os Parâmetros de Qualidade. Assim, desenvolveu-se um estudo quasi-experimental, longitudinal, prospetivo e analítico, entre 1 de Março a 30 de Junho de 2012, no qual se analisaram as Doses Unitárias, devolvidas pelos Serviços de Internamento da Unidade Local de Saúde de Matosinhos a funcionam com Distribuição Individual Diária em Dose Unitária, tendo-se recolhido as não conformes relativamente aos Parâmetros de Qualidade instituídos. Na 1.ª fase do estudo (Recolha de amostras), foram contabilizadas 337 amostras, correspondentes a uma perda de 471 €. Na 2.ª fase do estudo (Implementação de acções correctivas às amostras previamente recolhidas), foram rejeitadas definitivamente 129 amostras, correspondentes a uma perda de 198 €. Na 4.ª fase do estudo (Segunda recolha de amostras), foram contabilizadas 228 amostras, correspondentes a uma perda de 199 €. A 3.ª fase do estudo incluiu a divulgação dos resultados às enfermeiras chefes dos serviços envolvidos no estudo e a 5.ª fase a comparação dos resultados obtidos na 1.ª, 2.ª e 4.ª fase do estudo. As amostras foram recolhidas em 16 serviços de internamento, sendo os valores monetários mais relevantes associados ao Espessante para alimentos líquidos e ao Imipeno + Cilastatina. A prática observada e sujeita a estudo aumenta os riscos associados ao consumo do medicamento e os custos relativos ao Processo, sendo de eliminar. A sua monitorização deve constituir uma rotina, uma vez que houve mudança de comportamentos. A redução/eliminação desta prática não conforme conduz à integração de novas tarefas com valor acrescentado, com aumento da fiabilidade do Processo.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Paracetamol is among the most worldwide consumed pharmaceuticals. Although its occurrence in the environment is well documented, data about the presence of its metabolites and transformation products is very scarce. The present work describes the development of an analytical method for the simultaneous determination of paracetamol, its principal metabolite (paracetamol-glucuronide) and its main transformation product (p-aminophenol) based on solid phase extraction (SPE) and high performance liquid chromatography coupled to diode array detection (HPLC-DAD). The method was applied to analysis of river waters, showing to be suitable to be used in routine analysis. Different SPE sorbents were compared and the use of two Oasis WAX cartridges in tandem proved to be the most adequate approach for sample clean up and pre-concentration. Under optimized conditions, limits of detection in the range 40–67 ng/L were obtained, as well as mean recoveries between 60 and 110% with relative standard deviations (RSD) below 6%. Finally, the developed SPE-HPLC/DAD method was successfully applied to the analysis of the selected compounds in samples from seven rivers located in the north of Portugal. Nevertheless all the compounds were detected, it was the first time that paracetamol-glucuronide was found in river water at concentrations up to 3.57 μg/L.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The paper formulates a genetic algorithm that evolves two types of objects in a plane. The fitness function promotes a relationship between the objects that is optimal when some kind of interface between them occurs. Furthermore, the algorithm adopts an hexagonal tessellation of the two-dimensional space for promoting an efficient method of the neighbour modelling. The genetic algorithm produces special patterns with resemblances to those revealed in percolation phenomena or in the symbiosis found in lichens. Besides the analysis of the spacial layout, a modelling of the time evolution is performed by adopting a distance measure and the modelling in the Fourier domain in the perspective of fractional calculus. The results reveal a consistent, and easy to interpret, set of model parameters for distinct operating conditions.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

To avoid additional hardware deployment, indoor localization systems have to be designed in such a way that they rely on existing infrastructure only. Besides the processing of measurements between nodes, localization procedure can include the information of all available environment information. In order to enhance the performance of Wi-Fi based localization systems, the innovative solution presented in this paper considers also the negative information. An indoor tracking method inspired by Kalman filtering is also proposed.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Consider the problem of assigning real-time tasks on a heterogeneous multiprocessor platform comprising two different types of processors — such a platform is referred to as two-type platform. We present two linearithmic timecomplexity algorithms, SA and SA-P, each providing the follow- ing guarantee. For a given two-type platform and a given task set, if there exists a feasible task-to-processor-type assignment such that tasks can be scheduled to meet deadlines by allowing them to migrate only between processors of the same type, then (i) using SA, it is guaranteed to find such a feasible task-to- processor-type assignment where the same restriction on task migration applies but given a platform in which processors are 1+α/2 times faster and (ii) SA-P succeeds in finding 2 a feasible task-to-processor assignment where tasks are not allowed to migrate between processors but given a platform in which processors are 1+α/times faster, where 0<α≤1. The parameter α is a property of the task set — it is the maximum utilization of any task which is less than or equal to 1.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

It has been shown that in reality at least two general scenarios of data structuring are possible: (a) a self-similar (SS) scenario when the measured data form an SS structure and (b) a quasi-periodic (QP) scenario when the repeated (strongly correlated) data form random sequences that are almost periodic with respect to each other. In the second case it becomes possible to describe their behavior and express a part of their randomness quantitatively in terms of the deterministic amplitude–frequency response belonging to the generalized Prony spectrum. This possibility allows us to re-examine the conventional concept of measurements and opens a new way for the description of a wide set of different data. In particular, it concerns different complex systems when the ‘best-fit’ model pretending to be the description of the data measured is absent but the barest necessity of description of these data in terms of the reduced number of quantitative parameters exists. The possibilities of the proposed approach and detection algorithm of the QP processes were demonstrated on actual data: spectroscopic data recorded for pure water and acoustic data for a test hole. The suggested methodology allows revising the accepted classification of different incommensurable and self-affine spatial structures and finding accurate interpretation of the generalized Prony spectroscopy that includes the Fourier spectroscopy as a partial case.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In this paper we discuss challenges and design principles of an implementation of slot-based tasksplitting algorithms into the Linux 2.6.34 version. We show that this kernel version is provided with the required features for implementing such scheduling algorithms. We show that the real behavior of the scheduling algorithm is very close to the theoretical. We run and discuss experiments on 4-core and 24-core machines.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Multiprocessors, particularly in the form of multicores, are becoming standard building blocks for executing reliable software. But their use for applications with hard real-time requirements is non-trivial. Well-known realtime scheduling algorithms in the uniprocessor context (Rate-Monotonic [1] or Earliest-Deadline-First [1]) do not perform well on multiprocessors. For this reason the scientific community in the area of real-time systems has produced new algorithms specifically for multiprocessors. In the meanwhile, a proposal [2] exists for extending the Ada language with new basic constructs which can be used for implementing new algorithms for real-time scheduling; the family of task splitting algorithms is one of them which was emphasized in the proposal [2]. Consequently, assessing whether existing task splitting multiprocessor scheduling algorithms can be implemented with these constructs is paramount. In this paper we present a list of state-of-art task-splitting multiprocessor scheduling algorithms and, for each of them, we present detailed Ada code that uses the new constructs.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A MATLAB/SIMULINK-based simulator was employed for studies concerning the control of baker’s yeast fed-batch fermentation. Four control algorithms were implemented and compared: the classical PID control, two discrete versions- modified velocity and position algorithms, and a fuzzy law. The simulation package was seen to be an efficient tool for the simulation and tests of control strategies of the nonlinear process.