13 resultados para Parallel algorithm
em Scielo Saúde Pública - SP
Resumo:
In this paper we present an algorithm for the numerical simulation of the cavitation in the hydrodynamic lubrication of journal bearings. Despite the fact that this physical process is usually modelled as a free boundary problem, we adopted the equivalent variational inequality formulation. We propose a two-level iterative algorithm, where the outer iteration is associated to the penalty method, used to transform the variational inequality into a variational equation, and the inner iteration is associated to the conjugate gradient method, used to solve the linear system generated by applying the finite element method to the variational equation. This inner part was implemented using the element by element strategy, which is easily parallelized. We analyse the behavior of two physical parameters and discuss some numerical results. Also, we analyse some results related to the performance of a parallel implementation of the algorithm.
Resumo:
Classical serological screening assays for Chagas' disease are time consuming and subjective. The objective of the present work is to evaluate the enzyme immuno-assay (ELISA) methodology and to propose an algorithm for blood banks to be applied to Chagas' disease. Seven thousand, nine hundred and ninety nine blood donor samples were screened by both reverse passive hemagglutination (RPHA) and indirect immunofluorescence assay (IFA). Samples reactive on RPHA and/or IFA were submitted to supplementary RPHA, IFA and complement fixation (CFA) tests. This strategy allowed us to create a panel of 60 samples to evaluate the ELISA methodology from 3 different manufacturers. The sensitivity of the screening by IFA and the 3 different ELISA's was 100%. The specificity was better on ELISA methodology. For Chagas disease, ELISA seems to be the best test for blood donor screening, because it showed high sensitivity and specificity, it is not subjective and can be automated. Therefore, it was possible to propose an algorithm to screen samples and confirm donor results at the blood bank.
Resumo:
ABSTRACTThe Amazon várzeas are an important component of the Amazon biome, but anthropic and climatic impacts have been leading to forest loss and interruption of essential ecosystem functions and services. The objectives of this study were to evaluate the capability of the Landsat-based Detection of Trends in Disturbance and Recovery (LandTrendr) algorithm to characterize changes in várzeaforest cover in the Lower Amazon, and to analyze the potential of spectral and temporal attributes to classify forest loss as either natural or anthropogenic. We used a time series of 37 Landsat TM and ETM+ images acquired between 1984 and 2009. We used the LandTrendr algorithm to detect forest cover change and the attributes of "start year", "magnitude", and "duration" of the changes, as well as "NDVI at the end of series". Detection was restricted to areas identified as having forest cover at the start and/or end of the time series. We used the Support Vector Machine (SVM) algorithm to classify the extracted attributes, differentiating between anthropogenic and natural forest loss. Detection reliability was consistently high for change events along the Amazon River channel, but variable for changes within the floodplain. Spectral-temporal trajectories faithfully represented the nature of changes in floodplain forest cover, corroborating field observations. We estimated anthropogenic forest losses to be larger (1.071 ha) than natural losses (884 ha), with a global classification accuracy of 94%. We conclude that the LandTrendr algorithm is a reliable tool for studies of forest dynamics throughout the floodplain.
Resumo:
Background:Vascular remodeling, the dynamic dimensional change in face of stress, can assume different directions as well as magnitudes in atherosclerotic disease. Classical measurements rely on reference to segments at a distance, risking inappropriate comparison between dislike vessel portions.Objective:to explore a new method for quantifying vessel remodeling, based on the comparison between a given target segment and its inferred normal dimensions.Methods:Geometric parameters and plaque composition were determined in 67 patients using three-vessel intravascular ultrasound with virtual histology (IVUS-VH). Coronary vessel remodeling at cross-section (n = 27.639) and lesion (n = 618) levels was assessed using classical metrics and a novel analytic algorithm based on the fractional vessel remodeling index (FVRI), which quantifies the total change in arterial wall dimensions related to the estimated normal dimension of the vessel. A prediction model was built to estimate the normal dimension of the vessel for calculation of FVRI.Results:According to the new algorithm, “Ectatic” remodeling pattern was least common, “Complete compensatory” remodeling was present in approximately half of the instances, and “Negative” and “Incomplete compensatory” remodeling types were detected in the remaining. Compared to a traditional diagnostic scheme, FVRI-based classification seemed to better discriminate plaque composition by IVUS-VH.Conclusion:Quantitative assessment of coronary remodeling using target segment dimensions offers a promising approach to evaluate the vessel response to plaque growth/regression.
Resumo:
In Brazil, human and canine visceral leishmaniasis (CVL) caused byLeishmania infantum has undergone urbanisation since 1980, constituting a public health problem, and serological tests are tools of choice for identifying infected dogs. Until recently, the Brazilian zoonoses control program recommended enzyme-linked immunosorbent assays (ELISA) and indirect immunofluorescence assays (IFA) as the screening and confirmatory methods, respectively, for the detection of canine infection. The purpose of this study was to estimate the accuracy of ELISA and IFA in parallel or serial combinations. The reference standard comprised the results of direct visualisation of parasites in histological sections, immunohistochemical test, or isolation of the parasite in culture. Samples from 98 cases and 1,327 noncases were included. Individually, both tests presented sensitivity of 91.8% and 90.8%, and specificity of 83.4 and 53.4%, for the ELISA and IFA, respectively. When tests were used in parallel combination, sensitivity attained 99.2%, while specificity dropped to 44.8%. When used in serial combination (ELISA followed by IFA), decreased sensitivity (83.3%) and increased specificity (92.5%) were observed. Serial testing approach improved specificity with moderate loss in sensitivity. This strategy could partially fulfill the needs of public health and dog owners for a more accurate diagnosis of CVL.
Resumo:
It is presented a software developed with Delphi programming language to compute the reservoir's annual regulated active storage, based on the sequent-peak algorithm. Mathematical models used for that purpose generally require extended hydrological series. Usually, the analysis of those series is performed with spreadsheets or graphical representations. Based on that, it was developed a software for calculation of reservoir active capacity. An example calculation is shown by 30-years (from 1977 to 2009) monthly mean flow historical data, from Corrente River, located at São Francisco River Basin, Brazil. As an additional tool, an interface was developed to manage water resources, helping to manipulate data and to point out information that it would be of interest to the user. Moreover, with that interface irrigation districts where water consumption is higher can be analyzed as a function of specific seasonal water demands situations. From a practical application, it is possible to conclude that the program provides the calculation originally proposed. It was designed to keep information organized and retrievable at any time, and to show simulation on seasonal water demands throughout the year, contributing with the elements of study concerning reservoir projects. This program, with its functionality, is an important tool for decision making in the water resources management.
Management zones using fuzzy clustering based on spatial-temporal variability of soil and corn yield
Resumo:
Clustering soil and crop data can be used as a basis for the definition of management zones because the data are grouped into clusters based on the similar interaction of these variables. Therefore, the objective of this study was to identify management zones using fuzzy c-means clustering analysis based on the spatial and temporal variability of soil attributes and corn yield. The study site (18 by 250-m in size) was located in Jaboticabal, São Paulo/Brazil. Corn yield was measured in one hundred 4.5 by 10-m cells along four parallel transects (25 observations per transect) over five growing seasons between 2001 and 2010. Soil chemical and physical attributes were measured. SAS procedure MIXED was used to identify which variable(s) most influenced the spatial variability of corn yield over the five study years. Basis saturation (BS) was the variable that better related to corn yield, thus, semivariograms models were fitted for BS and corn yield and then, data values were krigged. Management Zone Analyst software was used to carry out the fuzzy c-means clustering algorithm. The optimum number of management zones can change over time, as well as the degree of agreement between the BS and corn yield management zone maps. Thus, it is very important take into account the temporal variability of crop yield and soil attributes to delineate management zones accurately.
Resumo:
The determination of the intersection curve between Bézier Surfaces may be seen as the composition of two separated problems: determining initial points and tracing the intersection curve from these points. The Bézier Surface is represented by a parametric function (polynomial with two variables) that maps a point in the tridimensional space from the bidimensional parametric space. In this article, it is proposed an algorithm to determine the initial points of the intersection curve of Bézier Surfaces, based on the solution of polynomial systems with the Projected Polyhedral Method, followed by a method for tracing the intersection curves (Marching Method with differential equations). In order to allow the use of the Projected Polyhedral Method, the equations of the system must be represented in terms of the Bernstein basis, and towards this goal it is proposed a robust and reliable algorithm to exactly transform a multivariable polynomial in terms of power basis to a polynomial written in terms of Bernstein basis .
Resumo:
This paper deals with the use of the conjugate gradient method of function estimation for the simultaneous identification of two unknown boundary heat fluxes in parallel plate channels. The fluid flow is assumed to be laminar and hydrodynamically developed. Temperature measurements taken inside the channel are used in the inverse analysis. The accuracy of the present solution approach is examined by using simulated measurements containing random errors, for strict cases involving functional forms with discontinuities and sharp-corners for the unknown functions. Three different types of inverse problems are addressed in the paper, involving the estimation of: (i) Spatially dependent heat fluxes; (ii) Time-dependent heat fluxes; and (iii) Time and spatially dependent heat fluxes.
Resumo:
In this paper we present a study of feasibility by using Cassino Parallel Manipulator (CaPaMan) as an earthquake simulator. We propose a suitable formulation to simulate the frequency, amplitude and acceleration magnitude of seismic motion by means of the movable platform motion by giving a suitable input motion. In this paper we have reported numerical simulations that simulate the three principal earthquake types for a seismic motion: one at the epicenter (having a vertical motion), another far from the epicenter (with the motion on a horizontal plane), and a combined general motion (with a vertical and horizontal motion).
Resumo:
We have developed a software called pp-Blast that uses the publicly available Blast package and PVM (parallel virtual machine) to partition a multi-sequence query across a set of nodes with replicated or shared databases. Benchmark tests show that pp-Blast running in a cluster of 14 PCs outperformed conventional Blast running in large servers. In addition, using pp-Blast and the cluster we were able to map all human cDNAs onto the draft of the human genome in less than 6 days. We propose here that the cost/benefit ratio of pp-Blast makes it appropriate for large-scale sequence analysis. The source code and configuration files for pp-Blast are available at http://www.ludwig.org.br/biocomp/tools/pp-blast.
Resumo:
Myocardial ischemia, as well as the induction agents used in anesthesia, may cause corrected QT interval (QTc) prolongation. The objective of this randomized, double-blind trial was to determine the effects of high- vs conventional-dose bolus rocuronium on QTc duration and the incidence of dysrhythmias following anesthesia induction and intubation. Fifty patients about to undergo coronary artery surgery were randomly allocated to receive conventional-dose (0.6 mg/kg, group C, n=25) or high-dose (1.2 mg/kg, group H, n=25) rocuronium after induction with etomidate and fentanyl. QTc, heart rate, and mean arterial pressure were recorded before induction (T0), after induction (T1), after rocuronium (just before laryngoscopy; T2), 2 min after intubation (T3), and 5 min after intubation (T4). The occurrence of dysrhythmias was recorded. In both groups, QTc was significantly longer at T3 than at baseline [475 vs 429 ms in group C (P=0.001), and 459 vs 434 ms in group H (P=0.005)]. The incidence of dysrhythmias in group C (28%) and in group H (24%) was similar. The QTc after high-dose rocuronium was not significantly longer than after conventional-dose rocuronium in patients about to undergo coronary artery surgery who were induced with etomidate and fentanyl. In both groups, compared with baseline, QTc was most prolonged at 2 min after intubation, suggesting that QTc prolongation may be due to the nociceptive stimulus of intubation.
Resumo:
Neoadjuvant chemotherapy has practical and theoretical advantages over adjuvant chemotherapy strategy in breast cancer (BC) management. Moreover, metronomic delivery has a more favorable toxicity profile. The present study examined the feasibility of neoadjuvant metronomic chemotherapy in two cohorts [HER2+ (TraQme) and HER2− (TAME)] of locally advanced BC. Twenty patients were prospectively enrolled (TraQme, n=9; TAME, n=11). Both cohorts received weekly paclitaxel at 100 mg/m2 during 8 weeks followed by weekly doxorubicin at 24 mg/m2 for 9 weeks in combination with oral cyclophosphamide at 100 mg/day (fixed dose). The HER2+ cohort received weekly trastuzumab. The study was interrupted because of safety issues. Thirty-six percent of patients in the TAME cohort and all patients from the TraQme cohort had stage III BC. Of note, 33% from the TraQme cohort and 66% from the TAME cohort displayed hormone receptor positivity in tumor tissue. The pathological complete response rates were 55% and 18% among patients enrolled in the TraQme and TAME cohorts, respectively. Patients in the TraQme cohort had more advanced BC stages at diagnosis, higher-grade pathological classification, and more tumors lacking hormone receptor expression, compared to the TAME cohort. The toxicity profile was also different. Two patients in the TraQme cohort developed pneumonitis, and in the TAME cohort we observed more hematological toxicity and hand-foot syndrome. The neoadjuvant metronomic chemotherapy regimen evaluated in this trial was highly effective in achieving a tumor response, especially in the HER2+ cohort. Pneumonitis was a serious, unexpected adverse event observed in this group. Further larger and randomized trials are warranted to evaluate the association between metronomic chemotherapy and trastuzumab treatment.