106 resultados para extraction methods.
Resumo:
This paper proposes two meta-heuristics (Genetic Algorithm and Evolutionary Particle Swarm Optimization) for solving a 15 bid-based case of Ancillary Services Dispatch in an Electricity Market. A Linear Programming approach is also included for comparison purposes. A test case based on the dispatch of Regulation Down, Regulation Up, Spinning Reserve and Non-Spinning Reserve services is used to demonstrate that the use of meta-heuristics is suitable for solving this kind of optimization problem. Faster execution times and lower computational resources requirements are the most relevant advantages of the used meta-heuristics when compared with the Linear Programming approach.
Resumo:
Electricity market players operating in a liberalized environment requires access to an adequate decision support tool, allowing them to consider all the business opportunities and take strategic decisions. Ancillary services represent a good negotiation opportunity that must be considered by market players. For this, decision support tool must include ancillary market simulation. This paper proposes two different methods (Linear Programming and Genetic Algorithm approaches) for ancillary services dispatch. The methodologies are implemented in MASCEM, a multi-agent based electricity market simulator. A test case based on California Independent System Operator (CAISO) data concerning the dispatch of Regulation Down, Regulation Up, Spinning Reserve and Non-Spinning Reserve services is included in this paper.
Resumo:
Objectives : The purpose of this article is to find out differences between surveys using paper and online questionnaires. The author has deep knowledge in the case of questions concerning opinions in the development of survey based research, e.g. the limits of postal and online questionnaires. Methods : In the physician studies carried out in 1995 (doctors graduated in 1982-1991), 2000 (doctors graduated in 1982-1996), 2005 (doctors graduated in 1982-2001), 2011 (doctors graduated in 1977-2006) and 457 family doctors in 2000, were used paper and online questionnaires. The response rates were 64%, 68%, 64%, 49% and 73%, respectively. Results : The results of the physician studies showed that there were differences between methods. These differences were connected with using paper-based questionnaire and online questionnaire and response rate. The online-based survey gave a lower response rate than the postal survey. The major advantages of online survey were short response time; very low financial resource needs and data were directly loaded in the data analysis software, thus saved time and resources associated with the data entry process. Conclusions : The current article helps researchers with planning the study design and choosing of the right data collection method.
Resumo:
In real optimization problems, usually the analytical expression of the objective function is not known, nor its derivatives, or they are complex. In these cases it becomes essential to use optimization methods where the calculation of the derivatives, or the verification of their existence, is not necessary: the Direct Search Methods or Derivative-free Methods are one solution. When the problem has constraints, penalty functions are often used. Unfortunately the choice of the penalty parameters is, frequently, very difficult, because most strategies for choosing it are heuristics strategies. As an alternative to penalty function appeared the filter methods. A filter algorithm introduces a function that aggregates the constrained violations and constructs a biobjective problem. In this problem the step is accepted if it either reduces the objective function or the constrained violation. This implies that the filter methods are less parameter dependent than a penalty function. In this work, we present a new direct search method, based on simplex methods, for general constrained optimization that combines the features of the simplex method and filter methods. This method does not compute or approximate any derivatives, penalty constants or Lagrange multipliers. The basic idea of simplex filter algorithm is to construct an initial simplex and use the simplex to drive the search. We illustrate the behavior of our algorithm through some examples. The proposed methods were implemented in Java.
Resumo:
The relentless discovery of cancer biomarkers demands improved methods for their detection. In this work, we developed protein imprinted polymer on three-dimensional gold nanoelectrode ensemble (GNEE) to detect epithelial ovarian cancer antigen-125 (CA 125), a protein biomarker associated with ovarian cancer. CA 125 is the standard tumor marker used to follow women during or after treatment for epithelial ovarian cancer. The template protein CA 125 was initially incorporated into the thin-film coating and, upon extraction of protein from the accessible surfaces on the thin film, imprints for CA 125 were formed. The fabrication and analysis of the CA 125 imprinted GNEE was done by using cyclic voltammetry (CV), differential pulse voltammetry (DPV) and electrochemical impedance spectroscopy (EIS) techniques. The surfaces of the very thin, protein imprinted sites on GNEE are utilized for immunospecific capture of CA 125 molecules, and the mass of bound on the electrode surface can be detected as a reduction in the faradic current from the redox marker. Under optimal conditions, the developed sensor showed good increments at the studied concentration range of 0.5–400 U mL−1. The lowest detection limit was found to be 0.5 U mL−1. Spiked human blood serum and unknown real serum samples were analyzed. The presence of non-specific proteins in the serum did not significantly affect the sensitivity of our assay. Molecular imprinting using synthetic polymers and nanomaterials provides an alternative approach to the trace detection of biomarker proteins.
Resumo:
This paper reports a novel application of microwave-assisted extraction (MAE) of polyphenols from brewer’s spent grains (BSG). A 24 orthogonal composite design was used to obtain the optimal conditions of MAE. The influence of the MAE operational parameters (extraction time, temperature, solvent volume and stirring speed) on the extraction yield of ferulic acid was investigated through response surface methodology. The results showed that the optimal conditions were 15 min extraction time, 100 °C extraction temperature, 20 mL of solvent, and maximum stirring speed. Under these conditions, the yield of ferulic acid was 1.31±0.04% (w/w), which was fivefold higher than that obtained with conventional solid–liquid extraction techniques. The developed new extraction method considerably reduces extraction time, energy and solvent consumption, while generating fewer wastes. HPLC-DADMS analysis indicated that other hydroxycinnamic acids and several ferulic acid dehydrodimers, as well as one dehydrotrimer were also present, confirming that BSG is a valuable source of antioxidant compounds.
Resumo:
This paper presents the study of the remediation of sandy soils containing six of the most common contaminants (benzene, toluene, ethylbenzene, xylene, trichloroethylene and perchloroethylene) using soil vapour extraction (SVE). The influence of soil water content on the process efficiency was evaluated considering the soil type and the contaminant. For artificially contaminated soils with negligible clay contents and natural organic matter it was concluded that: (i) all the remediation processes presented efficiencies above 92%; (ii) an increase of the soil water content led to a more time-consuming remediation; (iii) longer remediation periods were observed for contaminants with lower vapour pressures and lower water solubilities due to mass transfer limitations. Based on these results an easy and relatively fast procedure was developed for the prediction of the remediation times of real soils; 83% of the remediation times were predicted with relative deviations below 14%.
Resumo:
Soil vapor extraction (SVE) is an efficient, well-known and widely applied soil remediation technology. However, under certain conditions it cannot achieve the defined cleanup goals, requiring further treatment, for example, through bioremediation (BR). The sequential application of these technologies is presented as a valid option but is not yet entirely studied. This work presents the study of the remediation of ethylbenzene (EB)-contaminated soils, with different soil water and natural organic matter (NOMC) contents, using sequential SVE and BR. The obtained results allow the conclusion that: (1) SVE was sufficient to reach the cleanup goals in 63% of the experiments (all the soils with NOMC below 4%), (2) higher NOMCs led to longer SVE remediation times, (3) BR showed to be a possible and cost-effective option when EB concentrations were lower than 335 mg kgsoil −1, and (4) concentrations of EB above 438 mg kgsoil −1 showed to be inhibitory for microbial activity.
Resumo:
An accurate and sensitive method for determination of 18 polycyclic aromatic hydrocarbons (PAHs) (16 PAHs considered by USEPA as priority pollutants, dibenzo[a,l]pyrene and benzo[j]fluoranthene) in fish samples was validated. Analysis was performed by microwave-assisted extraction and liquid chromatography with photodiode array and fluorescence detection. Response surface methodology was used to find the optimal extraction parameters. Validation of the overall methodology was performed by spiking assays at four levels and using SRM 2977. Quantification limits ranging from 0.15–27.16 ng/g wet weight were obtained. The established method was applied in edible tissues of three commonly consumed and commercially valuable fish species (sardine, chub mackerel and horse mackerel) originated from Atlantic Ocean. Variable levels of naphthalene (1.03–2.95 ng/g wet weight), fluorene (0.34–1.09 ng/g wet weight) and phenanthrene (0.34–3.54 ng/g wet weight) were detected in the analysed samples. None of the samples contained detectable amounts of benzo[a]pyrene, the marker used for evaluating the occurrence and carcinogenic effects of PAHs in food.
Resumo:
A QuEChERS method for the extraction of ochratoxin A (OTA) from bread samples was evaluated. A factorial design (23) was used to find the optimal QuEChERS parameters (extraction time, extraction solvent volume and sample mass). Extracts were analysed by LC with fluorescence detection. The optimal extraction conditions were: 5 g of sample, 15 mL of acetonitrile and 3 min of agitation. The extraction procedure was validated by systematic recovery experiments at three levels. The recoveries obtained ranged from 94.8% (at 1.0 μg kg -1) to 96.6% (at 3.0 μg kg -1). The limit of quantification of the method was 0.05 μg kg -1. The optimised procedure was applied to 20 samples of different bread types (‘‘Carcaça’’, ‘‘Broa de Milho’’, and ‘‘Broa de Avintes’’) highly consumed in Portugal. None of the samples exceeded the established European legal limit of 3 μg kg -1.
Resumo:
Scientific evidence has shown an association between organochlorine compounds (OCC) exposure and human health hazards. Concerning this, OCC detection in human adipose samples has to be considered a public health priority. This study evaluated the efficacy of various solid-phase extraction (SPE) and cleanup methods for OCC determination in human adipose tissue. Octadecylsilyl endcapped (C18-E), benzenesulfonic acid modified silica cation exchanger (SA), poly (styrene-divinylbenzene (EN) and EN/RP18 SPE sorbents were evaluated. The relative sample cleanup provided by these SPE columns was evaluated using gas chromatography with electron capture detection (GC–ECD). The C18-E columns with strong homogenization were found to provide the most effective cleanup, removing the greatest amount of interfering substance, and simultaneously ensuring good analyte recoveries higher than 70%. Recoveries>70% with standard deviations (SD)<15% were obtained for all compounds under the selected conditions. Method detection limits were in the 0.003–0.009 mg/kg range. The positive samples were confirmed by gas chromatography coupled with tandem mass spectrometry (GC-MS/MS). The highest percentage found of the OCC in real samples corresponded to HCB, o,p′-DDT and methoxychlor, which were detected in 80 and 95% of samples analyzed respectively. Copyright © 2012 John Wiley & Sons, Ltd.
Resumo:
Environmental pollution continues to be an emerging study field, as there are thousands of anthropogenic compounds mixed in the environment whose possible mechanisms of toxicity and physiological outcomes are of great concern. Developing methods to access and prioritize the screening of these compounds at trace levels in order to support regulatory efforts is, therefore, very important. A methodology based on solid phase extraction followed by derivatization and gas chromatography-mass spectrometry analysis was developed for the assessment of four endocrine disrupting compounds (EDCs) in water matrices: bisphenol A, estrone, 17b-estradiol and 17a-ethinylestradiol. The study was performed, simultaneously, by two different laboratories in order to evaluate the robustness of the method and to increase the quality control over its application in routine analysis. Validation was done according to the International Conference on Harmonisation recommendations and other international guidelines with specifications for the GC-MS methodology. Matrix-induced chromatographic response enhancement was avoided by using matrix-standard calibration solutions and heteroscedasticity has been overtaken by a weighted least squares linear regression model application. Consistent evaluation of key analytical parameters such as extraction efficiency, sensitivity, specificity, linearity, limits of detection and quantification, precision, accuracy and robustness was done in accordance with standards established for acceptance. Finally, the application of the optimized method in the assessment of the selected analytes in environmental samples suggested that it is an expedite methodology for routine analysis of EDC residues in water matrices.
Resumo:
Introduction: Paper and thin layer chromatography methods are frequently used in Classic Nuclear Medicine for the determination of radiochemical purity (RCP) on radiopharmaceutical preparations. An aliquot of the radiopharmaceutical to be tested is spotted at the origin of a chromatographic strip (stationary phase), which in turn is placed in a chromatographic chamber in order to separate and quantify radiochemical species present in the radiopharmaceutical preparation. There are several methods for the RCP measurement, based on the use of equipment as dose calibrators, well scintillation counters, radiochromatografic scanners and gamma cameras. The purpose of this study was to compare these quantification methods for the determination of RCP. Material and Methods: 99mTc-Tetrofosmin and 99mTc-HDP are the radiopharmaceuticals chosen to serve as the basis for this study. For the determination of RCP of 99mTc-Tetrofosmin we used ITLC-SG (2.5 x 10 cm) and 2-butanone (99mTc-tetrofosmin Rf = 0.55, 99mTcO4- Rf = 1.0, other labeled impurities 99mTc-RH RF = 0.0). For the determination of RCP of 99mTc-HDP, Whatman 31ET and acetone was used (99mTc-HDP Rf = 0.0, 99mTcO4- Rf = 1.0, other labeled impurities RF = 0.0). After the development of the solvent front, the strips were allowed to dry and then imaged on the gamma camera (256x256 matrix; zoom 2; LEHR parallel-hole collimator; 5-minute image) and on the radiochromatogram scanner. Then, strips were cut in Rf 0.8 in the case of 99mTc-tetrofosmin and Rf 0.5 in the case of 99mTc-HDP. The resultant pieces were smashed in an assay tube (to minimize the effect of counting geometry) and counted in the dose calibrator and in the well scintillation counter (during 1 minute). The RCP was calculated using the formula: % 99mTc-Complex = [(99mTc-Complex) / (Total amount of 99mTc-labeled species)] x 100. Statistical analysis was done using the test of hypotheses for the difference between means in independent samples. Results:The gamma camera based method demonstrated higher operator-dependency (especially concerning the drawing of the ROIs) and the measures obtained using the dose calibrator are very sensitive to the amount of activity spotted in the chromatographic strip, so the use of a minimum of 3.7 MBq activity is essential to minimize quantification errors. Radiochromatographic scanner and well scintillation counter showed concordant results and demonstrated the higher level of precision. Conclusions: Radiochromatographic scanners and well scintillation counters based methods demonstrate to be the most accurate and less operator-dependant methods.
Resumo:
Introduction: Although relative uptake values aren’t the most important objective of a 99mTc-DMSA scan, they are important quantitative information. In most of the dynamic renal scintigraphies attenuation correction is essential if one wants to obtain a reliable result of the quantification process. Although in DMSA scans the absent of significant background and the lesser attenuation in pediatric patients, makes that this attenuation correction techniques are actually not applied. The geometric mean is the most common method, but that includes the acquisition of an anterior (extra) projection, which it is not acquired by a large number of NM departments. This method and the attenuation factors proposed by Tonnesen will be correlated with the absence of attenuation correction procedures. Material and Methods: Images from 20 individuals (aged 3 years +/- 2) were used and the two attenuation correction methods applied. The mean time of acquisition (time post DMSA administration) was 3.5 hours +/- 0.8h. Results: The absence of attenuation correction showed a good correlation with both attenuation methods (r=0.73 +/- 0.11) and the mean difference verified on the uptake values between the different methods were 4 +/- 3. The correlation was higher when the age was lower. The attenuation correction methods correlation was higher between them two than with the “no attenuation correction” method (r=0.82 +/- 0.8), and the mean differences of the uptake values were 2 +/- 2. Conclusion: The decision of not doing any kind of attenuation correction method can be justified by the minor differences verified on the relative kidney uptake values. Nevertheless, if it is recognized that there is a need for an accurate value of the relative kidney uptake, then an attenuation correction method should be used. Attenuation correction factors proposed by Tonnesen can be easily implemented and so become a practical and easy to implement alternative, namely when the anterior projection - needed for the geometric mean methodology – is not acquired.