964 resultados para bromine and iodine determination
Resumo:
[EN]We describe a method for studying the classicalexperiment of the simple pendulum, consisting of a body of magnetic material oscillating through a thin conducting coil (magnetic pendulum), which according to Faraday’s law of induction generates a fluctuating current in the coil that can be transferred into a periodic signal in an oscilloscope. The set up described here allows to study the motion of the pendulum beyond what is normally considered in more basic settings, including a detailed analysis of both small and large oscillations, and the determination of the value of the acceleration of gravity.
Resumo:
Several MCAO systems are under study to improve the angular resolution of the current and of the future generation large ground-based telescopes (diameters in the 8-40 m range). The subject of this PhD Thesis is embedded in this context. Two MCAO systems, in dierent realization phases, are addressed in this Thesis: NIRVANA, the 'double' MCAO system designed for one of the interferometric instruments of LBT, is in the integration and testing phase; MAORY, the future E-ELT MCAO module, is under preliminary study. These two systems takle the sky coverage problem in two dierent ways. The layer oriented approach of NIRVANA, coupled with multi-pyramids wavefront sensors, takes advantage of the optical co-addition of the signal coming from up to 12 NGS in a annular 2' to 6' technical FoV and up to 8 in the central 2' FoV. Summing the light coming from many natural sources permits to increase the limiting magnitude of the single NGS and to improve considerably the sky coverage. One of the two Wavefront Sensors for the mid- high altitude atmosphere analysis has been integrated and tested as a stand- alone unit in the laboratory at INAF-Osservatorio Astronomico di Bologna and afterwards delivered to the MPIA laboratories in Heidelberg, where was integrated and aligned to the post-focal optical relay of one LINC-NIRVANA arm. A number of tests were performed in order to characterize and optimize the system functionalities and performance. A report about this work is presented in Chapter 2. In the MAORY case, to ensure correction uniformity and sky coverage, the LGS-based approach is the current baseline. However, since the Sodium layer is approximately 10 km thick, the articial reference source looks elongated, especially when observed from the edge of a large aperture. On a 30-40 m class telescope, for instance, the maximum elongation varies between few arcsec and 10 arcsec, depending on the actual telescope diameter, on the Sodium layer properties and on the laser launcher position. The centroiding error in a Shack-Hartmann WFS increases proportionally to the elongation (in a photon noise dominated regime), strongly limiting the performance. To compensate for this effect a straightforward solution is to increase the laser power, i.e. to increase the number of detected photons per subaperture. The scope of Chapter 3 is twofold: an analysis of the performance of three dierent algorithms (Weighted Center of Gravity, Correlation and Quad-cell) for the instantaneous LGS image position measurement in presence of elongated spots and the determination of the required number of photons to achieve a certain average wavefront error over the telescope aperture. An alternative optical solution to the spot elongation problem is proposed in Section 3.4. Starting from the considerations presented in Chapter 3, a first order analysis of the LGS WFS for MAORY (number of subapertures, number of detected photons per subaperture, RON, focal plane sampling, subaperture FoV) is the subject of Chapter 4. An LGS WFS laboratory prototype was designed to reproduce the relevant aspects of an LGS SH WFS for the E-ELT and to evaluate the performance of different centroid algorithms in presence of elongated spots as investigated numerically and analytically in Chapter 3. This prototype permits to simulate realistic Sodium proles. A full testing plan for the prototype is set in Chapter 4.
Resumo:
Great strides have been made in the last few years in the pharmacological treatment of neuropsychiatric disorders, with the introduction into the therapy of several new and more efficient agents, which have improved the quality of life of many patients. Despite these advances, a large percentage of patients is still considered “non-responder” to the therapy, not drawing any benefits from it. Moreover, these patients have a peculiar therapeutic profile, due to the very frequent application of polypharmacy, attempting to obtain satisfactory remission of the multiple aspects of psychiatric syndromes. Therapy is heavily individualised and switching from one therapeutic agent to another is quite frequent. One of the main problems of this situation is the possibility of unwanted or unexpected pharmacological interactions, which can occur both during polypharmacy and during switching. Simultaneous administration of psychiatric drugs can easily lead to interactions if one of the administered compounds influences the metabolism of the others. Impaired CYP450 function due to inhibition of the enzyme is frequent. Other metabolic pathways, such as glucuronidation, can also be influenced. The Therapeutic Drug Monitoring (TDM) of psychotropic drugs is an important tool for treatment personalisation and optimisation. It deals with the determination of parent drugs and metabolites plasma levels, in order to monitor them over time and to compare these findings with clinical data. This allows establishing chemical-clinical correlations (such as those between administered dose and therapeutic and side effects), which are essential to obtain the maximum therapeutic efficacy, while minimising side and toxic effects. It is evident the importance of developing sensitive and selective analytical methods for the determination of the administered drugs and their main metabolites, in order to obtain reliable data that can correctly support clinical decisions. During the three years of Ph.D. program, some analytical methods based on HPLC have been developed, validated and successfully applied to the TDM of psychiatric patients undergoing treatment with drugs belonging to following classes: antipsychotics, antidepressants and anxiolytic-hypnotics. The biological matrices which have been processed were: blood, plasma, serum, saliva, urine, hair and rat brain. Among antipsychotics, both atypical and classical agents have been considered, such as haloperidol, chlorpromazine, clotiapine, loxapine, risperidone (and 9-hydroxyrisperidone), clozapine (as well as N-desmethylclozapine and clozapine N-oxide) and quetiapine. While the need for an accurate TDM of schizophrenic patients is being increasingly recognized by psychiatrists, only in the last few years the same attention is being paid to the TDM of depressed patients. This is leading to the acknowledgment that depression pharmacotherapy can greatly benefit from the accurate application of TDM. For this reason, the research activity has also been focused on first and second-generation antidepressant agents, like triciclic antidepressants, trazodone and m-chlorophenylpiperazine (m-cpp), paroxetine and its three main metabolites, venlafaxine and its active metabolite, and the most recent antidepressant introduced into the market, duloxetine. Among anxiolytics-hypnotics, benzodiazepines are very often involved in the pharmacotherapy of depression for the relief of anxious components; for this reason, it is useful to monitor these drugs, especially in cases of polypharmacy. The results obtained during these three years of Ph.D. program are reliable and the developed HPLC methods are suitable for the qualitative and quantitative determination of CNS drugs in biological fluids for TDM purposes.
Resumo:
Nano(bio)science and nano(bio)technology play a growing and tremendous interest both on academic and industrial aspects. They are undergoing rapid developments on many fronts such as genomics, proteomics, system biology, and medical applications. However, the lack of characterization tools for nano(bio)systems is currently considered as a major limiting factor to the final establishment of nano(bio)technologies. Flow Field-Flow Fractionation (FlFFF) is a separation technique that is definitely emerging in the bioanalytical field, and the number of applications on nano(bio)analytes such as high molar-mass proteins and protein complexes, sub-cellular units, viruses, and functionalized nanoparticles is constantly increasing. This can be ascribed to the intrinsic advantages of FlFFF for the separation of nano(bio)analytes. FlFFF is ideally suited to separate particles over a broad size range (1 nm-1 μm) according to their hydrodynamic radius (rh). The fractionation is carried out in an empty channel by a flow stream of a mobile phase of any composition. For these reasons, fractionation is developed without surface interaction of the analyte with packing or gel media, and there is no stationary phase able to induce mechanical or shear stress on nanosized analytes, which are for these reasons kept in their native state. Characterization of nano(bio)analytes is made possible after fractionation by interfacing the FlFFF system with detection techniques for morphological, optical or mass characterization. For instance, FlFFF coupling with multi-angle light scattering (MALS) detection allows for absolute molecular weight and size determination, and mass spectrometry has made FlFFF enter the field of proteomics. Potentialities of FlFFF couplings with multi-detection systems are discussed in the first section of this dissertation. The second and the third sections are dedicated to new methods that have been developed for the analysis and characterization of different samples of interest in the fields of diagnostics, pharmaceutics, and nanomedicine. The second section focuses on biological samples such as protein complexes and protein aggregates. In particular it focuses on FlFFF methods developed to give new insights into: a) chemical composition and morphological features of blood serum lipoprotein classes, b) time-dependent aggregation pattern of the amyloid protein Aβ1-42, and c) aggregation state of antibody therapeutics in their formulation buffers. The third section is dedicated to the analysis and characterization of structured nanoparticles designed for nanomedicine applications. The discussed results indicate that FlFFF with on-line MALS and fluorescence detection (FD) may become the unparallel methodology for the analysis and characterization of new, structured, fluorescent nanomaterials.
Resumo:
In this thesis, the influence of composition changes on the glass transition behavior of binary liquids in two and three spatial dimensions (2D/3D) is studied in the framework of mode-coupling theory (MCT).The well-established MCT equations are generalized to isotropic and homogeneous multicomponent liquids in arbitrary spatial dimensions. Furthermore, a new method is introduced which allows a fast and precise determination of special properties of glass transition lines. The new equations are then applied to the following model systems: binary mixtures of hard disks/spheres in 2D/3D, binary mixtures of dipolar point particles in 2D, and binary mixtures of dipolar hard disks in 2D. Some general features of the glass transition lines are also discussed. The direct comparison of the binary hard disk/sphere models in 2D/3D shows similar qualitative behavior. Particularly, for binary mixtures of hard disks in 2D the same four so-called mixing effects are identified as have been found before by Götze and Voigtmann for binary hard spheres in 3D [Phys. Rev. E 67, 021502 (2003)]. For instance, depending on the size disparity, adding a second component to a one-component liquid may lead to a stabilization of either the liquid or the glassy state. The MCT results for the 2D system are on a qualitative level in agreement with available computer simulation data. Furthermore, the glass transition diagram found for binary hard disks in 2D strongly resembles the corresponding random close packing diagram. Concerning dipolar systems, it is demonstrated that the experimental system of König et al. [Eur. Phys. J. E 18, 287 (2005)] is well described by binary point dipoles in 2D through a comparison between the experimental partial structure factors and those from computer simulations. For such mixtures of point particles it is demonstrated that MCT predicts always a plasticization effect, i.e. a stabilization of the liquid state due to mixing, in contrast to binary hard disks in 2D or binary hard spheres in 3D. It is demonstrated that the predicted plasticization effect is in qualitative agreement with experimental results. Finally, a glass transition diagram for binary mixtures of dipolar hard disks in 2D is calculated. These results demonstrate that at higher packing fractions there is a competition between the mixing effects occurring for binary hard disks in 2D and those for binary point dipoles in 2D.
Resumo:
The behaviour of a polymer depends strongly on the length- and time scale as well as on the temperature rnat which it is probed. In this work, I describe investigations of polymer surfaces using scanning probe rnmicroscopy with heatable probes. With these probes, surfaces can be heated within seconds down to rnmicroseconds. I introduce experiments for the local and fast determination of glass transition and melting rntemperatures. I developed a method which allows the determination of glass transition and melting rntemperatures on films with thicknesses below 100 nm: A background measurement on the substrate was rnperformed. The resulting curve was subtracted from the measurement on the polymer film. The rndifferential measurement on polystyrene films with thicknesses between 35 nm and 160 nm showed rncharacteristic signals at 95 ± 1 °C, in accordance with the glass transition of polystyrene. Pressing heated rnprobes into polymer films causes plastic deformation. Nanometer sized deformations are currently rninvestigated in novel concepts for high density data storage. A suitable medium for such a storage system rnhas to be easily indentable on one hand, but on the other hand it also has to be very stable towards rnsurface induced wear. For developing such a medium I investigated a new approach: A comparably soft rnmaterial, namely polystyrene, was protected with a thin but very hard layer made of plasma polymerized rnnorbornene. The resulting bilayered media were tested for surface stability and deformability. I showed rnthat the bilayered material combines the deformability of polystyrene with the surface stability of the rnplasma polymer, and that the material therefore is a very good storage medium. In addition we rninvestigated the glass transition temperature of polystyrene at timescales of 10 µs and found it to be rnapprox. 220 °C. The increase of this characteristic temperature of the polymer results from the short time rnat which the polymer was probed and reflects the well-known time-temperature superposition principle. rnHeatable probes were also used for the characterization of silverazide filled nanocapsules. The use of rnheatable probes allowed determining the decomposition temperature of the capsules from few rnnanograms of material. The measured decomposition temperatures ranged from 180 °C to 225 °C, in rnaccordance with literature values. The investigation of small amounts of sample was necessary due to the rnlimited availability of the material. Furthermore, investigating larger amounts of the capsules using rnconventional thermal gravimetric analysis could lead to contamination or even damage of the instrument. rnBesides the analysis of material parameters I used the heatable probes for the local thermal rndecomposition of pentacene precursor material in order to form nanoscale conductive structures. Here, rnthe thickness of the precursor layer was important for complete thermal decomposition. rnAnother aspect of my work was the investigation of redox active polymers - Poly-10-(4-vinylbenzyl)-10H-rnphenothiazine (PVBPT)- for data storage. Data is stored by changing the local conductivity of the material rnby applying a voltage between tip and surface. The generated structures were stable for more than 16 h. It rnwas shown that the presence of water is essential for succesfull patterning.
Resumo:
Since historical times, coastal areas throughout the eastern Mediterranean are exposed to tsunami hazard. For many decades the knowledge about palaeotsunamis was solely based on historical accounts. However, results from timeline analyses reveal different characteristics affecting the quality of the dataset (i.e. distribution of data, temporal thinning backward of events, local periodization phenomena) that emphasize the fragmentary character of the historical data. As an increasing number of geo-scientific studies give convincing examples of well dated tsunami signatures not reported in catalogues, the non-existing record is a major problem to palaeotsunami research. While the compilation of historical data allows a first approach in the identification of areas vulnerable to tsunamis, it must not be regarded as reliable for hazard assessment. Considering the increasing economic significance of coastal regions (e.g. for mass tourism) and the constantly growing coastal population, our knowledge on the local, regional and supraregional tsunami hazard along Mediterranean coasts has to be improved. For setting up a reliable tsunami risk assessment and developing risk mitigation strategies, it is of major importance (i) to identify areas under risk and (ii) to estimate the intensity and frequency of potential events. This approach is most promising when based on the analysis of palaeotsunami research seeking to detect areas of high palaeotsunami hazard, to calculate recurrence intervals and to document palaeotsunami destructiveness in terms of wave run-up, inundation and long-term coastal change. Within the past few years, geo-scientific studies on palaeotsunami events provided convincing evidence that throughout the Mediterranean ancient harbours were subject to strong tsunami-related disturbance or destruction. Constructed to protect ships from storm and wave activity, harbours provide especially sheltered and quiescent environments and thus turned out to be valuable geo-archives for tsunamigenic high-energy impacts on coastal areas. Directly exposed to the Hellenic Trench and extensive local fault systems, coastal areas in the Ionian Sea and the Gulf of Corinth hold a considerably high risk for tsunami events, respectively.Geo-scientific and geoarcheaological studies carried out in the environs of the ancient harbours of Krane (Cefalonia Island), Lechaion (Corinth, Gulf of Corinth) and Kyllini (western Peloponnese) comprised on-shore and near-shore vibracoring and subsequent sedimentological, geochemical and microfossil analyses of the recovered sediments. Geophysical methods like electrical resistivity tomography and ground penetrating radar were applied in order to detect subsurface structures and to verify stratigraphical patterns derived from vibracores over long distances. The overall geochronological framework of each study area is based on radiocarbon dating of biogenic material and age determination of diagnostic ceramic fragments. Results presented within this study provide distinct evidence of multiple palaeotsunami landfalls for the investigated areas. Tsunami signatures encountered in the environs of Krane, Lechaion and Kyllini include (i) coarse-grained allochthonous marine sediments intersecting silt-dominated quiescent harbour deposits and/or shallow marine environments, (ii) disturbed microfaunal assemblages and/or (iii) distinct geochemical fingerprints as well as (iv) geo-archaeological destruction layers and (v) extensive units of beachrock-type calcarenitic tsunamites. For Krane, geochronological data yielded termini ad or post quem (maximum ages) for tsunami event generations dated to 4150 ± 60 cal BC, ~ 3200 ± 110 cal BC, ~ 650 ± 110 cal BC, and ~ 930 ± 40 cal AD, respectively. Results for Lechaion suggest that the harbour was hit by strong tsunami impacts in the 8th-6th century BC, the 1st-2nd century AD and in the 6th century AD. At Kyllini, the harbour site was affected by tsunami impact in between the late 7th and early 4th cent. BC and between the 4th and 6th cent. AD. In case of Lechaion and Kyllini, the final destruction of the harbour facilities also seems to be related to the tsunami impact. Comparing the tsunami signals obtained for each study areas with geo-scientific data from palaeotsunami events from other sites indicates that the investigated harbour sites represent excellent geo-archives for supra-regional mega-tsunamis.
Resumo:
Nanoindentation is a valuable tool for characterization of biomaterials due to its ability to measure local properties in heterogeneous, small or irregularly shaped samples. However, applying nanoindentation to compliant, hydrated biomaterials leads to many challenges including adhesion between the nanoindenter tip and the sample. Although adhesion leads to overestimation of the modulus of compliant samples when analyzing nanoindentation data using traditional analysis techniques, most studies of biomaterials have ignored its effects. This paper demonstrates two methods for managing adhesion in nanoindentation analysis, the nano-JKR force curve method and the surfactant method, through application to two biomedically-relevant compliant materials, poly(dimethyl siloxane) (PDMS) elastomers and poly(ethylene glycol) (PEG) hydrogels. The nano-JKR force curve method accounts for adhesion during data analysis using equations based on the Johnson-Kendall-Roberts (JKR) adhesion model, while the surfactant method eliminates adhesion during data collection, allowing data analysis using traditional techniques. In this study, indents performed in air or water resulted in adhesion between the tip and the sample, while testing the same materials submerged in Optifree Express() contact lens solution eliminated tip-sample adhesion in most samples. Modulus values from the two methods were within 7% of each other, despite different hydration conditions and evidence of adhesion. Using surfactant also did not significantly alter the properties of the tested material, allowed accurate modulus measurements using commercial software, and facilitated nanoindentation testing in fluids. This technique shows promise for more accurate and faster determination of modulus values from nanoindentation of compliant, hydrated biological samples. Copyright 2013 Elsevier Ltd. All rights reserved.
Resumo:
The group studied 1,253 students from various types of schools chosen randomly from those in Prague and Budejovice in order to evaluate the life styles, prevailing value standards, attitudes and behavioural patterns of Czech adolescents. The respondents (including 614 men and 639 women with an average age of 16.4 years) completed questionnaires containing standard scales focusing on feelings about social life, conservative and authoritarian tendencies, levels of self-esteem, general health, eating attitudes and behaviour The adolescents showed a relatively high level of conformity with authoritarian, conservative tendencies and with a dictate of power, rigid conventionality, ethnocentrism and low inner tolerance of differences, their scores being higher than those found in Western European countries. These tendencies were stronger among students outside Prague and those attending vocational schools. As the level of education rose, the sense of fatality and social determination decreased, indicating a higher share of responsibility for events in the surrounding world. When changes of life style were considered, adolescents can be expected to adapt more easily to more risky, socially attractive and manifest models of attitudes and behaviour. On the one hand, adolescents were often involved in sports, and young women in particular often showed a extreme concern and care for their own bodies. On the other hand, one quarter of respondents smoked, one fifth reported serious problems with alcohol and one quarter had already had some experience with drugs. One third of young men and one quarter of young women reported regular consumption of alcohol, and 6.5 percent of men and 3.6 percent of women regularly smoked marihuana or hashish. For the majority of adolescents, life conditions and conformity seem to be more important than the sense of active choice and responsibility for one's own life.
Resumo:
We prospectively investigated urinary iodine concentration (UIC) in pregnant women and in female, non-pregnant controls in the canton of Berne, Switzerland, in 1992. Mean UIC of pregnant women [205 +/- 151 microg iodine/g creatinine (microg l/g Cr); no. = 153] steadily decreased from the first (236 +/- 180 microg l/g Cr; no. = 31) to the third trimester (183 +/- 111 microg l/g Cr, p < 0.0001; no. = 66) and differed significantly from that of the control group (91 +/- 37 microg l/g Cr, p < 0.0001; no. = 119). UIC increased 2.6-fold from levels indicating mild iodine deficiency in controls to the first trimester, demonstrating that high UIC during early gestation does not necessarily reflect a sufficient iodine supply to the overall population. Pregnancy is accompanied by important alterations in the regulation of thyroid function and iodine metabolism. Increased renal iodine clearance during pregnancy may explain increased UIC during early gestation, whereas increased thyroidal iodine clearance as well as the iodine shift from the maternal circulation to the growing fetal-placental unit, which both tend to lower the circulating serum levels of inorganic iodide, probably are the causes of the continuous decrease of UIC over the course of pregnancy. Mean UIC in our control group, as well as in one parallel and several consecutive investigations in the same region in the 1990s, was found to be below the actually recommended threshold, indicating a new tendency towards mild to moderate iodine deficiency. As salt is the main source of dietary iodine in Switzerland, its iodine concentration was therefore increased nationwide in 1998 for the fourth time, following increases in 1922, 1965 and 1980.
Resumo:
The detection rate of pulmonary emboli (PE) with computed tomography angiography (CTA) using either a standard or a low-dose protocol, combining reduced radiation exposure and iodine delivery rate, was retrospectively analyzed in a matched cohort of 120 patients.
Resumo:
Diarrhoea caused by Cryptosporidium parvum is a major problem in calves younger than 4 weeks of age. To date only a few compounds have been approved for prophylactic and none for therapeutic use. Nitazoxanide (NTZ) has proven its efficacy in vitro against C. parvum and is approved by FDA for the treatment of human cryptosporidiosis. In a first experimental study, 3 uninfected calves were treated with NTZ and pharmacokinetics was followed through blood samples. Serum samples of uninfected treated calves contained both NTZ metabolites (tizoxanide and tizoxanide glucuronide) and oral administration at 12 h intervals was considered as optimal. Three groups of three calves (1-3 days old) were then each inoculated with 1x10(7) oocysts of C. parvum (cattle genotype): the prophylactic group received 15 mg/kg body weight NTZ twice daily orally in milk from 1 day before to 8 days postinoculation (dpi). The therapeutic group received the same dosage of NTZ for 10 days from the appearance of diarrhoea (between 1 and 5 dpi). The control group was left untreated. All calves were monitored daily from day -1 to 28 dpi and faecal samples were collected for evaluation of consistency and for determination of oocyst numbers per gram (OPG) of faeces. Diarrhoea was observed in all calves within the first week. Neither prophylactic nor therapeutic use of NTZ improved the clinical appearance and calves of the therapeutic showed a longer diarrheic episode (p<0.05) with strong altered faecal consistency compared to the untreated control group. The number of days with oocyst excretion did not differ significantly between the groups. In 5 out of 6 infected and treated calves oocyst excretion stopped only after discontinuation of treatment. In the prophylactic and in the control group mean values of the sum of the daily OPG per calf (8.5x10(6) and 8.0x10(6), respectively) and of the mean daily number of OPG (0.3x10(6) and 0.3x10(6), respectively) were similar, while the therapeutic group showed significantly lower values (1.9x10(6) and 0.06x10(6), respectively, p<0.05). However oocyst determinations in this group may have been altered by the severe diarrhoea, diluting oocyst densities in the analysed faecal samples. In conclusion, these preliminary results about the first prophylactic and therapeutic use of NTZ in calves did not show the expected positive effect on the course of the Cryptosporidium-infection, neither on reducing the clinical severity, nor on oocyst excretion.
Resumo:
Self-Determination Theory (Deci and Ryan in Intrinsic motivation and self-determination in human behavior. Plenum Press, New York, 1985) suggests that certain experiences, such as competence, are equally beneficial to everyone’s well-being (universal hypothesis), whereas Motive Disposition Theory (McClelland in Human motivation. Scott, Foresman, Glenview, IL, 1985) predicts that some people, such as those with a high achievement motive, should benefit particularly from such experiences (matching hypothesis). Existing research on motives as moderators of the relationship between basic need satisfaction and positive outcomes supports both these seemingly inconsistent views. Focusing on the achievement motive, we sought to resolve this inconsistency by considering the specificity of the outcome variables. When predicting domain-specific well-being and flow, the achievement motive should interact with felt competence. However, when it comes to predicting general well-being and flow, felt competence should unfold its effects without being moderated by the achievement motive. Two studies confirmed these assumptions indicating that the universal and matching hypotheses are complementary rather than mutually exclusive.
Resumo:
PURPOSE Positron emission tomography (PET)∕computed tomography (CT) measurements on small lesions are impaired by the partial volume effect, which is intrinsically tied to the point spread function of the actual imaging system, including the reconstruction algorithms. The variability resulting from different point spread functions hinders the assessment of quantitative measurements in clinical routine and especially degrades comparability within multicenter trials. To improve quantitative comparability there is a need for methods to match different PET∕CT systems through elimination of this systemic variability. Consequently, a new method was developed and tested that transforms the image of an object as produced by one tomograph to another image of the same object as it would have been seen by a different tomograph. The proposed new method, termed Transconvolution, compensates for differing imaging properties of different tomographs and particularly aims at quantitative comparability of PET∕CT in the context of multicenter trials. METHODS To solve the problem of image normalization, the theory of Transconvolution was mathematically established together with new methods to handle point spread functions of different PET∕CT systems. Knowing the point spread functions of two different imaging systems allows determining a Transconvolution function to convert one image into the other. This function is calculated by convolving one point spread function with the inverse of the other point spread function which, when adhering to certain boundary conditions such as the use of linear acquisition and image reconstruction methods, is a numerically accessible operation. For reliable measurement of such point spread functions characterizing different PET∕CT systems, a dedicated solid-state phantom incorporating (68)Ge∕(68)Ga filled spheres was developed. To iteratively determine and represent such point spread functions, exponential density functions in combination with a Gaussian distribution were introduced. Furthermore, simulation of a virtual PET system provided a standard imaging system with clearly defined properties to which the real PET systems were to be matched. A Hann window served as the modulation transfer function for the virtual PET. The Hann's apodization properties suppressed high spatial frequencies above a certain critical frequency, thereby fulfilling the above-mentioned boundary conditions. The determined point spread functions were subsequently used by the novel Transconvolution algorithm to match different PET∕CT systems onto the virtual PET system. Finally, the theoretically elaborated Transconvolution method was validated transforming phantom images acquired on two different PET systems to nearly identical data sets, as they would be imaged by the virtual PET system. RESULTS The proposed Transconvolution method matched different PET∕CT-systems for an improved and reproducible determination of a normalized activity concentration. The highest difference in measured activity concentration between the two different PET systems of 18.2% was found in spheres of 2 ml volume. Transconvolution reduced this difference down to 1.6%. In addition to reestablishing comparability the new method with its parameterization of point spread functions allowed a full characterization of imaging properties of the examined tomographs. CONCLUSIONS By matching different tomographs to a virtual standardized imaging system, Transconvolution opens a new comprehensive method for cross calibration in quantitative PET imaging. The use of a virtual PET system restores comparability between data sets from different PET systems by exerting a common, reproducible, and defined partial volume effect.
Resumo:
OBJECTIVE We investigated the skeletal growth profile of female rats from birth to senescence (100weeks) on the basis of sequential radiometrical, hormonal and biochemical parameters. DESIGN Weaning rats entered the study which was divided into two sections: a) sequential measurements of vertebral and tibial growths and bone mineral density (BMD), estimation of mineral content of the entire skeleton (BMC) and chemical analysis of vertebral Ca; and b) determination of basal and pulsatile growth hormone (rGH), insulin-like growth hormone (IGF-I), estradiol (E2), parathyroid hormone (PTH), osteocalcin (OC) and urinary d-pyridinoline (dp) throughout the experimental period. RESULTS Vertebral and tibial growths ceased at week 25 whereas BMD and BMC as well as total vertebral Ca exhibited a peak bone mass at week 40. rGH pulsatile profiles were significantly higher in younger animals coinciding with the period of active growth and IGF-I peaked at 7weeks, slowly declining thereafter and stabilizing after week 60. OC and dp closely paralleled IGF-I coinciding with the period of enhanced skeletal growth, remaining thereafter in the low range indicative of reduced bone turnover. E2 increased during reproductive life but the lower values subsequently recorded were still in the physiological range, strongly suggesting a protective role of this steroid on bone remodeling. PTH followed a similar profile to E2, but the significance of this after completion of growth remains unclear. CONCLUSIONS Mechanisms governing skeletal growth in the female rat appear similar to those in humans. Bone progression and attainment of peak bone mass are under simultaneous control of rGH, IGF-I and calciotropic hormones and are modulated by E2. This steroid seems to protect the skeleton from resorption before senescence whereas the role of PTH in this context remains uncertain.