87 resultados para ROUTINE STORAGE
Resumo:
Objectives: To clarify the role of growth monitoring in primary school children, including obesity, and to examine issues that might impact on the effectiveness and cost-effectiveness of such programmes. Data sources: Electronic databases were searched up to July 2005. Experts in the field were also consulted. Review methods: Data extraction and quality assessment were performed on studies meeting the review's inclusion criteria. The performance of growth monitoring to detect disorders of stature and obesity was evaluated against National Screening Committee (NSC) criteria. Results: In the 31 studies that were included in the review, there were no controlled trials of the impact of growth monitoring and no studies of the diagnostic accuracy of different methods for growth monitoring. Analysis of the studies that presented a 'diagnostic yield' of growth monitoring suggested that one-off screening might identify between 1: 545 and 1: 1793 new cases of potentially treatable conditions. Economic modelling suggested that growth monitoring is associated with health improvements [ incremental cost per quality-adjusted life-year (QALY) of pound 9500] and indicated that monitoring was cost-effective 100% of the time over the given probability distributions for a willingness to pay threshold of pound 30,000 per QALY. Studies of obesity focused on the performance of body mass index against measures of body fat. A number of issues relating to human resources required for growth monitoring were identified, but data on attitudes to growth monitoring were extremely sparse. Preliminary findings from economic modelling suggested that primary prevention may be the most cost-effective approach to obesity management, but the model incorporated a great deal of uncertainty. Conclusions: This review has indicated the potential utility and cost-effectiveness of growth monitoring in terms of increased detection of stature-related disorders. It has also pointed strongly to the need for further research. Growth monitoring does not currently meet all NSC criteria. However, it is questionable whether some of these criteria can be meaningfully applied to growth monitoring given that short stature is not a disease in itself, but is used as a marker for a range of pathologies and as an indicator of general health status. Identification of effective interventions for the treatment of obesity is likely to be considered a prerequisite to any move from monitoring to a screening programme designed to identify individual overweight and obese children. Similarly, further long-term studies of the predictors of obesity-related co-morbidities in adulthood are warranted. A cluster randomised trial comparing growth monitoring strategies with no growth monitoring in the general population would most reliably determine the clinical effectiveness of growth monitoring. Studies of diagnostic accuracy, alongside evidence of effective treatment strategies, could provide an alternative approach. In this context, careful consideration would need to be given to target conditions and intervention thresholds. Diagnostic accuracy studies would require long-term follow-up of both short and normal children to determine sensitivity and specificity of growth monitoring.
Resumo:
Automatic indexing and retrieval of digital data poses major challenges. The main problem arises from the ever increasing mass of digital media and the lack of efficient methods for indexing and retrieval of such data based on the semantic content rather than keywords. To enable intelligent web interactions, or even web filtering, we need to be capable of interpreting the information base in an intelligent manner. For a number of years research has been ongoing in the field of ontological engineering with the aim of using ontologies to add such (meta) knowledge to information. In this paper, we describe the architecture of a system (Dynamic REtrieval Analysis and semantic metadata Management (DREAM)) designed to automatically and intelligently index huge repositories of special effects video clips, based on their semantic content, using a network of scalable ontologies to enable intelligent retrieval. The DREAM Demonstrator has been evaluated as deployed in the film post-production phase to support the process of storage, indexing and retrieval of large data sets of special effects video clips as an exemplar application domain. This paper provides its performance and usability results and highlights the scope for future enhancements of the DREAM architecture which has proven successful in its first and possibly most challenging proving ground, namely film production, where it is already in routine use within our test bed Partners' creative processes. (C) 2009 Published by Elsevier B.V.
Resumo:
It is generally assumed that the variability of neuronal morphology has an important effect on both the connectivity and the activity of the nervous system, but this effect has not been thoroughly investigated. Neuroanatomical archives represent a crucial tool to explore structure–function relationships in the brain. We are developing computational tools to describe, generate, store and render large sets of three–dimensional neuronal structures in a format that is compact, quantitative, accurate and readily accessible to the neuroscientist. Single–cell neuroanatomy can be characterized quantitatively at several levels. In computer–aided neuronal tracing files, a dendritic tree is described as a series of cylinders, each represented by diameter, spatial coordinates and the connectivity to other cylinders in the tree. This ‘Cartesian’ description constitutes a completely accurate mapping of dendritic morphology but it bears little intuitive information for the neuroscientist. In contrast, a classical neuroanatomical analysis characterizes neuronal dendrites on the basis of the statistical distributions of morphological parameters, e.g. maximum branching order or bifurcation asymmetry. This description is intuitively more accessible, but it only yields information on the collective anatomy of a group of dendrites, i.e. it is not complete enough to provide a precise ‘blueprint’ of the original data. We are adopting a third, intermediate level of description, which consists of the algorithmic generation of neuronal structures within a certain morphological class based on a set of ‘fundamental’, measured parameters. This description is as intuitive as a classical neuroanatomical analysis (parameters have an intuitive interpretation), and as complete as a Cartesian file (the algorithms generate and display complete neurons). The advantages of the algorithmic description of neuronal structure are immense. If an algorithm can measure the values of a handful of parameters from an experimental database and generate virtual neurons whose anatomy is statistically indistinguishable from that of their real counterparts, a great deal of data compression and amplification can be achieved. Data compression results from the quantitative and complete description of thousands of neurons with a handful of statistical distributions of parameters. Data amplification is possible because, from a set of experimental neurons, many more virtual analogues can be generated. This approach could allow one, in principle, to create and store a neuroanatomical database containing data for an entire human brain in a personal computer. We are using two programs, L–NEURON and ARBORVITAE, to investigate systematically the potential of several different algorithms for the generation of virtual neurons. Using these programs, we have generated anatomically plausible virtual neurons for several morphological classes, including guinea pig cerebellar Purkinje cells and cat spinal cord motor neurons. These virtual neurons are stored in an online electronic archive of dendritic morphology. This process highlights the potential and the limitations of the ‘computational neuroanatomy’ strategy for neuroscience databases.
Resumo:
Sensitive methods that are currently used to monitor proteolysis by plasmin in milk are limited due to 7 their high cost and lack of standardisation for quality assurance in the various dairy laboratories. In 8 this study, four methods, trinitrobenzene sulphonic acid (TNBS), reverse phase high pressure liquid 9 chromatography (RP-HPLC), gel electrophoresis and fluorescamine, were selected to assess their 10 suitability for the detection of proteolysis in milk by plasmin. Commercial UHT milk was incubated 11 with plasmin at 37 °C for one week. Clarification was achieved by isoelectric precipitation (pH 4·6 12 soluble extracts)or 6% (final concentration) trichloroacetic acid (TCA). The pH 4·6 and 6% TCA 13 soluble extracts of milk showed high correlations (R2 > 0·93) by the TNBS, fluorescamine and 14 RP-HPLC methods, confirming increased proteolysis during storage. For gel electrophoresis,15 extensive proteolysis was confirmed by the disappearance of α- and β-casein bands on the seventh 16 day, which was more evident in the highest plasmin concentration. This was accompanied by the 17 appearance of α- and β-casein proteolysis products with higher intensities than on previous days, 18 implying that more products had been formed as a result of casein breakdown. The fluorescamine 19 method had a lower detection limit compared with the other methods, whereas gel electrophoresis 20 was the best qualitative method for monitoring β-casein proteolysis products. Although HPLC was the 21 most sensitive, the TNBS method is recommended for use in routine laboratory analysis on the basis 22 of its accuracy, reliability and simplicity.
The microstratigraphy of middens: capturing daily routine in rubbish at Neolithic Çatalhöyük, Turkey
Resumo:
This study evaluated the effects of fat and sugar levels on the surface properties of Lactobacillus rhamnosus GG during storage in food model systems, simulating yogurt and ice cream, and related them with the ability of the bacterial cells to adhere to Caco-2 cells. Freeze-dried L. rhamnosus GG cells were added to the model food systems and stored for 7 days. The bacterial cells were analyzed for cell viability, hydrophobicity, ζ potential, and their ability to adhere to Caco-2 cells. The results indicated that the food type and its composition affected the surface and adhesion properties of the bacterial cells during storage, with yogurt being a better delivery vehicle than ice cream in terms of bacterial adhesion to Caco-2 cells. The most important factor influencing bacterial adhesion was the storage time rather than the levels of fats and sugars, indicating that conformational changes were taking place on the surface of the bacterial cells during storage.
Resumo:
Aims: Therapeutic limbal epithelial stem cells could be managed more efficiently if clinically validated batches were transported for ‘on-demand’ use. Materials & methods: In this study, corneal epithelial cell viability in calcium alginate hydrogels was examined under cell culture, ambient and chilled conditions for up to 7 days. Results: Cell viability improved as gel internal pore size increased, and was further enhanced with modification of the gel from a mass to a thin disc. Ambient storage conditions were optimal for supporting cell viability in gel discs. Cell viability in gel discs was significantly enhanced with increases in pore size mediated by hydroxyethyl cellulose. Conclusion: Our novel methodology of controlling alginate gel shape and pore size together provides a more practical and economical alternative to established corneal tissue/cell storage methods.
Resumo:
An active pharmaceutical ingredient (API) was found to dissociate from the highly crystalline hydrochloride form to the amorphous free base form, with consequent alterations to tablet properties. Here, a wet granulation manufacturing process has been investigated using in situ Fourier transform (FT)-Raman spectroscopic analyses of granules and tablets prepared with different granulating fluids and under different manufacturing conditions. Dosage form stability under a range of storage stresses was also investigated. Despite the spectral similarities between the two drug forms, low levels of API dissociation could be quantified in the tablets; the technique allowed discrimination of around 4% of the API content as the amorphous free base (i.e. less than 1% of the tablet compression weight). API dissociation was shown to be promoted by extended exposure to moisture. Aqueous granulating fluids and manufacturing delays between granulation and drying stages and storage of the tablets in open conditions at 40◦C/75% relative humidity (RH) led to dissociation. In contrast, non-aqueous granulating fluids, with no delay in processing and storage of the tablets in either sealed containers or at lower temperature/humidity prevented detectable dissociation. It is concluded that appropriate manufacturing process and storage conditions for the finished product involved minimising exposure to moisture of the API. Analysis of the drug using FT-Raman spectroscopy allowed rapid optimisation of the process whilst offering quantitative molecular information concerning the dissociation of the drug salt to the amorphous free base form.
Resumo:
Reduced flexibility of low carbon generation could pose new challenges for future energy systems. Both demand response and distributed storage may have a role to play in supporting future system balancing. This paper reviews how these technically different, but functionally similar approaches compare and compete with one another. Household survey data is used to test the effectiveness of price signals to deliver demand responses for appliances with a high degree of agency. The underlying unit of storage for different demand response options is discussed, with particular focus on the ability to enhance demand side flexibility in the residential sector. We conclude that a broad range of options, with different modes of storage, may need to be considered, if residential demand flexibility is to be maximised.
Resumo:
Acrylamide forms during cooking and processing predominately from the reaction of free asparagine and reducing sugars in the Maillard reaction. The identification of low free asparagine and reducing sugar varieties of crops is therefore an important target. In this study, nine varieties of potato (French fry varieties Maris Piper (from two suppliers), Pentland Dell, King Edward, Daisy, and Markies; and chipping varieties Lady Claire, Lady Rosetta, Saturna, and Hermes) grown in the United Kingdom in 2009 were analyzed at monthly intervals through storage from November 2009 to July 2010. Acrylamide formation was measured in heated flour and chips fried in oil. Analysis of variance revealed significant interactions between varieties nested within type (French fry and chipping) and storage time for most free amino acids, glucose, fructose, and acrylamide formation. Acrylamide formed in chips correlated significantly with acrylamide formed in flour and with chip color. There were significant correlations between glucose or total reducing sugar concentration and acrylamide formation in both variety types, but with fructose the correlation was much stronger for chipping than for French fry varieties. Conversely, there were significant correlations with acrylamide formation for both total free amino acid and free asparagine concentration in the French fry but not chipping varieties. The study showed the potential of variety selection for preventing unacceptable levels of acrylamide formation in potato products and the variety-dependent effect of long-term storage on acrylamide risk. It also highlighted the complex relationship between precursor concentration and acrylamide risk in potatoes.