894 resultados para work-related assessment
Resumo:
The relationship between organizational networks and employees' affect was examined in 2 organizations. In Study 1, social network analysis of work ties and job-related affect for 259 employees showed that affect converged within work interaction groups. Similarity of affect between employees depended on the presence of work ties and structural equivalence. Affect was also related to the size and density of employees' work networks. Study 2 used a 10-week diary study of 31 employees to examine a merger of 2 organizational divisions and found that negative changes in employees' affect were related to having fewer cross-divisional ties and to experiencing greater reductions in network density. The findings suggest that affect permeates through and is shaped by organizational networks.
Resumo:
New Approach’ Directives now govern the health and safety of most products whether destined for workplace or domestic use. These Directives have been enacted into UK law by various specific legislation principally relating to work equipment, machinery and consumer products. This research investigates whether the risk assessment approach used to ensure the safety of machinery may be applied to consumer products. Crucially, consumer products are subject to the Consumer Protection Act (CPA) 1987, where there is no direct reference to “assessing risk”. This contrasts with the law governing the safety of products used in the workplace, where risk assessment underpins the approach. New Approach Directives are supported by European harmonised standards, and in the case of machinery, further supported by the risk assessment standard, EN 1050. The system regulating consumer product safety is discussed, its key elements identified and a graphical model produced. This model incorporates such matters as conformity assessment, the system of regulation, near miss and accident reporting. A key finding of the research is that New Approach Directives have a common feature of specifying essential performance requirements that provide a hazard prompt-list that can form the basis for a risk assessment (the hazard identification stage). Drawing upon 272 prosecution cases, and with thirty examples examined in detail, this research provides evidence that despite the high degree of regulation, unsafe consumer products still find their way onto the market. The research presents a number of risk assessment tools to help Trading Standards Officers (TSOs) prioritise their work at the initial inspection stage when dealing with subsequent enforcement action.
Resumo:
This study concerns the application of a model of effective interpersonal relationships to problems arising from staff assessment at I.C.I. Ltd. Corporate Laboratory between 1972 and 1974. In collaboration with academic and industrial supervision, the study commenced with a survey of management and supervisor opinions about the effectiveness of current staff (work) relationships, with particular reference to the problem of recognising and developing creative potential. This survey emphasised a need to improve the relationships between staff in the staff assessment context. A survey of research into creativity emphasised the importance of the interpersonal environment for obtaining creative behaviour in an organisation context. A further survey of theories of how interpersonal behaviour related to personal creativity (therapeutic psychology) provided a model of effective interpersonal behaviour (Carkhuff, 1969) that could be applied to the organisation context of staff assessment. The objective of the project was redefined as a need to improve the conditions of interpersonal behaviour in relation to certain (career development) problems arising from staff assessment practices. In order to demonstrate the application of the model of effective interpersonal behaviour, the research student recorded interviews between himself and members of staff designed to develop and operate the dimensions of the model. Different samples of staff were used to develop the 'facilitative' and the 'action oriented' dimensions of bahaviour, and then for the operation of a helping programme (based on vocational guidance tests). These interactions have been analysed, according to the scales of measurement in the model ana the results are presented in case study form in this thesis. At each stage of the project, results and conclusions were presented to the sponsoring organisation (e.g. industrial supervisor) in order to assess their (subjective) opinion of relevance to the organ isation. Finally, recommendations on further actions towards general improvement of the work relationships in the laboratory were presented in a brief report to the sponsor.
Resumo:
Impaired postural control has been associated with poor reading skills, as well as with lower performance on measures of attention and motor control variables that frequently co-occur with reading difficulties. Measures of balance and motor control have been incorporated into several screening batteries for developmental dyslexia, but it is unclear whether the relationship between such skills and reading manifests as a behavioural continuum across the range of abilities or is restricted to groups of individuals with specific disorder phenotypes. Here were obtained measures of postural control alongside measures of reading, attention and general cognitive skills in a large sample of young adults (n = 100). Postural control was assessed using centre of pressure (CoP) measurements, obtained over 5 different task conditions. Our results indicate an absence of strong statistical relationships between balance measures with either reading, cognitive or attention measures across the sample as a whole. © 2014 Loras et al.
Resumo:
This paper critically evaluates the paradigm, theory, and methodology that dominate research on related party transactions (RPTs). RPTs have been debated in the literature whether they are a facet of conflict of interest between major and minor shareholders or they are normal efficient transactions that help the firms to achieve asset utilization. Literature has been widely interested in studying the association between corporate governance and RPTs especially that according to the agency theory it is assumed that corporate governance as a monitoring tool should impede the negative consequences of RPTs and ensure they are conducted to achieve better asset utilization.
Resumo:
Geospatial data have become a crucial input for the scientific community for understanding the environment and developing environmental management policies. The Global Earth Observation System of Systems (GEOSS) Clearinghouse is a catalogue and search engine that provides access to the Earth Observation metadata. However, metadata are often not easily understood by users, especially when presented in ISO XML encoding. Data quality included in the metadata is basic for users to select datasets suitable for them. This work aims to help users to understand the quality information held in metadata records and to provide the results to geospatial users in an understandable and comparable way. Thus, we have developed an enhanced tool (Rubric-Q) for visually assessing the metadata quality information and quantifying the degree of metadata population. Rubric-Q is an extension of a previous NOAA Rubric tool used as a metadata training and improvement instrument. The paper also presents a thorough assessment of the quality information by applying the Rubric-Q to all dataset metadata records available in the GEOSS Clearinghouse. The results reveal that just 8.7% of the datasets have some quality element described in the metadata, 63.4% have some lineage element documented, and merely 1.2% has some usage element described. © 2013 IEEE.
Resumo:
Purpose: To investigate the use of MRIA for quantitative characterisation of subretinal fibrosis secondary to nAMD. Methods: MRIA images of the posterior pole were acquired over 4 months from 20 eyes including those with inactive subretinal fibrosis and those being treated with ranibizumab for nAMD. Changes in morphology of the macula affected by nAMD were modelled and reflectance spectra at the MRIA acquisition wavelengths (507, 525, 552, 585, 596, 611 and 650nm) were computed using Monte Carlo simulation. Quantitative indicators of fibrosis were derived by matching image spectra to the model spectra of known morphological properties. Results: The model spectra were comparable to the image spectra, both normal and pathological. The key morphological changes that the model associated with nAMD were gliosis of the IS-OS junction, decrease in retinal blood and decrease in RPE melanin. However, these changes were not specific to fibrosis and none of the quantitative indicators showed a unique association with the degree of fibrosis. Moderate correlations were found with the clinical assessment, but not with the treatment program. Conclusion: MRIA can distinguish subretinal fibrosis from healthy tissue. The methods used show high sensitivity but low specificity, being unable to distinguish scarring from other abnormalities like atrophy. Quantification of scarring was not achieved with the wavelengths used due to the complex structural changes to retinal tissues in the process of nAMD. Further studies, incorporating other wavelengths, will establish whether MRIA has a role in the assessment of subretinal fibrosis in the context of retinal and choroidal pathology
Resumo:
Undergraduate programmes on construction management and other closely related built environment disciplines are currently taught and assessed on a modular basis. This is the case in the UK and in many other countries globally. However, it can be argued that professionally oriented programmes like these are better assessed on a non-modular basis, in order to produce graduates who can apply knowledge on different subject contents in cohesion to solve complex practical scenarios in their work environments. The examples of medical programmes where students are assessed on a non-modular basis can be cited as areas where this is already being done. A preliminary study was undertaken to explore the applicability of non-modular assessment within construction management undergraduate education. A selected sample of university academics was interviewed to gather their perspectives on applicability of non-modular assessment. General acceptance was observed among the academics involved that integrating non-modular assessment is applicable and will be beneficial. All academics stated that at least some form of non-modular assessment as being currently used in their programmes. Examples where cross-modular knowledge is assessed included comprehensive/multi-disciplinary project modules and creating larger modules to amalgamate a number of related subject areas. As opposed to a complete shift from modular to non-modular, an approach where non-modular assessment is integrated and its use further expanded within the current system is therefore suggested. This is due to the potential benefits associated with this form of assessment to professionally aligned built environment programmes
Resumo:
The safety of workers in nighttime roadway work zones has become a major concern for state transportation agencies due to the increase in the number of work zone fatalities. During the last decade, several studies have focused on the improvement of safety in nighttime roadway work zones; but the element that is still missing is a set of tools for translating the research results into practice. This paper discusses: 1) the importance of translating the research results related to the safety of workers and safety planning of nighttime work zones into practice, and 2) examples of tools that can be used for translating the results of such studies into practice. A tool that can propose safety recommendations in nighttime work zones and a web-based safety training tool for workers are presented in this paper. The tools were created as a component of a five-year research study on the assessment of the safety of nighttime roadway construction. The objectives of both tools are explained as well as their functionalities (i.e., what the tools can do for the users); their components (e.g., knowledge base, database, and interfaces); and their structures (i.e., how the components of the tools are organized to meet the objectives). Evaluations by the proposed users of each tool are also presented.
Resumo:
This study provides a comprehensive assessment of base rates for counterproductive work behaviors (CWB) and examines their relationship with personality and demographic variables. The Randomized-Response Technique (RRT) was employed in order to reduce the effects of social desirability. Base rates were calculated for 66 behaviors for a student and nationwide sample. Results revealed 15 significant behaviors for the student sample and 7 for the nationwide sample. In addition, low neuroticism was found to relate to higher reporting of counterproductive behavior for both groups. Also, low conscientiousness was related to higher reports of CWB in the student sample. Finally, CWB was found to differ based on ethnicity for the student sample such that Caucasians reported higher rates of CWB than Hispanics.
Resumo:
Acknowledgements We wish to express our gratitude to the National Geographic Society and the National Research Foundation of South Africa for funding the discovery, recovery, and analysis of the H. naledi material. The study reported here was also made possible by grants from the Social Sciences and Humanities Research Council of Canada, the Canada Foundation for Innovation, the British Columbia Knowledge Development Fund, the Canada Research Chairs Program, Simon Fraser University, the DST/NRF Centre of Excellence in Palaeosciences (COE-Pal), as well as by a Discovery Grant from the Natural Sciences and Engineering Research Council of Canada, a Young Scientist Development Grant from the Paleontological Scientific Trust (PAST), a Baldwin Fellowship from the L.S.B. Leakey Foundation, and a Seed Grant and a Cornerstone Faculty Fellowship from the Texas A&M University College of Liberal Arts. We would like to thank the South African Heritage Resource Agency for the permits necessary to work on the Rising Star site; the Jacobs family for granting access; Wilma Lawrence, Bonita De Klerk, Merrill Van der Walt, and Justin Mukanku for their assistance during all phases of the project; Lucas Delezene for valuable discussion on the dental characters of H. naledi. We would also like to thank Peter Schmid for the preparation of the Dinaledi fossil material; Yoel Rak for explaining in detail some of the characters used in previous studies; William Kimbel for drawing our attention to the possibility that there might be a problem with Dembo et al.’s (2015) codes for the two characters related to the articular eminence; Will Stein for helpful discussion about the Bayesian analyses; Mike Lee for his comments on this manuscript; John Hawks for his support in organizing the Rising Star workshop; and the associate editor and three anonymous reviewers for their valuable comments. We are grateful to S. Potze and the Ditsong Museum, B. Billings and the School of Anatomical Sciences at the University of the Witwatersrand, and B. Zipfel and the Evolutionary Studies Institute at the University of the Witwatersrand for providing access to the specimens in their care; the University of the Witwatersrand, the Evolutionary Studies Institute, and the South African National Centre of Excellence in PalaeoSciences for hosting a number of the authors while studying the material; and the Western Canada Research Grid for providing access to the high-performance computing facilities for the Bayesian analyses. Last but definitely not least, we thank the head of the Rising Star project, Lee Berger, for his leadership and support, and for encouraging us to pursue the study reported here.
Resumo:
Background: Dietary behaviour interventions have the potential to reduce diet-related disease. Ample opportunity exists to implement these interventions in the workplace. The overall aim is to assess the effectiveness and cost-effectiveness of complex dietary interventions focused on environmental dietary modification alone or in combination with nutrition education in large manufacturing workplace settings. Methods/design: A clustered controlled trial involving four large multinational manufacturing workplaces in Cork will be conducted. The complex intervention design has been developed using the Medical Research Council's framework and the National Institute for Health and Clinical Excellence (NICE) guidelines and will be reported using the TREND statement for the transparent reporting of evaluations with non-randomized designs. It will draw on a soft paternalistic 'nudge' theoretical perspective. It will draw on a soft paternalistic "nudge" theoretical perspective. Nutrition education will include three elements: group presentations, individual nutrition consultations and detailed nutrition information. Environmental dietary modification will consist of five elements: (a) restriction of fat, saturated fat, sugar and salt, (b) increase in fibre, fruit and vegetables, (c) price discounts for whole fresh fruit, (d) strategic positioning of healthier alternatives and (e) portion size control. No intervention will be offered in workplace A (control). Workplace B will receive nutrition education. Workplace C will receive nutrition education and environmental dietary modification. Workplace D will receive environmental dietary modification alone. A total of 448 participants aged 18 to 64 years will be selected randomly. All permanent, full-time employees, purchasing at least one main meal in the workplace daily, will be eligible. Changes in dietary behaviours, nutrition knowledge, health status with measurements obtained at baseline and at intervals of 3 to 4 months, 7 to 9 months and 13 to 16 months will be recorded. A process evaluation and cost-effectiveness economic evaluation will be undertaken. Discussion: A 'Food Choice at Work' toolbox (concise teaching kit to replicate the intervention) will be developed to inform and guide future researchers, workplace stakeholders, policy makers and the food industry. Trial registration: Current Controlled Trials, ISRCTN35108237.
Resumo:
X-ray computed tomography (CT) imaging constitutes one of the most widely used diagnostic tools in radiology today with nearly 85 million CT examinations performed in the U.S in 2011. CT imparts a relatively high amount of radiation dose to the patient compared to other x-ray imaging modalities and as a result of this fact, coupled with its popularity, CT is currently the single largest source of medical radiation exposure to the U.S. population. For this reason, there is a critical need to optimize CT examinations such that the dose is minimized while the quality of the CT images is not degraded. This optimization can be difficult to achieve due to the relationship between dose and image quality. All things being held equal, reducing the dose degrades image quality and can impact the diagnostic value of the CT examination.
A recent push from the medical and scientific community towards using lower doses has spawned new dose reduction technologies such as automatic exposure control (i.e., tube current modulation) and iterative reconstruction algorithms. In theory, these technologies could allow for scanning at reduced doses while maintaining the image quality of the exam at an acceptable level. Therefore, there is a scientific need to establish the dose reduction potential of these new technologies in an objective and rigorous manner. Establishing these dose reduction potentials requires precise and clinically relevant metrics of CT image quality, as well as practical and efficient methodologies to measure such metrics on real CT systems. The currently established methodologies for assessing CT image quality are not appropriate to assess modern CT scanners that have implemented those aforementioned dose reduction technologies.
Thus the purpose of this doctoral project was to develop, assess, and implement new phantoms, image quality metrics, analysis techniques, and modeling tools that are appropriate for image quality assessment of modern clinical CT systems. The project developed image quality assessment methods in the context of three distinct paradigms, (a) uniform phantoms, (b) textured phantoms, and (c) clinical images.
The work in this dissertation used the “task-based” definition of image quality. That is, image quality was broadly defined as the effectiveness by which an image can be used for its intended task. Under this definition, any assessment of image quality requires three components: (1) A well defined imaging task (e.g., detection of subtle lesions), (2) an “observer” to perform the task (e.g., a radiologists or a detection algorithm), and (3) a way to measure the observer’s performance in completing the task at hand (e.g., detection sensitivity/specificity).
First, this task-based image quality paradigm was implemented using a novel multi-sized phantom platform (with uniform background) developed specifically to assess modern CT systems (Mercury Phantom, v3.0, Duke University). A comprehensive evaluation was performed on a state-of-the-art CT system (SOMATOM Definition Force, Siemens Healthcare) in terms of noise, resolution, and detectability as a function of patient size, dose, tube energy (i.e., kVp), automatic exposure control, and reconstruction algorithm (i.e., Filtered Back-Projection– FPB vs Advanced Modeled Iterative Reconstruction– ADMIRE). A mathematical observer model (i.e., computer detection algorithm) was implemented and used as the basis of image quality comparisons. It was found that image quality increased with increasing dose and decreasing phantom size. The CT system exhibited nonlinear noise and resolution properties, especially at very low-doses, large phantom sizes, and for low-contrast objects. Objective image quality metrics generally increased with increasing dose and ADMIRE strength, and with decreasing phantom size. The ADMIRE algorithm could offer comparable image quality at reduced doses or improved image quality at the same dose (increase in detectability index by up to 163% depending on iterative strength). The use of automatic exposure control resulted in more consistent image quality with changing phantom size.
Based on those results, the dose reduction potential of ADMIRE was further assessed specifically for the task of detecting small (<=6 mm) low-contrast (<=20 HU) lesions. A new low-contrast detectability phantom (with uniform background) was designed and fabricated using a multi-material 3D printer. The phantom was imaged at multiple dose levels and images were reconstructed with FBP and ADMIRE. Human perception experiments were performed to measure the detection accuracy from FBP and ADMIRE images. It was found that ADMIRE had equivalent performance to FBP at 56% less dose.
Using the same image data as the previous study, a number of different mathematical observer models were implemented to assess which models would result in image quality metrics that best correlated with human detection performance. The models included naïve simple metrics of image quality such as contrast-to-noise ratio (CNR) and more sophisticated observer models such as the non-prewhitening matched filter observer model family and the channelized Hotelling observer model family. It was found that non-prewhitening matched filter observers and the channelized Hotelling observers both correlated strongly with human performance. Conversely, CNR was found to not correlate strongly with human performance, especially when comparing different reconstruction algorithms.
The uniform background phantoms used in the previous studies provided a good first-order approximation of image quality. However, due to their simplicity and due to the complexity of iterative reconstruction algorithms, it is possible that such phantoms are not fully adequate to assess the clinical impact of iterative algorithms because patient images obviously do not have smooth uniform backgrounds. To test this hypothesis, two textured phantoms (classified as gross texture and fine texture) and a uniform phantom of similar size were built and imaged on a SOMATOM Flash scanner (Siemens Healthcare). Images were reconstructed using FBP and a Sinogram Affirmed Iterative Reconstruction (SAFIRE). Using an image subtraction technique, quantum noise was measured in all images of each phantom. It was found that in FBP, the noise was independent of the background (textured vs uniform). However, for SAFIRE, noise increased by up to 44% in the textured phantoms compared to the uniform phantom. As a result, the noise reduction from SAFIRE was found to be up to 66% in the uniform phantom but as low as 29% in the textured phantoms. Based on this result, it clear that further investigation was needed into to understand the impact that background texture has on image quality when iterative reconstruction algorithms are used.
To further investigate this phenomenon with more realistic textures, two anthropomorphic textured phantoms were designed to mimic lung vasculature and fatty soft tissue texture. The phantoms (along with a corresponding uniform phantom) were fabricated with a multi-material 3D printer and imaged on the SOMATOM Flash scanner. Scans were repeated a total of 50 times in order to get ensemble statistics of the noise. A novel method of estimating the noise power spectrum (NPS) from irregularly shaped ROIs was developed. It was found that SAFIRE images had highly locally non-stationary noise patterns with pixels near edges having higher noise than pixels in more uniform regions. Compared to FBP, SAFIRE images had 60% less noise on average in uniform regions for edge pixels, noise was between 20% higher and 40% lower. The noise texture (i.e., NPS) was also highly dependent on the background texture for SAFIRE. Therefore, it was concluded that quantum noise properties in the uniform phantoms are not representative of those in patients for iterative reconstruction algorithms and texture should be considered when assessing image quality of iterative algorithms.
The move beyond just assessing noise properties in textured phantoms towards assessing detectability, a series of new phantoms were designed specifically to measure low-contrast detectability in the presence of background texture. The textures used were optimized to match the texture in the liver regions actual patient CT images using a genetic algorithm. The so called “Clustured Lumpy Background” texture synthesis framework was used to generate the modeled texture. Three textured phantoms and a corresponding uniform phantom were fabricated with a multi-material 3D printer and imaged on the SOMATOM Flash scanner. Images were reconstructed with FBP and SAFIRE and analyzed using a multi-slice channelized Hotelling observer to measure detectability and the dose reduction potential of SAFIRE based on the uniform and textured phantoms. It was found that at the same dose, the improvement in detectability from SAFIRE (compared to FBP) was higher when measured in a uniform phantom compared to textured phantoms.
The final trajectory of this project aimed at developing methods to mathematically model lesions, as a means to help assess image quality directly from patient images. The mathematical modeling framework is first presented. The models describe a lesion’s morphology in terms of size, shape, contrast, and edge profile as an analytical equation. The models can be voxelized and inserted into patient images to create so-called “hybrid” images. These hybrid images can then be used to assess detectability or estimability with the advantage that the ground truth of the lesion morphology and location is known exactly. Based on this framework, a series of liver lesions, lung nodules, and kidney stones were modeled based on images of real lesions. The lesion models were virtually inserted into patient images to create a database of hybrid images to go along with the original database of real lesion images. ROI images from each database were assessed by radiologists in a blinded fashion to determine the realism of the hybrid images. It was found that the radiologists could not readily distinguish between real and virtual lesion images (area under the ROC curve was 0.55). This study provided evidence that the proposed mathematical lesion modeling framework could produce reasonably realistic lesion images.
Based on that result, two studies were conducted which demonstrated the utility of the lesion models. The first study used the modeling framework as a measurement tool to determine how dose and reconstruction algorithm affected the quantitative analysis of liver lesions, lung nodules, and renal stones in terms of their size, shape, attenuation, edge profile, and texture features. The same database of real lesion images used in the previous study was used for this study. That database contained images of the same patient at 2 dose levels (50% and 100%) along with 3 reconstruction algorithms from a GE 750HD CT system (GE Healthcare). The algorithms in question were FBP, Adaptive Statistical Iterative Reconstruction (ASiR), and Model-Based Iterative Reconstruction (MBIR). A total of 23 quantitative features were extracted from the lesions under each condition. It was found that both dose and reconstruction algorithm had a statistically significant effect on the feature measurements. In particular, radiation dose affected five, three, and four of the 23 features (related to lesion size, conspicuity, and pixel-value distribution) for liver lesions, lung nodules, and renal stones, respectively. MBIR significantly affected 9, 11, and 15 of the 23 features (including size, attenuation, and texture features) for liver lesions, lung nodules, and renal stones, respectively. Lesion texture was not significantly affected by radiation dose.
The second study demonstrating the utility of the lesion modeling framework focused on assessing detectability of very low-contrast liver lesions in abdominal imaging. Specifically, detectability was assessed as a function of dose and reconstruction algorithm. As part of a parallel clinical trial, images from 21 patients were collected at 6 dose levels per patient on a SOMATOM Flash scanner. Subtle liver lesion models (contrast = -15 HU) were inserted into the raw projection data from the patient scans. The projections were then reconstructed with FBP and SAFIRE (strength 5). Also, lesion-less images were reconstructed. Noise, contrast, CNR, and detectability index of an observer model (non-prewhitening matched filter) were assessed. It was found that SAFIRE reduced noise by 52%, reduced contrast by 12%, increased CNR by 87%. and increased detectability index by 65% compared to FBP. Further, a 2AFC human perception experiment was performed to assess the dose reduction potential of SAFIRE, which was found to be 22% compared to the standard of care dose.
In conclusion, this dissertation provides to the scientific community a series of new methodologies, phantoms, analysis techniques, and modeling tools that can be used to rigorously assess image quality from modern CT systems. Specifically, methods to properly evaluate iterative reconstruction have been developed and are expected to aid in the safe clinical implementation of dose reduction technologies.
Resumo:
Computed tomography (CT) is a valuable technology to the healthcare enterprise as evidenced by the more than 70 million CT exams performed every year. As a result, CT has become the largest contributor to population doses amongst all medical imaging modalities that utilize man-made ionizing radiation. Acknowledging the fact that ionizing radiation poses a health risk, there exists the need to strike a balance between diagnostic benefit and radiation dose. Thus, to ensure that CT scanners are optimally used in the clinic, an understanding and characterization of image quality and radiation dose are essential.
The state-of-the-art in both image quality characterization and radiation dose estimation in CT are dependent on phantom based measurements reflective of systems and protocols. For image quality characterization, measurements are performed on inserts imbedded in static phantoms and the results are ascribed to clinical CT images. However, the key objective for image quality assessment should be its quantification in clinical images; that is the only characterization of image quality that clinically matters as it is most directly related to the actual quality of clinical images. Moreover, for dose estimation, phantom based dose metrics, such as CT dose index (CTDI) and size specific dose estimates (SSDE), are measured by the scanner and referenced as an indicator for radiation exposure. However, CTDI and SSDE are surrogates for dose, rather than dose per-se.
Currently there are several software packages that track the CTDI and SSDE associated with individual CT examinations. This is primarily the result of two causes. The first is due to bureaucracies and governments pressuring clinics and hospitals to monitor the radiation exposure to individuals in our society. The second is due to the personal concerns of patients who are curious about the health risks associated with the ionizing radiation exposure they receive as a result of their diagnostic procedures.
An idea that resonates with clinical imaging physicists is that patients come to the clinic to acquire quality images so they can receive a proper diagnosis, not to be exposed to ionizing radiation. Thus, while it is important to monitor the dose to patients undergoing CT examinations, it is equally, if not more important to monitor the image quality of the clinical images generated by the CT scanners throughout the hospital.
The purposes of the work presented in this thesis are threefold: (1) to develop and validate a fully automated technique to measure spatial resolution in clinical CT images, (2) to develop and validate a fully automated technique to measure image contrast in clinical CT images, and (3) to develop a fully automated technique to estimate radiation dose (not surrogates for dose) from a variety of clinical CT protocols.
Resumo:
With the new academic year structure encouraging more in-term assessment to replace end-of-year examinations one of the problems we face is assessing students and keeping track of individual student learning without overloading the students and staff with excessive assessment burdens.
In the School of Electronics, Electrical Engineering and Computer Science, we have constructed a system that allows students to self-assess their capability on a simple Yes/No/Don’t Know scale against fine grained learning outcomes for a module. As the term progresses students update their record as appropriately, including selecting a Learnt option to reflect improvements they have gained as part of their studies.
In the system each of the learning outcomes are linked to the relevant teaching session (lectures and labs) and to online resources that students can access at any time. Students can structure their own learning experience to their needs and preferences in order to attain the learning outcomes.
The system keeps a history of the student’s record, allowing the lecturer to observe how the students’ abilities progress over the term and to compare it to assessment results. The system also keeps of any of the resource links that student has clicked on and the related learning outcome.
The initial work is comparing the accuracy of the student self-assessments with their performance in the related questions in the traditional end-of-year examination.