985 resultados para exam-oriented assessment


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Disponer de información precisa y actualizada de inventario forestal es una pieza clave para mejorar la gestión forestal sostenible y para proponer y evaluar políticas de conservación de bosques que permitan la reducción de emisiones de carbono debidas a la deforestación y degradación forestal (REDD). En este sentido, la tecnología LiDAR ha demostrado ser una herramienta perfecta para caracterizar y estimar de forma continua y en áreas extensas la estructura del bosque y las principales variables de inventario forestal. Variables como la biomasa, el número de pies, el volumen de madera, la altura dominante, el diámetro o la altura media son estimadas con una calidad comparable a los inventarios tradicionales de campo. La presente tesis se centra en analizar la aplicación de los denominados métodos de masa de inventario forestal con datos LIDAR bajo diferentes condiciones y características de masa forestal (bosque templados puros y mixtos) y utilizando diferentes bases de datos LiDAR (información proveniente de vuelo nacionales e información capturada de forma específica). Como consecuencia de lo anterior, se profundiza en la generación de inventarios forestales continuos con LiDAR en grandes áreas. Los métodos de masa se basan en la búsqueda de relaciones estadísticas entre variables predictoras derivadas de la nube de puntos LiDAR y las variables de inventario forestal medidas en campo con el objeto de generar una cartografía continua de inventario forestal. El rápido desarrollo de esta tecnología en los últimos años ha llevado a muchos países a implantar programas nacionales de captura de información LiDAR aerotransportada. Estos vuelos nacionales no están pensados ni diseñados para fines forestales por lo que es necesaria la evaluación de la validez de esta información LiDAR para la descripción de la estructura del bosque y la medición de variables forestales. Esta información podría suponer una drástica reducción de costes en la generación de información continua de alta resolución de inventario forestal. En el capítulo 2 se evalúa la estimación de variables forestales a partir de la información LiDAR capturada en el marco del Plan Nacional de Ortofotografía Aérea (PNOA-LiDAR) en España. Para ello se compara un vuelo específico diseñado para inventario forestal con la información de la misma zona capturada dentro del PNOA-LiDAR. El caso de estudio muestra cómo el ángulo de escaneo, la pendiente y orientación del terreno afectan de forma estadísticamente significativa, aunque con pequeñas diferencias, a la estimación de biomasa y variables de estructura forestal derivadas del LiDAR. La cobertura de copas resultó más afectada por estos factores que los percentiles de alturas. Considerando toda la zona de estudio, la estimación de la biomasa con ambas bases de datos no presentó diferencias estadísticamente significativas. Las simulaciones realizadas muestran que las diferencias medias en la estimación de biomasa entre un vuelo específico y el vuelo nacional podrán superar el 4% en áreas abruptas, con ángulos de escaneo altos y cuando la pendiente de la ladera no esté orientada hacia la línea de escaneo. En el capítulo 3 se desarrolla un estudio en masas mixtas y puras de pino silvestre y haya, con un enfoque multi-fuente empleando toda la información disponible (vuelos LiDAR nacionales de baja densidad de puntos, imágenes satelitales Landsat y parcelas permanentes del inventario forestal nacional español). Se concluye que este enfoque multi-fuente es adecuado para realizar inventarios forestales continuos de alta resolución en grandes superficies. Los errores obtenidos en la fase de ajuste y de validación de los modelos de área basimétrica y volumen son similares a los registrados por otros autores (usando un vuelo específico y parcelas de campo específicas). Se observan errores mayores en la variable número de pies que los encontrados en la literatura, que pueden ser explicados por la influencia de la metodología de parcelas de radio variable en esta variable. En los capítulos 4 y 5 se evalúan los métodos de masa para estimar biomasa y densidad de carbono en bosques tropicales. Para ello se trabaja con datos del Parque Nacional Volcán Poás (Costa Rica) en dos situaciones diferentes: i) se dispone de una cobertura completa LiDAR del área de estudio (capitulo 4) y ii) la cobertura LiDAR completa no es técnica o económicamente posible y se combina una cobertura incompleta de LiDAR con imágenes Landsat e información auxiliar para la estimación de biomasa y carbono (capitulo 5). En el capítulo 4 se valida un modelo LiDAR general de estimación de biomasa aérea en bosques tropicales y se compara con los resultados obtenidos con un modelo ajustado de forma específica para el área de estudio. Ambos modelos están basados en la variable altura media de copas (TCH por sus siglas en inglés) derivada del modelo digital LiDAR de altura de la vegetación. Los resultados en el área de estudio muestran que el modelo general es una alternativa fiable al ajuste de modelos específicos y que la biomasa aérea puede ser estimada en una nueva zona midiendo en campo únicamente la variable área basimétrica (BA). Para mejorar la aplicación de esta metodología es necesario definir en futuros trabajos procedimientos adecuados de medición de la variable área basimétrica en campo (localización, tamaño y forma de las parcelas de campo). La relación entre la altura media de copas del LiDAR y el área basimétrica (Coeficiente de Stock) obtenida en el área de estudio varía localmente. Por tanto es necesario contar con más información de campo para caracterizar la variabilidad del Coeficiente de Stock entre zonas de vida y si estrategias como la estratificación pueden reducir los errores en la estimación de biomasa y carbono en bosques tropicales. En el capítulo 5 se concluye que la combinación de una muestra sistemática de información LiDAR con una cobertura completa de imagen satelital de moderada resolución (e información auxiliar) es una alternativa efectiva para la realización de inventarios continuos en bosques tropicales. Esta metodología permite estimar altura de la vegetación, biomasa y carbono en grandes zonas donde la captura de una cobertura completa de LiDAR y la realización de un gran volumen de trabajo de campo es económica o/y técnicamente inviable. Las alternativas examinadas para la predicción de biomasa a partir de imágenes Landsat muestran una ligera disminución del coeficiente de determinación y un pequeño aumento del RMSE cuando la cobertura de LiDAR es reducida de forma considerable. Los resultados indican que la altura de la vegetación, la biomasa y la densidad de carbono pueden ser estimadas en bosques tropicales de forma adecuada usando coberturas de LIDAR bajas (entre el 5% y el 20% del área de estudio). ABSTRACT The availability of accurate and updated forest data is essential for improving sustainable forest management, promoting forest conservation policies and reducing carbon emissions from deforestation and forest degradation (REDD). In this sense, LiDAR technology proves to be a clear-cut tool for characterizing forest structure in large areas and assessing main forest-stand variables. Forest variables such as biomass, stem volume, basal area, mean diameter, mean height, dominant height, and stem number can be thus predicted with better or comparable quality than with costly traditional field inventories. In this thesis, it is analysed the potential of LiDAR technology for the estimation of plot-level forest variables under a range of conditions (conifer & broadleaf temperate forests and tropical forests) and different LiDAR capture characteristics (nationwide LiDAR information vs. specific forest LiDAR data). This study evaluates the application of LiDAR-based plot-level methods in large areas. These methods are based on statistical relationships between predictor variables (derived from airborne data) and field-measured variables to generate wall to wall forest inventories. The fast development of this technology in recent years has led to an increasing availability of national LiDAR datasets, usually developed for multiple purposes throughout an expanding number of countries and regions. The evaluation of the validity of nationwide LiDAR databases (not designed specifically for forest purposes) is needed and presents a great opportunity for substantially reducing the costs of forest inventories. In chapter 2, the suitability of Spanish nationwide LiDAR flight (PNOA) to estimate forest variables is analyzed and compared to a specifically forest designed LiDAR flight. This study case shows that scan angle, terrain slope and aspect significantly affect the assessment of most of the LiDAR-derived forest variables and biomass estimation. Especially, the estimation of canopy cover is more affected than height percentiles. Considering the entire study area, biomass estimations from both databases do not show significant differences. Simulations show that differences in biomass could be larger (more than 4%) only in particular situations, such as steep areas when the slopes are non-oriented towards the scan lines and the scan angles are larger than 15º. In chapter 3, a multi-source approach is developed, integrating available databases such as nationwide LiDAR flights, Landsat imagery and permanent field plots from SNFI, with good resultos in the generation of wall to wall forest inventories. Volume and basal area errors are similar to those obtained by other authors (using specific LiDAR flights and field plots) for the same species. Errors in the estimation of stem number are larger than literature values as a consequence of the great influence that variable-radius plots, as used in SNFI, have on this variable. In chapters 4 and 5 wall to wall plot-level methodologies to estimate aboveground biomass and carbon density in tropical forest are evaluated. The study area is located in the Poas Volcano National Park (Costa Rica) and two different situations are analyzed: i) available complete LiDAR coverage (chapter 4) and ii) a complete LiDAR coverage is not available and wall to wall estimation is carried out combining LiDAR, Landsat and ancillary data (chapter 5). In chapter 4, a general aboveground biomass plot-level LiDAR model for tropical forest (Asner & Mascaro, 2014) is validated and a specific model for the study area is fitted. Both LiDAR plot-level models are based on the top-of-canopy height (TCH) variable that is derived from the LiDAR digital canopy model. Results show that the pantropical plot-level LiDAR methodology is a reliable alternative to the development of specific models for tropical forests and thus, aboveground biomass in a new study area could be estimated by only measuring basal area (BA). Applying this methodology, the definition of precise BA field measurement procedures (e.g. location, size and shape of the field plots) is decisive to achieve reliable results in future studies. The relation between BA and TCH (Stocking Coefficient) obtained in our study area in Costa Rica varied locally. Therefore, more field work is needed for assessing Stocking Coefficient variations between different life zones and the influence of the stratification of the study areas in tropical forests on the reduction of uncertainty. In chapter 5, the combination of systematic LiDAR information sampling and full coverage Landsat imagery (and ancillary data) prove to be an effective alternative for forest inventories in tropical areas. This methodology allows estimating wall to wall vegetation height, biomass and carbon density in large areas where full LiDAR coverage and traditional field work are technically and/or economically unfeasible. Carbon density prediction using Landsat imaginery shows a slight decrease in the determination coefficient and an increase in RMSE when harshly decreasing LiDAR coverage area. Results indicate that feasible estimates of vegetation height, biomass and carbon density can be accomplished using low LiDAR coverage areas (between 5% and 20% of the total area) in tropical locations.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

El auge y penetración de las nuevas tecnologías junto con la llamada Web Social están cambiando la forma en la que accedemos a la medicina. Cada vez más pacientes y profesionales de la medicina están creando y consumiendo recursos digitales de contenido clínico a través de Internet, surgiendo el problema de cómo asegurar la fiabilidad de estos recursos. Además, un nuevo concepto está apareciendo, el de pervasive healthcare o sanidad ubicua, motivado por pacientes que demandan un acceso a los servicios sanitarios en todo momento y en todo lugar. Este nuevo escenario lleva aparejado un problema de confianza en los proveedores de servicios sanitarios. Las plataformas de eLearning se están erigiendo como paradigma de esta nueva Medicina 2.0 ya que proveen un servicio abierto a la vez que controlado/supervisado a recursos digitales, y facilitan las interacciones y consultas entre usuarios, suponiendo una buena aproximación para esta sanidad ubicua. En estos entornos los problemas de fiabilidad y confianza pueden ser solventados mediante la implementación de mecanismos de recomendación de recursos y personas de manera confiable. Tradicionalmente las plataformas de eLearning ya cuentan con mecanismos de recomendación, si bien están más enfocados a la recomendación de recursos. Para la recomendación de usuarios es necesario acudir a mecanismos más elaborados como son los sistemas de confianza y reputación (trust and reputation) En ambos casos, tanto la recomendación de recursos como el cálculo de la reputación de los usuarios se realiza teniendo en cuenta criterios principalmente subjetivos como son las opiniones de los usuarios. En esta tesis doctoral proponemos un nuevo modelo de confianza y reputación que combina evaluaciones automáticas de los recursos digitales en una plataforma de eLearning, con las opiniones vertidas por los usuarios como resultado de las interacciones con otros usuarios o después de consumir un recurso. El enfoque seguido presenta la novedad de la combinación de una parte objetiva con otra subjetiva, persiguiendo mitigar el efecto de posibles castigos subjetivos por parte de usuarios malintencionados, a la vez que enriquecer las evaluaciones objetivas con información adicional acerca de la capacidad pedagógica del recurso o de la persona. El resultado son recomendaciones siempre adaptadas a los requisitos de los usuarios, y de la máxima calidad tanto técnica como educativa. Esta nueva aproximación requiere una nueva herramienta para su validación in-silico, al no existir ninguna aplicación que permita la simulación de plataformas de eLearning con mecanismos de recomendación de recursos y personas, donde además los recursos sean evaluados objetivamente. Este trabajo de investigación propone pues una nueva herramienta, basada en el paradigma de programación orientada a agentes inteligentes para el modelado de comportamientos complejos de usuarios en plataformas de eLearning. Además, la herramienta permite también la simulación del funcionamiento de este tipo de entornos dedicados al intercambio de conocimiento. La evaluación del trabajo propuesto en este documento de tesis se ha realizado de manera iterativa a lo largo de diferentes escenarios en los que se ha situado al sistema frente a una amplia gama de comportamientos de usuarios. Se ha comparado el rendimiento del modelo de confianza y reputación propuesto frente a dos modos de recomendación tradicionales: a) utilizando sólo las opiniones subjetivas de los usuarios para el cálculo de la reputación y por extensión la recomendación; y b) teniendo en cuenta sólo la calidad objetiva del recurso sin hacer ningún cálculo de reputación. Los resultados obtenidos nos permiten afirmar que el modelo desarrollado mejora la recomendación ofrecida por las aproximaciones tradicionales, mostrando una mayor flexibilidad y capacidad de adaptación a diferentes situaciones. Además, el modelo propuesto es capaz de asegurar la recomendación de nuevos usuarios entrando al sistema frente a la nula recomendación para estos usuarios presentada por el modo de recomendación predominante en otras plataformas que basan la recomendación sólo en las opiniones de otros usuarios. Por último, el paradigma de agentes inteligentes ha probado su valía a la hora de modelar plataformas virtuales complejas orientadas al intercambio de conocimiento, especialmente a la hora de modelar y simular el comportamiento de los usuarios de estos entornos. La herramienta de simulación desarrollada ha permitido la evaluación del modelo de confianza y reputación propuesto en esta tesis en una amplia gama de situaciones diferentes. ABSTRACT Internet is changing everything, and this revolution is especially present in traditionally offline spaces such as medicine. In recent years health consumers and health service providers are actively creating and consuming Web contents stimulated by the emergence of the Social Web. Reliability stands out as the main concern when accessing the overwhelming amount of information available online. Along with this new way of accessing the medicine, new concepts like ubiquitous or pervasive healthcare are appearing. Trustworthiness assessment is gaining relevance: open health provisioning systems require mechanisms that help evaluating individuals’ reputation in pursuit of introducing safety to these open and dynamic environments. Technical Enhanced Learning (TEL) -commonly known as eLearning- platforms arise as a paradigm of this Medicine 2.0. They provide an open while controlled/supervised access to resources generated and shared by users, enhancing what it is being called informal learning. TEL systems also facilitate direct interactions amongst users for consultation, resulting in a good approach to ubiquitous healthcare. The aforementioned reliability and trustworthiness problems can be faced by the implementation of mechanisms for the trusted recommendation of both resources and healthcare services providers. Traditionally, eLearning platforms already integrate recommendation mechanisms, although this recommendations are basically focused on providing an ordered classifications of resources. For users’ recommendation, the implementation of trust and reputation systems appears as the best solution. Nevertheless, both approaches base the recommendation on the information from the subjective opinions of other users of the platform regarding the resources or the users. In this PhD work a novel approach is presented for the recommendation of both resources and users within open environments focused on knowledge exchange, as it is the case of TEL systems for ubiquitous healthcare. The proposed solution adds the objective evaluation of the resources to the traditional subjective personal opinions to estimate the reputation of the resources and of the users of the system. This combined measure, along with the reliability of that calculation, is used to provide trusted recommendations. The integration of opinions and evaluations, subjective and objective, allows the model to defend itself against misbehaviours. Furthermore, it also allows ‘colouring’ cold evaluation values by providing additional quality information such as the educational capacities of a digital resource in an eLearning system. As a result, the recommendations are always adapted to user requirements, and of the maximum technical and educational quality. To our knowledge, the combination of objective assessments and subjective opinions to provide recommendation has not been considered before in the literature. Therefore, for the evaluation of the trust and reputation model defined in this PhD thesis, a new simulation tool will be developed following the agent-oriented programming paradigm. The multi-agent approach allows an easy modelling of independent and proactive behaviours for the simulation of users of the system, conforming a faithful resemblance of real users of TEL platforms. For the evaluation of the proposed work, an iterative approach have been followed, testing the performance of the trust and reputation model while providing recommendation in a varied range of scenarios. A comparison with two traditional recommendation mechanisms was performed: a) using only users’ past opinions about a resource and/or other users; and b) not using any reputation assessment and providing the recommendation considering directly the objective quality of the resources. The results show that the developed model improves traditional approaches at providing recommendations in Technology Enhanced Learning (TEL) platforms, presenting a higher adaptability to different situations, whereas traditional approaches only have good results under favourable conditions. Furthermore the promotion period mechanism implemented successfully helps new users in the system to be recommended for direct interactions as well as the resources created by them. On the contrary OnlyOpinions fails completely and new users are never recommended, while traditional approaches only work partially. Finally, the agent-oriented programming (AOP) paradigm has proven its validity at modelling users’ behaviours in TEL platforms. Intelligent software agents’ characteristics matched the main requirements of the simulation tool. The proactivity, sociability and adaptability of the developed agents allowed reproducing real users’ actions and attitudes through the diverse situations defined in the evaluation framework. The result were independent users, accessing to different resources and communicating amongst them to fulfil their needs, basing these interactions on the recommendations provided by the reputation engine.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper presents an online C compiler designed so that students can program their practical assignments in Programming courses. What is really innovative is the self-assessment of the exercises based on black-box tests and train students’ skill to test software. Moreover, this tool lets instructors, not only proposing and classifying practical exercises, but also evaluating automatically the efforts dedicated and the results obtained by the students. The system has been applied to the 1st-year students at the Industrial Engineering specialization at the Universidad Politecnica de Madrid. Results show that the students obtained better academic performance, reducing the failure rate in the practical exam considerably with respect to previous years, in addition that an anonymous survey proved that students are satisfied with the system because they get instant feedback about their programs.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Among the factors that affect the convergence towards the European Higher Education Area, university teaching staff's motivation is fundamental, and consequently, it is crucial to empirically know what this motivation depends on. In this context, one of the most relevant changes in the teacher-student relationship is assessment. In fact, the transition from a static assessment -focused on only one temporal point (final exam)- to a dynamic assessment, will require changes in thought and action, both on the part of teachers and students. In this line, the objective of this paper is to analyze the determinants of teaching staff's predisposition to the continuous assessment method. Specifically, we consider the following explanatory dimensions: teaching method used (which measures their degree of involvement with the ongoing adaptation process), type of subject (core, compulsory and optional), and teacher's personal characteristics (professional status and gender). The empirical application carried out at the University of Alicante uses Logit Models with Random Coefficients to capture heterogeneity, and shows that "cooperative learning" is a clear-cut determinant of "continuous assessment" as well as "continuous assessment plus final examination". Also, a conspicuous result, which in turn becomes a thought-provoking finding, is that professional status is highly relevant as a teacher's engagement is closely related to prospects of stability. Consequently, the most relevant implications from the results revolve around the way academic institutions can propose and implement inducement for their teaching staff.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Among the factors that affect the convergence towards the European Higher Education Area, university teaching staff's motivation is fundamental, and consequently, it is crucial to empirically know what this motivation depends on. In this context, one of the most relevant changes in the teacher-student relationship is assessment. In fact, the transition from a static assessment -focused on only one temporal point (final exam)- to a dynamic assessment, will require changes in thought and action, both on the part of teachers and students. In this line, the objective of this paper is to analyze the determinants of teaching staff's predisposition to the continuous assessment method. Specifically, we consider the following explanatory dimensions: teaching method used (which measures their degree of involvement with the ongoing adaptation process), type of subject (core, compulsory and optional), and teacher's personal characteristics (professional status and gender). The empirical application carried out at the University of Alicante uses Logit Models with Random Coefficients to capture heterogeneity, and shows that "cooperative learning" is a clear-cut determinant of "continuous assessment" as well as "continuous assessment plus final examination". Also, a conspicuous result, which in turn becomes a thought-provoking finding, is that professional status is highly relevant as a teacher's engagement is closely related to prospects of stability. Consequently, the most relevant implications from the results revolve around the way academic institutions can propose and implement inducement for their teaching staff.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Objective. To evaluate the association between nasal obstruction and (1) demographic factors, (2) medical history, (3) physical tests, and (4) nasal exam findings. Study Design. CASE SERIES: Methods. Chart review at a tertiary medical center. Results. Two hundred-forty consecutive patients (52.1 ± 17.5 years old, with a Nasal Obstruction Symptom Evaluation (NOSE) score of 32.0 ± 24.1) were included. Demographic factors and inferior turbinate sizes were not associated with NOSE score or Nasal Obstruction Visual Analog Scale (NO-VAS). A significant association was found between higher NOSE score on univariate analysis and positive history of nasal trauma (p = 0.0136), allergic rhinitis (p < 0.0001), use of nasal steroids (p = 0.0108), higher grade of external nasal deformity (p = 0.0149), higher internal nasal septal deviation grade (p = 0.0024), and narrow internal nasal valve angle (p < 0.0001). Multivariate analysis identified the following as independent predictors of high NOSE score: NO-VAS: ≥50 (Odds Ratio (OR) = 17.6 (95% CI 5.83-61.6), p < 0.0001), external nasal deformity: grades 2-4 (OR = 4.63 (95% CI 1.14-19.9), p = 0.0339), and allergic rhinitis: yes (OR = 5.5 (95% CI 1.77-18.7), p = 0.0041). Conclusion. Allergic rhinitis, NO-VAS score ≥ 50, and external nasal deformity (grades 2-4) were statistically significant independent predictors of high NOSE scores on multivariate analysis. Inferior turbinate size was not associated with NOSE scores or NO-VAS.

Relevância:

30.00% 30.00%

Publicador:

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This study concerns the application of a model of effective interpersonal relationships to problems arising from staff assessment at I.C.I. Ltd. Corporate Laboratory between 1972 and 1974. In collaboration with academic and industrial supervision, the study commenced with a survey of management and supervisor opinions about the effectiveness of current staff (work) relationships, with particular reference to the problem of recognising and developing creative potential. This survey emphasised a need to improve the relationships between staff in the staff assessment context. A survey of research into creativity emphasised the importance of the interpersonal environment for obtaining creative behaviour in an organisation context. A further survey of theories of how inter­personal behaviour related to personal creativity (therapeutic psychology) provided a model of effective interpersonal behaviour (Carkhuff, 1969) that could be applied to the organisation context of staff assessment. The objective of the project was redefined as a need to improve the conditions of interpersonal behaviour in relation to certain (career development) problems arising from staff assessment practices. In order to demonstrate the application of the model of effective interpersonal behaviour, the research student recorded interviews between himself and members of staff designed to develop and operate the dimensions of the model. Different samples of staff were used to develop the 'facilitative' and the 'action oriented' dimensions of bahaviour, and then for the operation of a helping programme (based on vocational guidance tests). These interactions have been analysed, according to the scales of measurement in the model ana the results are presented in case study form in this thesis. At each stage of the project, results and conclusions were presented to the sponsoring organisation (e.g. industrial supervisor) in order to assess their (subjective) opinion of relevance to the organ­ isation. Finally, recommendations on further actions towards general improvement of the work relationships in the laboratory were presented in a brief report to the sponsor.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Strategic sourcing has increased in importance in recent years, and now plays an important role in companies’ planning. The current volatility in supply markets means companies face multiple challenges involving lock-in situations, supplier bankruptcies or supply security issues. In addition, their exposure can increase due to natural disasters, as witnessed recently in the form of bird flu, volcanic ash and tsunamis. Therefore, the primary focus of this study is risk management in the context of strategic sourcing. The study presents a literature review on sourcing based on the 15 years from 1998–2012, and considers 131 academic articles. The literature describes strategic sourcing as a strategic, holistic process in managing supplier relationships, with a long-term focus on adding value to the company and realising competitive advantage. Few studies discovered the real risk impact and status of risk management in strategic sourcing, and evaluation across countries and industries was limited, with the construction sector particularly under-researched. This methodology is founded on a qualitative study of twenty cases across Ger-many and the United Kingdom from the construction sector and electronics manufacturing industries. While considering risk management in the context of strategic sourcing, the thesis takes into account six dimensions that cover trends in strategic sourcing, theoretical and practical sourcing models, risk management, supply and demand management, critical success factors and the strategic supplier evaluation. The study contributes in several ways. First, recent trends are traced and future needs identified across the research dimensions of countries, industries and companies. Second, it evaluates critical success factors in contemporary strategic sourcing. Third, it explores the application of theoretical and practical sourcing models in terms of effectiveness and sustainability. Fourth, based on the case study findings, a risk-oriented strategic sourcing framework and a model for strategic sourcing are developed. These are based on the validation of contemporary requirements and a critical evaluation of the existing situation. It contemplates the empirical findings and leads to a structured process to manage risk in strategic sourcing. The risk-oriented framework considers areas such as trends, corporate and sourcing strategy, critical success factors, strategic supplier selection criteria, risk assessment, reporting, strategy alignment and reporting. The proposed model highlights the essential dimensions in strategic sourcing and guides us to a new definition of strategic sourcing supported by this empirical study.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Undergraduate programmes on construction management and other closely related built environment disciplines are currently taught and assessed on a modular basis. This is the case in the UK and in many other countries globally. However, it can be argued that professionally oriented programmes like these are better assessed on a non-modular basis, in order to produce graduates who can apply knowledge on different subject contents in cohesion to solve complex practical scenarios in their work environments. The examples of medical programmes where students are assessed on a non-modular basis can be cited as areas where this is already being done. A preliminary study was undertaken to explore the applicability of non-modular assessment within construction management undergraduate education. A selected sample of university academics was interviewed to gather their perspectives on applicability of non-modular assessment. General acceptance was observed among the academics involved that integrating non-modular assessment is applicable and will be beneficial. All academics stated that at least some form of non-modular assessment as being currently used in their programmes. Examples where cross-modular knowledge is assessed included comprehensive/multi-disciplinary project modules and creating larger modules to amalgamate a number of related subject areas. As opposed to a complete shift from modular to non-modular, an approach where non-modular assessment is integrated and its use further expanded within the current system is therefore suggested. This is due to the potential benefits associated with this form of assessment to professionally aligned built environment programmes

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Small and Medium Enterprises (SMEs) play an important part in the economy of any country. Initially, a flat management hierarchy, quick response to market changes and cost competitiveness were seen as the competitive characteristics of an SME. Recently, in developed economies, technological capabilities (TCs) management- managing existing and developing or assimilating new technological capabilities for continuous process and product innovations, has become important for both large organisations and SMEs to achieve sustained competitiveness. Therefore, various technological innovation capability (TIC) models have been developed at firm level to assess firms‘ innovation capability level. These models output help policy makers and firm managers to devise policies for deepening a firm‘s technical knowledge generation, acquisition and exploitation capabilities for sustained technological competitive edge. However, in developing countries TCs management is more of TCs upgrading: acquisitions of TCs from abroad, and then assimilating, innovating and exploiting them. Most of the TIC models for developing countries delineate the level of TIC required as firms move from the acquisition to innovative level. However, these models do not provide tools for assessing the existing level of TIC of a firm and various factors affecting TIC, to help practical interventions for TCs upgrading of firms for improved or new processes and products. Recently, the Government of Pakistan (GOP) has realised the importance of TCs upgrading in SMEs-especially export-oriented, for their sustained competitiveness. The GOP has launched various initiatives with local and foreign assistance to identify ways and means of upgrading local SMEs capabilities. This research targets this gap and developed a TICs assessment model for identifying the existing level of TIC of manufacturing SMEs existing in clusters in Sialkot, Pakistan. SME executives in three different export-oriented clusters at Sialkot were interviewed to analyse technological capabilities development initiatives (CDIs) taken by them to develop and upgrade their firms‘ TCs. Data analysed at CDI, firm, cluster and cross-cluster level first helped classify interviewed firms as leader, follower and reactor, with leader firms claiming to introduce mostly new CDIs to their cluster. Second, the data analysis displayed that mostly interviewed leader firms exhibited ‗learning by interacting‘ and ‗learning by training‘ capabilities for expertise acquisition from customers and international consultants. However, these leader firms did not show much evidence of learning by using, reverse engineering and R&D capabilities, which according to the extant literature are necessary for upgrading existing TIC level and thus TCs of firm for better value-added processes and products. The research results are supported by extant literature on Sialkot clusters. Thus, in sum, a TIC assessment model was developed in this research which qualitatively identified interviewed firms‘ TIC levels, the factors affecting them, and is validated by existing literature on interviewed Sialkot clusters. Further, the research gives policy level recommendations for TIC and thus TCs upgrading at firm and cluster level for targeting better value-added markets.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Ebben a tanulmányban a Natura 2000 erdők közgazdasági kérdéseit jártuk körül az ökoszisztéma-szolgáltatások koncepciójának segítségével, nemzetközi és hazai szakirodalomra építve. Emellett a természetközeli erdőgazdálkodás fogalomkörét, s azon belül a folyamatos erdőborítást szolgáló erdőgazdálkodást vettük alapul. A következőkben néhány összegző megállapítást teszünk, és kijelölünk további kutatási irányokat. _____ This study has been prepared within the LIFEinFORESTS – Improved communication, cooperation and capacity building for preserving biodiversity in Natura 2000 forests (B2 action, LIFE13 INF/HU/001163) – project in the framework of LIFE+ Information and Communication under the contract signed with the Duna-Ipoly National Park Directorate. The main aim of the study is to summarize the international and Hungarian economic and environmental economic literature related to the Natura 2000 forests, and serve as a background study for the communication with and training of forest owners and users operating at Natura 2000 sites. The concept of ecosystem services (ESs) is used as an overall framework for the study. In our opinion it is able to show all the benefits provided by forests and can also help to reveal that the benefits of nature-oriented, continuous cover forest management (CCF) can exceed the benefits of traditional rotation forest management (RFM). The definition and the classification of the Millennium Ecosystem Assessment (MA, 2003, 2005) is used throughout the study, so provisioning, cultural, regulating and supporting services are distinguished.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

X-ray computed tomography (CT) imaging constitutes one of the most widely used diagnostic tools in radiology today with nearly 85 million CT examinations performed in the U.S in 2011. CT imparts a relatively high amount of radiation dose to the patient compared to other x-ray imaging modalities and as a result of this fact, coupled with its popularity, CT is currently the single largest source of medical radiation exposure to the U.S. population. For this reason, there is a critical need to optimize CT examinations such that the dose is minimized while the quality of the CT images is not degraded. This optimization can be difficult to achieve due to the relationship between dose and image quality. All things being held equal, reducing the dose degrades image quality and can impact the diagnostic value of the CT examination.

A recent push from the medical and scientific community towards using lower doses has spawned new dose reduction technologies such as automatic exposure control (i.e., tube current modulation) and iterative reconstruction algorithms. In theory, these technologies could allow for scanning at reduced doses while maintaining the image quality of the exam at an acceptable level. Therefore, there is a scientific need to establish the dose reduction potential of these new technologies in an objective and rigorous manner. Establishing these dose reduction potentials requires precise and clinically relevant metrics of CT image quality, as well as practical and efficient methodologies to measure such metrics on real CT systems. The currently established methodologies for assessing CT image quality are not appropriate to assess modern CT scanners that have implemented those aforementioned dose reduction technologies.

Thus the purpose of this doctoral project was to develop, assess, and implement new phantoms, image quality metrics, analysis techniques, and modeling tools that are appropriate for image quality assessment of modern clinical CT systems. The project developed image quality assessment methods in the context of three distinct paradigms, (a) uniform phantoms, (b) textured phantoms, and (c) clinical images.

The work in this dissertation used the “task-based” definition of image quality. That is, image quality was broadly defined as the effectiveness by which an image can be used for its intended task. Under this definition, any assessment of image quality requires three components: (1) A well defined imaging task (e.g., detection of subtle lesions), (2) an “observer” to perform the task (e.g., a radiologists or a detection algorithm), and (3) a way to measure the observer’s performance in completing the task at hand (e.g., detection sensitivity/specificity).

First, this task-based image quality paradigm was implemented using a novel multi-sized phantom platform (with uniform background) developed specifically to assess modern CT systems (Mercury Phantom, v3.0, Duke University). A comprehensive evaluation was performed on a state-of-the-art CT system (SOMATOM Definition Force, Siemens Healthcare) in terms of noise, resolution, and detectability as a function of patient size, dose, tube energy (i.e., kVp), automatic exposure control, and reconstruction algorithm (i.e., Filtered Back-Projection– FPB vs Advanced Modeled Iterative Reconstruction– ADMIRE). A mathematical observer model (i.e., computer detection algorithm) was implemented and used as the basis of image quality comparisons. It was found that image quality increased with increasing dose and decreasing phantom size. The CT system exhibited nonlinear noise and resolution properties, especially at very low-doses, large phantom sizes, and for low-contrast objects. Objective image quality metrics generally increased with increasing dose and ADMIRE strength, and with decreasing phantom size. The ADMIRE algorithm could offer comparable image quality at reduced doses or improved image quality at the same dose (increase in detectability index by up to 163% depending on iterative strength). The use of automatic exposure control resulted in more consistent image quality with changing phantom size.

Based on those results, the dose reduction potential of ADMIRE was further assessed specifically for the task of detecting small (<=6 mm) low-contrast (<=20 HU) lesions. A new low-contrast detectability phantom (with uniform background) was designed and fabricated using a multi-material 3D printer. The phantom was imaged at multiple dose levels and images were reconstructed with FBP and ADMIRE. Human perception experiments were performed to measure the detection accuracy from FBP and ADMIRE images. It was found that ADMIRE had equivalent performance to FBP at 56% less dose.

Using the same image data as the previous study, a number of different mathematical observer models were implemented to assess which models would result in image quality metrics that best correlated with human detection performance. The models included naïve simple metrics of image quality such as contrast-to-noise ratio (CNR) and more sophisticated observer models such as the non-prewhitening matched filter observer model family and the channelized Hotelling observer model family. It was found that non-prewhitening matched filter observers and the channelized Hotelling observers both correlated strongly with human performance. Conversely, CNR was found to not correlate strongly with human performance, especially when comparing different reconstruction algorithms.

The uniform background phantoms used in the previous studies provided a good first-order approximation of image quality. However, due to their simplicity and due to the complexity of iterative reconstruction algorithms, it is possible that such phantoms are not fully adequate to assess the clinical impact of iterative algorithms because patient images obviously do not have smooth uniform backgrounds. To test this hypothesis, two textured phantoms (classified as gross texture and fine texture) and a uniform phantom of similar size were built and imaged on a SOMATOM Flash scanner (Siemens Healthcare). Images were reconstructed using FBP and a Sinogram Affirmed Iterative Reconstruction (SAFIRE). Using an image subtraction technique, quantum noise was measured in all images of each phantom. It was found that in FBP, the noise was independent of the background (textured vs uniform). However, for SAFIRE, noise increased by up to 44% in the textured phantoms compared to the uniform phantom. As a result, the noise reduction from SAFIRE was found to be up to 66% in the uniform phantom but as low as 29% in the textured phantoms. Based on this result, it clear that further investigation was needed into to understand the impact that background texture has on image quality when iterative reconstruction algorithms are used.

To further investigate this phenomenon with more realistic textures, two anthropomorphic textured phantoms were designed to mimic lung vasculature and fatty soft tissue texture. The phantoms (along with a corresponding uniform phantom) were fabricated with a multi-material 3D printer and imaged on the SOMATOM Flash scanner. Scans were repeated a total of 50 times in order to get ensemble statistics of the noise. A novel method of estimating the noise power spectrum (NPS) from irregularly shaped ROIs was developed. It was found that SAFIRE images had highly locally non-stationary noise patterns with pixels near edges having higher noise than pixels in more uniform regions. Compared to FBP, SAFIRE images had 60% less noise on average in uniform regions for edge pixels, noise was between 20% higher and 40% lower. The noise texture (i.e., NPS) was also highly dependent on the background texture for SAFIRE. Therefore, it was concluded that quantum noise properties in the uniform phantoms are not representative of those in patients for iterative reconstruction algorithms and texture should be considered when assessing image quality of iterative algorithms.

The move beyond just assessing noise properties in textured phantoms towards assessing detectability, a series of new phantoms were designed specifically to measure low-contrast detectability in the presence of background texture. The textures used were optimized to match the texture in the liver regions actual patient CT images using a genetic algorithm. The so called “Clustured Lumpy Background” texture synthesis framework was used to generate the modeled texture. Three textured phantoms and a corresponding uniform phantom were fabricated with a multi-material 3D printer and imaged on the SOMATOM Flash scanner. Images were reconstructed with FBP and SAFIRE and analyzed using a multi-slice channelized Hotelling observer to measure detectability and the dose reduction potential of SAFIRE based on the uniform and textured phantoms. It was found that at the same dose, the improvement in detectability from SAFIRE (compared to FBP) was higher when measured in a uniform phantom compared to textured phantoms.

The final trajectory of this project aimed at developing methods to mathematically model lesions, as a means to help assess image quality directly from patient images. The mathematical modeling framework is first presented. The models describe a lesion’s morphology in terms of size, shape, contrast, and edge profile as an analytical equation. The models can be voxelized and inserted into patient images to create so-called “hybrid” images. These hybrid images can then be used to assess detectability or estimability with the advantage that the ground truth of the lesion morphology and location is known exactly. Based on this framework, a series of liver lesions, lung nodules, and kidney stones were modeled based on images of real lesions. The lesion models were virtually inserted into patient images to create a database of hybrid images to go along with the original database of real lesion images. ROI images from each database were assessed by radiologists in a blinded fashion to determine the realism of the hybrid images. It was found that the radiologists could not readily distinguish between real and virtual lesion images (area under the ROC curve was 0.55). This study provided evidence that the proposed mathematical lesion modeling framework could produce reasonably realistic lesion images.

Based on that result, two studies were conducted which demonstrated the utility of the lesion models. The first study used the modeling framework as a measurement tool to determine how dose and reconstruction algorithm affected the quantitative analysis of liver lesions, lung nodules, and renal stones in terms of their size, shape, attenuation, edge profile, and texture features. The same database of real lesion images used in the previous study was used for this study. That database contained images of the same patient at 2 dose levels (50% and 100%) along with 3 reconstruction algorithms from a GE 750HD CT system (GE Healthcare). The algorithms in question were FBP, Adaptive Statistical Iterative Reconstruction (ASiR), and Model-Based Iterative Reconstruction (MBIR). A total of 23 quantitative features were extracted from the lesions under each condition. It was found that both dose and reconstruction algorithm had a statistically significant effect on the feature measurements. In particular, radiation dose affected five, three, and four of the 23 features (related to lesion size, conspicuity, and pixel-value distribution) for liver lesions, lung nodules, and renal stones, respectively. MBIR significantly affected 9, 11, and 15 of the 23 features (including size, attenuation, and texture features) for liver lesions, lung nodules, and renal stones, respectively. Lesion texture was not significantly affected by radiation dose.

The second study demonstrating the utility of the lesion modeling framework focused on assessing detectability of very low-contrast liver lesions in abdominal imaging. Specifically, detectability was assessed as a function of dose and reconstruction algorithm. As part of a parallel clinical trial, images from 21 patients were collected at 6 dose levels per patient on a SOMATOM Flash scanner. Subtle liver lesion models (contrast = -15 HU) were inserted into the raw projection data from the patient scans. The projections were then reconstructed with FBP and SAFIRE (strength 5). Also, lesion-less images were reconstructed. Noise, contrast, CNR, and detectability index of an observer model (non-prewhitening matched filter) were assessed. It was found that SAFIRE reduced noise by 52%, reduced contrast by 12%, increased CNR by 87%. and increased detectability index by 65% compared to FBP. Further, a 2AFC human perception experiment was performed to assess the dose reduction potential of SAFIRE, which was found to be 22% compared to the standard of care dose.

In conclusion, this dissertation provides to the scientific community a series of new methodologies, phantoms, analysis techniques, and modeling tools that can be used to rigorously assess image quality from modern CT systems. Specifically, methods to properly evaluate iterative reconstruction have been developed and are expected to aid in the safe clinical implementation of dose reduction technologies.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Background: Over the past decade, annual heath exams have been de-emphasized for the general population but emphasized for adults with intellectual and developmental disabilities (IDD). The purpose of this project was to determine if there has been an increase in the uptake of the health exam among adults with IDD in Ontario, to what extent, and the effect on the quality of preventive care provided. Methods: Using administrative health data, the proportion of adults (18-64 years old) with IDD who received a health exam (long appointment, general assessment, and “true” health exam), a high value on the primary care quality composite score (PCQS), and a health exam or high PCQS each year was compared to the proportion in a propensity score matched sample of the general population. Negative binomial and segmented negative binomial regression controlling for age and sex were used to determine the relative risk of having a health exam/high PCQS/health exam or PCQS over time. Results: Pre joinpoint, the long appointment and general assessment health exam definitions saw a decrease and the “true” health exam saw an increase in the likelihood of adults having a health exam. Post joinpoint, all health exam definitions saw a decrease in the likelihood of adults having a health exam. Pre joinpoint, all PCQS measures (high PCQS, long appointment or high PCQS, “true” health exam or high PCQS) saw an increase in the likelihood for adults to achieve a high PCQS or high PCQS/have a health exam. Post joinpoint, all PCQS measures saw a decrease in the likelihood for adults to achieve a high PCQS or high PCQS/have a health exam. Achieving a high PCQS was strongly associated with having a health exam regardless of health exam definition or IDD status. Conclusions: Despite the publication of guidelines, only a small proportion of adults with IDD are receiving health exams. This indicates that the publication of guidelines alone was not sufficient to change practice. More targeted measures, such as the implementation of an IDD-specific health exam fee code, should be considered to increase the uptake of the health exam among adults with IDD.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Abstract. The performance objectives used for the formative assessment of com- plex skills are generally set through text-based analytic rubrics[1]. Moreover, video modeling examples are a widely applied method of observational learning, providing students with context-rich modeling examples of complex skills that act as an analogy for problem solving [1]. The purpose of this theoretical paper is to synthesize the components of video modeling and rubrics to support the formative assessment of complex skills. Based on theory, we argue that application of the developed Video Enhanced Rubrics (VER) fosters learners’ development of mental models, quality of provided feedback by various actors and finally, the learners mastery of complex skills.