14 resultados para Link managers
em Universidad Politécnica de Madrid
Resumo:
Cross‐lingual link discovery in the Web of Data
Resumo:
From the water management perspective, water scarcity is an unacceptable risk of facing water shortages to serve water demands in the near future. Water scarcity may be temporary and related to drought conditions or other accidental situation, or may be permanent and due to deeper causes such as excessive demand growth, lack of infrastructure for water storage or transport, or constraints in water management. Diagnosing the causes of water scarcity in complex water resources systems is a precondition to adopt effective drought risk management actions. In this paper we present four indices which have been developed to evaluate water scarcity. We propose a methodology for interpretation of index values that can lead to conclusions about the reliability and vulnerability of systems to water scarcity, as well as to diagnose their possible causes and to propose solutions. The described methodology was applied to the Ebro river basin, identifying existing and expected problems and possible solutions. System diagnostics, based exclusively on the analysis of index values, were compared with the known reality as perceived by system managers, validating the conclusions in all cases
Resumo:
De distintas formas y en diferentes momentos he intervenido activamente en cursos de informática para ejecutivos. Mi impresión es que tenían el mismo efecto sobre el auditorio que el que tiene representar una sesión de juegos malabares ante una asamblea de invidentes.
Resumo:
Although the aim of empirical software engineering is to provide evidence for selecting the appropriate technology, it appears that there is a lack of recognition of this work in industry. Results from empirical research only rarely seem to find their way to company decision makers. If information relevant for software managers is provided in reports on experiments, such reports can be considered as a source of information for them when they are faced with making decisions about the selection of software engineering technologies. To bridge this communication gap between researchers and professionals, we propose characterizing the information needs of software managers in order to show empirical software engineering researchers which information is relevant for decision-making and thus enable them to make this information available. We empirically investigated decision makers? information needs to identify which information they need to judge the appropriateness and impact of a software technology. We empirically developed a model that characterizes these needs. To ensure that researchers provide relevant information when reporting results from experiments, we extended existing reporting guidelines accordingly.We performed an experiment to evaluate our model with regard to its effectiveness. Software managers who read an experiment report according to the proposed model judged the technology?s appropriateness significantly better than those reading a report about the same experiment that did not explicitly address their information needs. Our research shows that information regarding a technology, the context in which it is supposed to work, and most importantly, the impact of this technology on development costs and schedule as well as on product quality is crucial for decision makers.
Resumo:
Auxin is associated with the regulation of virtually every aspect of plant growth and development. Many previous genetic and biochemical studies revealed that, among the proposed routes for the production of auxin, the so-called indole-3-pyruvic acid (IPA) pathway is the main source for indole-3-acetic acid (IAA) in plants. The IPA pathway involves the action of 2 classes of enzymes, tryptophan-pyruvate aminotransferases (TRYPTOPHAN AMINOTRANSFERASE OF ARABIDOPSIS 1(TAA1)/TRYPTOPHAN AMINOTRANSFERASE RELATED (TAR)) and flavin monooxygenases (YUCCA). Both enzyme classes appear to be encoded by small gene families in Arabidopsis consisting of 5 and 11 members, respectively. We recently showed that it is possible to induce transcript accumulation of 2 YUCCA genes, YUC8 and YUC9, by methyl jasmonate treatment. Both gene products were demonstrated to contribute to auxin biosynthesis in planta.1 Here we report that the overexpression of YUC8 as well as YUC9 led to strong lignification of plant aerial tissues. Furthermore, new evidence indicates that this abnormally strong secondary growth is linked to increased levels of ethylene production.
Resumo:
Sustainability managementand Sustainability Reporting (SR) practices have dramatically increased during the last two decades, raising important questions about the relationship between internal practices and external communication. Previous literature on SR has almost exclusively highlighted the role of institutional and stakeholder pressures in driving its adoption. However, as surveys among reporters also identify internal benefits of SR, its full role for company-level sustainability management remains unclear. In order to address this question, we develop a framework accounting for four SR configurations, stemming from different levels of relative importance of external and internal motives for SR. A multiple case study involving four large Spanish companies serves to illustrate the framework and to identify company-level factors that act both as enablers and barriers of SR internal relevance. We conclude that motivations for SR, along with such internal factors, decisively influence its contribution to sustainability management.
Resumo:
In Chile, during the last three decades there has been a strong decentralization process whose main objective has been to improve the management of schools through the transfer of responsibilities and resources of education and thus improve the outcome of learning, reducing equity gaps between schools and territories. This is how, there has been an evolution of school principals¿ professional profile from an administrative to a management approach, in which principals have become project managers of educational projects. From a competence model for school leaders, based on IPMA guidelines, the present article presents an analysis of best practices for school management, allowing to generate a link between competencies and school management, from the perspective of project management. Results showed that the different competence elements, have relative weights according to the different practice fields, and that this analysis can be considered as a strategic element in educational project planning and development.
Resumo:
Although the aim of empirical software engineering is to provide evidence for selecting the appropriate technology, it appears that there is a lack of recognition of this work in industry. Results from empirical research only rarely seem to find their way to company decision makers. If information relevant for software managers is provided in reports on experiments, such reports can be considered as a source of information for them when they are faced with making decisions about the selection of software engineering technologies. To bridge this communication gap between researchers and professionals, we propose characterizing the information needs of software managers in order to show empirical software engineering researchers which information is relevant for decision-making and thus enable them to make this information available. We empirically investigated decision makers? information needs to identify which information they need to judge the appropriateness and impact of a software technology. We empirically developed a model that characterizes these needs. To ensure that researchers provide relevant information when reporting results from experiments, we extended existing reporting guidelines accordingly. We performed an experiment to evaluate our model with regard to its effectiveness. Software managers who read an experiment report according to the proposed model judged the technology?s appropriateness significantly better than those reading a report about the same experiment that did not explicitly address their information needs. Our research shows that information regarding a technology, the context in which it is supposed to work, and most importantly, the impact of this technology on development costs and schedule as well as on product quality is crucial for decision makers.
Resumo:
The objective of this paper is to analyse the factors influencing tourists? choice of a destination and the role of High Speed Rail (HSR) systems in this choice. The methodology proposed consists in analysing two capitals in Europe, i.e. Paris and Madrid where HSR services are important, to investigate the factors influencing holidaymakers in choosing these cities, and the role of HSR in this choice. The main outcome of this paper is to show that several factors influence the choice of a tourist, like the presence of architectural sites, the quality of promotion of the destination itself, the presence of events, and also HSR services. However we found that the HSR system has affected the choice of Paris and Madrid in a different way. Concerning the French case study, HSR is considered a real transport mode alternative among tourists, therefore HSR is chosen to reach Paris as well as for revisiting it. On the other hand, Madrid is chosen by tourists irrespective on the presence of HSR, while HSR is chosen for reaching cities close to Madrid. Data collected from the two surveys have been used for a further quantitative analysis. Models have been specified and calibrated to identify the factors influencing holidaymakers to revisit Paris and Madrid and the role of HSR in this choice has been highlighted.
Resumo:
In Chile, during the last three decades there has been a strong decentralization process whose main objective has been to improve the management of schools through the transfer of responsibilities and resources of education and thus improve the outcome of learning, reducing equity gaps between schools and territories. This is how, there has been an evolution of school principals¿ professional profile from an administrative to a management approach, in which principals have become project managers of educational projects. From a competence model for school leaders, based on IPMA guidelines, the present article presents an analysis of best practices for school management, allowing to generate a link between competencies and school management, from the perspective of project management. Results showed that the different competence elements, have relative weights according to the different practice fields, and that this analysis can be considered as a strategic element in educational project planning and development.
Resumo:
Corporate Social Responsibility (CSR) strategies have a strong link with human resources policies. Not only because employees are one of the main stakeholders and because leaders’ style is directly related to the deployment of the strategy, but also, and with a growing importance, because a company culture aligned with CSR values could be a key competitive factor. The relationships among CSR values, employees’ commitment and productivity is one of the research lines of the GIOS (Grupo de Investigación de Organizaciones Sostenibles, Sustainable Organizations Research Group). Employees’ commitment management is one of the main challenges managers face, particularly in companies with a high proportion of knowledge workers. Many pieces of research indicate the direct relationship between employees’ commitment and company success. In this paper the results of a case study in REE (Red Eléctrica de España) identify some key variables to demonstrate that relationship. Based on commitment construct with the duality of emotional and rational commitment, and on the REE employee satisfaction survey, a direct relationship with organizational citizenship behaviour (OCB) variables appears. These OCB variables are an intermediate step with CSR values.From the results analysis of this survey a direct linear relationship can be seen between commitment and organizational citizenship behaviours. The relationships among emotional and rational commitment and OCB are examined separately with the conclusion being reached that there is a strong correlation in both cases. Moreover, the correlation between emotional commitment and OCB is somewhat stronger than that existing between rational commitment and OCB. it can also be seen how emotional commitment increases more strongly than rational commitment as organizational citizenship behaviours are gradually incorporated.
Resumo:
Las Field-Programmable Gate Arrays (FPGAs) SRAM se construyen sobre una memoria de configuración de tecnología RAM Estática (SRAM). Presentan múltiples características que las hacen muy interesantes para diseñar sistemas empotrados complejos. En primer lugar presentan un coste no-recurrente de ingeniería (NRE) bajo, ya que los elementos lógicos y de enrutado están pre-implementados (el diseño de usuario define su conexionado). También, a diferencia de otras tecnologías de FPGA, pueden ser reconfiguradas (incluso en campo) un número ilimitado de veces. Es más, las FPGAs SRAM de Xilinx soportan Reconfiguración Parcial Dinámica (DPR), la cual permite reconfigurar la FPGA sin interrumpir la aplicación. Finalmente, presentan una alta densidad de lógica, una alta capacidad de procesamiento y un rico juego de macro-bloques. Sin embargo, un inconveniente de esta tecnología es su susceptibilidad a la radiación ionizante, la cual aumenta con el grado de integración (geometrías más pequeñas, menores tensiones y mayores frecuencias). Esta es una precupación de primer nivel para aplicaciones en entornos altamente radiativos y con requisitos de alta confiabilidad. Este fenómeno conlleva una degradación a largo plazo y también puede inducir fallos instantáneos, los cuales pueden ser reversibles o producir daños irreversibles. En las FPGAs SRAM, los fallos inducidos por radiación pueden aparecer en en dos capas de arquitectura diferentes, que están físicamente superpuestas en el dado de silicio. La Capa de Aplicación (o A-Layer) contiene el hardware definido por el usuario, y la Capa de Configuración contiene la memoria de configuración y la circuitería de soporte. Los fallos en cualquiera de estas capas pueden hacer fracasar el sistema, lo cual puede ser ás o menos tolerable dependiendo de los requisitos de confiabilidad del sistema. En el caso general, estos fallos deben gestionados de alguna manera. Esta tesis trata sobre la gestión de fallos en FPGAs SRAM a nivel de sistema, en el contexto de sistemas empotrados autónomos y confiables operando en un entorno radiativo. La tesis se centra principalmente en aplicaciones espaciales, pero los mismos principios pueden aplicarse a aplicaciones terrenas. Las principales diferencias entre ambas son el nivel de radiación y la posibilidad de mantenimiento. Las diferentes técnicas para la gestión de fallos en A-Layer y C-Layer son clasificados, y sus implicaciones en la confiabilidad del sistema son analizados. Se proponen varias arquitecturas tanto para Gestores de Fallos de una capa como de doble-capa. Para estos últimos se propone una arquitectura novedosa, flexible y versátil. Gestiona las dos capas concurrentemente de manera coordinada, y permite equilibrar el nivel de redundancia y la confiabilidad. Con el objeto de validar técnicas de gestión de fallos dinámicas, se desarrollan dos diferentes soluciones. La primera es un entorno de simulación para Gestores de Fallos de C-Layer, basado en SystemC como lenguaje de modelado y como simulador basado en eventos. Este entorno y su metodología asociada permite explorar el espacio de diseño del Gestor de Fallos, desacoplando su diseño del desarrollo de la FPGA objetivo. El entorno incluye modelos tanto para la C-Layer de la FPGA como para el Gestor de Fallos, los cuales pueden interactuar a diferentes niveles de abstracción (a nivel de configuration frames y a nivel físico JTAG o SelectMAP). El entorno es configurable, escalable y versátil, e incluye capacidades de inyección de fallos. Los resultados de simulación para algunos escenarios son presentados y comentados. La segunda es una plataforma de validación para Gestores de Fallos de FPGAs Xilinx Virtex. La plataforma hardware aloja tres Módulos de FPGA Xilinx Virtex-4 FX12 y dos Módulos de Unidad de Microcontrolador (MCUs) de 32-bits de propósito general. Los Módulos MCU permiten prototipar Gestores de Fallos de C-Layer y A-Layer basados en software. Cada Módulo FPGA implementa un enlace de A-Layer Ethernet (a través de un switch Ethernet) con uno de los Módulos MCU, y un enlace de C-Layer JTAG con el otro. Además, ambos Módulos MCU intercambian comandos y datos a través de un enlace interno tipo UART. Al igual que para el entorno de simulación, se incluyen capacidades de inyección de fallos. Los resultados de pruebas para algunos escenarios son también presentados y comentados. En resumen, esta tesis cubre el proceso completo desde la descripción de los fallos FPGAs SRAM inducidos por radiación, pasando por la identificación y clasificación de técnicas de gestión de fallos, y por la propuesta de arquitecturas de Gestores de Fallos, para finalmente validarlas por simulación y pruebas. El trabajo futuro está relacionado sobre todo con la implementación de Gestores de Fallos de Sistema endurecidos para radiación. ABSTRACT SRAM-based Field-Programmable Gate Arrays (FPGAs) are built on Static RAM (SRAM) technology configuration memory. They present a number of features that make them very convenient for building complex embedded systems. First of all, they benefit from low Non-Recurrent Engineering (NRE) costs, as the logic and routing elements are pre-implemented (user design defines their connection). Also, as opposed to other FPGA technologies, they can be reconfigured (even in the field) an unlimited number of times. Moreover, Xilinx SRAM-based FPGAs feature Dynamic Partial Reconfiguration (DPR), which allows to partially reconfigure the FPGA without disrupting de application. Finally, they feature a high logic density, high processing capability and a rich set of hard macros. However, one limitation of this technology is its susceptibility to ionizing radiation, which increases with technology scaling (smaller geometries, lower voltages and higher frequencies). This is a first order concern for applications in harsh radiation environments and requiring high dependability. Ionizing radiation leads to long term degradation as well as instantaneous faults, which can in turn be reversible or produce irreversible damage. In SRAM-based FPGAs, radiation-induced faults can appear at two architectural layers, which are physically overlaid on the silicon die. The Application Layer (or A-Layer) contains the user-defined hardware, and the Configuration Layer (or C-Layer) contains the (volatile) configuration memory and its support circuitry. Faults at either layers can imply a system failure, which may be more ore less tolerated depending on the dependability requirements. In the general case, such faults must be managed in some way. This thesis is about managing SRAM-based FPGA faults at system level, in the context of autonomous and dependable embedded systems operating in a radiative environment. The focus is mainly on space applications, but the same principles can be applied to ground applications. The main differences between them are the radiation level and the possibility for maintenance. The different techniques for A-Layer and C-Layer fault management are classified and their implications in system dependability are assessed. Several architectures are proposed, both for single-layer and dual-layer Fault Managers. For the latter, a novel, flexible and versatile architecture is proposed. It manages both layers concurrently in a coordinated way, and allows balancing redundancy level and dependability. For the purpose of validating dynamic fault management techniques, two different solutions are developed. The first one is a simulation framework for C-Layer Fault Managers, based on SystemC as modeling language and event-driven simulator. This framework and its associated methodology allows exploring the Fault Manager design space, decoupling its design from the target FPGA development. The framework includes models for both the FPGA C-Layer and for the Fault Manager, which can interact at different abstraction levels (at configuration frame level and at JTAG or SelectMAP physical level). The framework is configurable, scalable and versatile, and includes fault injection capabilities. Simulation results for some scenarios are presented and discussed. The second one is a validation platform for Xilinx Virtex FPGA Fault Managers. The platform hosts three Xilinx Virtex-4 FX12 FPGA Modules and two general-purpose 32-bit Microcontroller Unit (MCU) Modules. The MCU Modules allow prototyping software-based CLayer and A-Layer Fault Managers. Each FPGA Module implements one A-Layer Ethernet link (through an Ethernet switch) with one of the MCU Modules, and one C-Layer JTAG link with the other. In addition, both MCU Modules exchange commands and data over an internal UART link. Similarly to the simulation framework, fault injection capabilities are implemented. Test results for some scenarios are also presented and discussed. In summary, this thesis covers the whole process from describing the problem of radiationinduced faults in SRAM-based FPGAs, then identifying and classifying fault management techniques, then proposing Fault Manager architectures and finally validating them by simulation and test. The proposed future work is mainly related to the implementation of radiation-hardened System Fault Managers.
Resumo:
En las últimas dos décadas, se ha puesto de relieve la importancia de los procesos de adquisición y difusión del conocimiento dentro de las empresas, y por consiguiente el estudio de estos procesos y la implementación de tecnologías que los faciliten ha sido un tema que ha despertado un creciente interés en la comunidad científica. Con el fin de facilitar y optimizar la adquisición y la difusión del conocimiento, las organizaciones jerárquicas han evolucionado hacia una configuración más plana, con estructuras en red que resulten más ágiles, disminuyendo la dependencia de una autoridad centralizada, y constituyendo organizaciones orientadas a trabajar en equipo. Al mismo tiempo, se ha producido un rápido desarrollo de las herramientas de colaboración Web 2.0, tales como blogs y wikis. Estas herramientas de colaboración se caracterizan por una importante componente social, y pueden alcanzar todo su potencial cuando se despliegan en las estructuras organizacionales planas. La Web 2.0 aparece como un concepto enfrentado al conjunto de tecnologías que existían a finales de los 90s basadas en sitios web, y se basa en la participación de los propios usuarios. Empresas del Fortune 500 –HP, IBM, Xerox, Cisco– las adoptan de inmediato, aunque no hay unanimidad sobre su utilidad real ni sobre cómo medirla. Esto se debe en parte a que no se entienden bien los factores que llevan a los empleados a adoptarlas, lo que ha llevado a fracasos en la implantación debido a la existencia de algunas barreras. Dada esta situación, y ante las ventajas teóricas que tienen estas herramientas de colaboración Web 2.0 para las empresas, los directivos de éstas y la comunidad científica muestran un interés creciente en conocer la respuesta a la pregunta: ¿cuáles son los factores que contribuyen a que los empleados de las empresas adopten estas herramientas Web 2.0 para colaborar? La respuesta a esta pregunta es compleja ya que se trata de herramientas relativamente nuevas en el contexto empresarial mediante las cuales se puede llevar a cabo la gestión del conocimiento en lugar del manejo de la información. El planteamiento que se ha llevado a cabo en este trabajo para dar respuesta a esta pregunta es la aplicación de los modelos de adopción tecnológica, que se basan en las percepciones de los individuos sobre diferentes aspectos relacionados con el uso de la tecnología. Bajo este enfoque, este trabajo tiene como objetivo principal el estudio de los factores que influyen en la adopción de blogs y wikis en empresas, mediante un modelo predictivo, teórico y unificado, de adopción tecnológica, con un planteamiento holístico a partir de la literatura de los modelos de adopción tecnológica y de las particularidades que presentan las herramientas bajo estudio y en el contexto especifico. Este modelo teórico permitirá determinar aquellos factores que predicen la intención de uso de las herramientas y el uso real de las mismas. El trabajo de investigación científica se estructura en cinco partes: introducción al tema de investigación, desarrollo del marco teórico, diseño del trabajo de investigación, análisis empírico, y elaboración de conclusiones. Desde el punto de vista de la estructura de la memoria de la tesis, las cinco partes mencionadas se desarrollan de forma secuencial a lo largo de siete capítulos, correspondiendo la primera parte al capítulo 1, la segunda a los capítulos 2 y 3, la tercera parte a los capítulos 4 y 5, la cuarta parte al capítulo 6, y la quinta y última parte al capítulo 7. El contenido del capítulo 1 se centra en el planteamiento del problema de investigación así como en los objetivos, principal y secundarios, que se pretenden cumplir a lo largo del trabajo. Así mismo, se expondrá el concepto de colaboración y su encaje con las herramientas colaborativas Web 2.0 que se plantean en la investigación y una introducción a los modelos de adopción tecnológica. A continuación se expone la justificación de la investigación, los objetivos de la misma y el plan de trabajo para su elaboración. Una vez introducido el tema de investigación, en el capítulo 2 se lleva a cabo una revisión de la evolución de los principales modelos de adopción tecnológica existentes (IDT, TRA, SCT, TPB, DTPB, C-TAM-TPB, UTAUT, UTAUT2), dando cuenta de sus fundamentos y factores empleados. Sobre la base de los modelos de adopción tecnológica expuestos en el capítulo 2, en el capítulo 3 se estudian los factores que se han expuesto en el capítulo 2 pero adaptados al contexto de las herramientas colaborativas Web 2.0. Con el fin de facilitar la comprensión del modelo final, los factores se agrupan en cuatro tipos: tecnológicos, de control, socio-normativos y otros específicos de las herramientas colaborativas. En el capítulo 4 se lleva a cabo la relación de los factores que son más apropiados para estudiar la adopción de las herramientas colaborativas y se define un modelo que especifica las relaciones entre los diferentes factores. Estas relaciones finalmente se convertirán en hipótesis de trabajo, y que habrá que contrastar mediante el estudio empírico. A lo largo del capítulo 5 se especifican las características del trabajo empírico que se lleva a cabo para contrastar las hipótesis que se habían enunciado en el capítulo 4. La naturaleza de la investigación es de carácter social, de tipo exploratorio, y se basa en un estudio empírico cuantitativo cuyo análisis se llevará a cabo mediante técnicas de análisis multivariante. En este capítulo se describe la construcción de las escalas del instrumento de medida, la metodología de recogida de datos, y posteriormente se presenta un análisis detallado de la población muestral, así como la comprobación de la existencia o no del sesgo atribuible al método de medida, lo que se denomina sesgo de método común (en inglés, Common Method Bias). El contenido del capítulo 6 corresponde al análisis de resultados, aunque previamente se expone la técnica estadística empleada, PLS-SEM, como herramienta de análisis multivariante con capacidad de análisis predictivo, así como la metodología empleada para validar el modelo de medida y el modelo estructural, los requisitos que debe cumplir la muestra, y los umbrales de los parámetros considerados. En la segunda parte del capítulo 6 se lleva a cabo el análisis empírico de los datos correspondientes a las dos muestras, una para blogs y otra para wikis, con el fin de validar las hipótesis de investigación planteadas en el capítulo 4. Finalmente, en el capítulo 7 se revisa el grado de cumplimiento de los objetivos planteados en el capítulo 1 y se presentan las contribuciones teóricas, metodológicas y prácticas derivadas del trabajo realizado. A continuación se exponen las conclusiones generales y detalladas por cada grupo de factores, así como las recomendaciones prácticas que se pueden extraer para orientar la implantación de estas herramientas en situaciones reales. Como parte final del capítulo se incluyen las limitaciones del estudio y se sugiere una serie de posibles líneas de trabajo futuras de interés, junto con los resultados de investigación parciales que se han obtenido durante el tiempo que ha durado la investigación. ABSTRACT In the last two decades, the relevance of knowledge acquisition and dissemination processes has been highlighted and consequently, the study of these processes and the implementation of the technologies that make them possible has generated growing interest in the scientific community. In order to ease and optimize knowledge acquisition and dissemination, hierarchical organizations have evolved to a more horizontal configuration with more agile net structures, decreasing the dependence of a centralized authority, and building team-working oriented organizations. At the same time, Web 2.0 collaboration tools such as blogs and wikis have quickly developed. These collaboration tools are characterized by a strong social component and can reach their full potential when they are deployed in horizontal organization structures. Web 2.0, based on user participation, arises as a concept to challenge the existing technologies of the 90’s which were based on websites. Fortune 500 companies – HP, IBM, Xerox, Cisco- adopted the concept immediately even though there was no unanimity about its real usefulness or how it could be measured. This is partly due to the fact that the factors that make the drivers for employees to adopt these tools are not properly understood, consequently leading to implementation failure due to the existence of certain barriers. Given this situation, and faced with theoretical advantages that these Web 2.0 collaboration tools seem to have for companies, managers and the scientific community are showing an increasing interest in answering the following question: Which factors contribute to the decision of the employees of a company to adopt the Web 2.0 tools for collaborative purposes? The answer is complex since these tools are relatively new in business environments. These tools allow us to move from an information Management approach to Knowledge Management. In order to answer this question, the chosen approach involves the application of technology adoption models, all of them based on the individual’s perception of the different aspects related to technology usage. From this perspective, this thesis’ main objective is to study the factors influencing the adoption of blogs and wikis in a company. This is done by using a unified and theoretical predictive model of technological adoption with a holistic approach that is based on literature of technological adoption models and the particularities that these tools presented under study and in a specific context. This theoretical model will allow us to determine the factors that predict the intended use of these tools and their real usage. The scientific research is structured in five parts: Introduction to the research subject, development of the theoretical framework, research work design, empirical analysis and drawing the final conclusions. This thesis develops the five aforementioned parts sequentially thorough seven chapters; part one (chapter one), part two (chapters two and three), part three (chapters four and five), parte four (chapters six) and finally part five (chapter seven). The first chapter is focused on the research problem statement and the objectives of the thesis, intended to be reached during the project. Likewise, the concept of collaboration and its link with the Web 2.0 collaborative tools is discussed as well as an introduction to the technology adoption models. Finally we explain the planning to carry out the research and get the proposed results. After introducing the research topic, the second chapter carries out a review of the evolution of the main existing technology adoption models (IDT, TRA, SCT, TPB, DTPB, C-TAM-TPB, UTAUT, UTAUT2), highlighting its foundations and factors used. Based on technology adoption models set out in chapter 2, the third chapter deals with the factors which have been discussed previously in chapter 2, but adapted to the context of Web 2.0 collaborative tools under study, blogs and wikis. In order to better understand the final model, the factors are grouped into four types: technological factors, control factors, social-normative factors and other specific factors related to the collaborative tools. The first part of chapter 4 covers the analysis of the factors which are more relevant to study the adoption of collaborative tools, and the second part proceeds with the theoretical model which specifies the relationship between the different factors taken into consideration. These relationships will become specific hypotheses that will be tested by the empirical study. Throughout chapter 5 we cover the characteristics of the empirical study used to test the research hypotheses which were set out in chapter 4. The nature of research is social, exploratory, and it is based on a quantitative empirical study whose analysis is carried out using multivariate analysis techniques. The second part of this chapter includes the description of the scales of the measuring instrument; the methodology for data gathering, the detailed analysis of the sample, and finally the existence of bias attributable to the measurement method, the "Bias Common Method" is checked. The first part of chapter 6 corresponds to the analysis of results. The statistical technique employed (PLS-SEM) is previously explained as a tool of multivariate analysis, capable of carrying out predictive analysis, and as the appropriate methodology used to validate the model in a two-stages analysis, the measurement model and the structural model. Futhermore, it is necessary to check the requirements to be met by the sample and the thresholds of the parameters taken into account. In the second part of chapter 6 an empirical analysis of the data is performed for the two samples, one for blogs and the other for wikis, in order to validate the research hypothesis proposed in chapter 4. Finally, in chapter 7 the fulfillment level of the objectives raised in chapter 1 is reviewed and the theoretical, methodological and practical conclusions derived from the results of the study are presented. Next, we cover the general conclusions, detailing for each group of factors including practical recommendations that can be drawn to guide implementation of these tools in real situations in companies. As a final part of the chapter the limitations of the study are included and a number of potential future researches suggested, along with research partial results which have been obtained thorough the research.
Resumo:
La presente tesis examina el uso potencial y actual de las técnicas de simulación visual aplicadas al campo de la gestión y la planificación del arbolado urbano. El estudio incluye las aplicaciones potenciales de las visualizaciones por ordenador, así como los beneficios que esto acarrearía. También se analizan las posibles barreras que surgirían de la implementación de esta nueva herramienta y se ofrece una lista de recomendaciones para superarlas. La investigación tiene un carácter exploratorio que utiliza una combinación de técnicas de investigación cuantitativas y cualitativas, dónde se emplean cuestionarios y entrevistas personales semi-estructuradas para estudiar y analizar las opiniones y reacciones de los gestores de arbolado urbano de los distritos de la ciudad de Londres (Reino Unido), denominados Tree Officers (LTOs). Para el desarrollo de la tesis se recopilaron y analizaron las respuestas al cuestionario del 41 por ciento de los LTOs pertenecientes al 88 por ciento de los distritos de Londres y se realizaron un total de 17 entrevistas personales. Los resultados del análisis estadístico de las respuestas del cuestionario y los análisis cruzados de las distintas variables se complementaron con las conclusiones obtenidas del análisis temático de los datos cualitativos recopilados durante las entrevistas. Los usos potenciales de las técnicas de simulación visual aplicadas a la gestión y planificación del arbolado urbano sugeridos fueron obtenidos combinado las conclusiones de, primero, la comparación de las cuestiones que los LTOs consideraron que más tiempo y recursos necesitaban y que actualmente no era posible resolverlas satisfactoriamente con las herramientas y los procesos disponibles, con la información acerca de cómo se habían empleado las visualizaciones en situaciones similares en otros campos tales como planificación urbano, el paisajismo o la gestión forestal. Segundo, se analizaron las reacciones y opiniones de los LTOs ante un conjunto de visualizaciones presentadas durante las entrevistas, desarrolladas ad hoc para mostrar un abanico representativo de ejemplos de utilización de las técnicas de simulación visual, que, a su vez se complementaron con los usos adicionales que los propios LTOs sugirieron tras ver las visualizaciones presentadas. Los resultados muestran que el uso actual de simulaciones visuales por parte de los LTOs es muy limitado pero si que reciben un gran número de visualizaciones de otros departamentos y como parte de la documentación presentada en las solicitudes de permisos para edificación o desarrollo urbanístico. Los resultados indican que las visualizaciones que son presentadas a los LTOs no son objetivas ni precisas por lo que se argumenta que esta situación es un factor importante que impide una toma de decisiones adecuada y una correcta transmisión de infracción al público y al resto de partes implicadas. Se sugiere la creación de un código que regule el uso de visualizaciones en el campo de la gestión y planificación del arbolado urbano. ABSTRACT This thesis examined the use of computer visualizations in urban forestry management and planning. Potential roles of visualizations were determined the benefits that its use would provide. Additionally, the possible barriers in the implementation of visualizations in urban forestry management and planning were also studied and recommendations on how to overcome them were provided The research conducted was an exploratory study using survey research methods and personal semi-structured interviews. The perspectives and reactions of London (UK) boroughs’ tree officers (LTOs) were analysed combining quantitative and qualitative research methods. The study surveyed 41 percent of all Tree Officers in London, obtaining responses from 88 percent of the boroughs and performed 17 personal interviews. Statistical analysis of the data and cross-variables analysis provided rich information that was then complemented with the conclusions from thematic analysis of the qualitative data from the interviews. Potential roles of visualizations were determined first by understanding the challenges that LTOs are facing today and comparing them with how visualizations have helped in similar situations in urban forestry and other related fields like landscape architecture, urban planning and forestry; second, the reactions of LTOs to a set of examples of proposed uses of visualizations were also complemented with the additional uses proposed by LTOs after seeing the visualizations. The visualizations were created ad hoc to show a variety of representative examples of the sue of visualization in urban forestry management and planning and were presented during the interviews to LTOS. Results show that the current production of visualizations is very reduced among tree officers but that they are frequent receptors of visualizations coming from other departments and as part of the documentation of planning applications. Findings show that the current visualizations that get to Tree Officers are biased and inaccurate and therefore it is argued the the current use of visualizations is a threat to legitimate informed decision making and public information. The development of a code for the use of visualizations in urban forestry management and planning is suggested.