915 resultados para Knowledge base maintenance
Resumo:
This talk will present an overview of the ongoing ERCIM project SMARTDOCS (SeMAntically-cReaTed DOCuments) which aims at automatically generating webpages from RDF data. It will particularly focus on the current issues and the investigated solutions in the different modules of the project, which are related to document planning, natural language generation and multimedia perspectives. The second part of the talk will be dedicated to the KODA annotation system, which is a knowledge-base-agnostic annotator designed to provide the RDF annotations required in the document generation process.
Resumo:
El presente trabajo tiene como finalidad analizar las características de generación y difusión del conocimiento en distintos distritos industriales o aquéllos dedicados a la industria de la confección. Se analizan dos distritos textiles localizados en las ciudades de Pergamino y Mar del Plata (Provincia de Buenos Aires, Argentina). Ellos poseen una importante masa crítica de talleres de confección ligados a empresas con marca y diseño propio, cuya localización dentro o fuera de los límites de la ciudad varía en ambos casos. La metodología de investigación es cuantitativa y de tipo descriptivo. Los resultados muestran que Pergamino ha construido una base más sólida de conocimiento que Mar del Plata, proveniente de diferentes orígenes y sendero evolutivo. Ello incide, como un círculo virtuoso, sobre el conocimiento arquitectural de las empresas y sobre sus estrategias empresariales. Sin embargo, si bien en Pergamino existe un flujo constante de saberes tecnológicos externos que permiten crear una base más elevada de conocimiento de las firmas, que se difunde en el distrito, no se vislumbra el surgimiento de nuevas empresas locales con marca y diseño. De esta manera persiste un tipo de organización donde las firmas núcleo,en su mayoría externas al territorio, son las que traccionan el dinamismo de la trama. The aim of this paper is to analyze the patterns of knowledge creation and diffusion in two textile industrial districts located in the Argentinean cities of Mar del Plata and Pergamino. Both of them have an important number of supplier manufacturing firms linked to fashion- oriented clothing subcontractors, which are located inside or outside the geographical boundaries of each city. The methodology research is quantitative and descriptive. The results show that Pergamino has developed over time a stronger knowledge base than Mar del Plata, which could be related to different historical circumstances and evolutionary paths. This has an importante effect on firms’ architectural knowledge and strategies. Even though in Pergamino there is a permanent flow of external technological know how, which increase its knowledge base and let the firms belonging to the cluster to share it, there is not new fashion-oriented clothing local firms, prevailing a type of industrial district where the dynamism of the economy depends on the hub firms located outside the local context.
Resumo:
El interés de este Estudio de Caso es investigar la manera en que la Misión de Paz de la ONU en Sierra Leona (UNAMSIL) redireccionó el programa Desarme, Desmovilización y Reintegración (DDR) hacia los niños soldados durante el post-conflicto en el país. Se analiza cómo a través de la coordinación de diferentes Agencias y Organizaciones Internacionales con UNAMSIL, el Gobierno y la Sociedad Civil hace posible que se refuerce el rol de los Interim Care Centers (ICCs) en donde se enfrentaron las necesidades esenciales de los menores excombatientes para lograr un mejor proceso de su desmovilización y reintegración. Se explica de igual forma, las diferentes herramientas que sirven para llevar a cabo la resolución del conflicto y la reconstrucción de la paz, enfocándose en el peacekeeping, peacemaking y peacebuilding, como mecanismos que ayudaron a crear un espacio seguro para los niños exsoldados. Por último, se exponen los alcances y límites de los ICCs con respecto a la reintegración de los menores excombatientes
Resumo:
La present tesi pretén recollir l'experiència viscuda en desenvolupar un sistema supervisor intel·ligent per a la millora de la gestió de plantes depuradores d'aigües residuals., implementar-lo en planta real (EDAR Granollers) i avaluar-ne el funcionament dia a dia amb situacions típiques de la planta. Aquest sistema supervisor combina i integra eines de control clàssic de les plantes depuradores (controlador automàtic del nivell d'oxigen dissolt al reactor biològic, ús de models descriptius del procés...) amb l'aplicació d'eines del camp de la intel·ligència artificial (sistemes basats en el coneixement, concretament sistemes experts i sistemes basats en casos, i xarxes neuronals). Aquest document s'estructura en 9 capítols diferents. Hi ha una primera part introductòria on es fa una revisió de l'estat actual del control de les EDARs i s'explica el perquè de la complexitat de la gestió d'aquests processos (capítol 1). Aquest capítol introductori juntament amb el capítol 2, on es pretén explicar els antecedents d'aquesta tesi, serveixen per establir els objectius d'aquest treball (capítol 3). A continuació, el capítol 4 descriu les peculiaritats i especificitats de la planta que s'ha escollit per implementar el sistema supervisor. Els capítols 5 i 6 del present document exposen el treball fet per a desenvolupar el sistema basat en regles o sistema expert (capítol 6) i el sistema basat en casos (capítol 7). El capítol 8 descriu la integració d'aquestes dues eines de raonament en una arquitectura multi nivell distribuïda. Finalment, hi ha una darrer capítol que correspon a la avaluació (verificació i validació), en primer lloc, de cadascuna de les eines per separat i, posteriorment, del sistema global en front de situacions reals que es donin a la depuradora
Resumo:
Una de las actuaciones posibles para la gestión de los residuos sólidos urbanos es la valorización energética, es decir la incineración con recuperación de energía. Sin embargo es muy importante controlar adecuadamente el proceso de incineración para evitar en lo posible la liberación de sustancias contaminantes a la atmósfera que puedan ocasionar problemas de contaminación industrial.Conseguir que tanto el proceso de incineración como el tratamiento de los gases se realice en condiciones óptimas presupone tener un buen conocimiento de las dependencias entre las variables de proceso. Se precisan métodos adecuados de medida de las variables más importantes y tratar los valores medidos con modelos adecuados para transformarlos en magnitudes de mando. Un modelo clásico para el control parece poco prometedor en este caso debido a la complejidad de los procesos, la falta de descripción cuantitativa y la necesidad de hacer los cálculos en tiempo real. Esto sólo se puede conseguir con la ayuda de las modernas técnicas de proceso de datos y métodos informáticos, tales como el empleo de técnicas de simulación, modelos matemáticos, sistemas basados en el conocimiento e interfases inteligentes. En [Ono, 1989] se describe un sistema de control basado en la lógica difusa aplicado al campo de la incineración de residuos urbanos. En el centro de investigación FZK de Karslruhe se están desarrollando aplicaciones que combinan la lógica difusa con las redes neuronales [Jaeschke, Keller, 1994] para el control de la planta piloto de incineración de residuos TAMARA. En esta tesis se plantea la aplicación de un método de adquisición de conocimiento para el control de sistemas complejos inspirado en el comportamiento humano. Cuando nos encontramos ante una situación desconocida al principio no sabemos como actuar, salvo por la extrapolación de experiencias anteriores que puedan ser útiles. Aplicando procedimientos de prueba y error, refuerzo de hipótesis, etc., vamos adquiriendo y refinando el conocimiento, y elaborando un modelo mental. Podemos diseñar un método análogo, que pueda ser implementado en un sistema informático, mediante el empleo de técnicas de Inteligencia Artificial.Así, en un proceso complejo muchas veces disponemos de un conjunto de datos del proceso que a priori no nos dan información suficientemente estructurada para que nos sea útil. Para la adquisición de conocimiento pasamos por una serie de etapas: - Hacemos una primera selección de cuales son las variables que nos interesa conocer. - Estado del sistema. En primer lugar podemos empezar por aplicar técnicas de clasificación (aprendizaje no supervisado) para agrupar los datos y obtener una representación del estado de la planta. Es posible establecer una clasificación, pero normalmente casi todos los datos están en una sola clase, que corresponde a la operación normal. Hecho esto y para refinar el conocimiento utilizamos métodos estadísticos clásicos para buscar correlaciones entre variables (análisis de componentes principales) y así poder simplificar y reducir la lista de variables. - Análisis de las señales. Para analizar y clasificar las señales (por ejemplo la temperatura del horno) es posible utilizar métodos capaces de describir mejor el comportamiento no lineal del sistema, como las redes neuronales. Otro paso más consiste en establecer relaciones causales entre las variables. Para ello nos sirven de ayuda los modelos analíticos - Como resultado final del proceso se pasa al diseño del sistema basado en el conocimiento. El objetivo principal es aplicar el método al caso concreto del control de una planta de tratamiento de residuos sólidos urbanos por valorización energética. En primer lugar, en el capítulo 2 Los residuos sólidos urbanos, se trata el problema global de la gestión de los residuos, dando una visión general de las diferentes alternativas existentes, y de la situación nacional e internacional en la actualidad. Se analiza con mayor detalle la problemática de la incineración de los residuos, poniendo especial interés en aquellas características de los residuos que tienen mayor importancia de cara al proceso de combustión.En el capítulo 3, Descripción del proceso, se hace una descripción general del proceso de incineración y de los distintos elementos de una planta incineradora: desde la recepción y almacenamiento de los residuos, pasando por los distintos tipos de hornos y las exigencias de los códigos de buena práctica de combustión, el sistema de aire de combustión y el sistema de humos. Se presentan también los distintos sistemas de depuración de los gases de combustión, y finalmente el sistema de evacuación de cenizas y escorias.El capítulo 4, La planta de tratamiento de residuos sólidos urbanos de Girona, describe los principales sistemas de la planta incineradora de Girona: la alimentación de residuos, el tipo de horno, el sistema de recuperación de energía, y el sistema de depuración de los gases de combustión Se describe también el sistema de control, la operación, los datos de funcionamiento de la planta, la instrumentación y las variables que son de interés para el control del proceso de combustión.En el capítulo 5, Técnicas utilizadas, se proporciona una visión global de los sistemas basados en el conocimiento y de los sistemas expertos. Se explican las diferentes técnicas utilizadas: redes neuronales, sistemas de clasificación, modelos cualitativos, y sistemas expertos, ilustradas con algunos ejemplos de aplicación.Con respecto a los sistemas basados en el conocimiento se analizan en primer lugar las condiciones para su aplicabilidad, y las formas de representación del conocimiento. A continuación se describen las distintas formas de razonamiento: redes neuronales, sistemas expertos y lógica difusa, y se realiza una comparación entre ellas. Se presenta una aplicación de las redes neuronales al análisis de series temporales de temperatura.Se trata también la problemática del análisis de los datos de operación mediante técnicas estadísticas y el empleo de técnicas de clasificación. Otro apartado está dedicado a los distintos tipos de modelos, incluyendo una discusión de los modelos cualitativos.Se describe el sistema de diseño asistido por ordenador para el diseño de sistemas de supervisión CASSD que se utiliza en esta tesis, y las herramientas de análisis para obtener información cualitativa del comportamiento del proceso: Abstractores y ALCMEN. Se incluye un ejemplo de aplicación de estas técnicas para hallar las relaciones entre la temperatura y las acciones del operador. Finalmente se analizan las principales características de los sistemas expertos en general, y del sistema experto CEES 2.0 que también forma parte del sistema CASSD que se ha utilizado.El capítulo 6, Resultados, muestra los resultados obtenidos mediante la aplicación de las diferentes técnicas, redes neuronales, clasificación, el desarrollo de la modelización del proceso de combustión, y la generación de reglas. Dentro del apartado de análisis de datos se emplea una red neuronal para la clasificación de una señal de temperatura. También se describe la utilización del método LINNEO+ para la clasificación de los estados de operación de la planta.En el apartado dedicado a la modelización se desarrolla un modelo de combustión que sirve de base para analizar el comportamiento del horno en régimen estacionario y dinámico. Se define un parámetro, la superficie de llama, relacionado con la extensión del fuego en la parrilla. Mediante un modelo linealizado se analiza la respuesta dinámica del proceso de incineración. Luego se pasa a la definición de relaciones cualitativas entre las variables que se utilizan en la elaboración de un modelo cualitativo. A continuación se desarrolla un nuevo modelo cualitativo, tomando como base el modelo dinámico analítico.Finalmente se aborda el desarrollo de la base de conocimiento del sistema experto, mediante la generación de reglas En el capítulo 7, Sistema de control de una planta incineradora, se analizan los objetivos de un sistema de control de una planta incineradora, su diseño e implementación. Se describen los objetivos básicos del sistema de control de la combustión, su configuración y la implementación en Matlab/Simulink utilizando las distintas herramientas que se han desarrollado en el capítulo anterior.Por último para mostrar como pueden aplicarse los distintos métodos desarrollados en esta tesis se construye un sistema experto para mantener constante la temperatura del horno actuando sobre la alimentación de residuos.Finalmente en el capítulo Conclusiones, se presentan las conclusiones y resultados de esta tesis.
Resumo:
La comunitat científica que treballa en Intel·ligència Artificial (IA) ha dut a terme una gran quantitat de treball en com la IA pot ajudar a les persones a trobar el que volen dins d'Internet. La idea dels sistemes recomanadors ha estat extensament acceptada pels usuaris. La tasca principal d'un sistema recomanador és localitzar ítems, fonts d'informació i persones relacionades amb els interessos i preferències d'una persona o d'un grup de persones. Això comporta la construcció de models d'usuari i l'habilitat d'anticipar i predir les preferències de l'usuari. Aquesta tesi està focalitzada en l'estudi de tècniques d'IA que millorin el rendiment dels sistemes recomanadors. Inicialment, s'ha dut a terme un anàlisis detallat de l'actual estat de l'art en aquest camp. Aquest treball ha estat organitzat en forma de taxonomia on els sistemes recomanadors existents a Internet es classifiquen en 8 dimensions generals. Aquesta taxonomia ens aporta una base de coneixement indispensable pel disseny de la nostra proposta. El raonament basat en casos (CBR) és un paradigma per aprendre i raonar a partir de la experiència adequat per sistemes recomanadors degut als seus fonaments en el raonament humà. Aquesta tesi planteja una nova proposta de CBR aplicat al camp de la recomanació i un mecanisme d'oblit per perfils basats en casos que controla la rellevància i edat de les experiències passades. Els resultats experimentals demostren que aquesta proposta adapta millor els perfils als usuaris i soluciona el problema de la utilitat que pateixen el sistemes basats en CBR. Els sistemes recomanadors milloren espectacularment la qualitat dels resultats quan informació sobre els altres usuaris és utilitzada quan es recomana a un usuari concret. Aquesta tesi proposa l'agentificació dels sistemes recomanadors per tal de treure profit de propietats interessants dels agents com ara la proactivitat, la encapsulació o l'habilitat social. La col·laboració entre agents es realitza a partir del mètode de filtratge basat en la opinió i del mètode col·laboratiu de filtratge a partir de confiança. Els dos mètodes es basen en un model social de confiança que fa que els agents siguin menys vulnerables als altres quan col·laboren. Els resultats experimentals demostren que els agents recomanadors col·laboratius proposats milloren el rendiment del sistema mentre que preserven la privacitat de les dades personals de l'usuari. Finalment, aquesta tesi també proposa un procediment per avaluar sistemes recomanadors que permet la discussió científica dels resultats. Aquesta proposta simula el comportament dels usuaris al llarg del temps basat en perfils d'usuari reals. Esperem que aquesta metodologia d'avaluació contribueixi al progrés d'aquesta àrea de recerca.
Resumo:
This paper reviews four approaches used to create rational tools to aid the planning and the management of the building design process and then proposes a fifth approach. The new approach that has been developed is based on the mechanical aspects of technology rather than subjective design issues. The knowledge base contains, for each construction technology, a generic model of the detailed design process. Each activity in the process is specified by its input and output information needs. By connecting the input demands of one technology with the output supply from another technology a map or network of design activity is formed. Thus, it is possible to structure a specific model from the generic knowledge base within a KBE system.
Resumo:
There is a lack of knowledge base in relation to experiences gained and lessons learnt from previously executed National Health Service (NHS) infrastructure projects in the UK. This is in part a feature of one-off construction projects, which typify healthcare infrastructure, and in part due to the absence of a suitable method for conveying such information. The complexity of infrastructure delivery process in the NHS makes the construction of healthcare buildings a formidable task. This is particularly the case for the NHS trusts who have little or no experience of construction projects. To facilitate understanding a most important aspect of the delivery process, which is the preparation of a capital investment proposal; steps taken in developing the business case for an NHS healthcare facility are examined. The context for such examination is provided by the planning process of a healthcare project, studied retrospectively. The process is analysed using a social science based method called ‘building stories’, developed at the University of California-Berkeley. By applying this method, stories or narratives are constructed around the data captured on the case study. The findings indicate that the business case process may be used to justify, rather than identify, trusts’ requirements. The study is useful for UK public sector clients as well as consultants and professionals who aim to participate in the delivery of healthcare infrastructure projects in the UK.
Resumo:
Diverticular disease (DD) is an age-related disorder of the large bowel which may affect half of the population over the age of 65 in the UK. This high prevalence ranks it as one of the most common bowel disorders in western nations. The majority of patients remain asymptomatic but there are associated life-threatening co-morbidities, which, given the large numbers of people with DD, translates into a considerable number of deaths per annum. Despite this public health burden, relatively little seems to be known about either the mechanisms of development or causality. In the 1970s, a model of DD formulated the concept that diverticula occur as a consequence of pressureinduced damage to the colon wall amongst those with a low intake of dietary fiber. In this review, we have examined the evidence regarding the influence of ageing, diet, inflammation and genetics on DD development. We argue that the evidence supporting the barotrauma hypothesis is largely anecdotal. We have also identified several gaps in the knowledge base which need to be filled before we can complete
Resumo:
It is problematic to use standard ontology tools when describing vague domains. Standard ontologies are designed to formally define one view of a domain, and although it is possible to define disagreeing statements, it is not advisable, as the resulting inferences could be incorrect. Two different solutions to the above problem in two different vague domains have been developed and are presented. The first domain is the knowledge base of conversational agents (chatbots). An ontological scripting language has been designed to access ontology data from within chatbot code. The solution developed is based on reifications of user statements. It enables a new layer of logics based on the different views of the users, enabling the body of knowledge to grow automatically. The second domain is competencies and competency frameworks. An ontological framework has been developed to model different competencies using the emergent standards. It enables comparison of competencies using a mix of linguistic logics and descriptive logics. The comparison results are non-binary, therefore not simple yes and no answers, highlighting the vague nature of the comparisons. The solution has been developed with small ontologies which can be added to and modified in order for the competency user to build a total picture that fits the user’s purpose. Finally these two approaches are viewed in the light of how they could aid future work in vague domains, further work in both domains is described and also in other domains such as the semantic web. This demonstrates two different approaches to achieve inferences using standard ontology tools in vague domains.
Resumo:
As the building industry proceeds in the direction of low impact buildings, research attention is being drawn towards the reduction of carbon dioxide emission and waste. Starting from design and construction to operation and demolition, various building materials are used throughout the whole building lifecycle involving significant energy consumption and waste generation. Building Information Modelling (BIM) is emerging as a tool that can support holistic design-decision making for reducing embodied carbon and waste production in the building lifecycle. This study aims to establish a framework for assessing embodied carbon and waste underpinned by BIM technology. On the basis of current research review, the framework is considered to include functional modules for embodied carbon computation. There are a module for waste estimation, a knowledge-base of construction and demolition methods, a repository of building components information, and an inventory of construction materials’ energy and carbon. Through both static 3D model visualisation and dynamic modelling supported by the framework, embodied energy (carbon), waste and associated costs can be analysed in the boundary of cradle-to-gate, construction, operation, and demolition. The proposed holistic modelling framework provides a possibility to analyse embodied carbon and waste from different building lifecycle perspectives including associated costs. It brings together existing segmented embodied carbon and waste estimation into a unified model, so that interactions between various parameters through the different building lifecycle phases can be better understood. Thus, it can improve design-decision support for optimal low impact building development. The applicability of this framework is anticipated being developed and tested on industrial projects in the near future.
Resumo:
International Perspective The development of GM technology continues to expand into increasing numbers of crops and conferred traits. Inevitably, the focus remains on the major field crops of soybean, maize, cotton, oilseed rape and potato with introduced genes conferring herbicide tolerance and/or pest resistance. Although there are comparatively few GM crops that have been commercialised to date, GM versions of 172 plant species have been grown in field trials in 31 countries. European Crops with Containment Issues Of the 20 main crops in the EU there are four for which GM varieties are commercially available (cotton, maize for animal feed and forage, and oilseed rape). Fourteen have GM varieties in field trials (bread wheat, barley, durum wheat, sunflower, oats, potatoes, sugar beet, grapes, alfalfa, olives, field peas, clover, apples, rice) and two have GM varieties still in development (rye, triticale). Many of these crops have hybridisation potential with wild and weedy relatives in the European flora (bread wheat, barley, oilseed rape, durum wheat, oats, sugar beet and grapes), with escapes (sunflower); and all have potential to cross-pollinate fields non-GM crops. Several fodder crops, forestry trees, grasses and ornamentals have varieties in field trials and these too may hybridise with wild relatives in the European flora (alfalfa, clover, lupin, silver birch, sweet chestnut, Norway spruce, Scots pine, poplar, elm, Agrostis canina, A. stolonifera, Festuca arundinacea, Lolium perenne, L. multiflorum, statice and rose). All these crops will require containment strategies to be in place if it is deemed necessary to prevent transgene movement to wild relatives and non-GM crops. Current Containment Strategies A wide variety of GM containment strategies are currently under development, with a particular focus on crops expressing pharmaceutical products. Physical containment in greenhouses and growth rooms is suitable for some crops (tomatoes, lettuce) and for research purposes. Aquatic bioreactors of some non-crop species (algae, moss, and duckweed) expressing pharmaceutical products have been adopted by some biotechnology companies. There are obvious limitations of the scale of physical containment strategies, addressed in part by the development of large underground facilities in the US and Canada. The additional resources required to grow plants underground incurs high costs that in the long term may negate any advantage of GM for commercial productioNatural genetic containment has been adopted by some companies through the selection of either non-food/feed crops (algae, moss, duckweed) as bio-pharming platforms or organisms with no wild relatives present in the local flora (safflower in the Americas). The expression of pharmaceutical products in leafy crops (tobacco, alfalfa, lettuce, spinach) enables growth and harvesting prior to and in the absence of flowering. Transgenically controlled containment strategies range in their approach and degree of development. Plastid transformation is relatively well developed but is not suited to all traits or crops and does not offer complete containment. Male sterility is well developed across a range of plants but has limitations in its application for fruit/seed bearing crops. It has been adopted in some commercial lines of oilseed rape despite not preventing escape via seed. Conditional lethality can be used to prevent flowering or seed development following the application of a chemical inducer, but requires 100% induction of the trait and sufficient application of the inducer to all plants. Equally, inducible expression of the GM trait requires equally stringent application conditions. Such a method will contain the trait but will allow the escape of a non-functioning transgene. Seed lethality (‘terminator’ technology) is the only strategy at present that prevents transgene movement via seed, but due to public opinion against the concept it has never been trialled in the field and is no longer under commercial development. Methods to control flowering and fruit development such as apomixis and cleistogamy will prevent crop-to-wild and wild-to-crop pollination, but in nature both of these strategies are complex and leaky. None of the genes controlling these traits have as yet been identified or characterised and therefore have not been transgenically introduced into crop species. Neither of these strategies will prevent transgene escape via seed and any feral apomicts that form are arguably more likely to become invasives. Transgene mitigation reduces the fitness of initial hybrids and so prevents stable introgression of transgenes into wild populations. However, it does not prevent initial formation of hybrids or spread to non-GM crops. Such strategies could be detrimental to wild populations and have not yet been demonstrated in the field. Similarly, auxotrophy prevents persistence of escapes and hybrids containing the transgene in an uncontrolled environment, but does not prevent transgene movement from the crop. Recoverable block of function, intein trans-splicing and transgene excision all use recombinases to modify the transgene in planta either to induce expression or to prevent it. All require optimal conditions and 100% accuracy to function and none have been tested under field conditions as yet. All will contain the GM trait but all will allow some non-native DNA to escape to wild populations or to non-GM crops. There are particular issues with GM trees and grasses as both are largely undomesticated, wind pollinated and perennial, thus providing many opportunities for hybridisation. Some species of both trees and grass are also capable of vegetative propagation without sexual reproduction. There are additional concerns regarding the weedy nature of many grass species and the long-term stability of GM traits across the life span of trees. Transgene stability and conferred sterility are difficult to trial in trees as most field trials are only conducted during the juvenile phase of tree growth. Bio-pharming of pharmaceutical and industrial compounds in plants Bio-pharming of pharmaceutical and industrial compounds in plants offers an attractive alternative to mammalian-based pharmaceutical and vaccine production. Several plantbased products are already on the market (Prodigene’s avidin, β-glucuronidase, trypsin generated in GM maize; Ventria’s lactoferrin generated in GM rice). Numerous products are in clinical trials (collagen, antibodies against tooth decay and non-Hodgkin’s lymphoma from tobacco; human gastric lipase, therapeutic enzymes, dietary supplements from maize; Hepatitis B and Norwalk virus vaccines from potato; rabies vaccines from spinach; dietary supplements from Arabidopsis). The initial production platforms for plant-based pharmaceuticals were selected from conventional crops, largely because an established knowledge base already existed. Tobacco and other leafy crops such as alfalfa, lettuce and spinach are widely used as leaves can be harvested and no flowering is required. Many of these crops can be grown in contained greenhouses. Potato is also widely used and can also be grown in contained conditions. The introduction of morphological markers may aid in the recognition and traceability of crops expressing pharmaceutical products. Plant cells or plant parts may be transformed and maintained in culture to produce recombinant products in a contained environment. Plant cells in suspension or in vitro, roots, root cells and guttation fluid from leaves may be engineered to secrete proteins that may be harvested in a continuous, non-destructive manner. Most strategies in this category remain developmental and have not been commercially adopted at present. Transient expression produces GM compounds from non-GM plants via the utilisation of bacterial or viral vectors. These vectors introduce the trait into specific tissues of whole plants or plant parts, but do not insert them into the heritable genome. There are some limitations of scale and the field release of such crops will require the regulation of the vector. However, several companies have several transiently expressed products in clinical and pre-clinical trials from crops raised in physical containment.
Resumo:
There is currently an increased interest of Government and Industry in the UK, as well as at the European Community level and International Agencies (i.e. Department of Energy, American International Energy Agency), to improve the performance and uptake of Ground Coupled Heat Pumps (GCHP), in order to meet the 2020 renewable energy target. A sound knowledge base is required to help inform the Government Agencies and advisory bodies; detailed site studies providing reliable data for model verification have an important role to play in this. In this study we summarise the effect of heat extraction by a horizontal ground heat exchanger (installed at 1 m depth) on the soil physical environment (between 0 and 1 m depth) for a site in the south of the UK. Our results show that the slinky influences the surrounding soil by significantly decreasing soil temperatures. Furthermore, soil moisture contents were lower for the GCHP soil profile, most likely due to temperature-gradient related soil moisture migration effects and a decreased hydraulic conductivity, the latter as a result of increased viscosity (caused by the lower temperatures for the GCHP soil profile). The effects also caused considerable differences in soil thermal properties. This is the first detailed mechanistic study conducted in the UK with the aim to understand the interactions between the soil, horizontal heat exchangers and the aboveground environment. An increased understanding of these interactions will help to achieve an optimum and sustainable use of the soil heat resources in the future. The results of this study will help to calibrate and verify a simulation model that will provide UK-wide recommendations to improve future GCHP uptake and performance, while safeguarding the soil physical resources.
Resumo:
The chapter starts from the premise that an historically- and institutionally-formed orientation to music education at primary level in European countries privileges a nineteenth century Western European music aesthetic, with its focus on formal characteristics such as melody and rhythm. While there is a move towards a multi-faceted understanding of musical ability, a discrete intelligence and willingness to accept musical styles or 'open-earedness', there remains a paucity of documented evidence of this in research at primary school level. To date there has been no study undertaken which has the potential to provide policy makers and practitioners with insights into the degree of homogeneity or universality in conceptions of musical ability within this educational sector. Against this background, a study was set up to explore the following research questions: 1. What conceptions of musical ability do primary teachers hold a) of themselves and; b) of their pupils? 2. To what extent are these conceptions informed by Western classical practices? A mixed methods approach was used which included survey questionnaire and semi-structured interview. Questionnaires have been sent to all classroom teachers in a random sample of primary schools in the South East of England. This was followed up with a series of semi-structured interviews with a sub-sample of respondents. The main ideas are concerned with the attitudes, beliefs and working theories held by teachers in contemporary primary school settings. By mapping the extent to which a knowledge base for teaching can be resistant to change in schools, we can problematise primary schools as sites for diversity and migration of cultural ideas. Alongside this, we can use the findings from the study undertaken in an English context as a starting point for further investigation into conceptions of music, musical ability and assessment held by practitioners in a variety of primary school contexts elsewhere in Europe; our emphasis here will be on the development of shared understanding in terms of policies and practices in music education. Within this broader framework, our study can have a significant impact internationally, with potential to inform future policy making, curriculum planning and practice.
Resumo:
The paper develops a more precise specification and understanding of the process of national-level knowledge accumulation and absorptive capabilities by applying the reasoning and evidence from the firm-level analysis pioneered by Cohen and Levinthal (1989, 1990). In doing so, we acknowledge that significant cross-border effects due to the role of both inward and outward FDI exist and that assimilation of foreign knowledge is not only confined to catching-up economies but is also carried out by countries at the frontier-sharing phase. We postulate a non-linear relationship between national absorptive capacity and the technological gap, due to the effects of the cumulative nature of the learning process and the increase in complexity of external knowledge as the country approaches the technological frontier. We argue that national absorptive capacity and the accumulation of knowledge stock are simultaneously determined. This implies that different phases of technological development require different strategies. During the catching-up phase, knowledge accumulation occurs predominately through the absorption of trade and/or inward FDI-related R&D spillovers. At the pre-frontier-sharing phase onwards, increases in the knowledge base occur largely through independent knowledge creation and actively accessing foreign-located technological spillovers, inter alia through outward FDI-related R&D, joint ventures and strategic alliances.