5 resultados para text analytic approaches
em AMS Tesi di Laurea - Alm@DL - Università di Bologna
Resumo:
Ontology design and population -core aspects of semantic technologies- re- cently have become fields of great interest due to the increasing need of domain-specific knowledge bases that can boost the use of Semantic Web. For building such knowledge resources, the state of the art tools for ontology design require a lot of human work. Producing meaningful schemas and populating them with domain-specific data is in fact a very difficult and time-consuming task. Even more if the task consists in modelling knowledge at a web scale. The primary aim of this work is to investigate a novel and flexible method- ology for automatically learning ontology from textual data, lightening the human workload required for conceptualizing domain-specific knowledge and populating an extracted schema with real data, speeding up the whole ontology production process. Here computational linguistics plays a fundamental role, from automati- cally identifying facts from natural language and extracting frame of relations among recognized entities, to producing linked data with which extending existing knowledge bases or creating new ones. In the state of the art, automatic ontology learning systems are mainly based on plain-pipelined linguistics classifiers performing tasks such as Named Entity recognition, Entity resolution, Taxonomy and Relation extraction [11]. These approaches present some weaknesses, specially in capturing struc- tures through which the meaning of complex concepts is expressed [24]. Humans, in fact, tend to organize knowledge in well-defined patterns, which include participant entities and meaningful relations linking entities with each other. In literature, these structures have been called Semantic Frames by Fill- 6 Introduction more [20], or more recently as Knowledge Patterns [23]. Some NLP studies has recently shown the possibility of performing more accurate deep parsing with the ability of logically understanding the structure of discourse [7]. In this work, some of these technologies have been investigated and em- ployed to produce accurate ontology schemas. The long-term goal is to collect large amounts of semantically structured information from the web of crowds, through an automated process, in order to identify and investigate the cognitive patterns used by human to organize their knowledge.
Resumo:
In this thesis, we explore three methods for the geometrico-static modelling of continuum parallel robots. Inspired by biological trunks, tentacles and snakes, continuum robot designs can reach confined spaces, manipulate objects in complex environments and conform to curvilinear paths in space. In addition, parallel continuum manipulators have the potential to inherit some of the compactness and compliance of continuum robots while retaining some of the precision, stability and strength of rigid-links parallel robots. Subsequently, the foundation of our work is performed on slender beam by applying the Cosserat rod theory, appropriate to model continuum robots. After that, three different approaches are developed on a case study of a planar parallel continuum robot constituted of two connected flexible links. We solve the forward and inverse geometrico-static problem namely by using (a) shooting methods to obtain a numerical solution, (b) an elliptic method to find a quasi-analytical solution, and (c) the Corde model to perform further model analysis. The performances of each of the studied methods are evaluated and their limits are highlighted. This thesis is divided as follows. Chapter one gives the introduction on the field of the continuum robotics and introduce the parallel continuum robots that is studied in this work. Chapter two describe the geometrico-static problem and gives the mathematical description of this problem. Chapter three explains the numerical approach with the shooting method and chapter four introduce the quasi-analytical solution. Then, Chapter five introduce the analytic method inspired by the Corde model and chapter six gives the conclusions of this work.
Resumo:
Worldwide companies currently make a significant effort in performing the materiality analysis, whose aim is to explain corporate sustainability in an annual report. Materiality reflects what are the most important social, economic and environmental issues for a company and its stakeholders. Many studies and standards have been proposed to establish what are the main steps to follow to identify the specific topics to be included in a sustainability report. However, few existing quantitative and structured approaches help understanding how to deal with the identified topics and how to prioritise them to effectively show the most valuable ones. Moreover, the use of traditional approaches involves a long-lasting and complex procedure where a lot of people have to be reached and interviewed and several companies' reports have to be read to extrapolate the material topics to be discussed in the sustainability report. This dissertation aims to propose an automated mechanism to gather stakeholders and the company's opinions identifying relevant issues. To accomplish this purpose, text mining techniques are exploited to analyse textual documents written by either a stakeholder or the reporting company. It is then extracted a measure of how much a document deals with some defined topics. This kind of information is finally manipulated to prioritise topics based on how the author's opinion matters. The entire work is based upon a real case study in the domain of telecommunications.
Resumo:
Di fronte alla concorrenza globale, la sopravvivenza di un'azienda manifatturiera dipende sempre più da come essa può progettare, gestire e strutturare al meglio il proprio sistema di produzione per far fronte alla diversità dei prodotti, per migliorare l'affidabilità di consegna e anche per ridurre i costi. In questo contesto, le aziende manifatturiere utilizzano spesso sistemi di produzione diversi, in base a ciò che richiede il mercato. Molto in generale, i sistemi produttivi possono essere classificati in due categorie principali: make-to-stock (MTS) e make-to-order (MTO), in base alla politica di risposta alla domanda del mercato. Nel nuovo contesto competitivo le aziende si sono trovate a dover produrre costantemente prodotti specifici e di alta qualità con costi unitari bassi e livelli di servizio elevati (ossia, tempi di consegna brevi). È chiaro, dunque, che una delle principali decisioni strategiche da prendere da parte delle aziende sia quella relativa alla ripartizione dei prodotti in MTS/MTO, ovvero quale prodotto o famiglia di prodotti può essere fabbricato per essere stoccato a magazzino (MTS), quale può essere prodotto su ordinazione (MTO) e quale dovrebbe essere fabbricato in base alla politica di produzione ibrida MTS/MTO. Gli ultimi anni hanno mostrato una serie di cambiamenti nella politica di produzione delle aziende, che si stanno gradualmente spostando sempre più verso la modalità̀ di produzione ibrida MTS/MTO. In particolare, questo elaborato si concentrerà sul delayed product differentiation (DPD), una particolare strategia produttiva ibrida, e ne verrà proposto un modello decisionale basato sul funzionamento dell’Analytic Network Process (ANP) implementato attraverso il software Superdecisions.
Resumo:
In this thesis we address a multi-label hierarchical text classification problem in a low-resource setting and explore different approaches to identify the best one for our case. The goal is to train a model that classifies English school exercises according to a hierarchical taxonomy with few labeled data. The experiments made in this work employ different machine learning models and text representation techniques: CatBoost with tf-idf features, classifiers based on pre-trained models (mBERT, LASER), and SetFit, a framework for few-shot text classification. SetFit proved to be the most promising approach, achieving better performance when during training only a few labeled examples per class are available. However, this thesis does not consider all the hierarchical taxonomy, but only the first two levels: to address classification with the classes at the third level further experiments should be carried out, exploring methods for zero-shot text classification, data augmentation, and strategies to exploit the hierarchical structure of the taxonomy during training.