820 resultados para Aleph Analytics


Relevância:

10.00% 10.00%

Publicador:

Resumo:

The convergence between the recent developments in sensing technologies, data science, signal processing and advanced modelling has fostered a new paradigm to the Structural Health Monitoring (SHM) of engineered structures, which is the one based on intelligent sensors, i.e., embedded devices capable of stream processing data and/or performing structural inference in a self-contained and near-sensor manner. To efficiently exploit these intelligent sensor units for full-scale structural assessment, a joint effort is required to deal with instrumental aspects related to signal acquisition, conditioning and digitalization, and those pertaining to data management, data analytics and information sharing. In this framework, the main goal of this Thesis is to tackle the multi-faceted nature of the monitoring process, via a full-scale optimization of the hardware and software resources involved by the {SHM} system. The pursuit of this objective has required the investigation of both: i) transversal aspects common to multiple application domains at different abstraction levels (such as knowledge distillation, networking solutions, microsystem {HW} architectures), and ii) the specificities of the monitoring methodologies (vibrations, guided waves, acoustic emission monitoring). The key tools adopted in the proposed monitoring frameworks belong to the embedded signal processing field: namely, graph signal processing, compressed sensing, ARMA System Identification, digital data communication and TinyML.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In questa tesi si trattano lo studio e la sperimentazione di un modello generativo retrieval-augmented, basato su Transformers, per il task di Abstractive Summarization su lunghe sentenze legali. La sintesi automatica del testo (Automatic Text Summarization) è diventata un task di Natural Language Processing (NLP) molto importante oggigiorno, visto il grandissimo numero di dati provenienti dal web e banche dati. Inoltre, essa permette di automatizzare un processo molto oneroso per gli esperti, specialmente nel settore legale, in cui i documenti sono lunghi e complicati, per cui difficili e dispendiosi da riassumere. I modelli allo stato dell’arte dell’Automatic Text Summarization sono basati su soluzioni di Deep Learning, in particolare sui Transformers, che rappresentano l’architettura più consolidata per task di NLP. Il modello proposto in questa tesi rappresenta una soluzione per la Long Document Summarization, ossia per generare riassunti di lunghe sequenze testuali. In particolare, l’architettura si basa sul modello RAG (Retrieval-Augmented Generation), recentemente introdotto dal team di ricerca Facebook AI per il task di Question Answering. L’obiettivo consiste nel modificare l’architettura RAG al fine di renderla adatta al task di Abstractive Long Document Summarization. In dettaglio, si vuole sfruttare e testare la memoria non parametrica del modello, con lo scopo di arricchire la rappresentazione del testo di input da riassumere. A tal fine, sono state sperimentate diverse configurazioni del modello su diverse tipologie di esperimenti e sono stati valutati i riassunti generati con diverse metriche automatiche.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Este estudo trata da comunicação face a face nas organizações sob diferentes abordagens teóricas. Considera a perspectiva da simultaneidade dos meios, já que as empresas utilizam diversos canais para dialogar com seus públicos de interesse. Leva em conta o fenômeno da midiatização, que reestrutura o modo como as pessoas se relacionam na sociedade contemporânea. O objetivo geral da pesquisa é sistematizar papeis potencialmente exercidos pela interação face a face e conhecer algumas circunstâncias que envolvem sua prática nas organizações. Por se tratar de uma tese teórica, a pesquisa bibliográfica se apresenta como um dos principais procedimentos metodológicos; análises de casos empíricos e um estudo de caso desenvolvido na Embrapa Pantanal constituem situações ilustrativas. Conclui-se que a comunicação face a face nas empresas ocorre de forma simultânea e combinada a outros canais de comunicação, porém, ela proporciona resultados práticos e filosóficos ainda pouco explorados. É rara a utilização estratégica de contatos presenciais como mecanismo para estabelecer relacionamentos, conhecer as reações alheias e ajustar a comunicação, aliar o discurso corporativo às práticas empresariais e avaliar o contexto onde se desenvolvem as interações, o que pode ser decisivo para a comunicação organizacional.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The proliferation of Web-based learning objects makes finding and evaluating resources a considerable hurdle for learners to overcome. While established learning analytics methods provide feedback that can aid learner evaluation of learning resources, the adequacy and reliability of these methods is questioned. Because engagement with online learning is different from other Web activity, it is important to establish pedagogically relevant measures that can aid the development of distinct, automated analysis systems. Content analysis is often used to examine online discussion in educational settings, but these instruments are rarely compared with each other which leads to uncertainty regarding their validity and reliability. In this study, participation in Massive Open Online Course (MOOC) comment forums was evaluated using four different analytical approaches: the Digital Artefacts for Learning Engagement (DiAL-e) framework, Bloom's Taxonomy, Structure of Observed Learning Outcomes (SOLO) and Community of Inquiry (CoI). Results from this study indicate that different approaches to measuring cognitive activity are closely correlated and are distinct from typical interaction measures. This suggests that computational approaches to pedagogical analysis may provide useful insights into learning processes.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The demographics of massive open online course (MOOC) analytics show that the great majority of learners are highly qualified professionals, and not, as originally envisaged, the global community of disadvantaged learners who have no access to good higher education. MOOC pedagogy fits well with the combination of instruction and peer community learning found in most professional development. A UNESCO study therefore set out to test the efficacy of an experimental course for teachers who need but do not receive high-quality continuing professional development, as a way of exploiting what MOOCs can do indirectly to serve disadvantaged students. The course was based on case studies around the world of information and communication technology (ICT) in primary education and was carried out to contribute to the UNESCO “Education For All” goal. It used a co-learning approach to engage the primary teaching community in exploring ways of using ICT in primary education. Course analytics, forums and participant surveys demonstrated that it worked well. The paper concludes by arguing that this technology has the power to tackle the large-scale educational problem of developing the primary-level teachers needed to meet the goal of universal education.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In Prior Analytics 1.1–22, Aristotle develops his proof system of non-modal and modal propositions. This system is given in the language of propositions, and Aristotle is concerned with establishing some properties and relations that the expressions of this language enjoy. However, modern scholarship has found some of his results inconsistent with positions defended elsewhere. The set of rules of inference of this system has also caused perplexity: there does not seem to be a single interpretation that validates all the rules which Aristotle is explicitly committed to using in his proofs. Some commentators have argued that these and other problems cannot be successfully addressed from the viewpoint of the traditional, ‘first-order’ interpretation of Aristotle’s syllogistic, whereby propositions are taken to involve quantification over individuals only. Accordingly, this interpretation not only is inadequate for formal analysis, but also stems from a misunderstanding of Aristotle’s ideas about quantification. On the contrary, in this study I purport to vindicate the adequacy and plausibility of the first-order interpretation. Together with some assumptions about the language of propositions and an appropriate regimentation, the first-order interpretation yields promising solutions to many of the problems raised by the modal syllogistic. Thus, I present a reconstruction of the language of propositions and a formal interpretation thereof which will prove respectful and responsive to most of the views endorsed by Aristotle in the ‘modal’ chapters of the Analytics.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The fourth industrial revolution is paving the way for Industrial Internet of Things applications where industrial assets (e.g., robotic arms, valves, pistons) are equipped with a large number of wireless devices (i.e., microcontroller boards that embed sensors and actuators) to enable a plethora of new applications, such as analytics, diagnostics, monitoring, as well as supervisory, and safety control use-cases. Nevertheless, current wireless technologies, such as Wi-Fi, Bluetooth, and even private 5G networks, cannot fulfill all the requirements set up by the Industry 4.0 paradigm, thus opening up new 6G-oriented research trends, such as the use of THz frequencies. In light of the above, this thesis provides (i) a broad overview of the main use-cases, requirements, and key enabling wireless technologies foreseen by the fourth industrial revolution, and (ii) proposes innovative contributions, both theoretical and empirical, to enhance the performance of current and future wireless technologies at different levels of the protocol stack. In particular, at the physical layer, signal processing techniques are being exploited to analyze two multiplexing schemes, namely Affine Frequency Division Multiplexing and Orthogonal Chirp Division Multiplexing, which seem promising for high-frequency wireless communications. At the medium access layer, three protocols for intra-machine communications are proposed, where one is based on LoRa at 2.4 GHz and the others work in the THz band. Different scheduling algorithms for private industrial 5G networks are compared, and two main proposals are described, i.e., a decentralized scheme that leverages machine learning techniques to better address aperiodic traffic patterns, and a centralized contention-based design that serves a federated learning industrial application. Results are provided in terms of numerical evaluations, simulation results, and real-world experiments. Several improvements over the state-of-the-art were obtained, and the description of up-and-running testbeds demonstrates the feasibility of some of the theoretical concepts when considering a real industry plant.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

I dati sono una risorsa di valore inestimabile per tutte le organizzazioni. Queste informazioni vanno da una parte gestite tramite i classici sistemi operazionali, dall’altra parte analizzate per ottenere approfondimenti che possano guidare le scelte di business. Uno degli strumenti fondamentali a supporto delle scelte di business è il data warehouse. Questo elaborato è il frutto di un percorso di tirocinio svolto con l'azienda Injenia S.r.l. Il focus del percorso era rivolto all'ottimizzazione di un data warehouse che l'azienda vende come modulo aggiuntivo di un software di nome Interacta. Questo data warehouse, Interacta Analytics, ha espresso nel tempo notevoli criticità architetturali e di performance. L’architettura attualmente usata per la creazione e la gestione dei dati all'interno di Interacta Analytics utilizza un approccio batch, pertanto, l’obiettivo cardine dello studio è quello di trovare soluzioni alternative batch che garantiscano un risparmio sia in termini economici che di tempo, esplorando anche la possibilità di una transizione ad un’architettura streaming. Gli strumenti da utilizzare in questa ricerca dovevano inoltre mantenersi in linea con le tecnologie utilizzate per Interacta, ossia i servizi della Google Cloud Platform. Dopo una breve dissertazione sul background teorico di questa area tematica, l'elaborato si concentra sul funzionamento del software principale e sulla struttura logica del modulo di analisi. Infine, si espone il lavoro sperimentale, innanzitutto proponendo un'analisi delle criticità principali del sistema as-is, dopodiché ipotizzando e valutando quattro ipotesi migliorative batch e due streaming. Queste, come viene espresso nelle conclusioni della ricerca, migliorano di molto le performance del sistema di analisi in termini di tempistiche di elaborazione, di costo totale e di semplicità dell'architettura, in particolare grazie all'utilizzo dei servizi serverless con container e FaaS della piattaforma cloud di Google.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Industry 4.0 refers to the 4th industrial revolution and at its bases, we can see the digitalization and the automation of the assembly line. The whole production process has improved and evolved thanks to the advances made in networking, and AI studies, which include of course machine learning, cloud computing, IoT, and other technologies that are finally being implemented into the industrial scenario. All these technologies have in common a need for faster, more secure, robust, and reliable communication. One of the many solutions for these demands is the use of mobile communication technologies in the industrial environment, but which technology is better suited for these demands? Of course, the answer isn’t as simple as it seems. The 4th industrial revolution has a never seen incomparable potential with respect to the previous ones, every factory, enterprise, or company have different network demands, and even in each of these infrastructures, the demands may diversify by sector, or by application. For example, in the health care industry, there may be e a need for increased bandwidth for the analysis of high-definition videos or, faster speeds in order to have analytics occur in real-time, and again another application might be higher security and reliability to protect patients’ data. As seen above, choosing the right technology for the right environment and application, considers many things, and the ones just stated are but a speck of dust with respect to the overall picture. In this thesis, we will investigate a comparison between the use of two of the available technologies in use for the industrial environment: Wi-Fi 6 and 5G Private Networks in the specific case of a steel factory.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Un ambiente sempre più interconnesso per facilitare la condivisione di dati, lo sviluppo di strumenti sempre più ricettivi, l’utilizzo di algoritmi sempre più mirati ed efficaci nel selezionare le giuste informazioni sono alcuni dei fattori chiave che hanno consentito e tuttora consentono la crescita, la gestione, il riutilizzo e la diffusione del patrimonio conoscitivo a disposizione delle organizzazioni. Il continuo aumento di risorse informatiche ha indotto le organizzazioni a rivedere il ruolo svolto dalla Business Intelligence, arricchendolo di strumenti e procedure nuove e creando ulteriori figure professionali. L’obiettivo di questo elaborato è fornire una panoramica della business intelligence, della sua origine e della rilevanza e utilità in ambito aziendale. Nel primo capitolo si tratta della disciplina della Business Intelligence, in particolare definizione, cenni storici e differenza con la Business Analytics. Si descrivono successivamente i sistemi informativi e i loro componenti per finire con l’architettura di una soluzione di BI. Nel secondo capitolo, si effettua una panoramica sui software di Business Intelligence sul mercato, dopo di che si presenta Microsoft Power BI di Microsoft, in particolare funzionalità e caratteristiche. Il terzo capitolo è relativo al progetto effettuato durante il periodo di tirocinio: l’implementazione di nuove funzionalità e analisi su un software BI sviluppato dall’azienda ospitante.