928 resultados para Android, Java, mobile, database, PHP, App, Json, design pattern, compatibilità, UML


Relevância:

100.00% 100.00%

Publicador:

Resumo:

The aim of the study reported in this paper was to evaluate the perceived benefit of video-recorded interviews as an alternative to the traditional lecture format in pre-service teacher education programs. Pre-service teachers traditionally struggle with the gap that they perceive exists between the theories taught at university and the practical competencies they require as teachers in primary and secondary school contexts. Additionally, they report that the shift from face-to-face to online delivery of course content, rapidly being adopted in the tertiary environment, seems to have often been made without associated shifts in format and pedagogy.

In the previously mentioned study, participants were drawn from cohorts of more than 300 pre-service teachers enrolled in an inclusive Education course in the Faculty of Education of an urban Australian university. The course was delivered both online and face-to-face on three campuses of the university to students in three different Education programs. In order to provide an alternative to the traditional lecture format, the developer of this course initiated and created video-recorded interview dialogues which were subsequently uploaded into the Faculty’s Learning Management System for student access. The interviews, conducted by the Course Coordinator, were held with a number of professionals with field experience relevant to key concepts of the course.

Data were collected from Student Evaluation of Teaching and Learning (SETL) questionnaires. The questionnaires were designed to gather both quantitative and qualitative data from the pre-service teachers about the extent to which the use of video-recorded interview dialogues enhanced their learning. In particular the questionnaires sought to ascertain whether, in the views of the pre-service teachers, this delivery method, first, engaged and interested them, and, second, assisted their learning through linking theory to practice. This paper provides a synthesis of the study’s findings and explores their implications for the delivery of learning experiences in both the online and on-campus modes of pre-service teacher education programs. A focus is placed on how video-recorded interviews can be used to enhance resource accessibility and to increase pre-service teachers’ engagement and learning in coursework.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This paper reports on the findings from a PhD study that investigated influences affecting early childhood teacher decision making at the preschool level in Victoria, Australia. A critical discourse analysis of preschool teacher interviews and curricula related policies found that developmentally appropriate practice was a dominant discourse influencing teacher decision making and, despite no government-mandated curricula framework at the time, teachers were held accountable for their curricula practice. Critical theory was the underpinning framework used for the study and it enabled taken-for-granted understandings to be uncovered within early childhood policies and teacher interviews. The concepts of power, authority, and accountability were examined in policy and in interview transcripts, and subsequently, various forms of control and obligation were brought to the forefront.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

An Android application uses a permission system to regulate the access to system resources and users' privacy-relevant information. Existing works have demonstrated several techniques to study the required permissions declared by the developers, but little attention has been paid towards used permissions. Besides, no specific permission combination is identified to be effective for malware detection. To fill these gaps, we have proposed a novel pattern mining algorithm to identify a set of contrast permission patterns that aim to detect the difference between clean and malicious applications. A benchmark malware dataset and a dataset of 1227 clean applications has been collected by us to evaluate the performance of the proposed algorithm. Valuable findings are obtained by analyzing the returned contrast permission patterns. © 2013 Elsevier B.V. All rights reserved.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In the last decade mobile wireless communications have witnessed an explosive growth in the user’s penetration rate and their widespread deployment around the globe. It is expected that this tendency will continue to increase with the convergence of fixed Internet wired networks with mobile ones and with the evolution to the full IP architecture paradigm. Therefore mobile wireless communications will be of paramount importance on the development of the information society of the near future. In particular a research topic of particular relevance in telecommunications nowadays is related to the design and implementation of mobile communication systems of 4th generation. 4G networks will be characterized by the support of multiple radio access technologies in a core network fully compliant with the Internet Protocol (all IP paradigm). Such networks will sustain the stringent quality of service (QoS) requirements and the expected high data rates from the type of multimedia applications to be available in the near future. The approach followed in the design and implementation of the mobile wireless networks of current generation (2G and 3G) has been the stratification of the architecture into a communication protocol model composed by a set of layers, in which each one encompasses some set of functionalities. In such protocol layered model, communications is only allowed between adjacent layers and through specific interface service points. This modular concept eases the implementation of new functionalities as the behaviour of each layer in the protocol stack is not affected by the others. However, the fact that lower layers in the protocol stack model do not utilize information available from upper layers, and vice versa, downgrades the performance achieved. This is particularly relevant if multiple antenna systems, in a MIMO (Multiple Input Multiple Output) configuration, are implemented. MIMO schemes introduce another degree of freedom for radio resource allocation: the space domain. Contrary to the time and frequency domains, radio resources mapped into the spatial domain cannot be assumed as completely orthogonal, due to the amount of interference resulting from users transmitting in the same frequency sub-channel and/or time slots but in different spatial beams. Therefore, the availability of information regarding the state of radio resources, from lower to upper layers, is of fundamental importance in the prosecution of the levels of QoS expected from those multimedia applications. In order to match applications requirements and the constraints of the mobile radio channel, in the last few years researches have proposed a new paradigm for the layered architecture for communications: the cross-layer design framework. In a general way, the cross-layer design paradigm refers to a protocol design in which the dependence between protocol layers is actively exploited, by breaking out the stringent rules which restrict the communication only between adjacent layers in the original reference model, and allowing direct interaction among different layers of the stack. An efficient management of the set of available radio resources demand for the implementation of efficient and low complexity packet schedulers which prioritize user’s transmissions according to inputs provided from lower as well as upper layers in the protocol stack, fully compliant with the cross-layer design paradigm. Specifically, efficiently designed packet schedulers for 4G networks should result in the maximization of the capacity available, through the consideration of the limitations imposed by the mobile radio channel and comply with the set of QoS requirements from the application layer. IEEE 802.16e standard, also named as Mobile WiMAX, seems to comply with the specifications of 4G mobile networks. The scalable architecture, low cost implementation and high data throughput, enable efficient data multiplexing and low data latency, which are attributes essential to enable broadband data services. Also, the connection oriented approach of Its medium access layer is fully compliant with the quality of service demands from such applications. Therefore, Mobile WiMAX seems to be a promising 4G mobile wireless networks candidate. In this thesis it is proposed the investigation, design and implementation of packet scheduling algorithms for the efficient management of the set of available radio resources, in time, frequency and spatial domains of the Mobile WiMAX networks. The proposed algorithms combine input metrics from physical layer and QoS requirements from upper layers, according to the crosslayer design paradigm. Proposed schedulers are evaluated by means of system level simulations, conducted in a system level simulation platform implementing the physical and medium access control layers of the IEEE802.16e standard.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Supervisor: Duarte Nuno Jardim Nunes

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In order to guarantee database consistency, a database system should synchronize operations of concurrent transactions. The database component responsible for such synchronization is the scheduler. A scheduler synchronizes operations belonging to different transactions by means of concurrency control protocols. Concurrency control protocols may present different behaviors: in general, a scheduler behavior can be classified as aggressive or conservative. This paper presents the Intelligent Transaction Scheduler (ITS), which has the ability to synchronize the execution of concurrent transactions in an adaptive manner. This scheduler adapts its behavior (aggressive or conservative), according to the characteristics of the computing environment in which it is inserted, using an expert system based on fuzzy logic. The ITS can implement different correctness criteria, such as conventional (syntactic) serializability and semantic serializability. In order to evaluate the performance of the ITS in relation to others schedulers with exclusively aggressive or conservative behavior, it was applied in a dynamic environment, such as a Mobile Database Community (MDBC). An MDBC simulator was developed and many sets of tests were run. The experimentation results, presented herein, prove the efficiency of the ITS in synchronizing transactions in a dynamic environment

Relevância:

100.00% 100.00%

Publicador:

Resumo:

O domínio alvo deste trabalho são os sistemas colaborativos distribuídos onde o foco está na troca dê mensagens entre usuários remotamente distribuídos. Nestes sistemas, há a necessidade das mensagens possuírem conteúdo multimídia e poderem ser entregues tanto a um usuário específico quanto a um grupo ou grupos de usuários. O objetivo deste trabalho é desenvolver um framework que facilite: a construção desse tipo de sistymas e diminua o tempo gasto com desenvolvimento através da técnica de reuso. Este trabalho apresenta o N2N Framework - Uma plataforma para desenvolvimento de Sistemas Colaborativos Distribuídos. O Framework foi concebido através da análise do comportamento de aplicações com características de multimídias colaborativas, como ambientes virtuais multi-usuários, chats, enquetes, e torcidas virtuais. O Framework foi implementado usando-se a plataforma Java. O N2N Framework facilita o design e implementação de sistemas colaborativos distribuídos, implementando a entrega das mensagens, e direcionando o desenvolvedor de aplicações para a preocupação com implementação de suas mensagens específicas e o processamento que delas decorre

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Pós-graduação em Agronomia (Energia na Agricultura) - FCA

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Nello sviluppo di sistemi informatici si sono affermate numerose tecnologie, che vanno utilizzate in modo combinato e, possibilmente sinergico. Da una parte, i sistemi di gestione di basi di dati relazionali consentono una gestione efficiente ed efficace di dati persistenti, condivisi e transazionali. Dall'altra, gli strumenti e i metodi orientati agli oggetti (linguaggi di programmazione, ma anche metodologie di analisi e progettazione) consentono uno sviluppo efficace della logica applicativa delle applicazioni. E’ utile in questo contesto spiegare che cosa s'intende per sistema informativo e sistema informatico. Sistema informativo: L'insieme di persone, risorse tecnologiche, procedure aziendali il cui compito è quello di produrre e conservare le informazioni che servono per operare nell'impresa e gestirla. Sistema informatico: L'insieme degli strumenti informatici utilizzati per il trattamento automatico delle informazioni, al fine di agevolare le funzioni del sistema informativo. Ovvero, il sistema informatico raccoglie, elabora, archivia, scambia informazione mediante l'uso delle tecnologie proprie dell'Informazione e della Comunicazione (ICT): calcolatori, periferiche, mezzi di comunicazione, programmi. Il sistema informatico è quindi un componente del sistema informativo. Le informazioni ottenute dall'elaborazione dei dati devono essere salvate da qualche parte, in modo tale da durare nel tempo dopo l'elaborazione. Per realizzare questo scopo viene in aiuto l'informatica. I dati sono materiale informativo grezzo, non (ancora) elaborato da chi lo riceve, e possono essere scoperti, ricercati, raccolti e prodotti. Sono la materia prima che abbiamo a disposizione o produciamo per costruire i nostri processi comunicativi. L'insieme dei dati è il tesoro di un'azienda e ne rappresenta la storia evolutiva. All'inizio di questa introduzione è stato accennato che nello sviluppo dei sistemi informatici si sono affermate diverse tecnologie e che, in particolare, l'uso di sistemi di gestione di basi di dati relazionali comporta una gestione efficace ed efficiente di dati persistenti. Per persistenza di dati in informatica si intende la caratteristica dei dati di sopravvivere all'esecuzione del programma che li ha creati. Se non fosse cosi, i dati verrebbero salvati solo in memoria RAM e sarebbero persi allo spegnimento del computer. Nella programmazione informatica, per persistenza si intende la possibilità di far sopravvivere strutture dati all'esecuzione di un programma singolo. Occorre il salvataggio in un dispositivo di memorizzazione non volatile, come per esempio su un file system o su un database. In questa tesi si è sviluppato un sistema che è in grado di gestire una base di dati gerarchica o relazionale consentendo l'importazione di dati descritti da una grammatica DTD. Nel capitolo 1 si vedranno più in dettaglio cosa di intende per Sistema Informativo, modello client-server e sicurezza dei dati. Nel capitolo 2 parleremo del linguaggio di programmazione Java, dei database e dei file XML. Nel capitolo 3 descriveremo un linguaggio di analisi e modellazione UML con esplicito riferimento al progetto sviluppato. Nel capitolo 4 descriveremo il progetto che è stato implementato e le tecnologie e tools utilizzati.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Recently in most of the industrial automation process an ever increasing degree of automation has been observed. This increasing is motivated by the higher requirement of systems with great performance in terms of quality of products/services generated, productivity, efficiency and low costs in the design, realization and maintenance. This trend in the growth of complex automation systems is rapidly spreading over automated manufacturing systems (AMS), where the integration of the mechanical and electronic technology, typical of the Mechatronics, is merging with other technologies such as Informatics and the communication networks. An AMS is a very complex system that can be thought constituted by a set of flexible working stations, one or more transportation systems. To understand how this machine are important in our society let considerate that every day most of us use bottles of water or soda, buy product in box like food or cigarets and so on. Another important consideration from its complexity derive from the fact that the the consortium of machine producers has estimated around 350 types of manufacturing machine. A large number of manufacturing machine industry are presented in Italy and notably packaging machine industry,in particular a great concentration of this kind of industry is located in Bologna area; for this reason the Bologna area is called “packaging valley”. Usually, the various parts of the AMS interact among them in a concurrent and asynchronous way, and coordinate the parts of the machine to obtain a desiderated overall behaviour is an hard task. Often, this is the case in large scale systems, organized in a modular and distributed manner. Even if the success of a modern AMS from a functional and behavioural point of view is still to attribute to the design choices operated in the definition of the mechanical structure and electrical electronic architecture, the system that governs the control of the plant is becoming crucial, because of the large number of duties associated to it. Apart from the activity inherent to the automation of themachine cycles, the supervisory system is called to perform other main functions such as: emulating the behaviour of traditional mechanical members thus allowing a drastic constructive simplification of the machine and a crucial functional flexibility; dynamically adapting the control strategies according to the different productive needs and to the different operational scenarios; obtaining a high quality of the final product through the verification of the correctness of the processing; addressing the operator devoted to themachine to promptly and carefully take the actions devoted to establish or restore the optimal operating conditions; managing in real time information on diagnostics, as a support of the maintenance operations of the machine. The kind of facilities that designers can directly find on themarket, in terms of software component libraries provides in fact an adequate support as regard the implementation of either top-level or bottom-level functionalities, typically pertaining to the domains of user-friendly HMIs, closed-loop regulation and motion control, fieldbus-based interconnection of remote smart devices. What is still lacking is a reference framework comprising a comprehensive set of highly reusable logic control components that, focussing on the cross-cutting functionalities characterizing the automation domain, may help the designers in the process of modelling and structuring their applications according to the specific needs. Historically, the design and verification process for complex automated industrial systems is performed in empirical way, without a clear distinction between functional and technological-implementation concepts and without a systematic method to organically deal with the complete system. Traditionally, in the field of analog and digital control design and verification through formal and simulation tools have been adopted since a long time ago, at least for multivariable and/or nonlinear controllers for complex time-driven dynamics as in the fields of vehicles, aircrafts, robots, electric drives and complex power electronics equipments. Moving to the field of logic control, typical for industrial manufacturing automation, the design and verification process is approached in a completely different way, usually very “unstructured”. No clear distinction between functions and implementations, between functional architectures and technological architectures and platforms is considered. Probably this difference is due to the different “dynamical framework”of logic control with respect to analog/digital control. As a matter of facts, in logic control discrete-events dynamics replace time-driven dynamics; hence most of the formal and mathematical tools of analog/digital control cannot be directly migrated to logic control to enlighten the distinction between functions and implementations. In addition, in the common view of application technicians, logic control design is strictly connected to the adopted implementation technology (relays in the past, software nowadays), leading again to a deep confusion among functional view and technological view. In Industrial automation software engineering, concepts as modularity, encapsulation, composability and reusability are strongly emphasized and profitably realized in the so-calledobject-oriented methodologies. Industrial automation is receiving lately this approach, as testified by some IEC standards IEC 611313, IEC 61499 which have been considered in commercial products only recently. On the other hand, in the scientific and technical literature many contributions have been already proposed to establish a suitable modelling framework for industrial automation. During last years it was possible to note a considerable growth in the exploitation of innovative concepts and technologies from ICT world in industrial automation systems. For what concerns the logic control design, Model Based Design (MBD) is being imported in industrial automation from software engineering field. Another key-point in industrial automated systems is the growth of requirements in terms of availability, reliability and safety for technological systems. In other words, the control system should not only deal with the nominal behaviour, but should also deal with other important duties, such as diagnosis and faults isolations, recovery and safety management. Indeed, together with high performance, in complex systems fault occurrences increase. This is a consequence of the fact that, as it typically occurs in reliable mechatronic systems, in complex systems such as AMS, together with reliable mechanical elements, an increasing number of electronic devices are also present, that are more vulnerable by their own nature. The diagnosis problem and the faults isolation in a generic dynamical system consists in the design of an elaboration unit that, appropriately processing the inputs and outputs of the dynamical system, is also capable of detecting incipient faults on the plant devices, reconfiguring the control system so as to guarantee satisfactory performance. The designer should be able to formally verify the product, certifying that, in its final implementation, it will perform itsrequired function guarantying the desired level of reliability and safety; the next step is that of preventing faults and eventually reconfiguring the control system so that faults are tolerated. On this topic an important improvement to formal verification of logic control, fault diagnosis and fault tolerant control results derive from Discrete Event Systems theory. The aimof this work is to define a design pattern and a control architecture to help the designer of control logic in industrial automated systems. The work starts with a brief discussion on main characteristics and description of industrial automated systems on Chapter 1. In Chapter 2 a survey on the state of the software engineering paradigm applied to industrial automation is discussed. Chapter 3 presentes a architecture for industrial automated systems based on the new concept of Generalized Actuator showing its benefits, while in Chapter 4 this architecture is refined using a novel entity, the Generalized Device in order to have a better reusability and modularity of the control logic. In Chapter 5 a new approach will be present based on Discrete Event Systems for the problemof software formal verification and an active fault tolerant control architecture using online diagnostic. Finally conclusive remarks and some ideas on new directions to explore are given. In Appendix A are briefly reported some concepts and results about Discrete Event Systems which should help the reader in understanding some crucial points in chapter 5; while in Appendix B an overview on the experimental testbed of the Laboratory of Automation of University of Bologna, is reported to validated the approach presented in chapter 3, chapter 4 and chapter 5. In Appendix C some components model used in chapter 5 for formal verification are reported.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

In der Herstellung fester Darreichungsformen umfasst die Granulierung einen komplexen Teilprozess mit hoher Relevanz für die Qualität des pharmazeutischen Produktes. Die Wirbelschichtgranulierung ist ein spezielles Granulierverfahren, welches die Teilprozesse Mischen, Agglomerieren und Trocknen in einem Gerät vereint. Durch die Kombination mehrerer Prozessstufen unterliegt gerade dieses Verfahren besonderen Anforderungen an ein umfassendes Prozessverständnis. Durch die konsequente Verfolgung des PAT- Ansatzes, welcher im Jahre 2004 durch die amerikanische Zulassungsbehörde (FDA) als Guideline veröffentlicht wurde, wurde der Grundstein für eine kontinuierliche Prozessverbesserung durch erhöhtes Prozessverständnis, für Qualitätserhöhung und Kostenreduktion gegeben. Die vorliegende Arbeit befasste sich mit der Optimierung der Wirbelschicht-Granulationsprozesse von zwei prozesssensiblen Arzneistoffformulierungen, unter Verwendung von PAT. rnFür die Enalapril- Formulierung, einer niedrig dosierten und hochaktiven Arzneistoffrezeptur, wurde herausgefunden, dass durch eine feinere Zerstäubung der Granulierflüssigkeit deutlich größere Granulatkörnchen erhalten werden. Eine Erhöhung der MassRatio verringert die Tröpfchengröße, dies führt zu größeren Granulaten. Sollen Enalapril- Granulate mit einem gewünschten D50-Kornverteilung zwischen 100 und 140 um hergestellt werden, dann muss die MassRatio auf hohem Niveau eingestellt werden. Sollen Enalapril- Granulate mit einem D50- Wert zwischen 80 und 120µm erhalten werden, so muss die MassRatio auf niedrigem Niveau eingestellt sein. Anhand der durchgeführten Untersuchungen konnte gezeigt werden, dass die MassRatio ein wichtiger Parameter ist und zur Steuerung der Partikelgröße der Enalapril- Granulate eingesetzt werden kann; unter der Voraussetzung dass alle anderen Prozessparameter konstant gehalten werden.rnDie Betrachtung der Schnittmengenplots gibt die Möglichkeit geeignete Einstellungen der Prozessparameter bzw. Einflussgrößen zu bestimmen, welche dann zu den gewünschten Granulat- und Tabletteneigenschaften führen. Anhand der Lage und der Größe der Schnittmenge können die Grenzen der Prozessparameter zur Herstellung der Enalapril- Granulate bestimmt werden. Werden die Grenzen bzw. der „Design Space“ der Prozessparameter eingehalten, kann eine hochwertige Produktqualität garantiert werden. rnUm qualitativ hochwertige Enalapril Tabletten mit der gewählten Formulierung herzustellen, sollte die Enalapril- Granulation mit folgenden Prozessparametern durchgeführt werden: niedrige Sprührate, hoher MassRatio, einer Zulufttemperatur von mindestens > 50 °C und einer effektiven Zuluftmenge < 180 Nm³/h. Wird hingegen eine Sprührate von 45 g/min und eine mittlere MassRatio von 4.54 eingestellt, so muss die effektive Zuluftmenge mindestens 200 Nm³/h und die Zulufttemperatur mindestens 60 °C betragen, um eine vorhersagbar hohe Tablettenqualität zu erhalten. Qualität wird in das Arzneimittel bereits während der Herstellung implementiert, indem die Prozessparameter bei der Enalapril- Granulierung innerhalb des „Design Space“ gehalten werden.rnFür die Metformin- Formulierung, einer hoch dosierten aber wenig aktiven Arzneistoffrezeptur wurde herausgefunden, dass sich der Wachstumsmechanismus des Feinanteils der Metformin- Granulate von dem Wachstumsmechanismus der D50- und D90- Kornverteilung unterscheidet. Der Wachstumsmechanismus der Granulate ist abhängig von der Partikelbenetzung durch die versprühten Flüssigkeitströpfchen und vom Größenverhältnis von Partikel zu Sprühtröpfchen. Der Einfluss der MassRatio ist für die D10- Kornverteilung der Granulate vernachlässigbar klein. rnMit Hilfe der Störgrößen- Untersuchungen konnte eine Regeleffizienz der Prozessparameter für eine niedrig dosierte (Enalapril)- und eine hoch dosierte (Metformin) Arzneistoffformulierung erarbeitet werden, wodurch eine weitgehende Automatisierung zur Verringerung von Fehlerquellen durch Nachregelung der Störgrößen ermöglicht wird. Es ergibt sich für die gesamte Prozesskette ein in sich geschlossener PAT- Ansatz. Die Prozessparameter Sprührate und Zuluftmenge erwiesen sich als am besten geeignet. Die Nachregelung mit dem Parameter Zulufttemperatur erwies sich als träge. rnFerner wurden in der Arbeit Herstellverfahren für Granulate und Tabletten für zwei prozesssensible Wirkstoffe entwickelt. Die Robustheit der Herstellverfahren gegenüber Störgrößen konnte demonstriert werden, wodurch die Voraussetzungen für eine Echtzeitfreigabe gemäß dem PAT- Gedanken geschaffen sind. Die Kontrolle der Qualität des Produkts findet nicht am Ende der Produktions- Prozesskette statt, sondern die Kontrolle wird bereits während des Prozesses durchgeführt und basiert auf einem besseren Verständnis des Produktes und des Prozesses. Außerdem wurde durch die konsequente Verfolgung des PAT- Ansatzes die Möglichkeit zur kontinuierlichen Prozessverbesserung, zur Qualitätserhöhung und Kostenreduktion gegeben und damit das ganzheitliche Ziel des PAT- Gedankens erreicht und verwirklicht.rn

Relevância:

100.00% 100.00%

Publicador:

Resumo:

La tesi consiste nello studiare e sviluppare servizi multimediali in sistemi di CrowdSensing, nello specifico si vuole sviluppare un sistema che permetta il trasferimento da client a server di file multimediali come il video nel sistema Participact dell'universita di Bologna.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

While designing systems and products requires a deep understanding of influences that achieve desirable performance, the need for an efficient and systematic decision-making approach drives the need for optimization strategies. This paper provides the motivation for this topic as well as a description of applications in Computing Center of Madrid city Council. Optimization applications can be found in almost all areas of engineering. Typical problems in process, working with a database, arise in query design, entity model design and concurrent processes. This paper proposes a solution to optimize a night process dealing with millions of records with an overall performance of about eight times in computation time.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Este trabajo corresponde con la implementación de componentes software dentro de la Plataforma COMPUTAPLEX, la cual tiene como objetivo facilitar a los investigadores la realización de tareas del proceso experimental de ingeniería de software. Uno de los aportes a esta plataforma tecnológica corresponde con el desarrolló de los componentes necesarios para la recuperación de datos experimentales disponibles en diversas fuentes de datos, para ello se hizo uso de un mecanismo capaz de unificar la extracción de información de MySQL, ficheros excel y ficheros SPSS. Con ello diferentes grupos de investigación asociados pueden compartir y tener acceso a repositorios experimentales que se mantienen tanto de manera local como externa. Por otra parte, se ha realizado un estudio de la tecnología de agentes en la que se describe sus definiciones, lenguajes de comunicación, especificación FIPA, JADE como implementación FIPA y parser XML. Además para este trabajo se ha definido e implementado una ontología de comunicación entre agentes, la misma que fue diseñada en la herramienta Protégé. En lo que se refiere al desarrollo de componentes se hizo uso de una amplía variedad de tecnologías que incluye lenguaje de programación Java, framework JADE para el desarrollo de agentes, librería JENA para manejo de ontologías, librería SAXParser para lectura de archivos XML y patrón de diseño Factory. Finalmente se describe la metodología de trabajo utilizada en el proyecto, la cual por medio de la realización de varios ciclos iterativos permitió obtener prototipos que poco a poco fueron cubriendo las necesidades del producto software.----ABSTRACT---- This work relates to the implementation of software components within the platform Computaplex, which aims to enable researchers to conduct experimental software engineering process tasks. One of the contributions to this platform technology corresponds to the development of components which are necessary for the recovery of experimental data available in different data sources, to archive this goal a mechanism able to unify the extraction of information from MySQL, Excel and SPSS files was made. Therefore, associated research groups can share and access experimental repositories that remain both locally and externally. Moreover, it has been conducted a study of agent technology in its definition is described, languages communication, FIPA, JADE and FIPA implementation and XML parser. In addition to this work, it has been defined and implemented an ontology for communication between agents, the same as was designed in the Protégé tool. In what refers to the development of components, a wide range of technologies have been made which includes Java programming language, framework JADE for agent development, JENA library for handling ontologies, SAXParser for reading XML files and Factory design pattern. Finally, describing the work methodology used in this project, which through the implementation of several iterative cycles allowed to obtain prototypes were gradually meeting the needs of the software product.