86 resultados para Metodologia de desenvolvimento de sistemas
Resumo:
Avaliação de usabilidade é um processo importante durante o desenvolvimento de um software, seja ele para um sistema web ou mobile. No caso de um sistema mobile, o custo é bastante oneroso, tanto no que se refere à mão de obra especializada, como de recursos tecnológicos utilizados, tornando essa avaliação de usabilidade ainda mais importante. Além disso, as versões webdos sistemas SIG da UFRN já estão consolidadas e com uma grande aceitação, fazendo com que suas versões mobile tenham, ainda mais, a preocupação de lançar um produto de boa qualidade para manter essa credibilidade dos sistemas SIG , tanto na UFRN, como nas demais instituições que utilizam esses sistemas. Com este trabalho, buscou-se identificar algumas diretrizes de interface que possam ser utilizadas no processo de desenvolvimento dos sistemas SIG Mobile, mais especificamente o SIGAA Mobile, de modo a facilitar o desenvolvimento de novas funcionalidades voltadas para estes sistemas. Para isso, foi realizada uma avaliação de usabilidade no Portal do Aluno do SIGAA Mobile Android, tomando como base padrões de interface já existentes na literatura. Posteriormente, foi aplicado um questionário com os usuários do sistema para coletar as opiniões e sugestões dos mesmos. De posse de todos esses dados coletados, foi possível definir algumas diretrizes de interface a serem seguidas como recomendações no processo de desenvolvimento dos sistemas SIG Mobile.
Resumo:
Nowadays, the search for new technologies that are able to follow the upcoming challenges in oil industry is a constant. Always trying properties improvements of the used materials, looking for the best performance and greater life time. Besides the search for technologies that show an improvement of performance, the search for materials environmentally correct along the whole production process. In Oil well cementing, this search for new technologies passes through the development of slurry systems that support these requests and that are also environmentally friendly. In this context, the use of geopolymer slurries is a great alternative route to cementing oil wells. Besides having good properties, comparable to Portland cement slurries, this alternative material releases much less CO2 gas in the production of their root materials when compared the production of Portland cement, which releases tons of CO2. In order to improve the properties of geopolymer slurries has been added Calcium Oxide, as observed in other studies that slurries where the Calcium is present the values of compressive strength is greater. The addition has been realized based in the CaO/SiO2 molar ratio of 0.05, 0.10 and 0.15. Have been performed compressive strength tests, thickening time, rheology and fliud loss control test of the slurries, following NBR 9831, as well as the physical chemical characterization of XRD, SEM and TG. Has been observed in most of the tests the slurries follow a tendency until the ratio of 0.10, which inverses in the ratio 0.15. This behavior can be explained by two phenomena that occur simultaneously, the first one is the break of the polymer chains and a consequent increase in molucules mobility, which prevails until the ratio of 0.1, and the second is possible approach of the chains due to the capacity of the calcium ions stabilize the charges of two different aluminum. There is only one linearity in the mechanical behavior that can be attributed to the appereance of the C-S-H phase. Based on this, it is concluded that the phenomenon of breaking the polymer chains predominates until the ratio of 0.1, causing an increase of the filtrate volume, lower rheological parameters and increasing thickening time. From the ratio of 0.15 the approach of the chains predominates, and the behavior is reversed
Resumo:
The multiphase flow occurrence in the oil and gas industry is common throughout fluid path, production, transportation and refining. The multiphase flow is defined as flow simultaneously composed of two or more phases with different properties and immiscible. An important computational tool for the design, planning and optimization production systems is multiphase flow simulation in pipelines and porous media, usually made by multiphase flow commercial simulators. The main purpose of the multiphase flow simulators is predicting pressure and temperature at any point at the production system. This work proposes the development of a multiphase flow simulator able to predict the dynamic pressure and temperature gradient in vertical, directional and horizontal wells. The prediction of pressure and temperature profiles was made by numerical integration using marching algorithm with empirical correlations and mechanistic model to predict pressure gradient. The development of this tool involved set of routines implemented through software programming Embarcadero C++ Builder® 2010 version, which allowed the creation of executable file compatible with Microsoft Windows® operating systems. The simulator validation was conduct by computational experiments and comparison the results with the PIPESIM®. In general, the developed simulator achieved excellent results compared with those obtained by PIPESIM and can be used as a tool to assist production systems development
Resumo:
The progressing cavity pump artificial lift system, PCP, is a main lift system used in oil production industry. As this artificial lift application grows the knowledge of it s dynamics behavior, the application of automatic control and the developing of equipment selection design specialist systems are more useful. This work presents tools for dynamic analysis, control technics and a specialist system for selecting lift equipments for this artificial lift technology. The PCP artificial lift system consists of a progressing cavity pump installed downhole in the production tubing edge. The pump consists of two parts, a stator and a rotor, and is set in motion by the rotation of the rotor transmitted through a rod string installed in the tubing. The surface equipment generates and transmits the rotation to the rod string. First, is presented the developing of a complete mathematical dynamic model of PCP system. This model is simplified for use in several conditions, including steady state for sizing PCP equipments, like pump, rod string and drive head. This model is used to implement a computer simulator able to help in system analysis and to operates as a well with a controller and allows testing and developing of control algorithms. The next developing applies control technics to PCP system to optimize pumping velocity to achieve productivity and durability of downhole components. The mathematical model is linearized to apply conventional control technics including observability and controllability of the system and develop design rules for PI controller. Stability conditions are stated for operation point of the system. A fuzzy rule-based control system are developed from a PI controller using a inference machine based on Mandami operators. The fuzzy logic is applied to develop a specialist system that selects PCP equipments too. The developed technics to simulate and the linearized model was used in an actual well where a control system is installed. This control system consists of a pump intake pressure sensor, an industrial controller and a variable speed drive. The PI control was applied and fuzzy controller was applied to optimize simulated and actual well operation and the results was compared. The simulated and actual open loop response was compared to validate simulation. A case study was accomplished to validate equipment selection specialist system
Resumo:
Due to the current need of the industry to integrate data of the beginning of production originating from of several sources and of transforming them in useful information for sockets of decisions, a search exists every time larger for systems of visualization of information that come to collaborate with that functionality. On the other hand, a common practice nowadays, due to the high competitiveness of the market, it is the development of industrial systems that possess characteristics of modularity, distribution, flexibility, scalability, adaptation, interoperability, reusability and access through web. Those characteristics provide an extra agility and a larger easiness in adapting to the frequent changes of demand of the market. Based on the arguments exposed above, this work consists of specifying a component-based architecture, with the respective development of a system based on that architecture, for the visualization of industrial data. The system was conceived to be capable to supply on-line information and, optionally, historical information of variables originating from of the beginning of production. In this work it is shown that the component-based architecture developed possesses the necessary requirements for the obtaining of a system robust, reliable and of easy maintenance, being, like this, in agreement with the industrial needs. The use of that architecture allows although components can be added, removed or updated in time of execution, through a manager of components through web, still activating more the adaptation process and updating of the system
Resumo:
Due to major progress of communication system in the last decades, need for more precise characterization of used components. The S-parameters modeling has been used to characterization, simulation and test of communication system. However, limitation of S-parameters to model nonlinear system has created new modeling systems that include the nonlinear characteristics. The polyharmonic distortion modeling is a characterizationg technique for nonlinear systems that has been growing up due to praticity and similarity with S-parameters. This work presents analysis the polyharmonic distortion modeling, the test bench development for simulation of planar structure and planar structure characterization with X-parameters
Resumo:
This work study of solar distillation feasibility in effluent of petroleum industry: produced water, making possible your reuse for irrigation of oleaginous cultures or fodder crops or in steam generation, as well the transport phenomena involved. The methodology for development of this project was to characterize the effluent to be treated and to accomplish physical and chemical analysis in the distilled, to build distillation equipment, concomitant operation of both equipments and implementation of data processing and economical evaluation. The methodology used for all parameters is outlined in APHA (1998) and sampling of the type compound. The feeding of distillation equipment was performed with treated effluent from UTPF of Guamaré. The temperature was monitored throughout the distillers and during the time of operation. The distillers feed occur, as a rule, for sifon. The distillers were operated by a period of 17 months between July 2007 and February 2009, in which 40 experiments were performed. The radiation and temperature datas were acquired in the INPE s site and the temperature inside of the distillers was registered by DATALOGGER Novus. The rates of condensation (mL / min) were determined by measuring of the flow in a graduate test tube of 10 mL and a chronometer. We used two simple solar effect distillers of passive type with different angles in coverage: 20 ° and 45 °. The results obtained in this study and the relevant discussions are divided into six topics: sample characterization and quality of distilled; construction of distillers; operation (data, temperature profile), climatic aspects, treatment of data and economical analysis. Results obtained can be inferred that: the energy loss by the adoption of vessel glass was not significant, however, complicates the logistics of maintenance the equipment on a large scale. In the other hand, the surface of the tub with a glass shield on the equipment deterioration, both devices showed similar performance, so there is not justified for use of equipment 450. With regard to the climatological study it was verified that the Natal city presents monthly medium radiation varying in a range between 350 and 600 W/m2, and medium of wind speed of 5 m / s. The medium humidity is around 70% and rainfall is very small. The regime of the system is transient and although it has been treated as a stationary system shows that the model accurately represents the distillers system's 20 degrees. The quality of the distilled with regard to the parameters evaluated in this study is consistent with the Class 3 waters of CONAMA (Resolution 357). Therefore we can conclude that solar distillation has viability for treat oilfield produced water when considered the technical and environmental aspects, although it is not economically viable
Resumo:
With hardware and software technologies advance, it s also happenning modifications in the development models of computational systems. New methodologies for user interface specification are being created with user interface description languages (UIDL). The UIDLs are a way to have a precise description in a language with more abstraction and independent of how will be implemented. A great problem is that even using these nowadays methodologies, we still have a big distance between the UIDLs and its design, what means, the distance between abstract and concrete. The tool BRIDGE (Interface Design Generator Environment) was created with the intention of being a linking bridge between a specification language (the Interactive Message Modeling Language IMML) and its implementation in Java, linking the abstract (specification) to the concrete (implementation). IMML is a language based on models, that allows the designer works in distinct abstraction levels, being each model a distinct abstraction level. IMML is a XML language, that uses the Semiotic Engineering concepts, that deals the computational system, with the user interface and its elements like a metacommunicative artifact, where these elements must to transmit a message to the user about what task must to be realized and the way to reach this goal. With BRIDGE, we intend to supply a lot of support to the design task, being the user interface prototipation the greater of them. BRIDGE allows the design becomes easier and more intuitive coming from an interface specification language
Resumo:
Motion estimation is the main responsible for data reduction in digital video encoding. It is also the most computational damanding step. H.264 is the newest standard for video compression and was planned to double the compression ratio achievied by previous standards. It was developed by the ITU-T Video Coding Experts Group (VCEG) together with the ISO/IEC Moving Picture Experts Group (MPEG) as the product of a partnership effort known as the Joint Video Team (JVT). H.264 presents novelties that improve the motion estimation efficiency, such as the adoption of variable block-size, quarter pixel precision and multiple reference frames. This work defines an architecture for motion estimation in hardware/software, using a full search algorithm, variable block-size and mode decision. This work consider the use of reconfigurable devices, soft-processors and development tools for embedded systems such as Quartus II, SOPC Builder, Nios II and ModelSim
Resumo:
The constant increase of complexity in computer applications demands the development of more powerful hardware support for them. With processor's operational frequency reaching its limit, the most viable solution is the use of parallelism. Based on parallelism techniques and the progressive growth in the capacity of transistors integration in a single chip is the concept of MPSoCs (Multi-Processor System-on-Chip). MPSoCs will eventually become a cheaper and faster alternative to supercomputers and clusters, and applications developed for these high performance systems will migrate to computers equipped with MP-SoCs containing dozens to hundreds of computation cores. In particular, applications in the area of oil and natural gas exploration are also characterized by the high processing capacity required and would benefit greatly from these high performance systems. This work intends to evaluate a traditional and complex application of the oil and gas industry known as reservoir simulation, developing a solution with integrated computational systems in a single chip, with hundreds of functional unities. For this, as the STORM (MPSoC Directory-Based Platform) platform already has a shared memory model, a new distributed memory model were developed. Also a message passing library has been developed folowing MPI standard
Resumo:
Middleware platforms have been widely used as an underlying infrastructure to the development of distributed applications. They provide distribution and heterogeneity transparency and a set of services that ease the construction of distributed applications. Nowadays, the middlewares accommodate an increasing variety of requirements to satisfy distinct application domains. This broad range of application requirements increases the complexity of the middleware, due to the introduction of many cross-cutting concerns in the architecture, which are not properly modularized by traditional programming techniques, resulting in a tangling and spread of theses concerns in the middleware code. The presence of these cross-cutting concerns limits the middleware scalability and aspect-oriented paradigm has been used successfully to improve the modularity, extensibility and customization capabilities of middleware. This work presents AO-OiL, an aspect-oriented (AO) middleware architecture, based on the AO middleware reference architecture. This middleware follows the philosophy that the middleware functionalities must be driven by the application requirements. AO-OiL consists in an AO refactoring of the OiL (Orb in Lua) middleware in order to separate basic and crosscutting concerns. The proposed architecture was implemented in Lua and RE-AspectLua. To evaluate the refactoring impact in the middleware architecture, this paper presents a comparative analysis of performance between AO-OiL and OiL
Resumo:
The use of increasingly complex software applications is demanding greater investment in the development of such systems to ensure applications with better quality. Therefore, new techniques are being used in Software Engineering, thus making the development process more effective. Among these new approaches, we highlight Formal Methods, which use formal languages that are strongly based on mathematics and have a well-defined semantics and syntax. One of these languages is Circus, which can be used to model concurrent systems. It was developed from the union of concepts from two other specification languages: Z, which specifies systems with complex data, and CSP, which is normally used to model concurrent systems. Circus has an associated refinement calculus, which can be used to develop software in a precise and stepwise fashion. Each step is justified by the application of a refinement law (possibly with the discharge of proof obligations). Sometimes, the same laws can be applied in the same manner in different developments or even in different parts of a single development. A strategy to optimize this calculus is to formalise these application as a refinement tactic, which can then be used as a single transformation rule. CRefine was developed to support the Circus refinement calculus. However, before the work presented here, it did not provide support for refinement tactics. The aim of this work is to provide tool support for refinement tactics. For that, we develop a new module in CRefine, which automates the process of defining and applying refinement tactics that are formalised in the tactic language ArcAngelC. Finally, we validate the extension by applying the new module in a case study, which used the refinement tactics in a refinement strategy for verification of SPARK Ada implementations of control systems. In this work, we apply our module in the first two phases of this strategy
Resumo:
The interval datatype applications in several areas is important to construct a interval type reusable, i.e., a interval constructor can be applied to any datatype and get intervals this datatype. Since the interval is, of certain form, a set of elements limited for two bounds, left and right, with a order notions, then it s reasonable that interval constructor enclose datatypes with partial order. On the order hand, what we want is work with interval of any datatype like this we work with this datatype then. it s important to guarantee the properties of the datatype when maps to interval of this datatype. Thus, the interval constructor get a theory to parametrized interval type, i.e., a interval with generics parameters (for example rational, real, complex). Sometimes, the interval application in some algebras doesn t guarantee the mainutenance of their properties, for example, when we use interval of real, that satisfies the field properties, it doesn t guarantee the distributivity propertie. A form to surpass this problem Santiago introduced the local equality theory that weakened the notion of strong equality, and thus, allowing some properties are local keeped, what can be discard before. The interval arithmetic generalization aim to apply the interval constructor on ordered algebras weakened for local equality with the purpose of the keep their properties. How the intervals are important in applications with continuous data, it s interesting specify that theory using a specification language that supply a system development using intervals of form disciplined, trustworth and safe. Currently, the algebraic specification language, based in math models, have been use to that intention often. We choose CASL (Common Algebraic Specification Language) among others languages because CASL has several characteristics excellent to parametrized interval type, such as, provide parcialiy and parametrization
Resumo:
Nowadays due to the security vulnerability of distributed systems, it is needed mechanisms to guarantee the security requirements of distributed objects communications. Middleware Platforms component integration platforms provide security functions that typically offer services for auditing, for guarantee messages protection, authentication, and access control. In order to support these functions, middleware platforms use digital certificates that are provided and managed by external entities. However, most middleware platforms do not define requirements to get, to maintain, to validate and to delegate digital certificates. In addition, most digital certification systems use X.509 certificates that are complex and have a lot of attributes. In order to address these problems, this work proposes a digital certification generic service for middleware platforms. This service provides flexibility via the joint use of public key certificates, to implement the authentication function, and attributes certificates to the authorization function. It also supports delegation. Certificate based access control is transparent for objects. The proposed service defines the digital certificate format, the store and retrieval system, certificate validation and support for delegation. In order to validate the proposed architecture, this work presents the implementation of the digital certification service for the CORBA middleware platform and a case study that illustrates the service functionalities
Resumo:
This work discusses the application of techniques of ensembles in multimodal recognition systems development in revocable biometrics. Biometric systems are the future identification techniques and user access control and a proof of this is the constant increases of such systems in current society. However, there is still much advancement to be developed, mainly with regard to the accuracy, security and processing time of such systems. In the search for developing more efficient techniques, the multimodal systems and the use of revocable biometrics are promising, and can model many of the problems involved in traditional biometric recognition. A multimodal system is characterized by combining different techniques of biometric security and overcome many limitations, how: failures in the extraction or processing the dataset. Among the various possibilities to develop a multimodal system, the use of ensembles is a subject quite promising, motivated by performance and flexibility that they are demonstrating over the years, in its many applications. Givin emphasis in relation to safety, one of the biggest problems found is that the biometrics is permanently related with the user and the fact of cannot be changed if compromised. However, this problem has been solved by techniques known as revocable biometrics, which consists of applying a transformation on the biometric data in order to protect the unique characteristics, making its cancellation and replacement. In order to contribute to this important subject, this work compares the performance of individual classifiers methods, as well as the set of classifiers, in the context of the original data and the biometric space transformed by different functions. Another factor to be highlighted is the use of Genetic Algorithms (GA) in different parts of the systems, seeking to further maximize their eficiency. One of the motivations of this development is to evaluate the gain that maximized ensembles systems by different GA can bring to the data in the transformed space. Another relevant factor is to generate revocable systems even more eficient by combining two or more functions of transformations, demonstrating that is possible to extract information of a similar standard through applying different transformation functions. With all this, it is clear the importance of revocable biometrics, ensembles and GA in the development of more eficient biometric systems, something that is increasingly important in the present day