21 resultados para Nationella prov
em Universidade Federal do Rio Grande do Norte(UFRN)
Resumo:
Study public space is studying the city and its dynamics through a representation point. The changes that are observed in these spaces and in the relations established there, or with it, are related to changes that occurred in the city as a whole, in the urban way of living. Study public space allied to urban sociability permits an even better focus on this relation between people in these spaces and with the spaces and in the consequences and fragile aspects this relation can impose to people and cities. This work is aimed towards this relation between public space and urban sociability. Through a conceptual/theoretical discussion aims, specifically, comprehend how to establish and what is the relation between urban space and urban sociability (appropriation/perception of place, tendencies of seclusion), from an isolated analysis of each one of these themes in the contemporaneous city. The Brazilian public space has peculiar characteristics, arising from its historical process of construction, also influenced by the public sphere fragile construction that permeates it, as well as more recent variables such as the sensation of insecurity and commodification of these places. The urban sociability influences and is influenced along all this process of significance and reframing of the public space. The conceptual discussion about each isolate variable provides the necessary coverage for discussion and analysis of the relation between them and the consequence of this relation in the city, such as the notation of relevant experiences of this process of revaluation of the public space. The hereby search is not through a path to reestablish the relation as it already existed, but also for an understanding of the dynamic as is established today and for existing possibilities for maintenance and appreciation of the relation between people and the city by believing in its importance to urban life
Resumo:
Because of social exclusion in Brazil and having as focus the digital inclusion, was started in Federal University of Rio Grande do Norte a project that could talk, at the same time, about concepts of collaborative learning and educational robotics , focused on children digitally excluded. In this context was created a methodology that approaches many subjects as technological elements (e. g. informatics and robotics) and school subjects (e. g. Portuguese, Mathematics, Geography, History), contextualized in everyday situations. We observed educational concepts of collaborative learning and the development of capacities from those students, as group work, logical knowledge and learning ability. This paper proposes an educational software for robotics teaching called RoboEduc, created to be used by children digitally excluded from primary school. Its introduction prioritizes a friendly interface, that makes the concepts of robotics and programming easy and fun to be taught. With this new tool, users without informatics or robotics previous knowledge are able to control a robot, previously set with Lego kits, or even program it to carry some activities out. This paper provides the implementation of the second version of the software. This version presents the control of the robot already used. After were implemented the different levels of programming linked to the many learning levels of the users and their different interfaces and functions. Nowadays, has been implemented the third version, with the improvement of each one of the mentioned stages. In order to validate, prove and test the efficience of the developed methodology to the RoboEduc, were made experiments, through practice of robotics, with children for fourth and fifth grades of primary school at the City School Professor Ascendino de Almeida, in the suburb of Natal (west zone), Rio Grande do Norte. As a preliminary result of the current technology, we verified that the use of robots associated with a well elaborated software can be spread to users that know very little about the subject, without the necessity of previous advanced technology knowledges. Therefore, they showed to be accessible and efficient tools in the process of digital inclusion
Resumo:
The sharing of knowledge and integration of data is one of the biggest challenges in health and essential contribution to improve the quality of health care. Since the same person receives care in various health facilities throughout his/her live, that information is distributed in different information systems which run on platforms of heterogeneous hardware and software. This paper proposes a System of Health Information Based on Ontologies (SISOnt) for knowledge sharing and integration of data on health, which allows to infer new information from the heterogeneous databases and knowledge base. For this purpose it was created three ontologies represented by the patterns and concepts proposed by the Semantic Web. The first ontology provides a representation of the concepts of diseases Secretariat of Health Surveillance (SVS) and the others are related to the representation of the concepts of databases of Health Information Systems (SIS), specifically the Information System of Notification of Diseases (SINAN) and the Information System on Mortality (SIM)
Resumo:
This paper describes the design, implementation and enforcement of a system for industrial process control based on fuzzy logic and developed using Java, with support for industrial communication protocol through the OPC (Ole for Process Control). Besides the java framework, the software is completely independent from other platforms. It provides friendly and functional tools for modeling, construction and editing of complex fuzzy inference systems, and uses these logical systems in control of a wide variety of industrial processes. The main requirements of the developed system should be flexibility, robustness, reliability and ease of expansion
Resumo:
This work treats of an implementation OFDMA baseband processor in hardware for LTE Downlink. The LTE or Long Term Evolution consist the last stage of development of the technology called 3G (Mobile System Third Generation) which offers an increasing in data rate and more efficiency and flexibility in transmission with application of advanced antennas and multiple carriers techniques. This technology applies in your physical layer the OFDMA technical (Orthogonal Frequency Division Multiple Access) for generation of signals and mapping of physical resources in downlink and has as base theoretical to OFDM multiple carriers technique (Orthogonal Frequency Division Multiplexing). With recent completion of LTE specifications, different hardware solutions have been developed, mainly, to the level symbol processing where the implementation of OFDMA processor in base band is commonly considered, because it is also considered a basic architecture of others important applications. For implementation of processor, the reconfigurable hardware offered by devices as FPGA are considered which shares not only to meet the high requirements of flexibility and adaptability of LTE as well as offers possibility of an implementation quick and efficient. The implementation of processor in reconfigurable hardware meets the specifications of LTE physical layer as well as have the flexibility necessary for to meet others standards and application which use OFDMA processor as basic architecture for your systems. The results obtained through of simulation and verification functional system approval the functionality and flexibility of processor implemented
Resumo:
Apresentamos, neste trabalho, com base na semântica cognitiva, uma análise do significado, em contexto, dos auxiliares modais poder, precisar e dever. Analisamos 120 textos produzidos por candidatos ao vestibular e por alunos do ensino fundamental, como resposta da questão número três da prova discursiva de Língua Portuguesa do vestibular 2005 da UFRN, que pede aos candidatos para explicitar a diferença de sentido entre três frases, observando o uso desses três verbos. Consideramos que um item lexical não é incorporado a uma representação lingüística semântica fixa, limitada e única, mas antes, é ligado a uma representação lingüística semântica flexível e aberta que provê acesso a muitas concepções e sistemas conceituais dependente de cada contexto determinado. Com base em seu significado, um item lexical evoca um grupo de domínios cognitivos, que por sua vez, apresentam um determinado conteúdo conceitual. Isto implica em afirmar que a rede de significados lexicais vai variar conforme o conhecimento de mundo de cada um (LANGACKER, 2000). A relevância deste trabalho é proporcionar uma contribuição para a descrição semântica do português
Resumo:
The Quadratic Minimum Spanning Tree Problem (QMST) is a version of the Minimum Spanning Tree Problem in which, besides the traditional linear costs, there is a quadratic structure of costs. This quadratic structure models interaction effects between pairs of edges. Linear and quadratic costs are added up to constitute the total cost of the spanning tree, which must be minimized. When these interactions are restricted to adjacent edges, the problem is named Adjacent Only Quadratic Minimum Spanning Tree (AQMST). AQMST and QMST are NP-hard problems that model several problems of transport and distribution networks design. In general, AQMST arises as a more suitable model for real problems. Although, in literature, linear and quadratic costs are added, in real applications, they may be conflicting. In this case, it may be interesting to consider these costs separately. In this sense, Multiobjective Optimization provides a more realistic model for QMST and AQMST. A review of the state-of-the-art, so far, was not able to find papers regarding these problems under a biobjective point of view. Thus, the objective of this Thesis is the development of exact and heuristic algorithms for the Biobjective Adjacent Only Quadratic Spanning Tree Problem (bi-AQST). In order to do so, as theoretical foundation, other NP-hard problems directly related to bi-AQST are discussed: the QMST and AQMST problems. Bracktracking and branch-and-bound exact algorithms are proposed to the target problem of this investigation. The heuristic algorithms developed are: Pareto Local Search, Tabu Search with ejection chain, Transgenetic Algorithm, NSGA-II and a hybridization of the two last-mentioned proposals called NSTA. The proposed algorithms are compared to each other through performance analysis regarding computational experiments with instances adapted from the QMST literature. With regard to exact algorithms, the analysis considers, in particular, the execution time. In case of the heuristic algorithms, besides execution time, the quality of the generated approximation sets is evaluated. Quality indicators are used to assess such information. Appropriate statistical tools are used to measure the performance of exact and heuristic algorithms. Considering the set of instances adopted as well as the criteria of execution time and quality of the generated approximation set, the experiments showed that the Tabu Search with ejection chain approach obtained the best results and the transgenetic algorithm ranked second. The PLS algorithm obtained good quality solutions, but at a very high computational time compared to the other (meta)heuristics, getting the third place. NSTA and NSGA-II algorithms got the last positions
Resumo:
On the last years, several middleware platforms for Wireless Sensor Networks (WSN) were proposed. Most of these platforms does not consider issues of how integrate components from generic middleware architectures. Many requirements need to be considered in a middleware design for WSN and the design, in this case, it is possibility to modify the source code of the middleware without changing the external behavior of the middleware. Thus, it is desired that there is a middleware generic architecture that is able to offer an optimal configuration according to the requirements of the application. The adoption of middleware based in component model consists of a promising approach because it allows a better abstraction, low coupling, modularization and management features built-in middleware. Another problem present in current middleware consists of treatment of interoperability with external networks to sensor networks, such as Web. Most current middleware lacks the functionality to access the data provided by the WSN via the World Wide Web in order to treat these data as Web resources, and they can be accessed through protocols already adopted the World Wide Web. Thus, this work presents the Midgard, a component-based middleware specifically designed for WSNs, which adopts the architectural patterns microkernel and REST. The microkernel architectural complements the component model, since microkernel can be understood as a component that encapsulates the core system and it is responsible for initializing the core services only when needed, as well as remove them when are no more needed. Already REST defines a standardized way of communication between different applications based on standards adopted by the Web and enables him to treat WSN data as web resources, allowing them to be accessed through protocol already adopted in the World Wide Web. The main goals of Midgard are: (i) to provide easy Web access to data generated by WSN, exposing such data as Web resources, following the principles of Web of Things paradigm and (ii) to provide WSN application developer with capabilities to instantiate only specific services required by the application, thus generating a customized middleware and saving node resources. The Midgard allows use the WSN as Web resources and still provide a cohesive and weakly coupled software architecture, addressing interoperability and customization. In addition, Midgard provides two services needed for most WSN applications: (i) configuration and (ii) inspection and adaptation services. New services can be implemented by others and easily incorporated into the middleware, because of its flexible and extensible architecture. According to the assessment, the Midgard provides interoperability between the WSN and external networks, such as web, as well as between different applications within a single WSN. In addition, we assessed the memory consumption, the application image size, the size of messages exchanged in the network, and response time, overhead and scalability on Midgard. During the evaluation, the Midgard proved satisfies their goals and shown to be scalable without consuming resources prohibitively
Resumo:
The development of smart card applications requires a high level of reliability. Formal methods provide means for this reliability to be achieved. The BSmart method and tool contribute to the development of smart card applications with the support of the B method, generating Java Card code from B specifications. For the development with BSmart to be effectively rigorous without overloading the user it is important to have a library of reusable components built in B. The goal of KitSmart is to provide this support. A first research about the composition of this library was a graduation work from Universidade Federal do Rio Grande do Norte, made by Thiago Dutra in 2006. This first version of the kit resulted in a specification of Java Card primitive types byte, short and boolean in B and the creation of reusable components for application development. This work provides an improvement of KitSmart with the addition of API Java Card specification made in B and a guide for the creation of new components. The API Java Card in B, besides being available to be used for development of applications, is also useful as a documentation of each API class. The reusable components correspond to modules to manipulate specific structures, such as date and time. These structures are not available for B or Java Card. These components for Java Card are generated from specifications formally verified in B. The guide contains quick reference on how to specify some structures and how some situations were adapted from object-orientation to the B Method. This work was evaluated through a case study made through the BSmart tool, that makes use of the KitSmart library. In this case study, it is possible to see the contribution of the components in a B specification. This kit should be useful for B method users and Java Card application developers
Resumo:
With the advance of the Cloud Computing paradigm, a single service offered by a cloud platform may not be enough to meet all the application requirements. To fulfill such requirements, it may be necessary, instead of a single service, a composition of services that aggregates services provided by different cloud platforms. In order to generate aggregated value for the user, this composition of services provided by several Cloud Computing platforms requires a solution in terms of platforms integration, which encompasses the manipulation of a wide number of noninteroperable APIs and protocols from different platform vendors. In this scenario, this work presents Cloud Integrator, a middleware platform for composing services provided by different Cloud Computing platforms. Besides providing an environment that facilitates the development and execution of applications that use such services, Cloud Integrator works as a mediator by providing mechanisms for building applications through composition and selection of semantic Web services that take into account metadata about the services, such as QoS (Quality of Service), prices, etc. Moreover, the proposed middleware platform provides an adaptation mechanism that can be triggered in case of failure or quality degradation of one or more services used by the running application in order to ensure its quality and availability. In this work, through a case study that consists of an application that use services provided by different cloud platforms, Cloud Integrator is evaluated in terms of the efficiency of the performed service composition, selection and adaptation processes, as well as the potential of using this middleware in heterogeneous computational clouds scenarios
Resumo:
The occurrence of problems related to the scattering and tangling phenomenon, such as the difficulty to do system maintenance, increasingly frequent. One way to solve this problem is related to the crosscutting concerns identification. To maximize its benefits, the identification must be performed from early stages of development process, but some works have reported that this has not been done in most of cases, making the system development susceptible to the errors incidence and prone to the refactoring later. This situation affects directly to the quality and cost of the system. PL-AOVgraph is a goal-oriented requirements modeling language which offers support to the relationships representation among requirements and provides separation of crosscutting concerns by crosscutting relationships representation. Therefore, this work presents a semi-automatic method to crosscutting concern identification in requirements specifications written in PL-AOVgraph. An adjacency matrix is used to identify the contributions relationships among the elements. The crosscutting concern identification is based in fan-out analysis of contribution relationships from the informations of adjacency matrix. When identified, the crosscutting relationships are created. And also, this method is implemented as a new module of ReqSys-MDD tool
Resumo:
Alongside the advances of technologies, embedded systems are increasingly present in our everyday. Due to increasing demand for functionalities, many tasks are split among processors, requiring more efficient communication architectures, such as networks on chip (NoC). The NoCs are structures that have routers with channel point-to-point interconnect the cores of system on chip (SoC), providing communication. There are several networks on chip in the literature, each with its specific characteristics. Among these, for this work was chosen the Integrated Processing System NoC (IPNoSyS) as a network on chip with different characteristics compared to general NoCs, because their routing components also accumulate processing function, ie, units have functional able to execute instructions. With this new model, packets are processed and routed by the router architecture. This work aims at improving the performance of applications that have repetition, since these applications spend more time in their execution, which occurs through repeated execution of his instructions. Thus, this work proposes to optimize the runtime of these structures by employing a technique of instruction-level parallelism, in order to optimize the resources offered by the architecture. The applications are tested on a dedicated simulator and the results compared with the original version of the architecture, which in turn, implements only packet level parallelism
Resumo:
The use of Multiple Input Multiple Output (MIMO) systems has permitted the recent evolution of wireless communication standards. The Spatial Multiplexing MIMO technique, in particular, provides a linear gain at the transmission capacity with the minimum between the numbers of transmit and receive antennas. To obtain a near capacity performance in SM-MIMO systems a soft decision Maximum A Posteriori Probability MIMO detector is necessary. However, such detector is too complex for practical solutions. Hence, the goal of a MIMO detector algorithm aimed for implementation is to get a good approximation of the ideal detector while keeping an acceptable complexity. Moreover, the algorithm needs to be mapped to a VLSI architecture with small area and high data rate. Since Spatial Multiplexing is a recent technique, it is argued that there is still much room for development of related algorithms and architectures. Therefore, this thesis focused on the study of sub optimum algorithms and VLSI architectures for broadband MIMO detector with soft decision. As a result, novel algorithms have been developed starting from proposals of optimizations for already established algorithms. Based on these results, new MIMO detector architectures with configurable modulation and competitive area, performance and data rate parameters are here proposed. The developed algorithms have been extensively simulated and the architectures were synthesized so that the results can serve as a reference for other works in the area
Resumo:
The Borborema Province, Northeastern Brazil, had its internal structure investigated by different geophysical methods like gravity, magnetics and seismics. Additionally, many geological studies were also carried out to define the structural domains of this province. Despite the plethora of studies, there are still many important open aspects about its evolution. Here, we study the velocity structure of S-wave in the crust using dispersion of surface waves. The dispersion of surface waves allows an estimate of the average thickness of the crust across the region between the stations. The inversion of the velocity structure was carried out using the inter-station dispersion of surface waves of Rayleigh and Love types. The teleseismic events are mainly from the edges of the South and North American plates. The period of data collection occurred between 2007 and 2010 and we selected 7 events with magnitude above 5.0 MW and up to 40 km depth. The difference between the events back-azimuths and the interstation path was not greater than 10. We also know the depth of the Moho, results from Receiver Functions (Novo Barbosa, 2008), and use those as constrains in inversion. Even using different parameterizations of models for the inversion, our results were very similar the mean profiles velocity structure of S-wave. In pairs of stations located in the Cear´a Central Domain Borborema the province, there are ranges of depths for which the velocities of S are very close. Most of the results in the profile near the Moho complicate their interpretation at that depth, coinciding with the geology of the region, where there are many shear zones. In particular, the profile that have the route Potiguar Bacia in inter-station, had low velocities in the crust. We combine these results to the results of gravimetry and magnetometry (Oliveira, 2008) and receptor function (Novo Barbosa, 2008). We finally, the first results on the behavior of the velocity structure of S-wave with depth in the Province Borborema
Resumo:
About 10% of faults involving the electrical system occurs in power transformers. Therefore, the protection applied to the power transformers is essential to ensure the continuous operation of this device and the efficiency of the electrical system. Among the protection functions applied to power transformers, the differential protection appears as one of the main schemes, presenting reliable discrimination between internal faults and external faults or inrush currents. However, when using the low frequency components of the differential currents flowing through the transformer, the main difficulty of the conventional methods of differential protection is the delay for detection of the events. However, internal faults, external faults and other disturbances related to the transformer operation present transient and can be appropriately detected by the wavelet transform. In this paper is proposed the development of a wavelet-based differential protection for detection and identification of external faults to the transformer, internal faults, and transformer energizing by using the wavelet coefficient energy of the differential currents. The obtained results reveal the advantages of using of the wavelet transform in the differential protection compared to conventional protection, since it provides reliability and speed in detection of these events.