325 resultados para Sistemas de computação virtual
Resumo:
The main aim of this investigation is to propose the notion of uniform and strong primeness in fuzzy environment. First, it is proposed and investigated the concept of fuzzy strongly prime and fuzzy uniformly strongly prime ideal. As an additional tool, the concept of t/m systems for fuzzy environment gives an alternative way to deal with primeness in fuzzy. Second, a fuzzy version of correspondence theorem and the radical of a fuzzy ideal are proposed. Finally, it is proposed a new concept of prime ideal for Quantales which enable us to deal with primeness in a noncommutative setting.
Resumo:
The main aim of this investigation is to propose the notion of uniform and strong primeness in fuzzy environment. First, it is proposed and investigated the concept of fuzzy strongly prime and fuzzy uniformly strongly prime ideal. As an additional tool, the concept of t/m systems for fuzzy environment gives an alternative way to deal with primeness in fuzzy. Second, a fuzzy version of correspondence theorem and the radical of a fuzzy ideal are proposed. Finally, it is proposed a new concept of prime ideal for Quantales which enable us to deal with primeness in a noncommutative setting.
Resumo:
The Traveling Salesman with Multiple Ridesharing (TSP-MR) is a type of the Capacitated Traveling Salesman, which presents the possibility of sharing seats with passengers taking advantage of the paths the salesman travels through his cycle. The salesman shares the cost of a path with the boarded passengers. This model can portray a real situation in which, for example, drivers are willing to share parts of a trip with tourists that wish to move between two locations visited by the driver’s route, accepting to share the vehicle with other individuals visiting other locations within the cycle. This work proposes a mathematical formulation for the problem, and an exact and metaheuristics algorithms for its solution, comparing them.
Resumo:
The Traveling Salesman with Multiple Ridesharing (TSP-MR) is a type of the Capacitated Traveling Salesman, which presents the possibility of sharing seats with passengers taking advantage of the paths the salesman travels through his cycle. The salesman shares the cost of a path with the boarded passengers. This model can portray a real situation in which, for example, drivers are willing to share parts of a trip with tourists that wish to move between two locations visited by the driver’s route, accepting to share the vehicle with other individuals visiting other locations within the cycle. This work proposes a mathematical formulation for the problem, and an exact and metaheuristics algorithms for its solution, comparing them.
Resumo:
Soft skills and teamwork practices were identi ed as the main de ciencies of recent graduates in computer courses. This issue led to a realization of a qualitative research aimed at investigating the challenges faced by professors of those courses in conducting, monitoring and assessing collaborative software development projects. Di erent challenges were reported by teachers, including di culties in the assessment of students both in the collective and individual levels. In this context, a quantitative research was conducted with the aim to map soft skill of students to a set of indicators that can be extracted from software repositories using data mining techniques. These indicators are aimed at measuring soft skills, such as teamwork, leadership, problem solving and the pace of communication. Then, a peer assessment approach was applied in a collaborative software development course of the software engineering major at the Federal University of Rio Grande do Norte (UFRN). This research presents a correlation study between the students' soft skills scores and indicators based on mining software repositories. This study contributes: (i) in the presentation of professors' perception of the di culties and opportunities for improving management and monitoring practices in collaborative software development projects; (ii) in investigating relationships between soft skills and activities performed by students using software repositories; (iii) in encouraging the development of soft skills and the use of software repositories among software engineering students; (iv) in contributing to the state of the art of three important areas of software engineering, namely software engineering education, educational data mining and human aspects of software engineering.
Resumo:
Soft skills and teamwork practices were identi ed as the main de ciencies of recent graduates in computer courses. This issue led to a realization of a qualitative research aimed at investigating the challenges faced by professors of those courses in conducting, monitoring and assessing collaborative software development projects. Di erent challenges were reported by teachers, including di culties in the assessment of students both in the collective and individual levels. In this context, a quantitative research was conducted with the aim to map soft skill of students to a set of indicators that can be extracted from software repositories using data mining techniques. These indicators are aimed at measuring soft skills, such as teamwork, leadership, problem solving and the pace of communication. Then, a peer assessment approach was applied in a collaborative software development course of the software engineering major at the Federal University of Rio Grande do Norte (UFRN). This research presents a correlation study between the students' soft skills scores and indicators based on mining software repositories. This study contributes: (i) in the presentation of professors' perception of the di culties and opportunities for improving management and monitoring practices in collaborative software development projects; (ii) in investigating relationships between soft skills and activities performed by students using software repositories; (iii) in encouraging the development of soft skills and the use of software repositories among software engineering students; (iv) in contributing to the state of the art of three important areas of software engineering, namely software engineering education, educational data mining and human aspects of software engineering.
Resumo:
The Information Technology (IT) is increasing his applicability to business, both private and public companies. It is necessary the adequate use of the new technologies and get cooperation and technology acceptance of the system. People tend to resist to the changes, contributing so that the technology is rejected or even it is not recognized as promoting of the changes. This study is relevant and aim to evaluate the impacts of new technologies, considering their users as fundamental factors in the change process. The survey analyzed the advantages and the barriers of the system use in three federal special judicial of Rio Grande do Norte, with data collected in May of 2007, through the application of questionnaires to thirty eight users of the virtual system CRETA. The users' perception was evaluated, under the optics of five variables: efficiency, image, agility, ease of use and quality. Starting from the obtained results, it was evidenced that the implementation of the system felt accordingly the expected and it reached the objectives intended that were: the velocity and efficiency in the path of the lawsuits, larger productivity, resulting in a better quality of the final work introduced to the citizen and proportionate an improvement in the organizational image of the judiciary power
Resumo:
Furthered mainly by new technologies, the expansion of distance education has created a demand for tools and methodologies to enhance teaching techniques based on proven pedagogical theories. Such methodologies must also be applied in the so-called Virtual Learning Environments. The aim of this work is to present a planning methodology based on known pedagogical theories which contributes to the incorporation of assessment in the process of teaching and learning. With this in mind, the pertinent literature was reviewed in order to identify the key pedagogical concepts needed to the definition of this methodology and a descriptive approach was used to establish current relations between this conceptual framework and distance education. As a result of this procedure, the Contents Map and the Dependence Map were specified and implemented, two teaching tools that promote the planning of a course by taking into account assessment still in this early stage. Inserted on Moodle, the developed tools were tested in a course of distance learning for practical observation of the involved concepts. It could be verified that the methodology proposed by the above-mentioned tools is in fact helpful in course planning and in strengthening educational assessment, placing the student as central element in the process of teaching and learning
Resumo:
In the recovering process of oil, rock heterogeneity has a huge impact on how fluids move in the field, defining how much oil can be recovered. In order to study this variability, percolation theory, which describes phenomena involving geometry and connectivity are the bases, is a very useful model. Result of percolation is tridimensional data and have no physical meaning until visualized in form of images or animations. Although a lot of powerful and sophisticated visualization tools have been developed, they focus on generation of planar 2D images. In order to interpret data as they would be in the real world, virtual reality techniques using stereo images could be used. In this work we propose an interactive and helpful tool, named ZSweepVR, based on virtual reality techniques that allows a better comprehension of volumetric data generated by simulation of dynamic percolation. The developed system has the ability to render images using two different techniques: surface rendering and volume rendering. Surface rendering is accomplished by OpenGL directives and volume rendering is accomplished by the Zsweep direct volume rendering engine. In the case of volumetric rendering, we implemented an algorithm to generate stereo images. We also propose enhancements in the original percolation algorithm in order to get a better performance. We applied our developed tools to a mature field database, obtaining satisfactory results. The use of stereoscopic and volumetric images brought valuable contributions for the interpretation and clustering formation analysis in percolation, what certainly could lead to better decisions about the exploration and recovery process in oil fields
Resumo:
The lava Platform is increasing1y being adopted in the development of distributed sys¬tems with higb user demando This kind of application is more complex because it needs beyond attending the functional requirements, to fulfil1 the pre-established performance parameters. This work makes a study on the Java Vutual Machine (JVM), approaching its intemal aspects and exploring the garbage collection strategies existing in the literature and used by the NM. It also presents a set of tools that helps in the job of optimizing applications and others that help in the monitoring of applications in the production envi¬ronment. Doe to the great amount of technologies that aim to solve problems which are common to the application layer, it becomes difficult to choose the one with best time response and less memory usage. This work presents a brief introduction to each one of tbe possible technologies and realize comparative tests through a statistical analysis of the response time and garbage collection activity random variables. The obtained results supply engineers and managers with a subside to decide which technologies to use in large applications through the knowledge of how they behave in their environments and the amount of resources that they consume. The relation between the productivity of the technology and its performance is also considered ao important factor in this choice
Resumo:
We propose a robotics simulation platform, named S-Educ, developed specifically for application in educational robotics, which can be used as an alternative or in association with robotics kits in classes involving the use of robotics. In the usually known approach, educational robotics uses robotics kits for classes which generally include interdisciplinary themes. The idea of this work is not to replace these kits, but to use the developed simulator as an alternative, where, for some reason, the traditional kits cannot be used, or even to use the platform in association with these kits. To develop the simulator, initially, we conducted research in the literature on the use of robotic simulators and robotic kits, facing the education sector, from which it was possible to define a set of features considered important for creating such a tool. Then, on the software development phase, the simulator S-Educ was implemented, taking into account the requirements and features defined in the design phase. Finally, to validate the platform, several tests were conducted with teachers, students and lay adults, in which it was used the simulator S-Educ, to evaluate its use in educational robotics classes. The results show that robotic simulator allows a reduction of financial costs, facilitate testing and reduce robot damage inherent to its use, in addition to other advantages. Furthermore, as a contribution to the community, the proposed tool can be used to increase adhesion of Brazilian schools to the methodologies of educational robotics or to robotics competitions
Resumo:
The Information Technology (IT) is increasing his applicability to business, both private and public companies. It is necessary the adequate use of the new technologies and get cooperation and technology acceptance of the system. People tend to resist to the changes, contributing so that the technology is rejected or even it is not recognized as promoting of the changes. This study is relevant and aim to evaluate the impacts of new technologies, considering their users as fundamental factors in the change process. The survey analyzed the advantages and the barriers of the system use in three federal special judicial of Rio Grande do Norte, with data collected in May of 2007, through the application of questionnaires to thirty eight users of the virtual system CRETA. The users' perception was evaluated, under the optics of five variables: efficiency, image, agility, ease of use and quality. Starting from the obtained results, it was evidenced that the implementation of the system felt accordingly the expected and it reached the objectives intended that were: the velocity and efficiency in the path of the lawsuits, larger productivity, resulting in a better quality of the final work introduced to the citizen and proportionate an improvement in the organizational image of the judiciary power
Resumo:
In this work, we propose a solution to solve the scalability problem found in collaborative, virtual and mixed reality environments of large scale, that use the hierarchical client-server model. Basically, we use a hierarchy of servers. When the capacity of a server is reached, a new server is created as a sun of the first one, and the system load is distributed between them (father and sun). We propose efficient tools and techniques for solving problems inherent to client-server model, as the definition of clusters of users, distribution and redistribution of users through the servers, and some mixing and filtering operations, that are necessary to reduce flow between servers. The new model was tested, in simulation, emulation and in interactive applications that were implemented. The results of these experimentations show enhancements in the traditional, previous models indicating the usability of the proposed in problems of all-to-all communications. This is the case of interactive games and other applications devoted to Internet (including multi-user environments) and interactive applications of the Brazilian Digital Television System, to be developed by the research group. Keywords: large scale virtual environments, interactive digital tv, distributed
Sistema de detecção e isolamento de falhas em sistemas dinâmicos baseado em identificação paramétrica
Resumo:
The present research aims at contributing to the area of detection and diagnosis of failure through the proposal of a new system architecture of detection and isolation of failures (FDI, Fault Detection and Isolation). The proposed architecture presents innovations related to the way the physical values monitored are linked to the FDI system and, as a consequence, the way the failures are detected, isolated and classified. A search for mathematical tools able to satisfy the objectives of the proposed architecture has pointed at the use of the Kalman Filter and its derivatives EKF (Extended Kalman Filter) and UKF (Unscented Kalman Filter). The use of the first one is efficient when the monitored process presents a linear relation among its physical values to be monitored and its out-put. The other two are proficient in case this dynamics is no-linear. After that, a short comparative of features and abilities in the context of failure detection concludes that the UFK system is a better alternative than the EKF one to compose the architecture of the FDI system proposed in case of processes of no-linear dynamics. The results shown in the end of the research refer to the linear and no-linear industrial processes. The efficiency of the proposed architecture may be observed since it has been applied to simulated and real processes. To conclude, the contributions of this thesis are found in the end of the text
Resumo:
Internet applications such as media streaming, collaborative computing and massive multiplayer are on the rise,. This leads to the need for multicast communication, but unfortunately group communications support based on IP multicast has not been widely adopted due to a combination of technical and non-technical problems. Therefore, a number of different application-layer multicast schemes have been proposed in recent literature to overcome the drawbacks. In addition, these applications often behave as both providers and clients of services, being called peer-topeer applications, and where participants come and go very dynamically. Thus, servercentric architectures for membership management have well-known problems related to scalability and fault-tolerance, and even peer-to-peer traditional solutions need to have some mechanism that takes into account member's volatility. The idea of location awareness distributes the participants in the overlay network according to their proximity in the underlying network allowing a better performance. Given this context, this thesis proposes an application layer multicast protocol, called LAALM, which takes into account the actual network topology in the assembly process of the overlay network. The membership algorithm uses a new metric, IPXY, to provide location awareness through the processing of local information, and it was implemented using a distributed shared and bi-directional tree. The algorithm also has a sub-optimal heuristic to minimize the cost of membership process. The protocol has been evaluated in two ways. First, through an own simulator developed in this work, where we evaluated the quality of distribution tree by metrics such as outdegree and path length. Second, reallife scenarios were built in the ns-3 network simulator where we evaluated the network protocol performance by metrics such as stress, stretch, time to first packet and reconfiguration group time