22 resultados para Parallel programming (computer science)
em Repositório Institucional UNESP - Universidade Estadual Paulista "Julio de Mesquita Filho"
Resumo:
This paper describes an innovative approach to develop the understanding about the relevance of mathematics to computer science. The mathematical subjects are introduced through an application-to-model scheme that lead computer science students to a better understanding of why they have to learn math and learn it effectively. Our approach consists of a single one semester course, taught at the first semester of the program, where the students are initially exposed to some typical computer applications. When they recognize the applications' complexity, the instructor gives the mathematical models supporting such applications, even before a formal introduction to the model in a math course. We applied this approach at Unesp (Brazil) and the results include a large reduction in the rate of students that abandon the college and better students in the final years of our program.
Resumo:
This paper describes an innovative approach to establish a CS curriculum, aiming flexibility and minimization of the time spent in the classrooms. This approach has been developed at the Paulista State University - Unesp - at São José do Rio Preto, and is producing very interesting results. The load reduction is achieved through a series of fundamental core and breadth courses that precede depth courses in specific areas. The flexibility comes as a side effect of the depth courses, which can be adapted without any changes in the core courses. In the following pages we fully describe our motivations, actions and results.
Resumo:
New information and communication technologies may be useful for providing more in-depth knowledge to students in many ways, whether through online multimedia educational material, or through online debates with colleagues, teachers and other area professionals in a synchronous or asynchronous manner. This paper focuses on participation in online discussion in e-learning courses for promoting learning. Although an important theoretical aspect, an analysis of literature reveals there are few studies evaluating the personal and social aspects of online course users in a quantitative manner. This paper aims to introduce a method for diagnosing inclusion and digital proficiency and other personal aspects of the student through a case study comparing Information System, Public Relations and Engineering students at a public university in Brazil. Statistical analysis and analysis of variances (ANOVA) were used as the methodology for data analysis in order to understand existing relations between the components of the proposed method. The survey methodology was also used, in its online format, as a research instrument. The method is based on using online questionnaires that diagnose digital proficiency and time management, level of extroversion and social skills of the students. According to the sample studied, there is no strong correlation between digital proficiency and individual characteristics tied to the use of time, level of extroversion and social skills of students. The differences in course grades for some components are partly due to subject 'Introduction to Economics' being offered to freshmen in Public Relations, whereas subject 'Economics in Engineering' is offered in the final semesters of Engineering and Information Systems courses. Therefore, the difference could be more tied to the respondent's age than to the course. Information Systems students were observed to be older, with access to computers and Internet at the workplace, compared to the other students who access the Internet more often from home. This paper presents a pilot study aimed at conducting a diagnosis that permits proposing actions for information and communication technology to contribute towards student education. Three levels of digital inclusion are described as a scale to measure whether information technology increases personal performance and professional knowledge and skills. This study may be useful for other readers interested in themes related to education in engineering. © 2013 IEEE.
Resumo:
Software Transactional Memory (STM) systems have poor performance under high contention scenarios. Since many transactions compete for the same data, most of them are aborted, wasting processor runtime. Contention management policies are typically used to avoid that, but they are passive approaches as they wait for an abort to happen so they can take action. More proactive approaches have emerged, trying to predict when a transaction is likely to abort so its execution can be delayed. Such techniques are limited, as they do not replace the doomed transaction by another or, when they do, they rely on the operating system for that, having little or no control on which transaction should run. In this paper we propose LUTS, a Lightweight User-Level Transaction Scheduler, which is based on an execution context record mechanism. Unlike other techniques, LUTS provides the means for selecting another transaction to run in parallel, thus improving system throughput. Moreover, it avoids most of the issues caused by pseudo parallelism, as it only launches as many system-level threads as the number of available processor cores. We discuss LUTS design and present three conflict-avoidance heuristics built around LUTS scheduling capabilities. Experimental results, conducted with STMBench7 and STAMP benchmark suites, show LUTS efficiency when running high contention applications and how conflict-avoidance heuristics can improve STM performance even more. In fact, our transaction scheduling techniques are capable of improving program performance even in overloaded scenarios. © 2011 Springer-Verlag.
Resumo:
Software transaction memory (STM) systems have been used as an approach to improve performance, by allowing the concurrent execution of atomic blocks. However, under high-contention workloads, STM-based systems can considerably degrade performance, as transaction conflict rate increases. Contention management policies have been used as a way to select which transaction to abort when a conflict occurs. In general, contention managers are not capable of avoiding conflicts, as they can only select which transaction to abort and the moment it should restart. Since contention managers act only after a conflict is detected, it becomes harder to effectively increase transaction throughput. More proactive approaches have emerged, aiming at predicting when a transaction is likely to abort, postponing its execution. Nevertheless, most of the proposed proactive techniques are limited, as they do not replace the doomed transaction by another or, when they do, they rely on the operating system for that, having little or no control on which transaction to run. This article proposes LUTS, a lightweight user-level transaction scheduler. Unlike other techniques, LUTS provides the means for selecting another transaction to run in parallel, thus improving system throughput. We discuss LUTS design and propose a dynamic conflict-avoidance heuristic built around its scheduling capabilities. Experimental results, conducted with the STAMP and STMBench7 benchmark suites, running on TinySTM and SwissTM, show how our conflict-avoidance heuristic can effectively improve STM performance on high contention applications. © 2012 Springer Science+Business Media, LLC.
Resumo:
Research on the micro-structural characterization of metal-matrix composites uses X-ray computed tomography to collect information about the interior features of the samples, in order to elucidate their exhibited properties. The tomographic raw data needs several steps of computational processing in order to eliminate noise and interference. Our experience with a program (Tritom) that handles these questions has shown that in some cases the processing steps take a very long time and that it is not easy for a Materials Science specialist to interact with Tritom in order to define the most adequate parameter values and the proper sequence of the available processing steps. For easing the use of Tritom, a system was built which addresses the aspects described before and that is based on the OpenDX visualization system. OpenDX visualization facilities constitute a great benefit to Tritom. The visual programming environment of OpenDX allows an easy definition of a sequence of processing steps thus fulfilling the requirement of an easy use by non-specialists on Computer Science. Also the possibility of incorporating external modules in a visual OpenDX program allows the researchers to tackle the aspect of reducing the long execution time of some processing steps. The longer processing steps of Tritom have been parallelized in two different types of hardware architectures (message-passing and shared-memory); the corresponding parallel programs can be easily incorporated in a sequence of processing steps defined in an OpenDX program. The benefits of our system are illustrated through an example where the tool is applied in the study of the sensitivity to crushing – and the implications thereof – of the reinforcements used in a functionally graded syntactic metallic foam.
Resumo:
This work describes a ludic proposal for programming learning of industrial robots to be developed by groups of engineering students. Two projects are presented: Tic-tac-toe Opponent Robot and Environmentalist Robot. The first project use competitive search techniques of the Artificial Intelligence, computational vision, electronic and pneumatic concepts for ability decision making for a robotic agent on the tic-tae-toe game. The second project consists of a game that contains a questions and answers database about environmental themes. An algorithm selects the group of questions to be answered by the player, analyses the answers and sends the result to a industrial robot through serial port. According with the player performance, the robot makes congratulation movements and giving a gift to the winner player. Otherwise, the robot makes movements, disapproving the player performance.
Resumo:
This article describes a methodological approach to conditional reasoning in online asynchronous learning environments such as Virtual-U VGroups, developed by SFU, BC, Canada, consistent with the notion of meaning implication: If part of a meaning C is embedded in B and a part of a meaning B is embedded in A, then A implies C in terms of meaning [Piaget 91]. A new transcript analysis technique was developed to assess the flows of conditional meaning implications and to identify the occurrence of hypotheses and connections among them in two human science graduate mixed-mode online courses offered in the summer/spring session of 1997 by SFU. Flows of conditional meaning implications were confronted with Virtual-U VGroups threads and results of the two courses were compared. Findings suggest that Virtual-U VGroups is a knowledge-building environment although the tree-like Virtual-U VGroups threads should be transformed into neuronal-like threads. Findings also suggest that formulating hypotheses together triggers a collaboratively problem-solving process that scaffolds knowledge-building in asynchronous learning environments: A pedagogical technique and an built-in tool for formulating hypotheses together are proposed. © Springer Pub. Co.
Resumo:
This paper presents the overall methodology that has been used to encode both the Brazilian Portuguese WordNet (WordNet.Br) standard language-independent conceptual-semantic relations (hyponymy, co-hyponymy, meronymy, cause, and entailment) and the so-called cross-lingual conceptual-semantic relations between different wordnets. Accordingly, after contextualizing the project and outlining the current lexical database structure and statistics, it describes the WordNet.Br editing GUI that was designed to aid the linguist in carrying out the tasks of building synsets, selecting sample sentences from corpora, writing synset concept glosses, and encoding both language-independent conceptual-semantic relations and cross-lingual conceptual-semantic relations between WordNet.Br and Princeton WordNet © Springer-Verlag Berlin Heidelberg 2006.
Resumo:
In this work an image pre-processing module has been developed to extract quantitative information from plantation images with various degrees of infestation. Four filters comprise this module: the first one acts on smoothness of the image, the second one removes image background enhancing plants leaves, the third filter removes isolated dots not removed by the previous filter, and the fourth one is used to highlight leaves' edges. At first the filters were tested with MATLAB, for a quick visual feedback of the filters' behavior. Then the filters were implemented in the C programming language. At last, the module as been coded in VHDL for the implementation on a Stratix II family FPGA. Tests were run and the results are shown in this paper. © 2008 Springer-Verlag Berlin Heidelberg.
Resumo:
One of the most important characteristics of intelligent activity is the ability to change behaviour according to many forms of feedback. Through learning an agent can interact with its environment to improve its performance over time. However, most of the techniques known that involves learning are time expensive, i.e., once the agent is supposed to learn over time by experimentation, the task has to be executed many times. Hence, high fidelity simulators can save a lot of time. In this context, this paper describes the framework designed to allow a team of real RoboNova-I humanoids robots to be simulated under USARSim environment. Details about the complete process of modeling and programming the robot are given, as well as the learning methodology proposed to improve robot's performance. Due to the use of a high fidelity model, the learning algorithms can be widely explored in simulation before adapted to real robots. © 2008 Springer-Verlag Berlin Heidelberg.
Resumo:
The Optimum-Path Forest (OPF) classifier is a recent and promising method for pattern recognition, with a fast training algorithm and good accuracy results. Therefore, the investigation of a combining method for this kind of classifier can be important for many applications. In this paper we report a fast method to combine OPF-based classifiers trained with disjoint training subsets. Given a fixed number of subsets, the algorithm chooses random samples, without replacement, from the original training set. Each subset accuracy is improved by a learning procedure. The final decision is given by majority vote. Experiments with simulated and real data sets showed that the proposed combining method is more efficient and effective than naive approach provided some conditions. It was also showed that OPF training step runs faster for a series of small subsets than for the whole training set. The combining scheme was also designed to support parallel or distributed processing, speeding up the procedure even more. © 2011 Springer-Verlag.
Resumo:
This work describes a control and supervision application takes into account the virtual instrumentation advantages to control and supervision industrial manufacturing stations belonging to the modular production system MPS® by Festo. These stations integrate sensors, actuators, conveyor belt and other industrial elements. The focus in this approach was to replace the use of programmable logic controllers by a computer equipped with a software application based on Labview and, together, performs the functions of traditional instruments and PLCs. The manufacturing stations had their processes modeled and simulated in Petri nets. After the models were implemented in Labview environment. Tests and previous similar works in MPS® installed in Automation Laboratory, at UNESP Sorocaba campus, showed the materials and methods used in this work allow the successful use of virtual instrumentation. The results indicate the technology as an advantageous approach for the automation of industrial processes, with gains in flexibility and reduction in project cost. © 2011 IEEE.
Resumo:
This paper proposes a new strategy to reduce the combinatorial search space of a mixed integer linear programming (MILP) problem. The construction phase of greedy randomized adaptive search procedure (GRASP-CP) is employed to reduce the domain of the integer variables of the transportation model of the transmission expansion planning (TM-TEP) problem. This problem is a MILP and very difficult to solve specially for large scale systems. The branch and bound (BB) algorithm is used to solve the problem in both full and the reduced search space. The proposed method might be useful to reduce the search space of those kinds of MILP problems that a fast heuristic algorithm is available for finding local optimal solutions. The obtained results using some real test systems show the efficiency of the proposed method. © 2012 Springer-Verlag.
Resumo:
This paper presents vectorized methods of construction and descent of quadtrees that can be easily adapted to message passing parallel computing. A time complexity analysis for the present approach is also discussed. The proposed method of tree construction requires a hash table to index nodes of a linear quadtree in the breadth-first order. The hash is performed in two steps: an internal hash to index child nodes and an external hash to index nodes in the same level (depth). The quadtree descent is performed by considering each level as a vector segment of a linear quadtree, so that nodes of the same level can be processed concurrently. © 2012 Springer-Verlag.