803 resultados para Multiprocessor computer architectures
Resumo:
Architectures based on Coordinated Atomic action (CA action) concepts have been used to build concurrent fault-tolerant systems. This conceptual model combines concurrent exception handling with action nesting to provide a general mechanism for both enclosing interactions among system components and coordinating forward error recovery measures. This article presents an architectural model to guide the formal specification of concurrent fault-tolerant systems. This architecture provides built-in Communicating Sequential Processes (CSPs) and predefined channels to coordinate exception handling of the user-defined components. Hence some safety properties concerning action scoping and concurrent exception handling can be proved by using the FDR (Failure Divergence Refinement) verification tool. As a result, a formal and general architecture supporting software fault tolerance is ready to be used and proved as users define components with normal and exceptional behaviors. (C) 2010 Elsevier B.V. All rights reserved.
Resumo:
This paper describes a new module of the expert system SISTEMAT used for the prediction of the skeletons of neolignans by (13)C NMR, (1)H NMR and botanical data obtained from the literature. SISTEMAT is composed of MACRONO, SISCONST, C13MACH, H1MACH and SISOCBOT programs, each analyzing data of the neolignan in question to predict the carbon skeleton of the compound. From these results, the global probability is computed and the most probable skeleton predicted. SISTEMAT predicted the skeletons of 75% of the 20 neolignans tested, in a rapid and simple procedure demonstrating its advantage for the structural elucidation of new compounds.
Resumo:
Lithium salt solutions of Li(CF3SO2)(2)N, LiTFSI, in a room-temperature ionic liquid (RTIL), 1-butyl-2,3-dimethyl-imidazolium cation, BMMI, and the (CF3SO2)(2)N-, bis(trifluoromethanesulfonyl)imide anion, [BMMI][TFSI], were prepared in different concentrations. Thermal properties, density, viscosity, ionic conductivity, and self-diffusion coefficients were determined at different temperatures for pure [BMMI][TFSI] and the lithium solutions. Raman spectroscopy measurements and computer simulations were also carried out in order to understand the microscopic origin of the observed changes in transport coefficients. Slopes of Walden plots for conductivity and fluidity, and the ratio between the actual conductivity and the Nernst-Einstein estimate for conductivity, decrease with increasing LiTFSI content. All of these studies indicated the formation of aggregates of different chemical nature, as it is corroborated by the Raman spectra. In addition, molecular dynamics (MD) simulations showed that the coordination of Li+ by oxygen atoms of TFSI anions changes with Li+ concentration producing a remarkable change of the RTIL structure with a concomitant reduction of diffusion coefficients of all species in the solutions.
Resumo:
This paper reports an expert system (SISTEMAT) developed for structural determination of diverse chemical classes of natural products, including lignans, based mainly on 13C NMR and 1H NMR data of these compounds. The system is composed of five programs that analyze specific data of a lignan and shows a skeleton probability for the compound. At the end of analyses, the results are grouped, the global probability is computed, and the most probable skeleton is exhibited to the user. SISTEMAT was able to properly predict the skeletons of 80% of the 30 lignans tested, demonstrating its advantage during the structural elucidation course in a short period of time.
Resumo:
Nursing school graduates are under pressure to pass the RN-NCLEX Exam on the first attempt since New York State monitors the results and uses them to evaluate the school’s nursing programs. Since the RN-NCLEX Exam is a standardized test, we sought a method to make our students better test takers. The use of on-line computer adaptive testing has raised our student’s standardized test scores at the end of the nursing course.
Resumo:
This paper examines the wage premium to computer use in Sweden in the early 1990’s. I use simple regression model and interaction terms in my paper to examine the effect of computer use at work. Although the data is only one-year cross-section data, my results clearly show a wage premium to computer use in Sweden. There are also interesting findings in my paper by using Swedish data. From the results, I find wage premium to be related to intensity of computer use at work.
Resumo:
In order to achieve the high performance, we need to have an efficient scheduling of a parallelprogram onto the processors in multiprocessor systems that minimizes the entire executiontime. This problem of multiprocessor scheduling can be stated as finding a schedule for ageneral task graph to be executed on a multiprocessor system so that the schedule length can be minimize [10]. This scheduling problem is known to be NP- Hard.In multi processor task scheduling, we have a number of CPU’s on which a number of tasksare to be scheduled that the program’s execution time is minimized. According to [10], thetasks scheduling problem is a key factor for a parallel multiprocessor system to gain betterperformance. A task can be partitioned into a group of subtasks and represented as a DAG(Directed Acyclic Graph), so the problem can be stated as finding a schedule for a DAG to beexecuted in a parallel multiprocessor system so that the schedule can be minimized. Thishelps to reduce processing time and increase processor utilization. The aim of this thesis workis to check and compare the results obtained by Bee Colony algorithm with already generatedbest known results in multi processor task scheduling domain.
Resumo:
In recent years, it has been observed that software clones and plagiarism are becoming an increased threat for one?s creativity. Clones are the results of copying and using other?s work. According to the Merriam – Webster dictionary, “A clone is one that appears to be a copy of an original form”. It is synonym to duplicate. Clones lead to redundancy of codes, but not all redundant code is a clone.On basis of this background knowledge ,in order to safeguard one?s idea and to avoid intentional code duplication for pretending other?s work as if their owns, software clone detection should be emphasized more. The objective of this paper is to review the methods for clone detection and to apply those methods for finding the extent of plagiarism occurrence among the Swedish Universities in Master level computer science department and to analyze the results.The rest part of the paper, discuss about software plagiarism detection which employs data analysis technique and then statistical analysis of the results.Plagiarism is an act of stealing and passing off the idea?s and words of another person?s as one?s own. Using data analysis technique, samples(Master level computer Science thesis report) were taken from various Swedish universities and processed in Ephorus anti plagiarism software detection. Ephorus gives the percentage of plagiarism for each thesis document, from this results statistical analysis were carried out using Minitab Software.The results gives a very low percentage of Plagiarism extent among the Swedish universities, which concludes that Plagiarism is not a threat to Sweden?s standard of education in computer science.This paper is based on data analysis, intelligence techniques, EPHORUS software plagiarism detection tool and MINITAB statistical software analysis.
Resumo:
The purpose of the calculations was to estimate the most suitable slopes and azimuths for three different positions per day of a solar panel in order to obtain the most possible energy from the PV panel compared with a stationary PV panel. The calculations were made in the computer program PV F-CHART.
Resumo:
It is known that despite companies’ efforts to improve the quality of their products, design and assembly defects results in large repair costs both in terms of repair and providing feedback to the origin of the defect. The purpose of this paper is to study these types of defects and the defect rates in design and assembly. The paper presents a web based questionnaire answered by 29 companies. The result shows that the defect rate (defects per product) spanned from 0.01 to 10. Also, design and assembly defects covered 46%, 23% respectively, of all occurred defects. A case study is also presented, performed at a company who recently implemented a modular architecture. In this company, defects from 5 700 integrated product architectures are compared with defects from 431 modular architectures. The average defect rate increased by 21.5% – from 0.65 to 0.79 – when a more modular architecture has been implemented. Furthermore, the study showed that the assembly defects have decreased while the design defects increased. The results presented in this paper will also support the development of the MPV (Module Property Verification) method which is briefly described.
Resumo:
Modular product architectures have generated numerous benefits for companies in terms of cost, lead-time and quality. The defined interfaces and the module’s properties decrease the effort to develop new product variants, and provide an opportunity to perform parallel tasks in design, manufacturing and assembly. The background of this thesis is that companies perform verifications (tests, inspections and controls) of products late, when most of the parts have been assembled. This extends the lead-time to delivery and ruins benefits from a modular product architecture; specifically when the verifications are extensive and the frequency of detected defects is high. Due to the number of product variants obtained from the modular product architecture, verifications must handle a wide range of equipment, instructions and goal values to ensure that high quality products can be delivered. As a result, the total benefits from a modular product architecture are difficult to achieve. This thesis describes a method for planning and performing verifications within a modular product architecture. The method supports companies by utilizing the defined modules for verifications already at module level, so called MPV (Module Property Verification). With MPV, defects are detected at an earlier point, compared to verification of a complete product, and the number of verifications is decreased. The MPV method is built up of three phases. In Phase A, candidate modules are evaluated on the basis of costs and lead-time of the verifications and the repair of defects. An MPV-index is obtained which quantifies the module and indicates if the module should be verified at product level or by MPV. In Phase B, the interface interaction between the modules is evaluated, as well as the distribution of properties among the modules. The purpose is to evaluate the extent to which supplementary verifications at product level is needed. Phase C supports a selection of the final verification strategy. The cost and lead-time for the supplementary verifications are considered together with the results from Phase A and B. The MPV method is based on a set of qualitative and quantitative measures and tools which provide an overview and support the achievement of cost and time efficient company specific verifications. A practical application in industry shows how the MPV method can be used, and the subsequent benefits
Resumo:
The rapid development of data transfer through internet made it easier to send the data accurate and faster to the destination. There are many transmission media to transfer the data to destination like e-mails; at the same time it is may be easier to modify and misuse the valuable information through hacking. So, in order to transfer the data securely to the destination without any modifications, there are many approaches like cryptography and steganography. This paper deals with the image steganography as well as with the different security issues, general overview of cryptography, steganography and digital watermarking approaches. The problem of copyright violation of multimedia data has increased due to the enormous growth of computer networks that provides fast and error free transmission of any unauthorized duplicate and possibly manipulated copy of multimedia information. In order to be effective for copyright protection, digital watermark must be robust which are difficult to remove from the object in which they are embedded despite a variety of possible attacks. The message to be send safe and secure, we use watermarking. We use invisible watermarking to embed the message using LSB (Least Significant Bit) steganographic technique. The standard LSB technique embed the message in every pixel, but my contribution for this proposed watermarking, works with the hint for embedding the message only on the image edges alone. If the hacker knows that the system uses LSB technique also, it cannot decrypt correct message. To make my system robust and secure, we added cryptography algorithm as Vigenere square. Whereas the message is transmitted in cipher text and its added advantage to the proposed system. The standard Vigenere square algorithm works with either lower case or upper case. The proposed cryptography algorithm is Vigenere square with extension of numbers also. We can keep the crypto key with combination of characters and numbers. So by using these modifications and updating in this existing algorithm and combination of cryptography and steganography method we develop a secure and strong watermarking method. Performance of this watermarking scheme has been analyzed by evaluating the robustness of the algorithm with PSNR (Peak Signal to Noise Ratio) and MSE (Mean Square Error) against the quality of the image for large amount of data. While coming to see results of the proposed encryption, higher value of 89dB of PSNR with small value of MSE is 0.0017. Then it seems the proposed watermarking system is secure and robust for hiding secure information in any digital system, because this system collect the properties of both steganography and cryptography sciences.
Resumo:
Background: Constructive alignment (CA) is a pedagogical approach that emphasizes the alignment between the intended learning outcomes (ILOs), teaching and learning activities (TLAs) and assessment tasks (ATs) as well as creation of a teaching/learning environment where students will be able to actively create their knowledge. Objectives: This paper aims at investigating the extent of constructively-aligned courses in Computer Engineering and Informatics department at Dalarna University, Sweden. This study is based on empirical observations of teacher’s perceptions of implementation of CA in their courses. Methods: Ten teachers (5 from each department) were asked to fill a paper-based questionnaire, which included a number of questions related to issues of implementing CA in courses. Results: Responses to the items of the questionnaire were mixed. Teachers clearly state the ILOs in their courses and try to align the TLAs and ATs to the ILOs. Computer Engineering teachers do not explicitly communicate the ILOs to the students as compared to Informatics teachers. In addition, Computer Engineering teachers stated that their students are less active in learning activities as compared to Informatics teachers. When asked about their subjective ratings of teaching methods all teachers stated that their current teaching is teacher-centered but they try to shift the focus of activity from them to the students. Conclusions: From teachers’ perspectives, the courses are partially constructively-aligned. Their courses are “aligned”, i.e. ILOs, TLAs and ATs are aligned to each other but they are not “constructive” since, according to them, there was a low student engagement in learning activities, especially in Computer Engineering department.