934 resultados para Additive Code


Relevância:

20.00% 20.00%

Publicador:

Resumo:

The present research problem is to study the existing encryption methods and to develop a new technique which is performance wise superior to other existing techniques and at the same time can be very well incorporated in the communication channels of Fault Tolerant Hard Real time systems along with existing Error Checking / Error Correcting codes, so that the intention of eaves dropping can be defeated. There are many encryption methods available now. Each method has got it's own merits and demerits. Similarly, many crypt analysis techniques which adversaries use are also available.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Embedded systems are usually designed for a single or a specified set of tasks. This specificity means the system design as well as its hardware/software development can be highly optimized. Embedded software must meet the requirements such as high reliability operation on resource-constrained platforms, real time constraints and rapid development. This necessitates the adoption of static machine codes analysis tools running on a host machine for the validation and optimization of embedded system codes, which can help meet all of these goals. This could significantly augment the software quality and is still a challenging field.Embedded systems are usually designed for a single or a specified set of tasks. This specificity means the system design as well as its hardware/software development can be highly optimized. Embedded software must meet the requirements such as high reliability operation on resource-constrained platforms, real time constraints and rapid development. This necessitates the adoption of static machine codes analysis tools running on a host machine for the validation and optimization of embedded system codes, which can help meet all of these goals. This could significantly augment the software quality and is still a challenging field.Embedded systems are usually designed for a single or a specified set of tasks. This specificity means the system design as well as its hardware/software development can be highly optimized. Embedded software must meet the requirements such as high reliability operation on resource-constrained platforms, real time constraints and rapid development. This necessitates the adoption of static machine codes analysis tools running on a host machine for the validation and optimization of embedded system codes, which can help meet all of these goals. This could significantly augment the software quality and is still a challenging field.Embedded systems are usually designed for a single or a specified set of tasks. This specificity means the system design as well as its hardware/software development can be highly optimized. Embedded software must meet the requirements such as high reliability operation on resource-constrained platforms, real time constraints and rapid development. This necessitates the adoption of static machine codes analysis tools running on a host machine for the validation and optimization of embedded system codes, which can help meet all of these goals. This could significantly augment the software quality and is still a challenging field.This dissertation contributes to an architecture oriented code validation, error localization and optimization technique assisting the embedded system designer in software debugging, to make it more effective at early detection of software bugs that are otherwise hard to detect, using the static analysis of machine codes. The focus of this work is to develop methods that automatically localize faults as well as optimize the code and thus improve the debugging process as well as quality of the code.Validation is done with the help of rules of inferences formulated for the target processor. The rules govern the occurrence of illegitimate/out of place instructions and code sequences for executing the computational and integrated peripheral functions. The stipulated rules are encoded in propositional logic formulae and their compliance is tested individually in all possible execution paths of the application programs. An incorrect sequence of machine code pattern is identified using slicing techniques on the control flow graph generated from the machine code.An algorithm to assist the compiler to eliminate the redundant bank switching codes and decide on optimum data allocation to banked memory resulting in minimum number of bank switching codes in embedded system software is proposed. A relation matrix and a state transition diagram formed for the active memory bank state transition corresponding to each bank selection instruction is used for the detection of redundant codes. Instances of code redundancy based on the stipulated rules for the target processor are identified.This validation and optimization tool can be integrated to the system development environment. It is a novel approach independent of compiler/assembler, applicable to a wide range of processors once appropriate rules are formulated. Program states are identified mainly with machine code pattern, which drastically reduces the state space creation contributing to an improved state-of-the-art model checking. Though the technique described is general, the implementation is architecture oriented, and hence the feasibility study is conducted on PIC16F87X microcontrollers. The proposed tool will be very useful in steering novices towards correct use of difficult microcontroller features in developing embedded systems.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Code clones are portions of source code which are similar to the original program code. The presence of code clones is considered as a bad feature of software as the maintenance of software becomes difficult due to the presence of code clones. Methods for code clone detection have gained immense significance in the last few years as they play a significant role in engineering applications such as analysis of program code, program understanding, plagiarism detection, error detection, code compaction and many more similar tasks. Despite of all these facts, several features of code clones if properly utilized can make software development process easier. In this work, we have pointed out such a feature of code clones which highlight the relevance of code clones in test sequence identification. Here program slicing is used in code clone detection. In addition, a classification of code clones is presented and the benefit of using program slicing in code clone detection is also mentioned in this work.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Optical Character Recognition plays an important role in Digital Image Processing and Pattern Recognition. Even though ambient study had been performed on foreign languages like Chinese and Japanese, effort on Indian script is still immature. OCR in Malayalam language is more complex as it is enriched with largest number of characters among all Indian languages. The challenge of recognition of characters is even high in handwritten domain, due to the varying writing style of each individual. In this paper we propose a system for recognition of offline handwritten Malayalam vowels. The proposed method uses Chain code and Image Centroid for the purpose of extracting features and a two layer feed forward network with scaled conjugate gradient for classification

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The present study investigates the benefits of stabilizing the stone mastic asphalt (SMA) mixture in flexible pavement with shredded waste plastic. Conventional (without plastic) and the stabilized SMA mixtures were subjected to performance tests including Marshall Stability, tensile strength and compressive strength tests. Triaxial tests were also conducted with varying percentage bitumen by weight of mineral aggregate (6% to 8%) and by varying percentage plastic by weight of mix (6% to 12% with an increment of 1%). Plastic content of 10% by weight of bitumen is recommended for the improvement of the performance of Stone Mastic Asphalt mixtures. 10% plastic content gives an increase in the stability, split tensile strength and compressive strength of about 64%, 18% and 75% respectively compared to the conventional SMA mix. Triaxial test results show a 44% increase in cohesion and 3% decrease in angle of shearing resistance showing an increase in the shear strength. The drain down value decreases with an increase in plastic content and the value is only 0.09 % at 10% plastic content and proves to be an effective stabilizing additive in SMA mixtures

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Modeling nonlinear systems using Volterra series is a century old method but practical realizations were hampered by inadequate hardware to handle the increased computational complexity stemming from its use. But interest is renewed recently, in designing and implementing filters which can model much of the polynomial nonlinearities inherent in practical systems. The key advantage in resorting to Volterra power series for this purpose is that nonlinear filters so designed can be made to work in parallel with the existing LTI systems, yielding improved performance. This paper describes the inclusion of a quadratic predictor (with nonlinearity order 2) with a linear predictor in an analog source coding system. Analog coding schemes generally ignore the source generation mechanisms but focuses on high fidelity reconstruction at the receiver. The widely used method of differential pnlse code modulation (DPCM) for speech transmission uses a linear predictor to estimate the next possible value of the input speech signal. But this linear system do not account for the inherent nonlinearities in speech signals arising out of multiple reflections in the vocal tract. So a quadratic predictor is designed and implemented in parallel with the linear predictor to yield improved mean square error performance. The augmented speech coder is tested on speech signals transmitted over an additive white gaussian noise (AWGN) channel.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We had previously shown that regularization principles lead to approximation schemes, as Radial Basis Functions, which are equivalent to networks with one layer of hidden units, called Regularization Networks. In this paper we show that regularization networks encompass a much broader range of approximation schemes, including many of the popular general additive models, Breiman's hinge functions and some forms of Projection Pursuit Regression. In the probabilistic interpretation of regularization, the different classes of basis functions correspond to different classes of prior probabilities on the approximating function spaces, and therefore to different types of smoothness assumptions. In the final part of the paper, we also show a relation between activation functions of the Gaussian and sigmoidal type.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Considers Sampling, Pulse Amplitude Modulation, Multiple Access, Quantisation, Pulse Coded Modulation, Manchester Line Coding, Amplitude Modulation, Double SideBand Suppressed Carrier Modulation, Quadrature Amplitude Modulation and M-ary Shift Keying.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Considers Huffman coding and arithmetic coding

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Code required to perform an import of module information from eFolio and output it into a single page. Within the context of eHandbook: this page is called via an iFrame to display module information as part of the handbook (and to avoid any duplication of content).

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This resource is for Health Scientists

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Notes, slides, links

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Memoria de máster (Universidad de Texas, 2010). Resumen basado en el de la publicación

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Raymond Saleilles (1855-1912) és considerat pels comparatistes com pare del dret comparat modern. De fet fou el màxim responsable de l'organització del primer congrés internacional de dret comparat a París el 1900. L'època marcada per la internacionalització dels intercanvis, la crisi del mètode jurídic francès i el repte de la codificació alemanya i suïssa motivà alguns juristes a buscar noves formes d'estudiar el dret. Saleilles defensava pel dret comparat un mètode, un objecte i un fi propis i autònoms diferents dels que fins ara tenien la dogmàtica jurídica o l'estudi de la legislació estrangera. El mètode tenia els seus fonaments en la sociologia i la història. L'objectiu del dret comparat havia de ser la construcció d'un ideal "droit commun de l'humanité civilisée". Aquest dret comú havia de ser construït no només per la doctrina sinó també per la jurisprudència i el legislador.