986 resultados para Job security
Resumo:
Le nombre important de véhicules sur le réseau routier peut entraîner des problèmes d'encombrement et de sécurité. Les usagers des réseaux routiers qui nous intéressent sont les camionneurs qui transportent des marchandises, pouvant rouler avec des véhicules non conformes ou emprunter des routes interdites pour gagner du temps. Le transport de matières dangereuses est réglementé et certains lieux, surtout les ponts et les tunnels, leur sont interdits d'accès. Pour aider à faire appliquer les lois en vigueur, il existe un système de contrôles routiers composé de structures fixes et de patrouilles mobiles. Le déploiement stratégique de ces ressources de contrôle mise sur la connaissance du comportement des camionneurs que nous allons étudier à travers l'analyse de leurs choix de routes. Un problème de choix de routes peut se modéliser en utilisant la théorie des choix discrets, elle-même fondée sur la théorie de l'utilité aléatoire. Traiter ce type de problème avec cette théorie est complexe. Les modèles que nous utiliserons sont tels, que nous serons amenés à faire face à des problèmes de corrélation, puisque plusieurs routes partagent probablement des arcs. De plus, puisque nous travaillons sur le réseau routier du Québec, le choix de routes peut se faire parmi un ensemble de routes dont le nombre est potentiellement infini si on considère celles ayant des boucles. Enfin, l'étude des choix faits par un humain n'est pas triviale. Avec l'aide du modèle de choix de routes retenu, nous pourrons calculer une expression de la probabilité qu'une route soit prise par le camionneur. Nous avons abordé cette étude du comportement en commençant par un travail de description des données collectées. Le questionnaire utilisé par les contrôleurs permet de collecter des données concernant les camionneurs, leurs véhicules et le lieu du contrôle. La description des données observées est une étape essentielle, car elle permet de présenter clairement à un analyste potentiel ce qui est accessible pour étudier les comportements des camionneurs. Les données observées lors d'un contrôle constitueront ce que nous appellerons une observation. Avec les attributs du réseau, il sera possible de modéliser le réseau routier du Québec. Une sélection de certains attributs permettra de spécifier la fonction d'utilité et par conséquent la fonction permettant de calculer les probabilités de choix de routes par un camionneur. Il devient alors possible d'étudier un comportement en se basant sur des observations. Celles provenant du terrain ne nous donnent pas suffisamment d'information actuellement et même en spécifiant bien un modèle, l'estimation des paramètres n'est pas possible. Cette dernière est basée sur la méthode du maximum de vraisemblance. Nous avons l'outil, mais il nous manque la matière première que sont les observations, pour continuer l'étude. L'idée est de poursuivre avec des observations de synthèse. Nous ferons des estimations avec des observations complètes puis, pour se rapprocher des conditions réelles, nous continuerons avec des observations partielles. Ceci constitue d'ailleurs un défi majeur. Nous proposons pour ces dernières, de nous servir des résultats des travaux de (Bierlaire et Frejinger, 2008) en les combinant avec ceux de (Fosgerau, Frejinger et Karlström, 2013). Bien qu'elles soient de nature synthétiques, les observations que nous utilisons nous mèneront à des résultats tels, que nous serons en mesure de fournir une proposition concrète qui pourrait aider à optimiser les décisions des responsables des contrôles routiers. En effet, nous avons réussi à estimer, sur le réseau réel du Québec, avec un seuil de signification de 0,05 les valeurs des paramètres d'un modèle de choix de routes discrets, même lorsque les observations sont partielles. Ces résultats donneront lieu à des recommandations sur les changements à faire dans le questionnaire permettant de collecter des données.
Resumo:
Conférence présentée à la Faculté de théologie et de sciences des religions de l’Université de Montréal le 11 septembre 2013. © Jean Duhaime
Resumo:
From the introduction: Mexico is in a state of siege. In recent years, organized crime and drug-related violence have escalated dramatically, taking innocent lives and leaving the country mired in bloodshed. The Mexican government, under the leadership of President Felipe Calderón, has responded in part by significantly extending the reach of its security operations, deploying thousands of federal police officers and military troops to combat the activities of drug cartels, and collaborating with the United States on an extensive regional security plan known as the Mérida Initiative. In the midst of the security crisis, however, the government has somewhat paradoxically adopted judicial reforms that protect human rights and civil liberties rather than erode them, specifically the presumption of innocence standard in criminal proceedings and the implementation of oral trials. Assuming that the new laws on the books will be applied in practice, these reforms represent an important commitment on the part of the government to uphold human rights and civil liberties. This is in stark contrast to the infamous judicial reforms in Colombia—the institutionalization of anonymous or “faceless” prosecutions in special courts—implemented after a surge in leftist and cartel brutality, and the murders of several prominent public and judicial officials in the 1980s.
Resumo:
The study is a close scrutiny of the process of investigation of offences in India along with an analysis of powers and functions of the investigating agency. The offences, which are prejudicial to sovereignty, integrity and security of the nation or to its friendly relations with foreign states, are generally called the offences against national security. Offences against national security being prejudicial to the very existence of the nation and its legal system, is a heinous and terrible one. As early as 1971 the Law Commission of India had pointed out the need of treating the offences relating to national security and their perpetrators on a totally different procedural footing. The recommendation that, all the offences coming under the said category ought to be brought under the purview of a single enactment so as to confront such offences effectively. The discrepancies in and inadequacies of the criminal justice system in India as much as they are related to the investigations of the offences against national security are examined and the reforms are also suggested. The quality of criminal justice is closely linked with the caliber of the prosecution system and many of the acquittals in courts can be ascribed not only to poor investigations but also to poor quality of prosecution.
Resumo:
The present research problem is to study the existing encryption methods and to develop a new technique which is performance wise superior to other existing techniques and at the same time can be very well incorporated in the communication channels of Fault Tolerant Hard Real time systems along with existing Error Checking / Error Correcting codes, so that the intention of eaves dropping can be defeated. There are many encryption methods available now. Each method has got it's own merits and demerits. Similarly, many crypt analysis techniques which adversaries use are also available.
Resumo:
Internet today has become a vital part of day to day life, owing to the revolutionary changes it has brought about in various fields. Dependence on the Internet as an information highway and knowledge bank is exponentially increasing so that a going back is beyond imagination. Transfer of critical information is also being carried out through the Internet. This widespread use of the Internet coupled with the tremendous growth in e-commerce and m-commerce has created a vital need for infonnation security.Internet has also become an active field of crackers and intruders. The whole development in this area can become null and void if fool-proof security of the data is not ensured without a chance of being adulterated. It is, hence a challenge before the professional community to develop systems to ensure security of the data sent through the Internet.Stream ciphers, hash functions and message authentication codes play vital roles in providing security services like confidentiality, integrity and authentication of the data sent through the Internet. There are several ·such popular and dependable techniques, which have been in use widely, for quite a long time. This long term exposure makes them vulnerable to successful or near successful attempts for attacks. Hence it is the need of the hour to develop new algorithms with better security.Hence studies were conducted on various types of algorithms being used in this area. Focus was given to identify the properties imparting security at this stage. By making use of a perception derived from these studies, new algorithms were designed. Performances of these algorithms were then studied followed by necessary modifications to yield an improved system consisting of a new stream cipher algorithm MAJE4, a new hash code JERIM- 320 and a new message authentication code MACJER-320. Detailed analysis and comparison with the existing popular schemes were also carried out to establish the security levels.The Secure Socket Layer (SSL) I Transport Layer Security (TLS) protocol is one of the most widely used security protocols in Internet. The cryptographic algorithms RC4 and HMAC have been in use for achieving security services like confidentiality and authentication in the SSL I TLS. But recent attacks on RC4 and HMAC have raised questions about the reliability of these algorithms. Hence MAJE4 and MACJER-320 have been proposed as substitutes for them. Detailed studies on the performance of these new algorithms were carried out; it has been observed that they are dependable alternatives.
Resumo:
Globalisation has many facets and its impact on labour is one of the most significant aspects.Though its influence is worldwide,it is much more significant in a transforming economy like India.The right of workers to social security is seen recognised under the Constitution of India and other welfare legislations.But,after adoption of the new economic policy of liberalisation and privatisation by the Government of India,the labour is exposed to new set of challenges.They are posed mainly due to economic restructuring affected in employment relationship,coupled with the increase in unprotected informal labour force.This study is an attempt to analyse the new challenges stemming up in employment relation,efficacy of the existing measures for social security of labour in the present economic condition and the suggestions for securing workers'right to social security in the trade regime.
Resumo:
A socio-economic research is required as an attempt to address the socio-economic issues facing small-scale fisheries. A study of the socio economic conditions of small-scale fishermen is a prerequisite for good design and successful implementation of effective assistance Programmes. It will provide an overall pidure of the structure, activities and standards of living of small-scale fisherfolk The study is confined to the coastal districts of Ernakulam, Thrissur and Malappuram districts. It also gives a picture of socio-economic conditions of the fisher folk in the study area. The variables that may depict the standard of living of the small-scale fisherfolk are occupational structure, family size, age structure, income, expenditure, education, housing and other social amenities. It attempts to see the asset creation of the fisherfolk with the help of government agencies, and the nature of savings and expenditure pattern of the fisherfolk. It also provides a picture of the indebtedness of the fisherfolk in the study area. The study analyses the schemes implemented by the government through its agencies, like Fisheries Department, Matsyaboard, and Matsyafed; and the awareness of fisherfolk regarding these schemes, their attitude and reactions, the extent of accessibility, and the viability of the schemes.
Resumo:
Modern computer systems are plagued with stability and security problems: applications lose data, web servers are hacked, and systems crash under heavy load. Many of these problems or anomalies arise from rare program behavior caused by attacks or errors. A substantial percentage of the web-based attacks are due to buffer overflows. Many methods have been devised to detect and prevent anomalous situations that arise from buffer overflows. The current state-of-art of anomaly detection systems is relatively primitive and mainly depend on static code checking to take care of buffer overflow attacks. For protection, Stack Guards and I-leap Guards are also used in wide varieties.This dissertation proposes an anomaly detection system, based on frequencies of system calls in the system call trace. System call traces represented as frequency sequences are profiled using sequence sets. A sequence set is identified by the starting sequence and frequencies of specific system calls. The deviations of the current input sequence from the corresponding normal profile in the frequency pattern of system calls is computed and expressed as an anomaly score. A simple Bayesian model is used for an accurate detection.Experimental results are reported which show that frequency of system calls represented using sequence sets, captures the normal behavior of programs under normal conditions of usage. This captured behavior allows the system to detect anomalies with a low rate of false positives. Data are presented which show that Bayesian Network on frequency variations responds effectively to induced buffer overflows. It can also help administrators to detect deviations in program flow introduced due to errors.
Resumo:
Assembly job shop scheduling problem (AJSP) is one of the most complicated combinatorial optimization problem that involves simultaneously scheduling the processing and assembly operations of complex structured products. The problem becomes even more complicated if a combination of two or more optimization criteria is considered. This thesis addresses an assembly job shop scheduling problem with multiple objectives. The objectives considered are to simultaneously minimizing makespan and total tardiness. In this thesis, two approaches viz., weighted approach and Pareto approach are used for solving the problem. However, it is quite difficult to achieve an optimal solution to this problem with traditional optimization approaches owing to the high computational complexity. Two metaheuristic techniques namely, genetic algorithm and tabu search are investigated in this thesis for solving the multiobjective assembly job shop scheduling problems. Three algorithms based on the two metaheuristic techniques for weighted approach and Pareto approach are proposed for the multi-objective assembly job shop scheduling problem (MOAJSP). A new pairing mechanism is developed for crossover operation in genetic algorithm which leads to improved solutions and faster convergence. The performances of the proposed algorithms are evaluated through a set of test problems and the results are reported. The results reveal that the proposed algorithms based on weighted approach are feasible and effective for solving MOAJSP instances according to the weight assigned to each objective criterion and the proposed algorithms based on Pareto approach are capable of producing a number of good Pareto optimal scheduling plans for MOAJSP instances.
Resumo:
The present research is carried out to understand how psychological empowerment, job satisfaction and job related stress are related.In banking sector, employees are less satisfied and less motivated than employees in other lines of work (Kelley, 1990; Bajpai, Naval and Deepak, 2004). The banking industry also suffers from high employee turnover rate (Branham, 2005; Nelson, 2007) and high level of stress (Chen and Lien, 2008). There are no adequate studies linking psychological empowerment and job satisfaction, stress, turnover etc. among employees of banking sector. Lack of psychological empowerment could be a reason for these problems faced by banking sector. Further majority of studies in psychological empowerment are carried out in manufacturing sector and studies in service sector are concentrated on hotel industry and hospitals. Empowerment takes different forms in different contexts (Zimmerman, 1995). In the light of above discussion, the present research is directed to explore the dimensions of psychological empowerment of employees in banking sector and to find out whether high psychological empowerment can increase job satisfaction and reduce job related stress among employees in banking sector
Resumo:
Any automatically measurable, robust and distinctive physical characteristic or personal trait that can be used to identify an individual or verify the claimed identity of an individual, referred to as biometrics, has gained significant interest in the wake of heightened concerns about security and rapid advancements in networking, communication and mobility. Multimodal biometrics is expected to be ultra-secure and reliable, due to the presence of multiple and independent—verification clues. In this study, a multimodal biometric system utilising audio and facial signatures has been implemented and error analysis has been carried out. A total of one thousand face images and 250 sound tracks of 50 users are used for training the proposed system. To account for the attempts of the unregistered signatures data of 25 new users are tested. The short term spectral features were extracted from the sound data and Vector Quantization was done using K-means algorithm. Face images are identified based on Eigen face approach using Principal Component Analysis. The success rate of multimodal system using speech and face is higher when compared to individual unimodal recognition systems
Resumo:
This study focuses on psychological empowerment of employees in banking sector because of the reasons stated below: Firstly, very little research has been conducted in understanding empowerment as a psychological construct. Majority of the studies have been conducted on the various empowerment practices in the organizations. Secondly, there is no empirical evidence that the empowerment practice will create a subjective feeling of empowerment within the individual. Employee empowerment will be effective only if the employees actually experience the empowerment. Even if the organizations have the empowerment practices like providing power and open communication it is not necessary that the employee is empowered. Empowerment describes only the condition of work environment. It does not describe employees’ response to these conditions. These responses form the basis for psychological empowerment
Resumo:
With this document, we provide a compilation of in-depth discussions on some of the most current security issues in distributed systems. The six contributions have been collected and presented at the 1st Kassel Student Workshop on Security in Distributed Systems (KaSWoSDS’08). We are pleased to present a collection of papers not only shedding light on the theoretical aspects of their topics, but also being accompanied with elaborate practical examples. In Chapter 1, Stephan Opfer discusses Viruses, one of the oldest threats to system security. For years there has been an arms race between virus producers and anti-virus software providers, with no end in sight. Stefan Triller demonstrates how malicious code can be injected in a target process using a buffer overflow in Chapter 2. Websites usually store their data and user information in data bases. Like buffer overflows, the possibilities of performing SQL injection attacks targeting such data bases are left open by unwary programmers. Stephan Scheuermann gives us a deeper insight into the mechanisms behind such attacks in Chapter 3. Cross-site scripting (XSS) is a method to insert malicious code into websites viewed by other users. Michael Blumenstein explains this issue in Chapter 4. Code can be injected in other websites via XSS attacks in order to spy out data of internet users, spoofing subsumes all methods that directly involve taking on a false identity. In Chapter 5, Till Amma shows us different ways how this can be done and how it is prevented. Last but not least, cryptographic methods are used to encode confidential data in a way that even if it got in the wrong hands, the culprits cannot decode it. Over the centuries, many different ciphers have been developed, applied, and finally broken. Ilhan Glogic sketches this history in Chapter 6.