914 resultados para 120402 Engineering Design Knowledge


Relevância:

30.00% 30.00%

Publicador:

Resumo:

Les défis conjoints du changement climatique d'origine anthropique et la diminution des réserves de combustibles fossiles sont le moteur de recherche intense pour des sources d'énergie alternatives. Une avenue attrayante est d'utiliser un processus biologique pour produire un biocarburant. Parmi les différentes options en matière de biocarburants, le bio-hydrogène gazeux est un futur vecteur énergétique attrayant en raison de son efficacité potentiellement plus élevé de conversion de puissance utilisable, il est faible en génération inexistante de polluants et de haute densité d'énergie. Cependant, les faibles rendements et taux de production ont été les principaux obstacles à l'application pratique des technologies de bio-hydrogène. Des recherches intensives sur bio-hydrogène sont en cours, et dans les dernières années, plusieurs nouvelles approches ont été proposées et étudiées pour dépasser ces inconvénients. À cette fin, l'objectif principal de cette thèse était d'améliorer le rendement en hydrogène moléculaire avec un accent particulier sur l'ingénierie métabolique et l’utilisation de bioprocédés à variables indépendantes. Une de nos hypothèses était que la production d’hydrogène pourrait être améliorée et rendue plus économiquement viable par ingénierie métabolique de souches d’Escherichia coli producteurs d’hydrogène en utilisant le glucose ainsi que diverses autres sources de carbone, y compris les pentoses. Les effets du pH, de la température et de sources de carbone ont été étudiés. La production maximale d'hydrogène a été obtenue à partir de glucose, à un pH initial de 6.5 et une température de 35°C. Les études de cinétiques de croissance ont montré que la μmax était 0.0495 h-1 avec un Ks de 0.0274 g L-1 lorsque le glucose est la seule source de carbone en milieu minimal M9. .Parmi les nombreux sucres et les dérivés de sucres testés, les rendements les plus élevés d'hydrogène sont avec du fructose, sorbitol et D-glucose; 1.27, 1.46 et 1.51 mol H2 mol-1 de substrat, respectivement. En outre, pour obtenir les interactions entre les variables importantes et pour atteindre une production maximale d'hydrogène, un design 3K factoriel complet Box-Behnken et la méthodologie de réponse de surface (RSM) ont été employées pour la conception expérimentale et l'analyse de la souche d'Escherichia coli DJT135. Le rendement en hydrogène molaire maximale de 1.69 mol H2 mol-1 de glucose a été obtenu dans les conditions optimales de 75 mM de glucose, à 35°C et un pH de 6.5. Ainsi, la RSM avec un design Box-Behken était un outil statistique utile pour atteindre des rendements plus élevés d'hydrogène molaires par des organismes modifiés génétiquement. Ensuite, l'expression hétérologue de l’hydrogénases soluble [Ni-Fe] de Ralstonia eutropha H16 (l'hydrogénase SH) a tenté de démontrer que la mise en place d'une voie capable de dériver l'hydrogène à partir de NADH pourrait surpasser le rendement stoechiométrique en hydrogène.. L’expression a été démontrée par des tests in vitro de l'activité enzymatique. Par ailleurs, l'expression de SH a restaurée la croissance en anaérobie de souches mutantes pour adhE, normalement inhibées en raison de l'incapacité de réoxyder le NADH. La mesure de la production d'hydrogène in vivo a montré que plusieurs souches modifiées métaboliquement sont capables d'utiliser l'hydrogénase SH pour dériver deux moles d’hydrogène par mole de glucose consommé, proche du maximum théorique. Une autre stratégie a montré que le glycérol brut pourrait être converti en hydrogène par photofermentation utilisant Rhodopseudomonas palustris par photofermentation. Les effets de la source d'azote et de différentes concentrations de glycérol brut sur ce processus ont été évalués. À 20 mM de glycérol, 4 mM glutamate, 6.1 mol hydrogène / mole de glycérol brut ont été obtenus dans des conditions optimales, un rendement de 87% de la théorie, et significativement plus élevés que ce qui a été réalisé auparavant. En prolongement de cette étude, l'optimisation des paramètres a également été utilisée. Dans des conditions optimales, une intensité lumineuse de 175 W/m2, 30 mM glycérol et 4.5 mM de glutamate, 6.69 mol hydrogène / mole de glycérol brut ont été obtenus, soit un rendement de 96% de la valeur théorique. La détermination de l'activité de la nitrogénase et ses niveaux d'expression ont montré qu'il y avait relativement peu de variation de la quantité de nitrogénase avec le changement des variables alors que l'activité de la nitrogénase variait considérablement, avec une activité maximale (228 nmol de C2H4/ml/min) au point central optimal. Dans la dernière section, la production d'hydrogène à partir du glucose via la photofermentation en une seule étape a été examinée avec la bactérie photosynthétique Rhodobacter capsulatus JP91 (hup-). La méthodologie de surface de réponse avec Box-Behnken a été utilisée pour optimiser les variables expérimentales de façon indépendante, soit la concentration de glucose, la concentration du glutamate et l'intensité lumineuse, ainsi que d'examiner leurs effets interactifs pour la maximisation du rendement en hydrogène moléculaire. Dans des conditions optimales, avec une intensité lumineuse de 175 W/m2, 35 mM de glucose, et 4.5 mM de glutamate,, un rendement maximal d'hydrogène de 5.5 (± 0.15) mol hydrogène /mol glucose, et un maximum d'activité de la nitrogénase de 246 (± 3.5) nmol C2H4/ml/min ont été obtenus. L'analyse densitométrique de l'expression de la protéine-Fe nitrogenase dans les différentes conditions a montré une variation significative de l'expression protéique avec un maximum au point central optimisé. Même dans des conditions optimales pour la production d'hydrogène, une fraction significative de la protéine Fe a été trouvée dans l'état ADP-ribosylée, suggérant que d'autres améliorations des rendements pourraient être possibles. À cette fin, un mutant amtB dérivé de Rhodobacter capsulatus JP91 (hup-) a été créé en utilisant le vecteur de suicide pSUP202. Les résultats expérimentaux préliminaires montrent que la souche nouvellement conçue métaboliquement, R. capsulatus DG9, produit 8.2 (± 0.06) mol hydrogène / mole de glucose dans des conditions optimales de cultures discontinues (intensité lumineuse, 175 W/m2, 35 mM de glucose et 4.5 mM glutamate). Le statut d'ADP-ribosylation de la nitrogénase-protéine Fe a été obtenu par Western Blot pour la souche R. capsulatus DG9. En bref, la production d'hydrogène est limitée par une barrière métabolique. La principale barrière métabolique est due au manque d'outils moléculaires possibles pour atteindre ou dépasser le rendement stochiométrique en bio-hydrogène depuis les dernières décennies en utilisant les microbes. À cette fin, une nouvelle approche d’ingénierie métabolique semble très prometteuse pour surmonter cette contrainte vers l'industrialisation et s'assurer de la faisabilité de la technologie de la production d'hydrogène. Dans la présente étude, il a été démontré que l’ingénierie métabolique de bactéries anaérobiques facultatives (Escherichia coli) et de bactéries anaérobiques photosynthétiques (Rhodobacter capsulatus et Rhodopseudomonas palustris) peuvent produire de l'hydrogène en tant que produit majeur à travers le mode de fermentation par redirection métabolique vers la production d'énergie potentielle. D'autre part, la méthodologie de surface de réponse utilisée dans cette étude représente un outil potentiel pour optimiser la production d'hydrogène en générant des informations appropriées concernant la corrélation entre les variables et des producteurs de bio-de hydrogène modifiés par ingénierie métabolique. Ainsi, un outil d'optimisation des paramètres représente une nouvelle avenue pour faire un pont entre le laboratoire et la production d'hydrogène à l'échelle industrielle en fournissant un modèle mathématique potentiel pour intensifier la production de bio-hydrogène. Par conséquent, il a été clairement mis en évidence dans ce projet que l'effort combiné de l'ingénierie métabolique et la méthodologie de surface de réponse peut rendre la technologie de production de bio-hydrogène potentiellement possible vers sa commercialisation dans un avenir rapproché.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Les logiciels sont en constante évolution, nécessitant une maintenance et un développement continus. Ils subissent des changements tout au long de leur vie, que ce soit pendant l'ajout de nouvelles fonctionnalités ou la correction de bogues. Lorsque les logiciels évoluent, leurs architectures ont tendance à se dégrader et deviennent moins adaptables aux nouvelles spécifications des utilisateurs. En effet, les architectures de ces logiciels deviennent plus complexes et plus difficiles à maintenir à cause des nombreuses dépendances entre les artefacts. Par conséquent, les développeurs doivent comprendre les dépendances entre les artefacts des logiciels pour prendre des mesures proactives qui facilitent les futurs changements et ralentissent la dégradation des architectures des logiciels. D'une part, le maintien d'un logiciel sans la compréhension des les dépendances entre ses artefacts peut conduire à l'introduction de défauts. D'autre part, lorsque les développeurs manquent de connaissances sur l'impact de leurs activités de maintenance, ils peuvent introduire des défauts de conception, qui ont un impact négatif sur l'évolution du logiciel. Ainsi, les développeurs ont besoin de mécanismes pour comprendre comment le changement d'un artefact impacte le reste du logiciel. Dans cette thèse, nous proposons trois contributions principales : La spécification de deux nouveaux patrons de changement et leurs utilisations pour fournir aux développeurs des informations utiles concernant les dépendances de co-changement. La spécification de la relation entre les patrons d'évolutions des artefacts et les fautes. La découverte de la relation entre les dépendances des anti-patrons et la prédisposition des différentes composantes d'un logiciel aux fautes.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Le 21e siècle est le berceau d’une conscientisation grandissante sur les impacts environnementaux des processus utilisés pour synthétiser des molécules cibles. Parmi les avancées qui ont marqué ces dernières décennies, il est également question de réduction de déchets, de conservation de l’énergie et de durabilité des innovations. Ces aspects constituent les lignes directrices de la chimie verte. De ce fait, il est impératif de développer des stratégies de synthèse dont les impacts environnementaux sont bénins. Dans ce mémoire nous présentons la synthèse, la caractérisation et l’étude des propriétés catalytiques en milieu aqueux d’un ligand composé d’une unité -cyclodextrine native, d’une unité imidazolium et d’une chaine alkyle à 12 carbones. Ce ligand hybride s’auto-assemble dans l’eau sous forme de micelles, permettant ainsi d’effectuer en sa présence des couplages de Suzuki-Miyaura dans l’eau, avec de bons rendements. La fonctionnalisation de la face primaire de la -cyclodextrine par un noyau alkyl-imidazolium, précurseur de ligand de type carbène N-hétérocyclique, a permis le développement d’un système catalytique vert et hautement recyclable. Dans un deuxième temps, nous présentons l’utilisation du même ligand hybride dans des couplages de Heck dans l’eau, démontrant ainsi la versatilité du ligand.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Les systèmes logiciels sont devenus de plus en plus répondus et importants dans notre société. Ainsi, il y a un besoin constant de logiciels de haute qualité. Pour améliorer la qualité de logiciels, l’une des techniques les plus utilisées est le refactoring qui sert à améliorer la structure d'un programme tout en préservant son comportement externe. Le refactoring promet, s'il est appliqué convenablement, à améliorer la compréhensibilité, la maintenabilité et l'extensibilité du logiciel tout en améliorant la productivité des programmeurs. En général, le refactoring pourra s’appliquer au niveau de spécification, conception ou code. Cette thèse porte sur l'automatisation de processus de recommandation de refactoring, au niveau code, s’appliquant en deux étapes principales: 1) la détection des fragments de code qui devraient être améliorés (e.g., les défauts de conception), et 2) l'identification des solutions de refactoring à appliquer. Pour la première étape, nous traduisons des régularités qui peuvent être trouvés dans des exemples de défauts de conception. Nous utilisons un algorithme génétique pour générer automatiquement des règles de détection à partir des exemples de défauts. Pour la deuxième étape, nous introduisons une approche se basant sur une recherche heuristique. Le processus consiste à trouver la séquence optimale d'opérations de refactoring permettant d'améliorer la qualité du logiciel en minimisant le nombre de défauts tout en priorisant les instances les plus critiques. De plus, nous explorons d'autres objectifs à optimiser: le nombre de changements requis pour appliquer la solution de refactoring, la préservation de la sémantique, et la consistance avec l’historique de changements. Ainsi, réduire le nombre de changements permets de garder autant que possible avec la conception initiale. La préservation de la sémantique assure que le programme restructuré est sémantiquement cohérent. De plus, nous utilisons l'historique de changement pour suggérer de nouveaux refactorings dans des contextes similaires. En outre, nous introduisons une approche multi-objective pour améliorer les attributs de qualité du logiciel (la flexibilité, la maintenabilité, etc.), fixer les « mauvaises » pratiques de conception (défauts de conception), tout en introduisant les « bonnes » pratiques de conception (patrons de conception).

Relevância:

30.00% 30.00%

Publicador:

Resumo:

En la actualidad, el uso de las tecnologías ha sido primordial para el avance de las sociedades, estas han permitido que personas sin conocimientos informáticos o usuarios llamados “no expertos” se interesen en su uso, razón por la cual los investigadores científicos se han visto en la necesidad de producir estudios que permitan la adaptación de sistemas, a la problemática existente dentro del ámbito informático. Una necesidad recurrente de todo usuario de un sistema es la gestión de la información, la cual se puede administrar por medio de una base de datos y lenguaje específico, como lo es el SQL (Structured Query Language), pero esto obliga al usuario sin conocimientos a acudir a un especialista para su diseño y construcción, lo cual se ve reflejado en costos y métodos complejos, entonces se plantea una pregunta ¿qué hacer cuando los proyectos son pequeñas y los recursos y procesos son limitados? Teniendo como base la investigación realizada por la universidad de Washington[39], donde sintetizan sentencias SQL a partir de ejemplos de entrada y salida, se pretende con esta memoria automatizar el proceso y aplicar una técnica diferente de aprendizaje, para lo cual utiliza una aproximación evolucionista, donde la aplicación de un algoritmo genético adaptado origina sentencias SQL válidas que responden a las condiciones establecidas por los ejemplos de entrada y salida dados por el usuario. Se obtuvo como resultado de la aproximación, una herramienta denominada EvoSQL que fue validada en este estudio. Sobre los 28 ejercicios empleados por la investigación [39], 23 de los cuales se obtuvieron resultados perfectos y 5 ejercicios sin éxito, esto representa un 82.1% de efectividad. Esta efectividad es superior en un 10.7% al establecido por la herramienta desarrollada en [39] SQLSynthesizer y 75% más alto que la herramienta siguiente más próxima Query by Output QBO[31]. El promedio obtenido en la ejecución de cada ejercicio fue de 3 minutos y 11 segundos, este tiempo es superior al establecido por SQLSynthesizer; sin embargo, en la medida un algoritmo genético supone la existencia de fases que amplían los rangos de tiempos, por lo cual el tiempo obtenido es aceptable con relación a las aplicaciones de este tipo. En conclusión y según lo anteriormente expuesto, se obtuvo una herramienta automática con una aproximación evolucionista, con buenos resultados y un proceso simple para el usuario “no experto”.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Recent advances in tissue engineering and regenerative medicine have shown that controlling cells microenvironment during growth is a key element to the development of successful therapeutic system. To achieve such control, researchers have first proposed the use of polymeric scaffolds that were able to support cellular growth and, to a certain extent, favor cell organization and tissue structure. With nowadays availability of a large pool of stem cell lines, such approach has appeared to be rather limited since it does not offer the fine control of the cell micro-environment in space and time (4D). Therefore, researchers are currently focusing their efforts on developing strategies that include active compound delivery systems in order to add a fourth dimension to the design of 3D scaffolds. This review will focus on recent concepts and applications of 2D and 3D techniques that have been used to control the load and release of active compounds used to promote cell differentiation and proliferation in or out of a scaffold. We will first present recent advances in the design of 2D polymeric scaffolds and the different techniques that have been used to deposit molecular cues and cells in a controlled fashion. We will continue presenting the recent advances made in the design of 3D scaffolds based on hydrogels as well as polymeric fibers and we will finish by presenting some of the research avenues that are still to be explored.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The rapid developments in fields such as fibre optic communication engineering and integrated optical electronics have expanded the interest and have increased the expectations about guided wave optics, in which optical waveguides and optical fibres play a central role. The technology of guided wave photonics now plays a role in generating information (guided-wave sensors) and processing information (spectral analysis, analog-to-digital conversion and other optical communication schemes) in addition to its original application of transmitting information (fibre optic communication). Passive and active polymer devices have generated much research interest recently because of the versatility of the fabrication techniques and the potential applications in two important areas – short distant communication network and special functionality optical devices such as amplifiers, switches and sensors. Polymer optical waveguides and fibres are often designed to have large cores with 10-1000 micrometer diameter to facilitate easy connection and splicing. Large diameter polymer optical fibres being less fragile and vastly easier to work with than glass fibres, are attractive in sensing applications. Sensors using commercial plastic optical fibres are based on ideas already used in silica glass sensors, but exploiting the flexible and cost effective nature of the plastic optical fibre for harsh environments and throw-away sensors. In the field of Photonics, considerable attention is centering on the use of polymer waveguides and fibres, as they have a great potential to create all-optical devices. By attaching organic dyes to the polymer system we can incorporate a variety of optical functions. Organic dye doped polymer waveguides and fibres are potential candidates for solid state gain media. High power and high gain optical amplification in organic dye-doped polymer waveguide amplifier is possible due to extremely large emission cross sections of dyes. Also, an extensive choice of organic dye dopants is possible resulting in amplification covering a wide range in the visible region.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Analog-to digital Converters (ADC) have an important impact on the overall performance of signal processing system. This research is to explore efficient techniques for the design of sigma-delta ADC,specially for multi-standard wireless tranceivers. In particular, the aim is to develop novel models and algorithms to address this problem and to implement software tools which are avle to assist the designer's decisions in the system-level exploration phase. To this end, this thesis presents a framework of techniques to design sigma-delta analog to digital converters.A2-2-2 reconfigurable sigma-delta modulator is proposed which can meet the design specifications of the three wireless communication standards namely GSM,WCDMA and WLAN. A sigma-delta modulator design tool is developed using the Graphical User Interface Development Environment (GUIDE) In MATLAB.Genetic Algorithm(GA) based search method is introduced to find the optimum value of the scaling coefficients and to maximize the dynamic range in a sigma-delta modulator.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Indian marine engineers are renowned for employment globally due to their knowledge, skill and reliability. This praiseworthy status has been achieved mainly due to the systematic training imparted to marine engineering cadets. However, in an era of advancing technology, marine engineering training has to remain dynamic to imbibe latest technology as well as to meet the demands of the shipping industry. New subjects of studies have to be included in the curriculum in a timely manner taking into consideration the industry requirements and best practices in shipping. Technical competence of marine engineers also has to be subjected to changes depending upon the needs of the ever growing and over regulated shipping industry. Besides. certain soft skills are to be developed and improved amongst the marine engineers in order to alter or amend the personality traits leading to their career success.If timely corrective action is taken. Indian marine engineers can be in still greater demand for employment in global maritime field. In order to enhance the employability of our mmine engineers by improving their quality, a study of marine engineers in general and class IV marine engineers in particular was conducted based on three distinct surveys, viz., survey among senior marine engineers, survey among employers of marine engineers and survey of class IV marine engineers themselves.The surveys have been planned and questionnaires have been designed to focus the study of marine engineer officer class IV from the point of view of the three distinct groups of maritime personnels. As a result of this, the strength and weakness of class IV marine engineers are identified with regard to their performance on board ships, acquisition of necessary technical skills. employability and career success. The criteria of essential qualities of a marine engineer are classified as academic, technical, social, psychological. physical, mental, emergency responsive, communicative and leadership, and have been assessed for a practicing marine engineer by statistical analysis of data collected from surveys. These are assessed for class IV marine engineers from the point of view of senior marine engineers and employers separately. The Endings are delineated and graphically depicted in this thesis.Besides. six pertinent personality traits of a marine engineer viz. self esteem. learning style. decision making. motivation. team work and listening self inventory have been subjected to study and their correlation with career success have been established wherever possible. This is carried out to develop a theoretical framework to understand what leads a marine engineer to his career attainment. This enables the author to estimate the personality strengths and weaknesses of a serving marine engineer and eventually to deduce possible corrective measures or modifications in marine engineering training in India.Maritime training is largely based on International Conventions on Standard of Training. Certification and Watch keeping for Seafarers 1995. its associated Code and Merchant Shipping (STCW for Seafarers) Rules 1998. Further, Maritime Education, Training and Assessment (META) Manual was subjected to a critical scrutiny and relevant Endings of thc surveys arc superimposed on the existing rule requirement and curriculum. Views of senior marine engineers and executives of various shipping companies are taken into account before arriving at the revision of syllabus of marine engineering courses. Modifications in the pattern of workshop and sea service for graduate mechanical engineering trainees are recommended. Desirable age brackets of junior engineers and chief engineers. use of Training and Assessment Record book (TAR Book) during training etc. have also been evaluated.As a result of the pedagogic introspection of the existing system of marine engineering training in India. in this thesis, a revised pattern of workshop training of six months duration for graduate mechanical engineers. revised pattern of sea service training of one year duration and modified now diagram incorporating the above have been arrived at. Effects of various personality traits on career success have been established along with certain findings for improvement of desirable personality traits of marine engineers.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Internet today has become a vital part of day to day life, owing to the revolutionary changes it has brought about in various fields. Dependence on the Internet as an information highway and knowledge bank is exponentially increasing so that a going back is beyond imagination. Transfer of critical information is also being carried out through the Internet. This widespread use of the Internet coupled with the tremendous growth in e-commerce and m-commerce has created a vital need for infonnation security.Internet has also become an active field of crackers and intruders. The whole development in this area can become null and void if fool-proof security of the data is not ensured without a chance of being adulterated. It is, hence a challenge before the professional community to develop systems to ensure security of the data sent through the Internet.Stream ciphers, hash functions and message authentication codes play vital roles in providing security services like confidentiality, integrity and authentication of the data sent through the Internet. There are several ·such popular and dependable techniques, which have been in use widely, for quite a long time. This long term exposure makes them vulnerable to successful or near successful attempts for attacks. Hence it is the need of the hour to develop new algorithms with better security.Hence studies were conducted on various types of algorithms being used in this area. Focus was given to identify the properties imparting security at this stage. By making use of a perception derived from these studies, new algorithms were designed. Performances of these algorithms were then studied followed by necessary modifications to yield an improved system consisting of a new stream cipher algorithm MAJE4, a new hash code JERIM- 320 and a new message authentication code MACJER-320. Detailed analysis and comparison with the existing popular schemes were also carried out to establish the security levels.The Secure Socket Layer (SSL) I Transport Layer Security (TLS) protocol is one of the most widely used security protocols in Internet. The cryptographic algorithms RC4 and HMAC have been in use for achieving security services like confidentiality and authentication in the SSL I TLS. But recent attacks on RC4 and HMAC have raised questions about the reliability of these algorithms. Hence MAJE4 and MACJER-320 have been proposed as substitutes for them. Detailed studies on the performance of these new algorithms were carried out; it has been observed that they are dependable alternatives.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Information and communication technologies are the tools that underpin the emerging “Knowledge Society”. Exchange of information or knowledge between people and through networks of people has always taken place. But the ICT has radically changed the magnitude of this exchange, and thus factors such as timeliness of information and information dissemination patterns have become more important than ever.Since information and knowledge are so vital for the all round human development, libraries and institutions that manage these resources are indeed invaluable. So, the Library and Information Centres have a key role in the acquisition, processing, preservation and dissemination of information and knowledge. ln the modern context, library is providing service based on different types of documents such as manuscripts, printed, digital, etc. At the same time, acquisition, access, process, service etc. of these resources have become complicated now than ever before. The lCT made instrumental to extend libraries beyond the physical walls of a building and providing assistance in navigating and analyzing tremendous amounts of knowledge with a variety of digital tools. Thus, modern libraries are increasingly being re-defined as places to get unrestricted access to information in many formats and from many sources.The research was conducted in the university libraries in Kerala State, India. lt was identified that even though the information resources are flooding world over and several technologies have emerged to manage the situation for providing effective services to its clientele, most of the university libraries in Kerala were unable to exploit these technologies at maximum level. Though the libraries have automated many of their functions, wide gap prevails between the possible services and provided services. There are many good examples world over in the application of lCTs in libraries for the maximization of services and many such libraries have adopted the principles of reengineering and re-defining as a management strategy. Hence this study was targeted to look into how effectively adopted the modern lCTs in our libraries for maximizing the efficiency of operations and services and whether the principles of re-engineering and- redefining can be applied towards this.Data‘ was collected from library users, viz; student as well as faculty users; library ,professionals and university librarians, using structured questionnaires. This has been .supplemented by-observation of working of the libraries, discussions and interviews with the different types of users and staff, review of literature, etc. Personal observation of the organization set up, management practices, functions, facilities, resources, utilization of information resources and facilities by the users, etc. of the university libraries in Kerala have been made. Statistical techniques like percentage, mean, weighted mean, standard deviation, correlation, trend analysis, etc. have been used to analyse data.All the libraries could exploit only a very few possibilities of modern lCTs and hence they could not achieve effective Universal Bibliographic Control and desired efficiency and effectiveness in services. Because of this, the users as well as professionals are dissatisfied. Functional effectiveness in acquisition, access and process of information resources in various formats, development and maintenance of OPAC and WebOPAC, digital document delivery to remote users, Web based clearing of library counter services and resources, development of full-text databases, digital libraries and institutional repositories, consortia based operations for e-journals and databases, user education and information literacy, professional development with stress on lCTs, network administration and website maintenance, marketing of information, etc. are major areas need special attention to improve the situation. Finance, knowledge level on ICTs among library staff, professional dynamism and leadership, vision and support of the administrators and policy makers, prevailing educational set up and social environment in the state, etc. are some of the major hurdles in reaping the maximum possibilities of lCTs by the university libraries in Kerala. The principles of Business Process Re-engineering are found suitable to effectively apply to re-structure and redefine the operations and service system of the libraries. Most of the conventional departments or divisions prevailing in the university libraries were functioning as watertight compartments and their existing management system was more rigid to adopt the principles of change management. Hence, a thorough re-structuring of the divisions was indicated. Consortia based activities and pooling and sharing of information resources was advocated to meet the varied needs of the users in the main campuses and off campuses of the universities, affiliated colleges and remote stations. A uniform staff policy similar to that prevailing in CSIR, DRDO, ISRO, etc. has been proposed by the study not only in the university libraries in kerala but for the entire country.Restructuring of Lis education,integrated and Planned development of school,college,research and public library systems,etc.were also justified for reaping maximum benefits of the modern ICTs.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Data mining is one of the hottest research areas nowadays as it has got wide variety of applications in common man’s life to make the world a better place to live. It is all about finding interesting hidden patterns in a huge history data base. As an example, from a sales data base, one can find an interesting pattern like “people who buy magazines tend to buy news papers also” using data mining. Now in the sales point of view the advantage is that one can place these things together in the shop to increase sales. In this research work, data mining is effectively applied to a domain called placement chance prediction, since taking wise career decision is so crucial for anybody for sure. In India technical manpower analysis is carried out by an organization named National Technical Manpower Information System (NTMIS), established in 1983-84 by India's Ministry of Education & Culture. The NTMIS comprises of a lead centre in the IAMR, New Delhi, and 21 nodal centres located at different parts of the country. The Kerala State Nodal Centre is located at Cochin University of Science and Technology. In Nodal Centre, they collect placement information by sending postal questionnaire to passed out students on a regular basis. From this raw data available in the nodal centre, a history data base was prepared. Each record in this data base includes entrance rank ranges, reservation, Sector, Sex, and a particular engineering. From each such combination of attributes from the history data base of student records, corresponding placement chances is computed and stored in the history data base. From this data, various popular data mining models are built and tested. These models can be used to predict the most suitable branch for a particular new student with one of the above combination of criteria. Also a detailed performance comparison of the various data mining models is done.This research work proposes to use a combination of data mining models namely a hybrid stacking ensemble for better predictions. A strategy to predict the overall absorption rate for various branches as well as the time it takes for all the students of a particular branch to get placed etc are also proposed. Finally, this research work puts forward a new data mining algorithm namely C 4.5 * stat for numeric data sets which has been proved to have competent accuracy over standard benchmarking data sets called UCI data sets. It also proposes an optimization strategy called parameter tuning to improve the standard C 4.5 algorithm. As a summary this research work passes through all four dimensions for a typical data mining research work, namely application to a domain, development of classifier models, optimization and ensemble methods.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

With the increase in population, housing and construction of various facilities have been a problem with urbanization. Having exhausted all the trouble free hand, man is nowon the lookout for techniques to improve areas which were originally considered uninhabitable. Thus this study is based on the nature and engineering behavior of soft clays covering long stretches of coastal line and methods to improve their geotechnical properties .The main aim of the present investigation is to study in detail the physical and engineering behavior of the marine clays of Cochin. While it is well known that the marine clays have been posing numerous problems to foundation engineers all along, the relevant literature reveals that no systematic and comprehensive study has been attempted to date. The: knowledge gained through the study is suitably used to improve these properties with appropriate additives.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The demand for new telecommunication services requiring higher capacities, data rates and different operating modes have motivated the development of new generation multi-standard wireless transceivers. A multi-standard design often involves extensive system level analysis and architectural partitioning, typically requiring extensive calculations. In this research, a decimation filter design tool for wireless communication standards consisting of GSM, WCDMA, WLANa, WLANb, WLANg and WiMAX is developed in MATLAB® using GUIDE environment for visual analysis. The user can select a required wireless communication standard, and obtain the corresponding multistage decimation filter implementation using this toolbox. The toolbox helps the user or design engineer to perform a quick design and analysis of decimation filter for multiple standards without doing extensive calculation of the underlying methods.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A marine isolate of jáÅêçÅçÅÅìë MCCB 104 has been identified as an aquaculture probiotic antagonistic to sáÄêáç. In the present study different carbon and nitrogen sources and growth factors in a mineral base medium were optimized for enhanced biomass production and antagonistic activity against the target pathogen, sáÄêáç=Ü~êîÉóá, following response surface methodology (RSM). Accordingly the minimum and maximum limits of the selected variables were determined and a set of fifty experiments programmed employing central composite design (CCD) of RSM for the final optimization. The response surface plots of biomass showed similar pattern with that of antagonistic activity, which indicated a strong correlation between the biomass and antagonism. The optimum concentration of the carbon sources, nitrogen sources, and growth factors for both biomass and antagonistic activity were glucose (17.4 g/L), lactose (17 g/L), sodium chloride (16.9 g/L), ammonium chloride (3.3 g/L), and mineral salts solution (18.3 mL/L). © KSBB