912 resultados para Optics in computing


Relevância:

80.00% 80.00%

Publicador:

Resumo:

This research provides a description of the process followed in order to assemble a "Social Accounting Matrix" for Spain corresponding to the year 2000 (SAMSP00). As argued in the paper, this process attempts to reconcile ESA95 conventions with requirements of applied general equilibrium modelling. Particularly, problems related to the level of aggregation of net taxation data, and to the valuation system used for expressing the monetary value of input-output transactions have deserved special attention. Since the adoption of ESA95 conventions, input-output transactions have been preferably valued at basic prices, which impose additional difficulties on modellers interested in computing applied general equilibrium models. This paper addresses these difficulties by developing a procedure that allows SAM-builders to change the valuation system of input-output transactions conveniently. In addition, this procedure produces new data related to net taxation information.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Introduction: Therapeutic drug monitoring (TDM) aims at optimizing treatment by individualizing dosage regimen based on measurement of blood concentrations. Maintaining concentrations within a target range requires pharmacokinetic and clinical capabilities. Bayesian calculation represents a gold standard in TDM approach but requires computing assistance. In the last decades computer programs have been developed to assist clinicians in this assignment. The aim of this benchmarking was to assess and compare computer tools designed to support TDM clinical activities.¦Method: Literature and Internet search was performed to identify software. All programs were tested on common personal computer. Each program was scored against a standardized grid covering pharmacokinetic relevance, user-friendliness, computing aspects, interfacing, and storage. A weighting factor was applied to each criterion of the grid to consider its relative importance. To assess the robustness of the software, six representative clinical vignettes were also processed through all of them.¦Results: 12 software tools were identified, tested and ranked. It represents a comprehensive review of the available software's characteristics. Numbers of drugs handled vary widely and 8 programs offer the ability to the user to add its own drug model. 10 computer programs are able to compute Bayesian dosage adaptation based on a blood concentration (a posteriori adjustment) while 9 are also able to suggest a priori dosage regimen (prior to any blood concentration measurement), based on individual patient covariates, such as age, gender, weight. Among those applying Bayesian analysis, one uses the non-parametric approach. The top 2 software emerging from this benchmark are MwPharm and TCIWorks. Other programs evaluated have also a good potential but are less sophisticated (e.g. in terms of storage or report generation) or less user-friendly.¦Conclusion: Whereas 2 integrated programs are at the top of the ranked listed, such complex tools would possibly not fit all institutions, and each software tool must be regarded with respect to individual needs of hospitals or clinicians. Interest in computing tool to support therapeutic monitoring is still growing. Although developers put efforts into it the last years, there is still room for improvement, especially in terms of institutional information system interfacing, user-friendliness, capacity of data storage and report generation.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Abstract in English : Ubiquitous Computing is the emerging trend in computing systems. Based on this observation this thesis proposes an analysis of the hardware and environmental constraints that rule pervasive platforms. These constraints have a strong impact on the programming of such platforms. Therefore solutions are proposed to facilitate this programming both at the platform and node levels. The first contribution presented in this document proposes a combination of agentoriented programming with the principles of bio-inspiration (Phylogenesys, Ontogenesys and Epigenesys) to program pervasive platforms such as the PERvasive computing framework for modeling comPLEX virtually Unbounded Systems platform. The second contribution proposes a method to program efficiently parallelizable applications on each computing node of this platform. Résumé en Français : Basée sur le constat que les calculs ubiquitaires vont devenir le paradigme de programmation dans les années à venir, cette thèse propose une analyse des contraintes matérielles et environnementale auxquelles sont soumises les plateformes pervasives. Ces contraintes ayant un impact fort sur la programmation des plateformes. Des solutions sont donc proposées pour faciliter cette programmation tant au niveau de l'ensemble des noeuds qu'au niveau de chacun des noeuds de la plateforme. La première contribution présentée dans ce document propose d'utiliser une alliance de programmation orientée agent avec les grands principes de la bio-inspiration (Phylogénèse, Ontogénèse et Épigénèse). Ceci pour répondres aux contraintes de programmation de plateformes pervasives comme la plateforme PERvasive computing framework for modeling comPLEX virtually Unbounded Systems . La seconde contribution propose quant à elle une méthode permettant de programmer efficacement des applications parallélisable sur chaque noeud de calcul de la plateforme

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This research provides a description of the process followed in order to assemble a "Social Accounting Matrix" for Spain corresponding to the year 2000 (SAMSP00). As argued in the paper, this process attempts to reconcile ESA95 conventions with requirements of applied general equilibrium modelling. Particularly, problems related to the level of aggregation of net taxation data, and to the valuation system used for expressing the monetary value of input-output transactions have deserved special attention. Since the adoption of ESA95 conventions, input-output transactions have been preferably valued at basic prices, which impose additional difficulties on modellers interested in computing applied general equilibrium models. This paper addresses these difficulties by developing a procedure that allows SAM-builders to change the valuation system of input-output transactions conveniently. In addition, this procedure produces new data related to net taxation information.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Exposure to solar ultraviolet (UV) radiation is the main causative factor for skin cancer. UV exposure depends on environmental and individual factors, but individual exposure data remain scarce. UV irradiance is monitored via different techniques including ground measurements and satellite observations. However it is difficult to translate such observations into human UV exposure or dose because of confounding factors (shape of the exposed surface, shading, behavior, etc.) A collaboration between public health institutions, a meteorological office and an institute specialized in computing techniques developed a model predicting the dose and distribution of UV exposure on the basis of ground irradiation and morphological data. Standard 3D computer graphics techniques were adapted to develop this tool, which estimates solar exposure of a virtual manikin depicted as a triangle mesh surface. The amount of solar energy received by various body locations is computed for direct, diffuse and reflected radiation separately. The radiation components are deduced from corresponding measurements of UV irradiance, and the related UV dose received by each triangle of the virtual manikin is computed accounting for shading by other body parts and eventual protection measures. The model was verified with dosimetric measurements (n=54) in field conditions using a foam manikin as surrogate for an exposed individual. Dosimetric results were compared to the model predictions. The model predicted exposure to solar UV adequately. The symmetric mean absolute percentage error was 13%. Half of the predictions were within 17% range of the measurements. This model allows assessing outdoor occupational and recreational UV exposures, without necessitating time-consuming individual dosimetry, with numerous potential uses in skin cancer prevention and research. Using this tool, we investigated solar UV exposure patterns with respect to the relative contribution of the direct, diffuse and reflected radiation. We assessed exposure doses for various body parts and exposure scenarios of a standing individual (static and dynamic postures). As input, the model used erythemally-weighted ground irradiance data measured in 2009 at Payerne, Switzerland. A year-round daily exposure (8 am to 5 pm) without protection was assumed. For most anatomical sites, mean daily doses were high (typically 6.2-14.6 SED) and exceeded recommended exposure values. Direct exposure was important during specific periods (e.g. midday during summer), but contributed moderately to the annual dose, ranging from 15 to 24% for vertical and horizontal body parts, respectively. Diffuse irradiation explained about 80% of the cumulative annual exposure dose. Acute diffuse exposures were also obtained for cloudy summer days. The importance of diffuse UV radiation should not be underestimated when advocating preventive measures. Messages focused on avoiding acute direct exposures may be of limited efficiency to prevent skin cancers associated with chronic exposure (e.g., squamous cell carcinomas).

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The aim of this research paper is to present a macroscopic study about the feasibility and the efficiency of mobile devices in computing Least-Cost Path (LCP). This kind of artifact must work in off-line mode and must allow to load data from a mountain zone like digital terrain models and meteorological data.The research strategy has two steps:- First of all, we need to identify the set of software components in order to implement them inside the IT artifact. This set of components should have to be able to do LCP calculations, visualize results and present a well adapted human interface. The main goal of this first steep is to demonstrate the feasibility of a mobile geographic information system by following the ¿Design & Creation¿ research strategy.- In a second time, the goal is to evaluate the reliability and usability of this IT artifact by an ¿Experiments¿ research approach. In this step we want to characterize the behavior of the artifact in terms of fidelity and LCP process speed. This evaluation will be carried out by some external users.During the reading of this paper, we will see that this kind of geographic information system (the IT artifact) has the minimal requirements needed to carry out LCP calculations in mobile devices although it has several limitations and constraints in terms of useability and reliability. We will point out qualitative and quantitative elements related to the IT artifact performances while doing this kind of computations.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The aim of this research paper is to present a macroscopic study about the feasibility and the efficiency of mobile devices in computing least-cost path (LCP). This kind of artifact must work in off-line mode and must allow to load data from a mountain zone like digital terrain models and meteorological data.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

For some years, Chemistry teachers have used scientific visualization software of molecular models in computing rooms and chemistry laboratories for educational purposes. However, its application in classrooms has been limited. This article describes the integration and use of computer programs for scientific molecular visualization in a traditional classroom. We consider that the improvement of technical aspects of their application and use (usability) has a direct effect on students' understanding of molecular structures (including students' extrinsic motivation), among other factors. Consequently, we developed a guide for the integration of hardware and software of molecular visualization for its use in the classroom.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This thesis comprises seven peer-reviewed articles and examines systems and applications suitable for increasing Future Force Warrior performance, minimizing collateral damage, improving situational awareness and Common Operational Picture. Based on a literature study, missing functionalities of Future Force Warrior were identified and new ideas, concepts and solutions were created as part of early stages of Systems of Systems creation. These introduced ideas have not yet been implemented or tested in combat and for this reason benefit analyses are excluded. The main results of this thesis include the following: A new networking concept, Wireless Polling Sensor Network, which is a swarm of a few Unmanned Aerial Vehicles forming an ad-hoc network and polling a large number of fixed sensor nodes. The system is more robust in a military environment than traditional Wireless Sensor Networks. A Business Process approach to Service Oriented Architecture in a tactical setting is a concept for scheduling and sharing limited resources. New components to military Service Oriented Architecture have been introduced in the thesis. Other results of the thesis include an investigation of the use of Free Space Optics in tactical communications, a proposal for tracking neutral forces, a system for upgrading simple collaboration tools for command, control and collaboration purposes, a three-level hierarchy of Future Force Warrior, and methods for reducing incidents of fratricide.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Cloud computing is a practically relevant paradigm in computing today. Testing is one of the distinct areas where cloud computing can be applied. This study addressed the applicability of cloud computing for testing within organizational and strategic contexts. The study focused on issues related to the adoption, use and effects of cloudbased testing. The study applied empirical research methods. The data was collected through interviews with practitioners from 30 organizations and was analysed using the grounded theory method. The research process consisted of four phases. The first phase studied the definitions and perceptions related to cloud-based testing. The second phase observed cloud-based testing in real-life practice. The third phase analysed quality in the context of cloud application development. The fourth phase studied the applicability of cloud computing in the gaming industry. The results showed that cloud computing is relevant and applicable for testing and application development, as well as other areas, e.g., game development. The research identified the benefits, challenges, requirements and effects of cloud-based testing; and formulated a roadmap and strategy for adopting cloud-based testing. The study also explored quality issues in cloud application development. As a special case, the research included a study on applicability of cloud computing in game development. The results can be used by companies to enhance the processes for managing cloudbased testing, evaluating practical cloud-based testing work and assessing the appropriateness of cloud-based testing for specific testing needs.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Previous studies on the determinants of the choice of college major have assumed a constant probability of success across majors or a constant earnings stream across majors. Our model disregards these two restrictive assumptions in computing an expected earnings variable to explain the probability that a student will choose a specific major among four choices of concentrations. The construction of an expected earnings variable requires information on the student s perceived probability of success, the predicted earnings of graduates in all majors and the student s expected earnings if he (she) fails to complete a college program. Using data from the National Longitudinal Survey of Youth, we evaluate the chances of success in all majors for all the individuals in the sample. Second, the individuals' predicted earnings of graduates in all majors are obtained using Rumberger and Thomas's (1993) regression estimates from a 1987 Survey of Recent College Graduates. Third, we obtain idiosyncratic estimates of earnings alternative of not attending college or by dropping out with a condition derived from our college major decision-making model applied to our sample of college students. Finally, with a mixed multinominal logit model, we explain the individuals' choice of a major. The results of the paper show that the expected earnings variable is essential in the choice of a college major. There are, however, significant differences in the impact of expected earnings by gender and race.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Le domaine des énergies est au cœur des préoccupations technologiques, politiques et économiques de notre société moderne. Ce domaine nécessite une compréhension minimale du concept scientifique de l’énergie. Elle est selon nous essentielle à toute formation citoyenne. Nous avons dans un premier temps, à partir de considérations théoriques et pratiques, examiné pourquoi ce domaine si important dans notre société technologique est si peu abordé dans le cursus scolaire québécois? Pourquoi se contente-t-on d’un enseignement théorique et discursif de ce concept? Pourquoi, au contraire de tout enseignement scientifique, n’a-t-on pas envisagé de situations d’apprentissages en laboratoire pour l’étude des énergies? Dans un deuxième temps, nous avons proposé une idée de solution concrète et réaliste pour répondre à l’ensemble de ces questions. Une solution qui invite les élèves à s’investir de manière constructive dans des activités de laboratoire afin de s’approprier ces concepts. Pour ce faire, nous avons conçu des variables globales énergies qui ont permis aux élèves de les mesurer et d’expérimenter facilement des transformations énergétiques. Cette recherche de développement technologique en éducation consiste donc à profiter des nouveaux développements technologiques de l’informatique et de la micro-électronique pour concevoir, réaliser et mettre à l’essai un environnement informatisé d’apprentissage en laboratoire pour les sciences et la technologie. Par ce que l’énergie est au confluent de trois domaines, cet environnement a été conçu pour supporter dans une même activité l’apprentissage des mathématiques, des sciences et de la technologie. Cette intégration recommandée par les nouveaux programmes est, selon nous, essentielle à la compréhension des concepts liés à l’énergie et à ses transformations. Par cette activité d’apprentissage multidisciplinaire, nous voulons, via une approche empirique et concrète, aborder ces problèmes de transformations énergétiques afin de donner aux élèves la capacité de perfectionner les prototypes qu’ils construisent en technologie de manière à améliorer leurs performances. Nous avons montré que cette démarche technoscientifique, assimilable à la conception d’un schème expérimental en sciences, favorise la compréhension des concepts liés aux énergies et à leurs transformations. Ce développement, ouvert à l’investigation scientifique, apporte un bénéfice didactique, non seulement, pour des enseignants en exercices et des étudiants-maîtres, mais aussi pour des élèves de 5ème année du niveau secondaire, ce que nous avons démontré dans une mise à l’essai empirique.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

The rapid developments in fields such as fibre optic communication engineering and integrated optical electronics have expanded the interest and have increased the expectations about guided wave optics, in which optical waveguides and optical fibres play a central role. The technology of guided wave photonics now plays a role in generating information (guided-wave sensors) and processing information (spectral analysis, analog-to-digital conversion and other optical communication schemes) in addition to its original application of transmitting information (fibre optic communication). Passive and active polymer devices have generated much research interest recently because of the versatility of the fabrication techniques and the potential applications in two important areas – short distant communication network and special functionality optical devices such as amplifiers, switches and sensors. Polymer optical waveguides and fibres are often designed to have large cores with 10-1000 micrometer diameter to facilitate easy connection and splicing. Large diameter polymer optical fibres being less fragile and vastly easier to work with than glass fibres, are attractive in sensing applications. Sensors using commercial plastic optical fibres are based on ideas already used in silica glass sensors, but exploiting the flexible and cost effective nature of the plastic optical fibre for harsh environments and throw-away sensors. In the field of Photonics, considerable attention is centering on the use of polymer waveguides and fibres, as they have a great potential to create all-optical devices. By attaching organic dyes to the polymer system we can incorporate a variety of optical functions. Organic dye doped polymer waveguides and fibres are potential candidates for solid state gain media. High power and high gain optical amplification in organic dye-doped polymer waveguide amplifier is possible due to extremely large emission cross sections of dyes. Also, an extensive choice of organic dye dopants is possible resulting in amplification covering a wide range in the visible region.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

This paper underlines a methodology for translating text from English into the Dravidian language, Malayalam using statistical models. By using a monolingual Malayalam corpus and a bilingual English/Malayalam corpus in the training phase, the machine automatically generates Malayalam translations of English sentences. This paper also discusses a technique to improve the alignment model by incorporating the parts of speech information into the bilingual corpus. Removing the insignificant alignments from the sentence pairs by this approach has ensured better training results. Pre-processing techniques like suffix separation from the Malayalam corpus and stop word elimination from the bilingual corpus also proved to be effective in training. Various handcrafted rules designed for the suffix separation process which can be used as a guideline in implementing suffix separation in Malayalam language are also presented in this paper. The structural difference between the English Malayalam pair is resolved in the decoder by applying the order conversion rules. Experiments conducted on a sample corpus have generated reasonably good Malayalam translations and the results are verified with F measure, BLEU and WER evaluation metrics

Relevância:

80.00% 80.00%

Publicador:

Resumo:

In Statistical Machine Translation from English to Malayalam, an unseen English sentence is translated into its equivalent Malayalam sentence using statistical models. A parallel corpus of English-Malayalam is used in the training phase. Word to word alignments has to be set among the sentence pairs of the source and target language before subjecting them for training. This paper deals with certain techniques which can be adopted for improving the alignment model of SMT. Methods to incorporate the parts of speech information into the bilingual corpus has resulted in eliminating many of the insignificant alignments. Also identifying the name entities and cognates present in the sentence pairs has proved to be advantageous while setting up the alignments. Presence of Malayalam words with predictable translations has also contributed in reducing the insignificant alignments. Moreover, reduction of the unwanted alignments has brought in better training results. Experiments conducted on a sample corpus have generated reasonably good Malayalam translations and the results are verified with F measure, BLEU and WER evaluation metrics.