970 resultados para QUT Speaker Identity Verification System
Resumo:
Traditionally, a country's electoral system requires the voter to vote at a specific day and place, which conflicts with the mobility usually seen in modern live styles. Thus, the widespread of Internet (mobile) broadband access can be seen as an opportunity to deal with this mobility problem, i.e. the adoption of an Internet voting system can make the live of voter's much more convenient; however, a widespread Internet voting systems adoption relies on the ability to develop trustworthy systems, i.e. systems that are verifiable and preserve the voter's privacy. Building such a system is still an open research problem. Our contribution is a new Internet voting system: EVIV, a highly sound End-to-end Verifiable Internet Voting system, which offers full voter's mobility and preserves the voter's privacy from the vote casting PC even if the voter votes from a public PC, such as a PC at a cybercafe or at a public library. Additionally, EVIV has private vote verification mechanisms, in which the voter just has to perform a simple match of two small strings (4-5 alphanumeric characters), that detect and protect against vote manipulations both at the insecure vote client platform and at the election server side. (c) 2012 Elsevier Ltd. All rights reserved.
The use of non-standard CT conversion ramps for Monte Carlo verification of 6 MV prostate IMRT plans
Resumo:
Monte Carlo (MC) dose calculation algorithms have been widely used to verify the accuracy of intensity-modulated radiotherapy (IMRT) dose distributions computed by conventional algorithms due to the ability to precisely account for the effects of tissue inhomogeneities and multileaf collimator characteristics. Both algorithms present, however, a particular difference in terms of dose calculation and report. Whereas dose from conventional methods is traditionally computed and reported as the water-equivalent dose (Dw), MC dose algorithms calculate and report dose to medium (Dm). In order to compare consistently both methods, the conversion of MC Dm into Dw is therefore necessary. This study aims to assess the effect of applying the conversion of MC-based Dm distributions to Dw for prostate IMRT plans generated for 6 MV photon beams. MC phantoms were created from the patient CT images using three different ramps to convert CT numbers into material and mass density: a conventional four material ramp (CTCREATE) and two simplified CT conversion ramps: (1) air and water with variable densities and (2) air and water with unit density. MC simulations were performed using the BEAMnrc code for the treatment head simulation and the DOSXYZnrc code for the patient dose calculation. The conversion of Dm to Dw by scaling with the stopping power ratios of water to medium was also performed in a post-MC calculation process. The comparison of MC dose distributions calculated in conventional and simplified (water with variable densities) phantoms showed that the effect of material composition on dose-volume histograms (DVH) was less than 1% for soft tissue and about 2.5% near and inside bone structures. The effect of material density on DVH was less than 1% for all tissues through the comparison of MC distributions performed in the two simplified phantoms considering water. Additionally, MC dose distributions were compared with the predictions from an Eclipse treatment planning system (TPS), which employed a pencil beam convolution (PBC) algorithm with Modified Batho Power Law heterogeneity correction. Eclipse PBC and MC calculations (conventional and simplified phantoms) agreed well (<1%) for soft tissues. For femoral heads, differences up to 3% were observed between the DVH for Eclipse PBC and MC calculated in conventional phantoms. The use of the CT conversion ramp of water with variable densities for MC simulations showed no dose discrepancies (0.5%) with the PBC algorithm. Moreover, converting Dm to Dw using mass stopping power ratios resulted in a significant shift (up to 6%) in the DVH for the femoral heads compared to the Eclipse PBC one. Our results show that, for prostate IMRT plans delivered with 6 MV photon beams, no conversion of MC dose from medium to water using stopping power ratio is needed. In contrast, MC dose calculations using water with variable density may be a simple way to solve the problem found using the dose conversion method based on the stopping power ratio.
Resumo:
The premise of this paper is that a model for communicating the national value system must start from a strategy aimed at the identification, the cultivation and communication of values that give consistency to the value system. The analysis concentrates on the elements of such strategies and on the implications of applying a value communication program on the identity architecture of the community. The paper will also discuss the role of the national value system in the context of the emerging global culture, where the individual has the power to create his/her own hybrid cultural model.
Resumo:
Dissertação apresentada na Faculdade de Ciências e Tecnologia da Universidade Nova de Lisboa para a obtenção do Grau de Mestre em Engenharia Informática.
Resumo:
Dissertação apresentada na Faculdade de Ciências e Tecnologia da Universidade Nova de Lisboa para a obtenção do grau de Mestre em Engenharia Informática.
Resumo:
Os dispositivos móveis são pessoais, intransmissíveis e cada vez mais utilizados, tornando-se assim numa boa ferramenta para a realização de um conjunto de serviços na indústria hoteleira. Entre esses serviços que necessitam da identificação pessoal, encontram-se a possibilidade do cliente reservar um quarto ou utilizar o serviço de quartos. Atualmente é muito utilizado, nos locais de alojamento, um smart card que possibilite ao cliente ter acesso a alguns dos serviços disponíveis. O objetivo deste documento é apresentar uma alternativa ao sistema de cartões, utilizando para o efeito, dispositivos móveis. De modo a garantir a segurança e uma utilização semelhante ao sistema de cartões existentes foi utilizada a tecnologia NFC (Near Field Communication) que, ao permitir o modo de emulação de cartão, facilita a transação do sistema de smart card existente, para o da utilização de dispositivos móveis na realização das mesmas funções. Mais concretamente, será abordada a utilização de smartphones para o processo de abertura de portas. Para que exista uma melhor compreensão e para que haja um conhecimento das suas capacidades e limites foram estudados casos de uso da tecnologia NFC. Este documento apresenta ainda os processos de desenvolvimento de uma aplicação nativa para o sistema operativo Android, cujo objetivo é proporcionar ao cliente de um local de alojamento um novo modo de acesso ao quarto, utilizando a tecnologia NFC. Para além desta funcionalidade a aplicação permite ainda ao utilizador fazer reservas, fazer o check-in, fazer o check-out entre outras. Posteriormente serão apresentadas as conclusões e possíveis trabalhos futuros.
Resumo:
Poster presented in The 28th GI/ITG International Conference on Architecture of Computing Systems (ARCS 2015). 24 to 26, Mar, 2015. Porto, Portugal.
Resumo:
Presented at SEMINAR "ACTION TEMPS RÉEL:INFRASTRUCTURES ET SERVICES SYSTÉMES". 10, Apr, 2015. Brussels, Belgium.
Resumo:
work presented in the context of the European Master’s program in Computational Logic, as the partial requirement for obtaining Master of Science degree in Computational Logic
Resumo:
Dissertação para obtenção do Grau de Mestre em Engenharia Informática
Resumo:
In a reconfigurable system, the response to contextual or internal change may trigger reconfiguration events which, on their turn, activate scripts that change the system׳s architecture at runtime. To be safe, however, such reconfigurations are expected to obey the fundamental principles originally specified by its architect. This paper introduces an approach to ensure that such principles are observed along reconfigurations by verifying them against concrete specifications in a suitable logic. Architectures, reconfiguration scripts, and principles are specified in Archery, an architectural description language with formal semantics. Principles are encoded as constraints, which become formulas of a two-layer graded hybrid logic, where the upper layer restricts reconfigurations, and the lower layer constrains the resulting configurations. Constraints are verified by translating them into logic formulas, which are interpreted over models derived from Archery specifications of architectures and reconfigurations. Suitable notions of bisimulation and refinement, to which the architect may resort to compare configurations, are given, and their relationship with modal validity is discussed.
Resumo:
Identificación y caracterización del problema. Uno de los problemas más importantes asociados con la construcción de software es la corrección del mismo. En busca de proveer garantías del correcto funcionamiento del software, han surgido una variedad de técnicas de desarrollo con sólidas bases matemáticas y lógicas conocidas como métodos formales. Debido a su naturaleza, la aplicación de métodos formales requiere gran experiencia y conocimientos, sobre todo en lo concerniente a matemáticas y lógica, por lo cual su aplicación resulta costosa en la práctica. Esto ha provocado que su principal aplicación se limite a sistemas críticos, es decir, sistemas cuyo mal funcionamiento puede causar daños de magnitud, aunque los beneficios que sus técnicas proveen son relevantes a todo tipo de software. Poder trasladar los beneficios de los métodos formales a contextos de desarrollo de software más amplios que los sistemas críticos tendría un alto impacto en la productividad en tales contextos. Hipótesis. Contar con herramientas de análisis automático es un elemento de gran importancia. Ejemplos de esto son varias herramientas potentes de análisis basadas en métodos formales, cuya aplicación apunta directamente a código fuente. En la amplia mayoría de estas herramientas, la brecha entre las nociones a las cuales están acostumbrados los desarrolladores y aquellas necesarias para la aplicación de estas herramientas de análisis formal sigue siendo demasiado amplia. Muchas herramientas utilizan lenguajes de aserciones que escapan a los conocimientos y las costumbres usuales de los desarrolladores. Además, en muchos casos la salida brindada por la herramienta de análisis requiere cierto manejo del método formal subyacente. Este problema puede aliviarse mediante la producción de herramientas adecuadas. Otro problema intrínseco a las técnicas automáticas de análisis es cómo se comportan las mismas a medida que el tamaño y complejidad de los elementos a analizar crece (escalabilidad). Esta limitación es ampliamente conocida y es considerada crítica en la aplicabilidad de métodos formales de análisis en la práctica. Una forma de atacar este problema es el aprovechamiento de información y características de dominios específicos de aplicación. Planteo de objetivos. Este proyecto apunta a la construcción de herramientas de análisis formal para contribuir a la calidad, en cuanto a su corrección funcional, de especificaciones, modelos o código, en el contexto del desarrollo de software. Más precisamente, se busca, por un lado, identificar ambientes específicos en los cuales ciertas técnicas de análisis automático, como el análisis basado en SMT o SAT solving, o el model checking, puedan llevarse a niveles de escalabilidad superiores a los conocidos para estas técnicas en ámbitos generales. Se intentará implementar las adaptaciones a las técnicas elegidas en herramientas que permitan su uso a desarrolladores familiarizados con el contexto de aplicación, pero no necesariamente conocedores de los métodos o técnicas subyacentes. Materiales y métodos a utilizar. Los materiales a emplear serán bibliografía relevante al área y equipamiento informático. Métodos. Se emplearán los métodos propios de la matemática discreta, la lógica y la ingeniería de software. Resultados esperados. Uno de los resultados esperados del proyecto es la individualización de ámbitos específicos de aplicación de métodos formales de análisis. Se espera que como resultado del desarrollo del proyecto surjan herramientas de análisis cuyo nivel de usabilidad sea adecuado para su aplicación por parte de desarrolladores sin formación específica en los métodos formales utilizados. Importancia del proyecto. El principal impacto de este proyecto será la contribución a la aplicación práctica de técnicas formales de análisis en diferentes etapas del desarrollo de software, con la finalidad de incrementar su calidad y confiabilidad. A crucial factor for software quality is correcteness. Traditionally, formal approaches to software development concentrate on functional correctness, and tackle this problem basically by being based on well defined notations founded on solid mathematical grounds. This makes formal methods better suited for analysis, due to their precise semantics, but they are usually more complex, and require familiarity and experience with the manipulation of mathematical definitions. So, their acceptance by software engineers is rather restricted, and formal methods applications have been confined to critical systems. Nevertheless, it is obvious that the advantages that formal methods provide apply to any kind of software system. It is accepted that appropriate software tool support for formal analysis is essential, if one seeks providing support for software development based on formal methods. Indeed, some of the relatively recent sucesses of formal methods are accompanied by good quality tools that automate powerful analysis mechanisms, and are even integrated in widely used development environments. Still, most of these tools either concentrate on code analysis, and in many cases are still far from being simple enough to be employed by software engineers without experience in formal methods. Another important problem for the adoption of tool support for formal methods is scalability. Automated software analysis is intrinsically complex, and thus techniques do not scale well in the general case. In this project, we will attempt to identify particular modelling, design, specification or coding activities in software development processes where to apply automated formal analysis techniques. By focusing in very specific application domains, we expect to find characteristics that might be exploited to increase the scalability of the corresponding analyses, compared to the general case.
Resumo:
Magdeburg, Univ., Fak. für Informatik, Diss., 2015
Resumo:
Based on homology with GLUT1-5, we have isolated a cDNA for a novel glucose transporter, GLUTX1. This cDNA encodes a protein of 478 amino acids that shows between 29 and 32% identity with rat GLUT1-5 and 32-36% identity with plant and bacterial hexose transporters. Unlike GLUT1-5, GLUTX1 has a short extracellular loop between transmembrane domain (TM) 1 and TM2 and a long extracellular loop between TM9 and TM10 that contains the only N-glycosylation site. When expressed in Xenopus oocytes, GLUTX1 showed strong transport activity only after suppression of a dileucine internalization motif present in the amino-terminal region. Transport activity was inhibited by cytochalasin B and partly competed by D-fructose and D-galactose. The Michaelis-Menten constant for glucose was approximately 2 mM. When translated in reticulocytes lysates, GLUTX1 migrates as a 35-kDa protein that becomes glycosylated in the presence of microsomal membranes. Western blot analysis of GLUTX1 transiently expressed in HEK293T cells revealed a diffuse band with a molecular mass of 37-50 kDa that could be converted to a approximately 35-kDa polypeptide following enzymatic deglycosylation. Immunofluorescence microscopy detection of GLUTX1 transfected into HEK293T cells showed an intracellular staining. Mutation of the dileucine internalization motif induced expression of GLUTX1 at the cell surface. GLUTX1 mRNA was detected in testis, hypothalamus, cerebellum, brainstem, hippocampus, and adrenal gland. We hypothesize that, in a similar fashion to GLUT4, in vivo cell surface expression of GLUTX1 may be inducible by a hormonal or other stimulus.
Resumo:
Reticulitermes santonensis is a subterranean termite that invades urban areas in France and elsewhere where it causes damage to human-built structures. We investigated the breeding system, colony and population genetic structure, and mode of dispersal of two French populations of R. santonensis. Termite workers were sampled from 43 and 31 collection points, respectively, from a natural population in west-central France (in and around the island of Oleron) and an urban population (Paris). Ten to 20 workers per collection point were genotyped at nine variable microsatellite loci to determine colony identity and to infer colony breeding structure. There was a total of 26 colonies, some of which were spatially expansive, extending up to 320 linear metres. Altogether, the analysis of genotype distribution, F-statistics and relatedness coefficients suggested that all colonies were extended families headed by numerous neotenics (nonwinged precocious reproductives) probably descended from pairs of primary (winged) reproductives. Isolation by distance among collection points within two large colonies from both populations suggested spatially separated reproductive centres with restricted movement of workers and neotenics. There was a moderate level of genetic differentiation (F(ST) = 0.10) between the Oleron and Paris populations, and the number of alleles was significantly higher in Oleron than in Paris, as expected if the Paris population went through bottlenecks when it was introduced from western France. We hypothesize that the diverse and flexible breeding systems found in subterranean termites pre-adapt them to invade new or marginal habitats. Considering that R. santonensis may be an introduced population of the North American species R. flavipes, a breeding system consisting primarily of extended family colonies containing many neotenic reproductives may facilitate human-mediated spread and establishment of R. santonensis in urban areas with harsh climates.