44 resultados para Massimo, Vittorio.
em Queensland University of Technology - ePrints Archive
Resumo:
What happens when patterns become all pervasive? When pattern contagiously corrupts and saturates adjacent objects, artefacts and surfaces; blurring internal and external environment and dissolving any single point of perspective or static conception of space. Mark Taylor ruminates on the possibilities of relentless patterning in interior space in both a historic and a contemporary context.
Resumo:
A data-driven background dataset refinement technique was recently proposed for SVM based speaker verification. This method selects a refined SVM background dataset from a set of candidate impostor examples after individually ranking examples by their relevance. This paper extends this technique to the refinement of the T-norm dataset for SVM-based speaker verification. The independent refinement of the background and T-norm datasets provides a means of investigating the sensitivity of SVM-based speaker verification performance to the selection of each of these datasets. Using refined datasets provided improvements of 13% in min. DCF and 9% in EER over the full set of impostor examples on the 2006 SRE corpus with the majority of these gains due to refinement of the T-norm dataset. Similar trends were observed for the unseen data of the NIST 2008 SRE.
Resumo:
This paper presents Scatter Difference Nuisance Attribute Projection (SD-NAP) as an enhancement to NAP for SVM-based speaker verification. While standard NAP may inadvertently remove desirable speaker variability, SD-NAP explicitly de-emphasises this variability by incorporating a weighted version of the between-class scatter into the NAP optimisation criterion. Experimental evaluation of SD-NAP with a variety of SVM systems on the 2006 and 2008 NIST SRE corpora demonstrate that SD-NAP provides improved verification performance over standard NAP in most cases, particularly at the EER operating point.
Resumo:
Spatial organization of Ge islands, grown by physical vapor deposition, on prepatterned Si(001) substrates has been investigated. The substrates were patterned prior to Ge deposition by nanoindentation. Characterization of Ge dots is performed by atomic force microscopy and scanning electron microscopy. The nanoindents act as trapping sites, allowing ripening of Ge islands at those locations during subsequent deposition and diffusion of Ge on the surface. The results show that island ordering is intrinsically linked to the nucleation and growth at indented sites and it strongly depends on pattern parameters.
Resumo:
Eigen-based techniques and other monolithic approaches to face recognition have long been a cornerstone in the face recognition community due to the high dimensionality of face images. Eigen-face techniques provide minimal reconstruction error and limit high-frequency content while linear discriminant-based techniques (fisher-faces) allow the construction of subspaces which preserve discriminatory information. This paper presents a frequency decomposition approach for improved face recognition performance utilising three well-known techniques: Wavelets; Gabor / Log-Gabor; and the Discrete Cosine Transform. Experimentation illustrates that frequency domain partitioning prior to dimensionality reduction increases the information available for classification and greatly increases face recognition performance for both eigen-face and fisher-face approaches.
Resumo:
This paper proposes a semi-supervised intelligent visual surveillance system to exploit the information from multi-camera networks for the monitoring of people and vehicles. Modules are proposed to perform critical surveillance tasks including: the management and calibration of cameras within a multi-camera network; tracking of objects across multiple views; recognition of people utilising biometrics and in particular soft-biometrics; the monitoring of crowds; and activity recognition. Recent advances in these computer vision modules and capability gaps in surveillance technology are also highlighted.
Resumo:
The strain-induced self-assembly of suitable semiconductor pairs is an attractive natural route to nanofabrication. To bring to fruition their full potential for actual applications, individual nanostructures need to be combined into ordered patterns in which the location of each single unit is coupled with others and the surrounding environment. Within the Ge/Si model system, we analyze a number of examples of bottom-up strategies in which the shape, positioning, and actual growth mode of epitaxial nanostructures are tailored by manipulating the intrinsic physical processes of heteroepitaxy. The possibility of controlling elastic interactions and, hence, the configuration of self-assembled quantum dots by modulating surface orientation with the miscut angle is discussed. We focus on the use of atomic steps and step bunching as natural templates for nanodot clustering. Then, we consider several different patterning techniques which allow one to harness the natural self-organization dynamics of the system, such as: scanning tunneling nanolithography, focused ion beam and nanoindentation patterning. By analyzing the evolution of the dot assembly by scanning probe microscopy, we follow the pathway which leads to lateral ordering, discussing the thermodynamic and kinetic effects involved in selective nucleation on patterned substrates.
Resumo:
This position paper provides an overview of work conducted and an outlook of future directions within the field of Information Retrieval (IR) that aims to develop novel models, methods and frameworks inspired by Quantum Theory (QT).
Resumo:
Semiconductor epitaxial nanostructures have been recently proposed as the key building blocks of many innovative applications in materials science and technology. To bring their tremendous potential to fruition, a fine control of nanostructure size and placement is necessary. We present a detailed investigation of the self-ordering process in the prototype case of Ge/Si heteroepitaxy. Starting from a bottom-up strategy (step-bunching instabilities), our analysis moves to lithographic techniques (scanning tunneling lithography, nanomechanical stamping, focused ion beam patterning) with the aim of developing a hybrid approach in which the exogenous intervention is specifically designed to suit and harness the natural self-organization dynamics of the system.
Resumo:
Nowadays, Workflow Management Systems (WfMSs) and, more generally, Process Management Systems (PMPs) are process-aware Information Systems (PAISs), are widely used to support many human organizational activities, ranging from well-understood, relatively stable and structures processes (supply chain management, postal delivery tracking, etc.) to processes that are more complicated, less structured and may exhibit a high degree of variation (health-care, emergency management, etc.). Every aspect of a business process involves a certain amount of knowledge which may be complex depending on the domain of interest. The adequate representation of this knowledge is determined by the modeling language used. Some processes behave in a way that is well understood, predictable and repeatable: the tasks are clearly delineated and the control flow is straightforward. Recent discussions, however, illustrate the increasing demand for solutions for knowledge-intensive processes, where these characteristics are less applicable. The actors involved in the conduct of a knowledge-intensive process have to deal with a high degree of uncertainty. Tasks may be hard to perform and the order in which they need to be performed may be highly variable. Modeling knowledge-intensive processes can be complex as it may be hard to capture at design-time what knowledge is available at run-time. In realistic environments, for example, actors lack important knowledge at execution time or this knowledge can become obsolete as the process progresses. Even if each actor (at some point) has perfect knowledge of the world, it may not be certain of its beliefs at later points in time, since tasks by other actors may change the world without those changes being perceived. Typically, a knowledge-intensive process cannot be adequately modeled by classical, state of the art process/workflow modeling approaches. In some respect there is a lack of maturity when it comes to capturing the semantic aspects involved, both in terms of reasoning about them. The main focus of the 1st International Workshop on Knowledge-intensive Business processes (KiBP 2012) was investigating how techniques from different fields, such as Artificial Intelligence (AI), Knowledge Representation (KR), Business Process Management (BPM), Service Oriented Computing (SOC), etc., can be combined with the aim of improving the modeling and the enactment phases of a knowledge-intensive process. The 1st International Workshop on Knowledge-intensive Business process (KiBP 2012) was held as part of the program of the 2012 Knowledge Representation & Reasoning International Conference (KR 2012) in Rome, Italy, in June 2012. The workshop was hosted by the Dipartimento di Ingegneria Informatica, Automatica e Gestionale Antonio Ruberti of Sapienza Universita di Roma, with financial support of the University, through grant 2010-C26A107CN9 TESTMED, and the EU Commission through the projects FP7-25888 Greener Buildings and FP7-257899 Smart Vortex. This volume contains the 5 papers accepted and presented at the workshop. Each paper was reviewed by three members of the internationally renowned Program Committee. In addition, a further paper was invted for inclusion in the workshop proceedings and for presentation at the workshop. There were two keynote talks, one by Marlon Dumas (Institute of Computer Science, University of Tartu, Estonia) on "Integrated Data and Process Management: Finally?" and the other by Yves Lesperance (Department of Computer Science and Engineering, York University, Canada) on "A Logic-Based Approach to Business Processes Customization" completed the scientific program. We would like to thank all the Program Committee members for the valuable work in selecting the papers, Andrea Marrella for his valuable work as publication and publicity chair of the workshop, and Carola Aiello and the consulting agency Consulta Umbria for the organization of this successful event.
Resumo:
At present, many approaches have been proposed for deformable face alignment with varying degrees of success. However, the common drawback to nearly all these approaches is the inaccurate landmark registrations. The registration errors which occur are predominantly heterogeneous (i.e. low error for some frames in a sequence and higher error for others). In this paper we propose an approach for simultaneously aligning an ensemble of deformable face images stemming from the same subject given noisy heterogeneous landmark estimates. We propose that these initial noisy landmark estimates can be used as an “anchor” in conjunction with known state-of-the-art objectives for unsupervised image ensemble alignment. Impressive alignment performance is obtained using well known deformable face fitting algorithms as “anchors.
Resumo:
Abstract. In recent years, sparse representation based classification(SRC) has received much attention in face recognition with multipletraining samples of each subject. However, it cannot be easily applied toa recognition task with insufficient training samples under uncontrolledenvironments. On the other hand, cohort normalization, as a way of mea-suring the degradation effect under challenging environments in relationto a pool of cohort samples, has been widely used in the area of biometricauthentication. In this paper, for the first time, we introduce cohort nor-malization to SRC-based face recognition with insufficient training sam-ples. Specifically, a user-specific cohort set is selected to normalize theraw residual, which is obtained from comparing the test sample with itssparse representations corresponding to the gallery subject, using poly-nomial regression. Experimental results on AR and FERET databases show that cohort normalization can bring SRC much robustness against various forms of degradation factors for undersampled face recognition.
Resumo:
The geology/reservoir program of the Queensland Geothermal Energy Centre of Excellence (QGECE) has the mission to improve the existing knowledge and develop new innovative scientific approaches for the identification of geothermal resources in Australia, with a particular focus on Queensland. Specifically, the QGECE geology/reservoir program is currently (1) producing a comprehensive geochemical dataset for high heat producing rocks, (2) conducting detailed mineralogical and geochronological studies of granites and hydrothermal alteration minerals, and ; (3) investigating the Cooper Basin representing a superb natural laboratory for understanding of radiogenic heat enrichment process and possible involvement of mantle heat flow. Seven research projects have been established, which are being conducted largely as PhD studies. In the preliminary studies, high quality and valuable results were obtained to address the research topics of understanding the causes and timing of heat producing element enrichment.
Resumo:
We show that SiGe islands are transformed into nanoholes and rings by annealing treatments only and without Si capping. Rings are produced by a rapid flash heating at temperatures higher than the melting point of Ge, whereas nanoholes are produced by several minute annealing. The rings are markedly rich in Si with respect to the pristine islands, suggesting that the evolution path from islands to rings is driven by the selective dissolution of Ge occurring at high temperature.
Resumo:
In recent years face recognition systems have been applied in various useful applications, such as surveillance, access control, criminal investigations, law enforcement, and others. However face biometric systems can be highly vulnerable to spoofing attacks where an impostor tries to bypass the face recognition system using a photo or video sequence. In this paper a novel liveness detection method, based on the 3D structure of the face, is proposed. Processing the 3D curvature of the acquired data, the proposed approach allows a biometric system to distinguish a real face from a photo, increasing the overall performance of the system and reducing its vulnerability. In order to test the real capability of the methodology a 3D face database has been collected simulating spoofing attacks, therefore using photographs instead of real faces. The experimental results show the effectiveness of the proposed approach.