942 resultados para Debugging in computer science


Relevância:

100.00% 100.00%

Publicador:

Resumo:

The present work proposes a method based on CLV (Clustering around Latent Variables) for identifying groups of consumers in L-shape data. This kind of datastructure is very common in consumer studies where a panel of consumers is asked to assess the global liking of a certain number of products and then, preference scores are arranged in a two-way table Y. External information on both products (physicalchemical description or sensory attributes) and consumers (socio-demographic background, purchase behaviours or consumption habits) may be available in a row descriptor matrix X and in a column descriptor matrix Z respectively. The aim of this method is to automatically provide a consumer segmentation where all the three matrices play an active role in the classification, getting homogeneous groups from all points of view: preference, products and consumer characteristics. The proposed clustering method is illustrated on data from preference studies on food products: juices based on berry fruits and traditional cheeses from Trentino. The hedonic ratings given by the consumer panel on the products under study were explained with respect to the product chemical compounds, sensory evaluation and consumer socio-demographic information, purchase behaviour and consumption habits.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

[EN]In this paper, we address the challenge of gender classi - cation using large databases of images with two goals. The rst objective is to evaluate whether the error rate decreases compared to smaller databases. The second goal is to determine if the classi er that provides the best classi cation rate for one database, improves the classi cation results for other databases, that is, the cross-database performance.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

[EN]In this paper, we focus on gender recognition in challenging large scale scenarios. Firstly, we review the literature results achieved for the problem in large datasets, and select the currently hardest dataset: The Images of Groups. Secondly, we study the extraction of features from the face and its local context to improve the recognition accuracy. Diff erent descriptors, resolutions and classfii ers are studied, overcoming previous literature results, reaching an accuracy of 89.8%.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

[EN]Re-identi fication is commonly accomplished using appearance features based on salient points and color information. In this paper, we make an study on the use of di fferent features exclusively obtained from depth images captured with RGB-D cameras. The results achieved, using simple geometric features extracted in a top-view setup, seem to provide useful descriptors for the re-identi fication task.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

[EN]This paper focuses on four different initialization methods for determining the initial shape for the AAM algorithm and their particular performance in two different classification tasks with respect to either the facial expression DaFEx database and to the real world data obtained from a robot’s point of view.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

[ES]This paper describes some simple but useful computer vision techniques for human-robot interaction. First, an omnidirectional camera setting is described that can detect people in the surroundings of the robot, giving their angular positions and a rough estimate of the distance. The device can be easily built with inexpensive components. Second, we comment on a color-based face detection technique that can alleviate skin-color false positives. Third, a simple head nod and shake detector is described, suitable for detecting affirmative/negative, approval/dissaproval, understanding/disbelief head gestures.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

A prevalent claim is that we are in knowledge economy. When we talk about knowledge economy, we generally mean the concept of “Knowledge-based economy” indicating the use of knowledge and technologies to produce economic benefits. Hence knowledge is both tool and raw material (people’s skill) for producing some kind of product or service. In this kind of environment economic organization is undergoing several changes. For example authority relations are less important, legal and ownership-based definitions of the boundaries of the firm are becoming irrelevant and there are only few constraints on the set of coordination mechanisms. Hence what characterises a knowledge economy is the growing importance of human capital in productive processes (Foss, 2005) and the increasing knowledge intensity of jobs (Hodgson, 1999). Economic processes are also highly intertwined with social processes: they are likely to be informal and reciprocal rather than formal and negotiated. Another important point is also the problem of the division of labor: as economic activity becomes mainly intellectual and requires the integration of specific and idiosyncratic skills, the task of dividing the job and assigning it to the most appropriate individuals becomes arduous, a “supervisory problem” (Hogdson, 1999) emerges and traditional hierarchical control may result increasingly ineffective. Not only specificity of know how makes it awkward to monitor the execution of tasks, more importantly, top-down integration of skills may be difficult because ‘the nominal supervisors will not know the best way of doing the job – or even the precise purpose of the specialist job itself – and the worker will know better’ (Hogdson,1999). We, therefore, expect that the organization of the economic activity of specialists should be, at least partially, self-organized. The aim of this thesis is to bridge studies from computer science and in particular from Peer-to-Peer Networks (P2P) to organization theories. We think that the P2P paradigm well fits with organization problems related to all those situation in which a central authority is not possible. We believe that P2P Networks show a number of characteristics similar to firms working in a knowledge-based economy and hence that the methodology used for studying P2P Networks can be applied to organization studies. Three are the main characteristics we think P2P have in common with firms involved in knowledge economy: - Decentralization: in a pure P2P system every peer is an equal participant, there is no central authority governing the actions of the single peers; - Cost of ownership: P2P computing implies shared ownership reducing the cost of owing the systems and the content, and the cost of maintaining them; - Self-Organization: it refers to the process in a system leading to the emergence of global order within the system without the presence of another system dictating this order. These characteristics are present also in the kind of firm that we try to address and that’ why we have shifted the techniques we adopted for studies in computer science (Marcozzi et al., 2005; Hales et al., 2007 [39]) to management science.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The aim of the present work is to contribute to a better understanding of the relation between organization theory and management practice. It is organized as a collection of two papers, a theoretical and conceptual contribution and an ethnographic study. The first paper is concerned with systematizing different literatures inside and outside the field of organization studies that deal with the theory-practice relation. After identifying a series of positions to the theory-practice debate and unfolding some of their implicit assumptions and limitations, a new position called entwinement is developed in order to overcome status quo through reconciliation and integration. Accordingly, the paper proposes to reconceptualize theory and practice as a circular iterative process of action and cognition, science and common-sense enacted in the real world both by organization scholars and practitioners according to purposes at hand. The second paper is the ethnographic study of an encounter between two groups of expert academics and practitioners occasioned by a one-year executive business master in an international business school. The research articulates a process view of the knowledge exchange between management academics and practitioners in particular and between individuals belonging to different communities of practice, in general, and emphasizes its dynamic, relational and transformative mechanisms. Findings show that when they are given the chance to interact, academics and practitioners set up local provisional relations that enable them to act as change intermediaries vis-a-vis each other’s worlds, without tying themselves irremediably to each other and to the scenarios they conjointly projected during the master’s experience. Finally, the study shows that provisional relations were accompanied by a recursive shift in knowledge modes. While interacting, academics passed from theory to practical theorizing, practitioners passed from an involved practical mode to a reflexive and quasi-theoretical one, and then, as exchanges proceeded, the other way around.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The subject of this work concerns the study of the immigration phenomenon, with emphasis on the aspects related to the integration of an immigrant population in a hosting one. Aim of this work is to show the forecasting ability of a recent finding where the behavior of integration quantifiers was analyzed and investigated with a mathematical model of statistical physics origins (a generalization of the monomer dimer model). After providing a detailed literature review of the model, we show that not only such a model is able to identify the social mechanism that drives a particular integration process, but it also provides correct forecast. The research reported here proves that the proposed model of integration and its forecast framework are simple and effective tools to reduce uncertainties about how integration phenomena emerge and how they are likely to develop in response to increased migration levels in the future.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Questa tesi si focalizza sullo studio dei modelli fisico-matematici attualmente in uso per la simulazione di fluidi al calcolatore con l’obiettivo di fornire nozioni di base e avanzate sull’utilizzo di tali metodi. La trattazione ha lo scopo di facilitare la comprensione dei principi su cui si fonda la simulazione di fluidi e rappresenta una base per la creazione di un proprio simulatore. E’ possibile studiare le caratteristiche di un fluido in movimento mediante due approcci diversi, l’approccio lagrangiano e l’approccio euleriano. Mentre l’approccio lagrangiano ha lo scopo di conoscere il valore, nel tempo, di una qualsiasi proprietà di ciascuna particella che compone il fluido, l’approccio euleriano, fissato uno o più punti del volume di spazio occupato da quest’ultimo, vuole studiare quello che accade, nel tempo, in quei punti. In particolare, questa tesi approfondisce lo studio delle equazioni di Navier-Stokes, approcciandosi al problema in maniera euleriana. La soluzione numerica del sistema di equazioni differenziali alle derivate parziali derivante dalle equazioni sopracitate, approssima la velocità del fluido, a partire dalla quale è possibile risalire a tutte le grandezze che lo caratterizzano. Attenzione viene riservata anche ad un modello facente parte dell’approccio semi-lagrangiano, il Lattice Boltzmann, considerato una via di mezzo tra i metodi puramente euleriani e quelli lagrangiani, che si basa sulla soluzione dell’equazione di Boltzmann mediante modelli di collisione di particelle. Infine, analogamente al metodo di Lattice Boltzmann, viene trattato il metodo Smoothed Particles Hydrodynamics, tipicamente lagrangiano, secondo il quale solo le proprietà delle particelle comprese dentro il raggio di una funzione kernel, centrata nella particella di interesse, influenzano il valore della particella stessa. Un resoconto pratico della teoria trattata viene dato mediante delle simulazioni realizzate tramite il software Blender 2.76b.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Computed tomography based navigation for endoscopic sinus surgery is inflationary used despite of major public concern about iatrogenic radiation induced cancer risk. Studies on dose reduction for CAS-CT are almost nonexistent. We validate the use of radiation dose reduced CAS-CT for clinically applied surface registration.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Delineating brain tumor boundaries from magnetic resonance images is an essential task for the analysis of brain cancer. We propose a fully automatic method for brain tissue segmentation, which combines Support Vector Machine classification using multispectral intensities and textures with subsequent hierarchical regularization based on Conditional Random Fields. The CRF regularization introduces spatial constraints to the powerful SVM classification, which assumes voxels to be independent from their neighbors. The approach first separates healthy and tumor tissue before both regions are subclassified into cerebrospinal fluid, white matter, gray matter and necrotic, active, edema region respectively in a novel hierarchical way. The hierarchical approach adds robustness and speed by allowing to apply different levels of regularization at different stages. The method is fast and tailored to standard clinical acquisition protocols. It was assessed on 10 multispectral patient datasets with results outperforming previous methods in terms of segmentation detail and computation times.