868 resultados para Debugging in computer science.


Relevância:

100.00% 100.00%

Publicador:

Resumo:

A prevalent claim is that we are in knowledge economy. When we talk about knowledge economy, we generally mean the concept of “Knowledge-based economy” indicating the use of knowledge and technologies to produce economic benefits. Hence knowledge is both tool and raw material (people’s skill) for producing some kind of product or service. In this kind of environment economic organization is undergoing several changes. For example authority relations are less important, legal and ownership-based definitions of the boundaries of the firm are becoming irrelevant and there are only few constraints on the set of coordination mechanisms. Hence what characterises a knowledge economy is the growing importance of human capital in productive processes (Foss, 2005) and the increasing knowledge intensity of jobs (Hodgson, 1999). Economic processes are also highly intertwined with social processes: they are likely to be informal and reciprocal rather than formal and negotiated. Another important point is also the problem of the division of labor: as economic activity becomes mainly intellectual and requires the integration of specific and idiosyncratic skills, the task of dividing the job and assigning it to the most appropriate individuals becomes arduous, a “supervisory problem” (Hogdson, 1999) emerges and traditional hierarchical control may result increasingly ineffective. Not only specificity of know how makes it awkward to monitor the execution of tasks, more importantly, top-down integration of skills may be difficult because ‘the nominal supervisors will not know the best way of doing the job – or even the precise purpose of the specialist job itself – and the worker will know better’ (Hogdson,1999). We, therefore, expect that the organization of the economic activity of specialists should be, at least partially, self-organized. The aim of this thesis is to bridge studies from computer science and in particular from Peer-to-Peer Networks (P2P) to organization theories. We think that the P2P paradigm well fits with organization problems related to all those situation in which a central authority is not possible. We believe that P2P Networks show a number of characteristics similar to firms working in a knowledge-based economy and hence that the methodology used for studying P2P Networks can be applied to organization studies. Three are the main characteristics we think P2P have in common with firms involved in knowledge economy: - Decentralization: in a pure P2P system every peer is an equal participant, there is no central authority governing the actions of the single peers; - Cost of ownership: P2P computing implies shared ownership reducing the cost of owing the systems and the content, and the cost of maintaining them; - Self-Organization: it refers to the process in a system leading to the emergence of global order within the system without the presence of another system dictating this order. These characteristics are present also in the kind of firm that we try to address and that’ why we have shifted the techniques we adopted for studies in computer science (Marcozzi et al., 2005; Hales et al., 2007 [39]) to management science.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The aim of the present work is to contribute to a better understanding of the relation between organization theory and management practice. It is organized as a collection of two papers, a theoretical and conceptual contribution and an ethnographic study. The first paper is concerned with systematizing different literatures inside and outside the field of organization studies that deal with the theory-practice relation. After identifying a series of positions to the theory-practice debate and unfolding some of their implicit assumptions and limitations, a new position called entwinement is developed in order to overcome status quo through reconciliation and integration. Accordingly, the paper proposes to reconceptualize theory and practice as a circular iterative process of action and cognition, science and common-sense enacted in the real world both by organization scholars and practitioners according to purposes at hand. The second paper is the ethnographic study of an encounter between two groups of expert academics and practitioners occasioned by a one-year executive business master in an international business school. The research articulates a process view of the knowledge exchange between management academics and practitioners in particular and between individuals belonging to different communities of practice, in general, and emphasizes its dynamic, relational and transformative mechanisms. Findings show that when they are given the chance to interact, academics and practitioners set up local provisional relations that enable them to act as change intermediaries vis-a-vis each other’s worlds, without tying themselves irremediably to each other and to the scenarios they conjointly projected during the master’s experience. Finally, the study shows that provisional relations were accompanied by a recursive shift in knowledge modes. While interacting, academics passed from theory to practical theorizing, practitioners passed from an involved practical mode to a reflexive and quasi-theoretical one, and then, as exchanges proceeded, the other way around.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The subject of this work concerns the study of the immigration phenomenon, with emphasis on the aspects related to the integration of an immigrant population in a hosting one. Aim of this work is to show the forecasting ability of a recent finding where the behavior of integration quantifiers was analyzed and investigated with a mathematical model of statistical physics origins (a generalization of the monomer dimer model). After providing a detailed literature review of the model, we show that not only such a model is able to identify the social mechanism that drives a particular integration process, but it also provides correct forecast. The research reported here proves that the proposed model of integration and its forecast framework are simple and effective tools to reduce uncertainties about how integration phenomena emerge and how they are likely to develop in response to increased migration levels in the future.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Questa tesi si focalizza sullo studio dei modelli fisico-matematici attualmente in uso per la simulazione di fluidi al calcolatore con l’obiettivo di fornire nozioni di base e avanzate sull’utilizzo di tali metodi. La trattazione ha lo scopo di facilitare la comprensione dei principi su cui si fonda la simulazione di fluidi e rappresenta una base per la creazione di un proprio simulatore. E’ possibile studiare le caratteristiche di un fluido in movimento mediante due approcci diversi, l’approccio lagrangiano e l’approccio euleriano. Mentre l’approccio lagrangiano ha lo scopo di conoscere il valore, nel tempo, di una qualsiasi proprietà di ciascuna particella che compone il fluido, l’approccio euleriano, fissato uno o più punti del volume di spazio occupato da quest’ultimo, vuole studiare quello che accade, nel tempo, in quei punti. In particolare, questa tesi approfondisce lo studio delle equazioni di Navier-Stokes, approcciandosi al problema in maniera euleriana. La soluzione numerica del sistema di equazioni differenziali alle derivate parziali derivante dalle equazioni sopracitate, approssima la velocità del fluido, a partire dalla quale è possibile risalire a tutte le grandezze che lo caratterizzano. Attenzione viene riservata anche ad un modello facente parte dell’approccio semi-lagrangiano, il Lattice Boltzmann, considerato una via di mezzo tra i metodi puramente euleriani e quelli lagrangiani, che si basa sulla soluzione dell’equazione di Boltzmann mediante modelli di collisione di particelle. Infine, analogamente al metodo di Lattice Boltzmann, viene trattato il metodo Smoothed Particles Hydrodynamics, tipicamente lagrangiano, secondo il quale solo le proprietà delle particelle comprese dentro il raggio di una funzione kernel, centrata nella particella di interesse, influenzano il valore della particella stessa. Un resoconto pratico della teoria trattata viene dato mediante delle simulazioni realizzate tramite il software Blender 2.76b.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Computed tomography based navigation for endoscopic sinus surgery is inflationary used despite of major public concern about iatrogenic radiation induced cancer risk. Studies on dose reduction for CAS-CT are almost nonexistent. We validate the use of radiation dose reduced CAS-CT for clinically applied surface registration.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Delineating brain tumor boundaries from magnetic resonance images is an essential task for the analysis of brain cancer. We propose a fully automatic method for brain tissue segmentation, which combines Support Vector Machine classification using multispectral intensities and textures with subsequent hierarchical regularization based on Conditional Random Fields. The CRF regularization introduces spatial constraints to the powerful SVM classification, which assumes voxels to be independent from their neighbors. The approach first separates healthy and tumor tissue before both regions are subclassified into cerebrospinal fluid, white matter, gray matter and necrotic, active, edema region respectively in a novel hierarchical way. The hierarchical approach adds robustness and speed by allowing to apply different levels of regularization at different stages. The method is fast and tailored to standard clinical acquisition protocols. It was assessed on 10 multispectral patient datasets with results outperforming previous methods in terms of segmentation detail and computation times.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The biggest challenge facing software developers today is how to gracefully evolve complex software systems in the face of changing requirements. We clearly need software systems to be more dynamic, compositional and model-centric, but instead we continue to build systems that are static, baroque and inflexible. How can we better build change-enabled systems in the future? To answer this question, we propose to look back to one of the most successful systems to support change, namely Smalltalk. We briefly introduce Smalltalk with a few simple examples, and draw some lessons for software evolution. Smalltalk's simplicity, its reflective design, and its highly dynamic nature all go a long way towards enabling change in Smalltalk applications. We then illustrate how these lessons work in practice by reviewing a number of research projects that support software evolution by exploiting Smalltalk's design. We conclude by summarizing open issues and challenges for change-enabled systems of the future.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

Background Increasing attention is being paid to improvement in undergraduate science, technology, engineering, and mathematics (STEM) education through increased adoption of research-based instructional strategies (RBIS), but high-quality measures of faculty instructional practice do not exist to monitor progress. Purpose/Hypothesis The measure of how well an implemented intervention follows the original is called fidelity of implementation. This theory was used to address the research questions: What is the fidelity of implementation of selected RBIS in engineering science courses? That is, how closely does engineering science classroom practice reflect the intentions of the original developers? Do the critical components that characterize an RBIS discriminate between engineering science faculty members who claimed use of the RBIS and those who did not? Design/Method A survey of 387 U.S. faculty teaching engineering science courses (e.g., statics, circuits, thermodynamics) included questions about class time spent on 16 critical components and use of 11 corresponding RBIS. Fidelity was quantified as the percentage of RBIS users who also spent time on corresponding critical components. Discrimination between users and nonusers was tested using chi square. Results Overall fidelity of the 11 RBIS ranged from 11% to 80% of users spending time on all required components. Fidelity was highest for RBIS with one required component: case-based teaching, just-in-time teaching, and inquiry learning. Thirteen of 16 critical components discriminated between users and nonusers for all RBIS to which they were mapped. Conclusions Results were consistent with initial mapping of critical components to RBIS. Fidelity of implementation is a potentially useful framework for future work in STEM undergraduate education.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

The progress of wet age-related macular degeneration can now be controlled by intravitreal drug injection. This approach requires repeated injections, which could be avoided by delivering the drug to the retina. Intraocular implants are a promising solution for drug delivery near the retina. Currently, their accurate placement is challenging, and they can only be removed after a vitrectomy. In this paper, we introduce an approach for minimally invasive retinal drug delivery using magnetic intraocular inserts. We briefly discuss the electromagnetic-control system for magnetic implants and then focus on evaluating their ability to move in the vitreous humor. The mobility of magnetic intraocular implants is estimated in vitro with synthesized vitreous humors, and ex vivo with experiments on cadaver porcine eyes. Preliminary results show that with such magnetic implants a vitrectomy can be avoided.

Relevância:

100.00% 100.00%

Publicador:

Resumo:

This study explored how academics' beliefs about teaching and learning influenced their teaching in engineering science courses typically taught in the second or third year of 4-year engineering undergraduate degrees. Data were collected via a national survey of 166 U. S. statics instructors and interviews at two different institutions with 17 instructors of engineering science courses such as thermodynamics, circuits and statics. The study identified a number of common beliefs about how to best support student learning of these topics; each is discussed in relation to the literature about student development and learning. Specific recommendations are given for educational developers to encourage use of research-based instructional strategies in these courses.