951 resultados para DSA,problemi matematici,didattica personalizzata e individualizzata,consegne modificate
Resumo:
L'argomento principale della mia tesi è la LIM, lavagna interattiva multimediale. Ho suddiviso la tesi in due parti: - nella prima descrivo i nativi digitali e in contrapposizione gli immigrati digitali. - nella seconda parte faccio riferimento ad un caso di studio da me effettuato in una scuola primaria
Resumo:
Nel 1837 il matematico A.F. Möbius definì la funzione aritmetica mu(n) che vale 0 se n è divisibile per il quadrato di un numero primo, (-1)^k se n è il prodotto di k primi distinti e \mu(1)=1. Essa ricopre un ruolo di fondamentale importanza per quanto riguarda la distribuzione dei numeri primi, nonché per la sua duttilità nella risoluzione di diversi problemi di conteggio grazie alla formula di inversione di Möbius, che può essere pensata come un analogo formale del teorema fondamentale del calcolo integrale. Una sorprendente varietà di problemi di calcolo combinatorio si rivelano essere nient'altro che casi particolari di un problema più generale che riguarda la possibilità di invertire una somma fatta sugli elementi di un insieme parzialmente ordinato. L'obiettivo di questo elaborato è quello di illustrare come sia possibile generalizzare il concetto di funzione aritmetica estendendolo a quello di funzione di un'algebra di incidenza. Le algebre di incidenza hanno catturato l'interesse di svariati matematici a partire dagli anni '60 del secolo scorso, e si svilupparono come ambiente naturale nel quale generalizzare la formula di inversione di Mobius. La funzione di Möbius della teoria dei numeri, definita originariamente sull'insieme dei numeri interi positivi ordinato per divisibilità, può quindi essere definita su generici insiemi parzialmente ordinati.
Resumo:
All'interno di questo elaborato vengono analizzati gli effetti positivi e negativi dell'introduzione delle nuove tecnologie nel sistema scolastico all'interno della società odierna, definita società della conoscenza; vengono descritte la lavagna interattiva multimediale e successivamente due software didattici, proponendo al termine una proposta di innovazione attuabile nella scuola secondaria di secondo grado, basata su un mio intervento riguardante l'unità didattica della "Retta nel piano cartesiano" e della "Parabola". Il filo conduttore è quello dell'analisi critica nei confronti delle potenzialità che la rivoluzione digitale porta con sé, cercando di riflettere su quello che potrebbe essere il giusto compromesso tra la nostra scuola e i tempi odierni.
Resumo:
In the paper we present some of the most significant tendencies in Italian LS teaching in the past thirty years (1976 -2006). All of the simplifications concern Italian Verbal Moods: imperative, gerundio presente and gerundio passato, past and present infinitive, congiuntivo trapassato and condizionale composto. In our analysis we want to compare certain features present in the most recent Italian LS grammars with those of the seventies and nineties of the past century.
Resumo:
La maggior parte degli strumenti per la didattica dei sistemi operativi sono basati sull'architettura MIPS, che è ampiamente superata. Questo lavoro introduce μARM, un emulatore, basato su architettura ARM7TDMI adeguato all'insegnamento a livello universitario. Inoltre viene presentato JaeOS, un manuale di specifiche per un sistema operativo multi-strato che supporti esecuzione multi-processo, memoria virtuale, sincronizzazione dei thread, gestione di dispositivi esterni ed un file system. I progetti tradizionali, come OSP2 o OS/161, forniscono un buon quantitativo di codice già sviluppato agli studenti, i quali devono modificare i moduli del sistema già implementati ed aggiungerne di nuovi. Con μARM/JaeOS gli studenti sono sottoposti ad un'esperienza pedagogicamente differente, partendo dal solo emulatore hardware e finendo con un sistema operativo, interamente sviluppato da loro, in grado di eseguire programmi sviluppati dagli studenti stessi.
Resumo:
Frontline employee behaviours are recognised as vital for achieving a competitive advantage for service organisations. The services marketing literature has comprehensively examined ways to improve frontline employee behaviours in service delivery and recovery. However, limited attention has been paid to frontline employee behaviours that favour customers in ways that go against organisational norms or rules. This study examines these behaviours by introducing a behavioural concept of Customer-Oriented Deviance (COD). COD is defined as, “frontline employees exhibiting extra-role behaviours that they perceive to defy existing expectations or prescribed rules of higher authority through service adaptation, communication and use of resources to benefit customers during interpersonal service encounters.” This thesis develops a COD measure and examines the key determinants of these behaviours from a frontline employee perspective. Existing research on similar behaviours that has originated in the positive deviance and pro-social behaviour domains has limitations and is considered inadequate to examine COD in the services context. The absence of a well-developed body of knowledge on non-conforming service behaviours has implications for both theory and practice. The provision of ‘special favours’ increases customer satisfaction but the over-servicing of customers is also counterproductive for the service delivery and costly for the organisation. Despite these implications of non-conforming service behaviours, there is little understanding about the nature of these behaviours and its key drivers. This research builds on inadequacies in prior research on positive deviance, pro-social and pro-customer literature to develop the theoretical foundation of COD. The concept of positive deviance which has predominantly been used to study organisational behaviours is applied within a services marketing setting. Further, it addresses previous limitations in pro-social and pro-customer behavioural literature that has examined limited forms of behaviours with no clear understanding on the nature of these behaviours. Building upon these literature streams, this research adopts a holistic approach towards the conceptualisation of COD. It addresses previous shortcomings in the literature by providing a well bounded definition, developing a psychometrically sound measure of COD and a conceptually well-founded model of COD. The concept of COD was examined across three separate studies and based on the theoretical foundations of role theory and social identity theory. Study 1 was exploratory and based on in-depth interviews using the Critical Incident Technique (CIT). The aim of Study 1 was to understand the nature of COD and qualitatively identify its key drivers. Thematic analysis was conducted to analyse the data and the two potential dimensions of COD behaviours of Deviant Service Adaptation (DSA) and Deviant Service Communication (DSC) were revealed in the analysis. In addition, themes representing the potential influences of COD were broadly classified as individual factors, situational factors, and organisational factors. Study 2 was a scale development procedure that involved the generation and purification of items for the measure based on two student samples working in customer service roles (Pilot sample, N=278; Initial validation sample, N=231). The results for the reliability and Exploratory Factor Analyses (EFA) on the pilot sample suggested the scale had poor psychometric properties. As a result, major revisions were made in terms of item wordings and new items were developed based on the literature to reflect a new dimension, Deviant Use of Resources (DUR). The revised items were tested on the initial validation sample with the EFA analysis suggesting a four-factor structure of COD. The aim of Study 3 was to further purify the COD measure and test for nomological validity based on its theoretical relationships with key antecedents and similar constructs (key correlates). The theoretical model of COD consisting of nine hypotheses was tested on a retail and hospitality sample of frontline employees (Retail N=311; Hospitality N=305) of a market research panel using an online survey. The data was analysed using Structural Equation Modelling (SEM). The results provided support for a re-specified second-order three-factor model of COD which consists of 11 items. Overall, the COD measure was found to be reliable and valid, demonstrating convergent validity, discriminant validity and marginal partial invariance for the factor loadings. The results showed support for nomological validity, although the antecedents had differing impact on COD across samples. Specifically, empathy and perspective-taking, role conflict, and job autonomy significantly influenced COD in the retail sample, whereas empathy and perspective-taking, risk-taking propensity and role conflict were significant predictors in the hospitality sample. In addition, customer orientation-selling orientation, the altruistic dimension of organisational citizenship behaviours, workplace deviance, and social desirability responding were found to correlate with COD. This research makes several contributions to theory. First, the findings of this thesis extend the literature on positive deviance, pro-social and pro-customer behaviours. Second, the research provides an empirically tested model which describes the antecedents of COD. Third, this research contributes by providing a reliable and valid measure of COD. Finally, the research investigates the differential effects of the key antecedents in different service sectors on COD. The research findings also contribute to services marketing practice. Based on the research findings, service practitioners can better understand the phenomenon of COD and utilise the measurement tool to calibrate COD levels within their organisations. Knowledge on the key determinants of COD will help improve recruitment and training programs and drive internal initiatives within the firm.
Resumo:
This paper describes a vision-based airborne collision avoidance system developed by the Australian Research Centre for Aerospace Automation (ARCAA) under its Dynamic Sense-and-Act (DSA) program. We outline the system architecture and the flight testing undertaken to validate the system performance under realistic collision course scenarios. The proposed system could be implemented in either manned or unmanned aircraft, and represents a step forward in the development of a “sense-and-avoid” capability equivalent to human “see-and-avoid”.
Resumo:
While frontline employees (FLEs) are known to bend the rules or act in non-conforming ways for customers, the phenomenon of FLEs over-servicing customers is not well understood. This paper proposes a behavioural concept termed customer-oriented deviance (COD) and a conceptual model of its key drivers. Using a qualitative study involving 22 in-depth interviews with FLEs, the analysis reveals three categories of COD behaviours: deviant service adaptation (DSA), deviant service communication (DSC), and deviant use of resources (DUR). The drivers of COD are categorised as individual (risk-taking, service aptitude, and pro-social moral values), situational (resource availability, social capita with customers, legitimacy of customer problems, and avoidance of hassles), and organisational (unconducive service climate and anticipated rewards). This paper contributes to understanding how and why FLEs over-service customers and extends current research by exploring multiple categories of behaviours within a services marketing context.
Resumo:
We describe a short signature scheme that is strongly existentially unforgeable under an adaptive chosen message attack in the standard security model. Our construction works in groups equipped with an efficient bilinear map, or, more generally, an algorithm for the Decision Diffie-Hellman problem. The security of our scheme depends on a new intractability assumption we call Strong Diffie-Hellman (SDH), by analogy to the Strong RSA assumption with which it shares many properties. Signature generation in our system is fast and the resulting signatures are as short as DSA signatures for comparable security. We give a tight reduction proving that our scheme is secure in any group in which the SDH assumption holds, without relying on the random oracle model.
Resumo:
Much interest surrounds the effect of extracellular matrix (ECM) elasticity on cell behavior. Here we present a rapid method for measuring the elasticity of synthetic ECM substrates based on indentation of the substrate with a ferromagnetic sphere and optical tracking of the resulting deformation. We find that this method yields order-of-magnitude agreement with atomic force microscopy elasticity measurements, but that the degree of this agreement depends strongly on sphere density and gel elasticity. In its regime of greatest accuracy, we envision that this method may be used for high-throughput characterization of ECM substrates in cell biological studies.
Resumo:
Standard signature schemes are usually designed only to achieve weak unforgeability – i.e. preventing forgery of signatures on new messages not previously signed. However, most signature schemes are randomised and allow many possible signatures for a single message. In this case, it may be possible to produce a new signature on a previously signed message. Some applications require that this type of forgery also be prevented – this requirement is called strong unforgeability. At PKC2006, Boneh Shen and Waters presented an efficient transform based on any randomised trapdoor hash function which converts a weakly unforgeable signature into a strongly unforgeable signature and applied it to construct a strongly unforgeable signature based on the CDH problem. However, the transform of Boneh et al only applies to a class of so-called partitioned signatures. Although many schemes fall in this class, some do not, for example the DSA signature. Hence it is natural to ask whether one can obtain a truly generic efficient transform based on any randomised trapdoor hash function which converts any weakly unforgeable signature into a strongly unforgeable one. We answer this question in the positive by presenting a simple modification of the Boneh-Shen-Waters transform. Our modified transform uses two randomised trapdoor hash functions.
Resumo:
Total strain controlled low cycle fatigue tests on 316L(N) stainless steel have been conducted in air at various strain rates in the temperature range of 773-873 K to identify the operative time-dependent mechanisms and to understand their influence on the cyclic deformation and fracture behaviour of the alloy. The cyclic stress response at all the testing conditions was marked by an initial hardening followed by stress saturation. A negative strain rate stress response is observed under specific testing conditions which is attributed to dynamic strain ageing (DSA). Transmission electron microscopy studies reveal that there is an increase in the dislocation density and enhanced slip planarity in the DSA regime. Fatigue life is found to decrease with a decrease in strain rate. The degradation in fatigue resistance is attributed to the detrimental effects associated with DSA and oxidation. Quantitative measurement of secondary cracks indicate that both transgranular and intergranular cracking are accelerated predominantly under conditions conducive to DSA.
Resumo:
An experimental investigation into the dynamic strain ageing (DSA) of a wrought Ni-base superalloy 720Li was conducted. Characteristics of jerky, flow have been studied at intermediate temperatures of 350, 400 and 450 degrees C at strain-rates between 10(-3) and 10(-5) s(-1). Serrations of Type C are predominant within the temperature/strain-rate range explored. The major characteristics of the serrations-i.e. (a) critical plastic strain for onset of serrations, epsilon(c); (b) average stress decrement, Delta sigma(avg); and (c) strain increment between serrations. Delta epsilon(BS)-have been examined at selected temperatures and strain-rates. Negative strain-rate sensitivity was observed in the DSA regime. However. temperature did not influence tensile properties such as yield strength, ultimate strength. elongation, reduction in area, and work hardening rate or fracture features in DSA regime. Analysis of the results Suggests that locking of the mobile dislocations by substitutional alloying elements is responsible for the DSA in alloy 720Li.
Resumo:
Water temperature measurements from Wivenhoe Dam offer a unique opportunity for studying fluctuations of temperatures in a subtropical dam as a function of time and depth. Cursory examination of the data indicate a complicated structure across both time and depth. We propose simplifying the task of describing these data by breaking the time series at each depth into physically meaningful components that individually capture daily, subannual, and annual (DSA) variations. Precise definitions for each component are formulated in terms of a wavelet-based multiresolution analysis. The DSA components are approximately pairwise uncorrelated within a given depth and between different depths. They also satisfy an additive property in that their sum is exactly equal to the original time series. Each component is based upon a set of coefficients that decomposes the sample variance of each time series exactly across time and that can be used to study both time-varying variances of water temperature at each depth and time-varying correlations between temperatures at different depths. Each DSA component is amenable for studying a certain aspect of the relationship between the series at different depths. The daily component in general is weakly correlated between depths, including those that are adjacent to one another. The subannual component quantifies seasonal effects and in particular isolates phenomena associated with the thermocline, thus simplifying its study across time. The annual component can be used for a trend analysis. The descriptive analysis provided by the DSA decomposition is a useful precursor to a more formal statistical analysis.
Resumo:
Technological development of fast multi-sectional, helical computed tomography (CT) scanners has allowed computed tomography perfusion (CTp) and angiography (CTA) in evaluating acute ischemic stroke. This study focuses on new multidetector computed tomography techniques, namely whole-brain and first-pass CT perfusion plus CTA of carotid arteries. Whole-brain CTp data is acquired during slow infusion of contrast material to achieve constant contrast concentration in the cerebral vasculature. From these data quantitative maps are constructed of perfused cerebral blood volume (pCBV). The probability curve of cerebral infarction as a function of normalized pCBV was determined in patients with acute ischemic stroke. Normalized pCBV, expressed as a percentage of contralateral normal brain pCBV, was determined in the infarction core and in regions just inside and outside the boundary between infarcted and noninfarcted brain. Corresponding probabilities of infarction were 0.99, 0.96, and 0.11, R² was 0.73, and differences in perfusion between core and inner and outer bands were highly significant. Thus a probability of infarction curve can help predict the likelihood of infarction as a function of percentage normalized pCBV. First-pass CT perfusion is based on continuous cine imaging over a selected brain area during a bolus injection of contrast. During its first passage, contrast material compartmentalizes in the intravascular space, resulting in transient tissue enhancement. Functional maps such as cerebral blood flow (CBF), and volume (CBV), and mean transit time (MTT) are then constructed. We compared the effects of three different iodine concentrations (300, 350, or 400 mg/mL) on peak enhancement of normal brain tissue and artery and vein, stratified by region-of-interest (ROI) location, in 102 patients within 3 hours of stroke onset. A monotonic increasing peak opacification was evident at all ROI locations, suggesting that CTp evaluation of patients with acute stroke is best performed with the highest available concentration of contrast agent. In another study we investigated whether lesion volumes on CBV, CBF, and MTT maps within 3 hours of stroke onset predict final infarct volume, and whether all these parameters are needed for triage to intravenous recombinant tissue plasminogen activator (IV-rtPA). The effect of IV-rtPA on the affected brain by measuring salvaged tissue volume in patients receiving IV-rtPA and in controls was investigated also. CBV lesion volume did not necessarily represent dead tissue. MTT lesion volume alone can serve to identify the upper size limit of the abnormally perfused brain, and those with IV-rtPA salvaged more brain than did controls. Carotid CTA was compared with carotid DSA in grading of stenosis in patients with stroke symptoms. In CTA, the grade of stenosis was determined by means of axial source and maximum intensity projection (MIP) images as well as a semiautomatic vessel analysis. CTA provides an adequate, less invasive alternative to conventional DSA, although tending to underestimate clinically relevant grades of stenosis.