998 resultados para Software Acquisition
Resumo:
Die Unterrichtsmaterialien beschreiben einen Unterrichtsvorschlag für einen kontinuierlich computergestützten Stochastikunterricht in der gymnasialen Oberstufe. Verwendet wird die Software Fathom. Die Unterrichtsmaterialien wurden bislang in fünf verschiedenen Kasseler Oberstufenkursen (Jahrgangsstufe 12 bzw. 13) erprobt und dabei jeweils weiter entwickelt. Ausgearbeitet sind drei Unterrichtseinheiten: 1. „Einstieg in die Stochastik mit Simulationen“ 2. Unterrichtseinheit „Binomialverteilung“ 3. Unterrichtseinheit „Testen von Hypothesen“ Zu jeder der drei Unterrichtseinheiten existieren eine didaktische Beschreibung, die Arbeitsmaterialien (Arbeitsblätter, Merkblätter zur Ergebnissicherung, dynamische Lernumgebungen) sowie Musterlösungen zu den Aufgaben. Für alle Aufgaben mit Verwendung der Software Fathom stehen die zugehörigen Fathom-Dateien zur Verfügung. Alle Fathom-Dateien (Lernumgebungen und Musterlösungen der Aufgaben) erhalten Sie als *.zip-Datei zum Download.
Resumo:
This thesis in Thermal Flow Drilling and Flowtap in thin metal sheet and pipes of copper and copper alloys had as objectives to know the comportment of copper and copper alloys sheet metal during the Thermal Flow Drill processes with normal tools, to know the best Speed and Feed machine data for the best bushing quality, to known the best Speed for Form Tapping processes and to know the best bush long in pure copper pipes for water solar interchange equipment. Thermal Flow Drilling (TFD) and Form Tapping (FT) is one of the research lines of the Institute of Production and Logistics (IPL) at University of Kassel. At December 1995, a work meeting of IPL, Santa Catarina University, Brazil, Buenos Aires University, Argentine, Tarapacá University (UTA), Chile members and the CEO of Flowdrill B.V. was held in Brazil. The group decided that the Manufacturing Laboratory (ML) of UTA would work with pure copper and brass alloys sheet metal and pure copper pipes in order to develop a water interchange solar heater. The Flowdrill BV Company sent tools to Tarapacá University in 1996. In 1999 IPL and the ML carried out an ALECHILE research project promoted by the DAAD and CONICyT in copper sheet metal and copper pipes and sheet metal a-brass alloys. The normal tools are lobed, conical tungsten carbide tool. When rotated at high speed and pressed with high axial force into sheet metal or thin walled tube generated heat softens the metal and allows the drill to feed forward produce a hole and simultaneously form a bushing from the displacement material. In the market exist many features but in this thesis is used short and longs normal tools of TFD. For reach the objectives it was takes as references four qualities of the frayed end bushing, where the best one is the quality class I. It was used pure copper and a-brass alloys sheet metals, with different thickness. It was used different TFD drills diameter for four thread type, from M-5 to M10. Similar to the Aluminium sheet metals studies it was used the predrilling processes with HSS drills around 30% of the TFD diameter (1,5 – 3,0 mm D). In the next step is used only 2,0 mm thick metal sheet, and 9,2 mm TFD diameter for M-10 thread. For the case of pure commercial copper pipes is used for ¾” inch diameter and 12, 8 mm (3/8”) TFD drill for holes for 3/8” pipes and different normal HSS drills for predrilling processes. The chemical sheet metal characteristics were takes as reference for the material behaviour. The Chilean pure copper have 99,35% of Cu and 0,163% of Zinc and the Chilean a-brass alloys have 75,6% of Cu and 24,0% of Zinc. It is used two German a-brass alloys; Nº1 have 61,6% of Cu, 36,03 % of Zinc and 2,2% of Pb and the German a-brass alloys Nº2 have 63,1% of Cu, 36,7% of Zinc and 0% of Pb. The equipments used were a HAAS CNC milling machine centre, a Kistler dynamometer, PC Pentium II, Acquisition card, TESTPOINT and XAct software, 3D measurement machine, micro hardness, universal test machine, and metallographic microscope. During the test is obtained the feed force and momentum curves that shows the material behaviour with TFD processes. In general it is take three phases. It was possible obtain the best machining data for the different sheet of copper and a-brass alloys thick of Chilean materials and bush quality class I. In the case of a-brass alloys, the chemical components and the TFD processes temperature have big influence. The temperature reach to 400º Celsius during the TFD processes and the a-brass alloys have some percents of Zinc the bush quality is class I. But when the a-brass alloys have some percents of Lead who have 200º C melting point is not possible to obtain a bush, because the Lead gasify and the metallographic net broke. During the TFD processes the recrystallization structures occur around the Copper and a-brass alloy bush, who gives more hardness in these zones. When the threads were produce with Form Tapping processes with Flowtap tools, this hardness amount gives a high limit load of the thread when hey are tested in a special support that was developed for it. For eliminated the predrilling processes with normal HSS drills it was developed a compound tool. With this new tool it was possible obtain the best machining data for quality class I bush. For the copper pipes it is made bush without predrilling and the quality class IV was obtained. When it is was used predrilling processes, quality classes I bush were obtained. Then with different HSS drill diameter were obtained different long bush, where were soldering with four types soldering materials between pipes with 3/8” in a big one as ¾”. Those soldering unions were tested by traction test and all the 3/8” pipes broken, and the soldering zone doesn’t have any problem. Finally were developed different solar water interchange heaters and tested. As conclusions, the present Thesis shows that the Thermal Flow Drilling in thinner metal sheets of cooper and cooper alloys needs a predrilling process for frayed end quality class I bushings, similar to thinner sheets of aluminium bushes. The compound tool developed could obtain quality class I bushings and excludes predrilling processes. The bush recrystalization, product of the friction between the tool and the material, the hardness grows and it is advantageous for the Form Tapping. The methodology developed for commercial copper pipes permits to built water solar interchange heaters.
Resumo:
DIADEM, created by THOMSON-CSF, is a methodology for specifying and developing user interfaces. It improves productivity of the interface development process as well as quality of the interface. The method provides support to user interface development in three aspects. (1) DIADEM defines roles of people involved and their tasks and organises the sequence of activities. (2) It provides graphical formalisms supporting information exchange between people. (3) It offers a basic set of rules for optimum human-machine interfaces. The use of DIADEM in three areas (process control, sales support, and multimedia presentation) was observed and evaluated by our laboratory in the European project DIAMANTA (ESPRIT P20507). The method provides an open procedure that leaves room for adaptation to a specific application and environment. This paper gives an overview of DIADEM and shows how to extend formalisms for developing multimedia interfaces.
Resumo:
Self-adaptive software provides a profound solution for adapting applications to changing contexts in dynamic and heterogeneous environments. Having emerged from Autonomic Computing, it incorporates fully autonomous decision making based on predefined structural and behavioural models. The most common approach for architectural runtime adaptation is the MAPE-K adaptation loop implementing an external adaptation manager without manual user control. However, it has turned out that adaptation behaviour lacks acceptance if it does not correspond to a user’s expectations – particularly for Ubiquitous Computing scenarios with user interaction. Adaptations can be irritating and distracting if they are not appropriate for a certain situation. In general, uncertainty during development and at run-time causes problems with users being outside the adaptation loop. In a literature study, we analyse publications about self-adaptive software research. The results show a discrepancy between the motivated application domains, the maturity of examples, and the quality of evaluations on the one hand and the provided solutions on the other hand. Only few publications analysed the impact of their work on the user, but many employ user-oriented examples for motivation and demonstration. To incorporate the user within the adaptation loop and to deal with uncertainty, our proposed solutions enable user participation for interactive selfadaptive software while at the same time maintaining the benefits of intelligent autonomous behaviour. We define three dimensions of user participation, namely temporal, behavioural, and structural user participation. This dissertation contributes solutions for user participation in the temporal and behavioural dimension. The temporal dimension addresses the moment of adaptation which is classically determined by the self-adaptive system. We provide mechanisms allowing users to influence or to define the moment of adaptation. With our solution, users can have full control over the moment of adaptation or the self-adaptive software considers the user’s situation more appropriately. The behavioural dimension addresses the actual adaptation logic and the resulting run-time behaviour. Application behaviour is established during development and does not necessarily match the run-time expectations. Our contributions are three distinct solutions which allow users to make changes to the application’s runtime behaviour: dynamic utility functions, fuzzy-based reasoning, and learning-based reasoning. The foundation of our work is a notification and feedback solution that improves intelligibility and controllability of self-adaptive applications by implementing a bi-directional communication between self-adaptive software and the user. The different mechanisms from the temporal and behavioural participation dimension require the notification and feedback solution to inform users on adaptation actions and to provide a mechanism to influence adaptations. Case studies show the feasibility of the developed solutions. Moreover, an extensive user study with 62 participants was conducted to evaluate the impact of notifications before and after adaptations. Although the study revealed that there is no preference for a particular notification design, participants clearly appreciated intelligibility and controllability over autonomous adaptations.
Resumo:
Se presenta experiencia educativa que propone la utilizaci??n de las nuevas tecnolog??as para la creaci??n de materiales did??cticos en educaci??n infantil. Se realiza en el CPR Los R??os en Arenas del Rey, Granada. Los objetivos son: conocer en qu?? consiste el proyecto JClic; aprender a instalar correctamente, todos los componentes del programa JClic; conocer cu??les son sus diferentes componentes y en qu?? consiste cada uno de ellos; conocer los diversos tipos de actividades que podemos crear con esta herramienta, as?? como sus caracter??sticas y modalidades; crear diversas actividades utilizando la herramienta JClic autor; crear nuevos proyectos con JClic autor; abrir y modificar proyectos ya existentes; organizar las diferentes actividades creadas o ya existentes; abrir y ejecutar actividades; crear recursos para la utilizaci??n en educaci??n infantil.
Resumo:
Adecuación del software de los ciclos formativos de la Familia de Administración del IES La Sagra a los utilizados por las empresas de la comarca.
Resumo:
Numerosos programas de los que se presentan en este CD han sido probados y experimentados, entre otros, por el profesorado del C.P. Vicente Aleixandre, de Fortuna (Murcia) durante el curso escolar 2002/2003
Resumo:
We present an unsupervised learning algorithm that acquires a natural-language lexicon from raw speech. The algorithm is based on the optimal encoding of symbol sequences in an MDL framework, and uses a hierarchical representation of language that overcomes many of the problems that have stymied previous grammar-induction procedures. The forward mapping from symbol sequences to the speech stream is modeled using features based on articulatory gestures. We present results on the acquisition of lexicons and language models from raw speech, text, and phonetic transcripts, and demonstrate that our algorithm compares very favorably to other reported results with respect to segmentation performance and statistical efficiency.
Resumo:
This report represents research conducted at the Massachusetts Institute of Technology under the Lean Aerospace Initiative (LAI) program. The research focused on identifying Acquisition Intrapreneurs, viewed and defined for the purpose of this research as, individuals within the acquisition profession who take direct responsibility for turning ideas into products through assertive risk taking. The basis for this research stems from the agile acquisition push for “breeding innovators” to achieve a leaner and more responsive approach to the design, build, test and fielding of warfighting systems.
Resumo:
The memory hierarchy is the main bottleneck in modern computer systems as the gap between the speed of the processor and the memory continues to grow larger. The situation in embedded systems is even worse. The memory hierarchy consumes a large amount of chip area and energy, which are precious resources in embedded systems. Moreover, embedded systems have multiple design objectives such as performance, energy consumption, and area, etc. Customizing the memory hierarchy for specific applications is a very important way to take full advantage of limited resources to maximize the performance. However, the traditional custom memory hierarchy design methodologies are phase-ordered. They separate the application optimization from the memory hierarchy architecture design, which tend to result in local-optimal solutions. In traditional Hardware-Software co-design methodologies, much of the work has focused on utilizing reconfigurable logic to partition the computation. However, utilizing reconfigurable logic to perform the memory hierarchy design is seldom addressed. In this paper, we propose a new framework for designing memory hierarchy for embedded systems. The framework will take advantage of the flexible reconfigurable logic to customize the memory hierarchy for specific applications. It combines the application optimization and memory hierarchy design together to obtain a global-optimal solution. Using the framework, we performed a case study to design a new software-controlled instruction memory that showed promising potential.
Resumo:
Analizar la formación recibida en software por parte de los coordinadores de nuevas tecnologías en los centros educativos públicos no universitarios de Asturias ajustándose a la dicotomía libre/privado. Conocer el tipo de software empleado por estos coordinadores en los diferentes contextos: aula, gestión docente y personal. Valorar su opinión sobre qué software es el más indicado para aplicar en el sistema educativo. Conocer las razones usadas para la selección de uno u otro tipo de software. Explicar, en su caso, las diferencias encontradas con las variables de identificación (sexo, edad, años de experiencia como coordinador de nuevas tecnologías, etc.). Explorar las líneas de un proceso formativo en software libre con profesionales de la educación para identificar las dificultades y oportunidades de la formación en este campo. El trabajo se estructura en dos apartados, en primer lugar, una fundamentación teórica compuesta por dos capítulos en los que se revisan los elementos teóricos de la investigación y en segundo lugar, el estudio de campo donde se aclaran todos los aspectos técnicos del proceso investigador, sus objetivos, sus conclusiones y las líneas de actuación propuestas. En la parte teórica, se abordan las relaciones que se establecen entre el desarrollo tecnológico y social y el modo en que esta relación incide en los discursos sobre la inclusión de las nuevas tecnologías de la información y de la comunicación en la escuela. Se hace una revisión de esta relación entre las TIC y el sistema educativo, que concluye con la esquematización de las racionalidades curriculares existentes y el papel que, los medios en general, y el software en particular, ocupan en ella. Se analizan los conceptos de alfabetización digital, software libre y sistema operativo, y se explica el concepto de código abierto y el programa Linux y, finalmente, se hace un repaso a las argumentaciones contrarias al software libre intentando desenmascarar aquellas que transmiten mitos o falsedades. Se exponen las razones que justifican la difusión del software libre en el sistema educativo. Tras la fundamentación teórica de la investigación, en el apartado, estudio de campo, se describe éste, la metodología y las conclusiones de la investigación. Se ha utilizado una metodología cuantitativa en el caso de la primera parte de la investigación y metodologías cualitativas en el caso de la experiencia de formación en software libre. La técnica de investigación ha sido la encuesta, con la realización de un cuestionario. La muestra la integraron 307 sujetos, de los que sólo un 38,8 por ciento respondió al cuestionario. El perfil tipo de los sujetos muestrales es: varón de 47 años de edad con 22 de experiencia en la educación y 4 como coordinador de nuevas tecnologías. Se concluye que a pesar de que la mayor parte de los sujetos encuestados manifiestan que el software libre es el que se debe emplear en el sistema educativo su uso en la escuela asturiana está lejos de ser una realidad, tanto en el aula como en otros ámbitos, siendo el software privado el que ocupa un lugar prioritario favorecido por las propias políticas de la administración. Finalmente, se realiza una propuesta de actuación encaminada a la elaboración de un plan de implementación de software libre en el sistema educativo del que se apuntan procedimientos a corto, medio y largo plazo.
Resumo:
In the eighties, John Aitchison (1986) developed a new methodological approach for the statistical analysis of compositional data. This new methodology was implemented in Basic routines grouped under the name CODA and later NEWCODA inMatlab (Aitchison, 1997). After that, several other authors have published extensions to this methodology: Marín-Fernández and others (2000), Barceló-Vidal and others (2001), Pawlowsky-Glahn and Egozcue (2001, 2002) and Egozcue and others (2003). (...)