733 resultados para Design-Build-Test, Project-Based-Learning
Resumo:
The study takes as its object the evaluation process of a graduation discipline at the School of Arts, Sciences and Humanities, University of São Paulo (USP), developed through the methodology of Problem Based Learning (PBL). We consider three distinct and complementary aspects: procedural review (partial and final) student peer assessment and self-assessment. The sample consisted of 17 students from the undergraduate courses. Data collection occurred in two stages through questionnaires with open and closed questions. The qualitativeinterpretative analysis allowed two comparisons: cross (intra-step) and longitudinal (inter-steps). The results point to a trend of students evaluating objective aspects of the work as well as to overcome some difficulties in the development of the project.
Resumo:
Education is an essential part of the overall development of a human being. There are opportunities of knowledge production and teaching and learning in many environments, so education is not limited to the school environment. A possible field of teacher's performance takes place in the context of educational social projects, therefore, the importance of knowing and evaluating this environment. Rise to concern about how important and what the reflection of the social projects in the lives of children who often find themselves in a social risk to the margins of society and the research is justified by considering the community education as an opportunity to change lives and society itself. Thus, a qualitative study of descriptive type in a social project based on the Waldorf was developed, aiming to know how the project was carried out and verify the importance of it in a poor neighborhood of the city of Botucatu-SP
Resumo:
OBJETIVO: Apresentar a experiência de uma instituição brasileira no ensino da Reumatologia na graduação médica, cujo projeto pedagógico é estruturado em metodologias ativas de aprendizado. MÉTODOS: Realizou-se um estudo descritivo, com abordagem qualitativa dos conteúdos referentes à Reumatologia no curso de Medicina do Centro Universitário do Estado do Pará (Cesupa). RESULTADOS: O sistema musculoesquelético é abordado no segundo e sétimo semestres, nos módulos referentes ao sistema locomotor e Clínica Médica II, respectivamente, sendo que cada etapa e cada atividade apresentam objetivos específicos mínimos. Além do conteúdo teórico, no sétimo semestre, os alunos realizam atendimentos no ambulatório de Reumatologia, quando existe maior ênfase na elaboração correta de anamnese e exame físico. No internato, os alunos retornam ao ambulatório de Reumatologia e, neste momento, as habilidades de diagnóstico, de investigação e de terapêutica são as mais exigidas. CONCLUSÃO: Ainda há muito para evoluir em busca de um modelo ideal para o ensino da Reumatologia, porém, cumprindo as principais recomendações disponíveis para a boa prática do ensino na graduação, podemos proporcionar ao futuro médico conhecimento, habilidade e experiências capazes de ajudá-lo na condução desses pacientes.
Resumo:
Piezoresistive sensors are commonly made of a piezoresistive membrane attached to a flexible substrate, a plate. They have been widely studied and used in several applications. It has been found that the size, position and geometry of the piezoresistive membrane may affect the performance of the sensors. Based on this remark, in this work, a topology optimization methodology for the design of piezoresistive plate-based sensors, for which both the piezoresistive membrane and the flexible substrate disposition can be optimized, is evaluated. Perfect coupling conditions between the substrate and the membrane based on the `layerwise' theory for laminated plates, and a material model for the piezoresistive membrane based on the solid isotropic material with penalization model, are employed. The design goal is to obtain the configuration of material that maximizes the sensor sensitivity to external loading, as well as the stiffness of the sensor to particular loads, which depend on the case (application) studied. The proposed approach is evaluated by studying two distinct examples: the optimization of an atomic force microscope probe and a pressure sensor. The results suggest that the performance of the sensors can be improved by using the proposed approach.
Resumo:
Quantitative structure – activity relationships (QSARs) developed to evaluate percentage of inhibition of STa-stimulated (Escherichia coli) cGMP accumulation in T84 cells are calculated by the Monte Carlo method. This endpoint represents a measure of biological activity of a substance against diarrhea. Statistical quality of the developed models is quite good. The approach is tested using three random splits of data into the training and test sets. The statistical characteristics for three splits are the following: (1) n = 20, r2 = 0.7208, q2 = 0.6583, s = 16.9, F = 46 (training set); n = 11, r2 = 0.8986, s = 14.6 (test set); (2) n = 19, r2 = 0.6689, q2 = 0.5683, s = 17.6, F = 34 (training set); n = 12, r2 = 0.8998, s = 12.1 (test set); and (3) n = 20, r2 = 0.7141, q2 = 0.6525, s = 14.7, F = 45 (training set); n = 11, r2 = 0.8858, s = 19.5 (test set). Based on the proposed here models hypothetical compounds which can be useful agents against diarrhea are suggested.
Resumo:
Galaxy clusters occupy a special position in the cosmic hierarchy as they are the largest bound structures in the Universe. There is now general agreement on a hierarchical picture for the formation of cosmic structures, in which galaxy clusters are supposed to form by accretion of matter and merging between smaller units. During merger events, shocks are driven by the gravity of the dark matter in the diffuse barionic component, which is heated up to the observed temperature. Radio and hard-X ray observations have discovered non-thermal components mixed with the thermal Intra Cluster Medium (ICM) and this is of great importance as it calls for a “revision” of the physics of the ICM. The bulk of present information comes from the radio observations which discovered an increasing number of Mpcsized emissions from the ICM, Radio Halos (at the cluster center) and Radio Relics (at the cluster periphery). These sources are due to synchrotron emission from ultra relativistic electrons diffusing through µG turbulent magnetic fields. Radio Halos are the most spectacular evidence of non-thermal components in the ICM and understanding the origin and evolution of these sources represents one of the most challenging goal of the theory of the ICM. Cluster mergers are the most energetic events in the Universe and a fraction of the energy dissipated during these mergers could be channelled into the amplification of the magnetic fields and into the acceleration of high energy particles via shocks and turbulence driven by these mergers. Present observations of Radio Halos (and possibly of hard X-rays) can be best interpreted in terms of the reacceleration scenario in which MHD turbulence injected during these cluster mergers re-accelerates high energy particles in the ICM. The physics involved in this scenario is very complex and model details are difficult to test, however this model clearly predicts some simple properties of Radio Halos (and resulting IC emission in the hard X-ray band) which are almost independent of the details of the adopted physics. In particular in the re-acceleration scenario MHD turbulence is injected and dissipated during cluster mergers and thus Radio Halos (and also the resulting hard X-ray IC emission) should be transient phenomena (with a typical lifetime <» 1 Gyr) associated with dynamically disturbed clusters. The physics of the re-acceleration scenario should produce an unavoidable cut-off in the spectrum of the re-accelerated electrons, which is due to the balance between turbulent acceleration and radiative losses. The energy at which this cut-off occurs, and thus the maximum frequency at which synchrotron radiation is produced, depends essentially on the efficiency of the acceleration mechanism so that observations at high frequencies are expected to catch only the most efficient phenomena while, in principle, low frequency radio surveys may found these phenomena much common in the Universe. These basic properties should leave an important imprint in the statistical properties of Radio Halos (and of non-thermal phenomena in general) which, however, have not been addressed yet by present modellings. The main focus of this PhD thesis is to calculate, for the first time, the expected statistics of Radio Halos in the context of the re-acceleration scenario. In particular, we shall address the following main questions: • Is it possible to model “self-consistently” the evolution of these sources together with that of the parent clusters? • How the occurrence of Radio Halos is expected to change with cluster mass and to evolve with redshift? How the efficiency to catch Radio Halos in galaxy clusters changes with the observing radio frequency? • How many Radio Halos are expected to form in the Universe? At which redshift is expected the bulk of these sources? • Is it possible to reproduce in the re-acceleration scenario the observed occurrence and number of Radio Halos in the Universe and the observed correlations between thermal and non-thermal properties of galaxy clusters? • Is it possible to constrain the magnetic field intensity and profile in galaxy clusters and the energetic of turbulence in the ICM from the comparison between model expectations and observations? Several astrophysical ingredients are necessary to model the evolution and statistical properties of Radio Halos in the context of re-acceleration model and to address the points given above. For these reason we deserve some space in this PhD thesis to review the important aspects of the physics of the ICM which are of interest to catch our goals. In Chapt. 1 we discuss the physics of galaxy clusters, and in particular, the clusters formation process; in Chapt. 2 we review the main observational properties of non-thermal components in the ICM; and in Chapt. 3 we focus on the physics of magnetic field and of particle acceleration in galaxy clusters. As a relevant application, the theory of Alfv´enic particle acceleration is applied in Chapt. 4 where we report the most important results from calculations we have done in the framework of the re-acceleration scenario. In this Chapter we show that a fraction of the energy of fluid turbulence driven in the ICM by the cluster mergers can be channelled into the injection of Alfv´en waves at small scales and that these waves can efficiently re-accelerate particles and trigger Radio Halos and hard X-ray emission. The main part of this PhD work, the calculation of the statistical properties of Radio Halos and non-thermal phenomena as expected in the context of the re-acceleration model and their comparison with observations, is presented in Chapts.5, 6, 7 and 8. In Chapt.5 we present a first approach to semi-analytical calculations of statistical properties of giant Radio Halos. The main goal of this Chapter is to model cluster formation, the injection of turbulence in the ICM and the resulting particle acceleration process. We adopt the semi–analytic extended Press & Schechter (PS) theory to follow the formation of a large synthetic population of galaxy clusters and assume that during a merger a fraction of the PdV work done by the infalling subclusters in passing through the most massive one is injected in the form of magnetosonic waves. Then the processes of stochastic acceleration of the relativistic electrons by these waves and the properties of the ensuing synchrotron (Radio Halos) and inverse Compton (IC, hard X-ray) emission of merging clusters are computed under the assumption of a constant rms average magnetic field strength in emitting volume. The main finding of these calculations is that giant Radio Halos are naturally expected only in the more massive clusters, and that the expected fraction of clusters with Radio Halos is consistent with the observed one. In Chapt. 6 we extend the previous calculations by including a scaling of the magnetic field strength with cluster mass. The inclusion of this scaling allows us to derive the expected correlations between the synchrotron radio power of Radio Halos and the X-ray properties (T, LX) and mass of the hosting clusters. For the first time, we show that these correlations, calculated in the context of the re-acceleration model, are consistent with the observed ones for typical µG strengths of the average B intensity in massive clusters. The calculations presented in this Chapter allow us to derive the evolution of the probability to form Radio Halos as a function of the cluster mass and redshift. The most relevant finding presented in this Chapter is that the luminosity functions of giant Radio Halos at 1.4 GHz are expected to peak around a radio power » 1024 W/Hz and to flatten (or cut-off) at lower radio powers because of the decrease of the electron re-acceleration efficiency in smaller galaxy clusters. In Chapt. 6 we also derive the expected number counts of Radio Halos and compare them with available observations: we claim that » 100 Radio Halos in the Universe can be observed at 1.4 GHz with deep surveys, while more than 1000 Radio Halos are expected to be discovered in the next future by LOFAR at 150 MHz. This is the first (and so far unique) model expectation for the number counts of Radio Halos at lower frequency and allows to design future radio surveys. Based on the results of Chapt. 6, in Chapt.7 we present a work in progress on a “revision” of the occurrence of Radio Halos. We combine past results from the NVSS radio survey (z » 0.05 − 0.2) with our ongoing GMRT Radio Halos Pointed Observations of 50 X-ray luminous galaxy clusters (at z » 0.2−0.4) and discuss the possibility to test our model expectations with the number counts of Radio Halos at z » 0.05 − 0.4. The most relevant limitation in the calculations presented in Chapt. 5 and 6 is the assumption of an “averaged” size of Radio Halos independently of their radio luminosity and of the mass of the parent clusters. This assumption cannot be released in the context of the PS formalism used to describe the formation process of clusters, while a more detailed analysis of the physics of cluster mergers and of the injection process of turbulence in the ICM would require an approach based on numerical (possible MHD) simulations of a very large volume of the Universe which is however well beyond the aim of this PhD thesis. On the other hand, in Chapt.8 we report our discovery of novel correlations between the size (RH) of Radio Halos and their radio power and between RH and the cluster mass within the Radio Halo region, MH. In particular this last “geometrical” MH − RH correlation allows us to “observationally” overcome the limitation of the “average” size of Radio Halos. Thus in this Chapter, by making use of this “geometrical” correlation and of a simplified form of the re-acceleration model based on the results of Chapt. 5 and 6 we are able to discuss expected correlations between the synchrotron power and the thermal cluster quantities relative to the radio emitting region. This is a new powerful tool of investigation and we show that all the observed correlations (PR − RH, PR − MH, PR − T, PR − LX, . . . ) now become well understood in the context of the re-acceleration model. In addition, we find that observationally the size of Radio Halos scales non-linearly with the virial radius of the parent cluster, and this immediately means that the fraction of the cluster volume which is radio emitting increases with cluster mass and thus that the non-thermal component in clusters is not self-similar.
Resumo:
Loaded with 16% of the world’s population, India is a challenged country. More than a third of its citizens live below the poverty line - on less than a dollar a day. These people have no proper electricity, no proper drinking water supply, no proper sanitary facilities and well over 40% are illiterates. More than 65% live in rural areas and 60% earn their livelihood from agriculture. Only a meagre 3.63% have access to telephone and less than 1% have access to a computer. Therefore, providing access to timely information on agriculture, weather, social, health care, employment, fishing, is of utmost importance to improve the conditions of rural poor. After some introductive chapters, whose function is to provide a comprehensive framework – both theoretical and practical – of the current rural development policies and of the media situation in India and Uttar Pradesh, my dissertation presents the findings of the pilot project entitled “Enhancing development support to rural masses through community media activity”, launched in 2005 by the Department of Mass Communication and Journalism of the Faculty of Arts of the University of Lucknow (U.P.) and by the local NGO Bharosa. The project scope was to involve rural people and farmers from two villages of the district of Lucknow (namely Kumhrava and Barhi Gaghi) in a three-year participatory community media project, based on the creation, implementation and use of a rural community newspaper and a rural community internet centre. Community media projects like this one have been rarely carried out in India because the country has no proper community media tradition: therefore the development of the project has been a challenge for the all stakeholders involved.
Resumo:
Vor dem Hintergrund sich wandelnder (medialer) Lebenswelten von Schülerinnen und Schülern gewinnen Bestimmungsversuche um medienpädagogische Handlungskompetenzen von Lehrpersonen an Bedeutung. Der Erwerb medienpädagogischer Kompetenz, verstanden als dynamisches Zusammenspiel von domänenspezifischem Wissen und anwendungsorientiertem Können, wird in der vorliegenden Arbeit als wesentliches Lernziel der medienpädagogischen (Aus-)Bildung bestimmt. Als ein Weg zur Förderung medienpädagogischer Handlungskompetenz wird von der Autorin auf der Folie konstruktivistischer Vorstellungen über das Lehren und Lernen die Methode der Problemorientierung vorgeschlagen. Im ersten Teil der Arbeit werden Modelle und Konzepte diskutiert, die Bausteine für ein Modell medienpädagogischer Kompetenz liefern. Im zweiten Teil wird eine empirische Untersuchung zum Erwerb medienpädagogischer Handlungskompetenz auf der Basis eines von der Autorin erarbeiteten Modells vorgestellt und die Ergebnisse diskutiert. Eine kompetenztheoretische Annäherung erfolgt auf der Basis zweier Konzepte. Dies sind die Ausführungen zu einem Konzept kommunikativer Kompetenz nach Jürgen Habermas sowie dessen Überführung in die Medienpädagogik durch Dieter Baacke. Ferner wird die rezente bildungspolitische Kompetenzdebatte in Anbindung an Franz E. Weinert analysiert. Es folgt eine Zusammenschau über die methodischen Konzepte zur Erfassung von Kompetenzen in der Erziehungswissenschaft und deren Anwendbarkeit für die medienpädagogische Kompetenzforschung. Die gegenwärtig vorliegenden Entwürfe zu einer inhaltlichen Bestimmung medienpädagogischer Kompetenzen werden besprochen (Sigrid Blömeke, Werner Sesink, International Society for Technology in Education). Im Rekurs auf konstruktivistische lerntheoretische Überlegungen erfährt das problemorientierte Lernen beim Aufbau von Kompetenzen eine enorme Aufwertung. In der Arbeit wird insbesondere den Arbeiten von David Jonassen zu einer konstruktivistisch-instruktionistischen Herangehensweise bei der Gestaltung problemorientierter Lernumgebungen eine große Bedeutung zugesprochen (vgl. auch Ansätze des Goal-based Scenarios/Roger Schank und des Learning by Design/Janet Kolodner). Im zweiten Teil wird die Interventionsstudie im Kontrollgruppendesign vorgestellt. Anhand eines Modells medienpädagogischer Kompetenz, dass auf den Dimensionen Wissen einerseits und Können andererseits basiert, wurden Studierende (n=59) in einem Pre-Posttestverfahren auf diese Dimensionen getestet. Die Studierenden der Interventionsgruppe (n=30) arbeiteten über ein Semester mit einer problemorientierten Lernanwendung, die Studierenden der Kontrollgruppe (n=29) in einem klassischen Seminarsetting. Hauptergebnis der Untersuchung ist es, das die Intervention zu einem messbaren Lernerfolg beim medienpädagogischen Können führte. In der Diskussion der Ergebnisse werden Empfehlungen zur Gestaltung problemorientierter Lernumgebungen formuliert. Die Chancen einer Orientierung an problemorientierten Lernsettings für das Lernen an Hochschulen werden herausgestellt.
Resumo:
Con l’avvento di Internet, potentissimo strumento tecnologico di diffusione di informazioni e di comunicazione a distanza, anche le modalità di apprendimento sono cambiate: persino nelle scuole si tende a non utilizzare più i classici libri di testo, ma ad utilizzare dispositivi dai quali scaricare in formato elettronico, libri, dispense, test, video ed ogni altro genere di materiale di apprendimento, dando vita a un vero e proprio nuovo modo di apprendere chiamato E-learning, più veloce, comodo e ricco di alternative rispetto al vecchio modello offline che si presentava sottoforma di floppy inizialmente e poi di CD-ROM. E-learning significa, electronic based learning, ed è appunto una vera e propria metodologia di didattica che sfrutta e viene facilitata da risorse e servizi disponibili e accessibili virtualmente in rete. Al momento vi sono numerose piattaforme di E-learning, una delle quali è il nucleo di questa tesi, ovvero il tool autore AContent. Questo documento di tesi, infatti, raccoglie la descrizione della progettazione e della fase implementativa della gestione delle politiche di copyright per il tool AContent. L’obbiettivo è quello di rendere possibile l’assegnazione di un copyright a qualsiasi tipo di materiale didattico venga creato, caricato e/o condiviso sulla piattaforma in questione. Pertanto l’idea è stata quella di dare la possibilità di scegliere fra più copyright preimpostati, utilizzando degli standard di licenze riguardanti i diritti d’autore, lasciando anche l’opportunità di inserire la propria politica.
Resumo:
La ricerca ha per oggetto il progetto del Foro Bonaparte a Milano redatto da Giovanni Antonio Antolini a seguito del decreto del 23 giugno 1800 che stabiliva l’abbattimento delle mura del Castello Sforzesco. Si affronta il tema dell’architettura avendo come obiettivo una lettura critica di tale progetto servendosi dell’analisi compositiva come strumento in grado di stabilire i rapporti che intercorrono tra la città, l’architettura e il tipo. Attraverso lo studio del progetto urbano la ricerca conferma l’ipotesi per la quale la grande forma totalizzante, perentoria e assoluta è capace di mutare la struttura urbana, offrendo un nuovo modello con cui rinnovare la città. L’ambizione di Antolini di veder realizzata l’opera è destinata a svanire nell’arco di pochi anni, ma il progetto per il Foro Bonaparte continuerà per lungo tempo ad evocare la sua idea innovativa fino ai giorni nostri. Sebbene l’opera sia destinata a rimanere un’architettura solo disegnata, le varie pubblicazioni continuano a circolare nelle accademie prima e nelle università successivamente, costituendo un importante patrimonio di studio e di ricerca per generazioni di architetti, fino alla riscoperta avvenuta con Aldo Rossi e Manfredo Tafuri negli anni sessanta del secolo scorso. Dalle lezioni formulate nelle architetture del passato è possibile avanzare nuove ipotesi e alimentare riflessioni e dibattiti sul ruolo dell’architettura nella città contemporanea. La ricerca si occupa del progetto d’architettura per offrire un ulteriore contributo al tema, attraverso una lettura di carattere compositivo, supportata da una serie di schemi e disegni di studio necessari per completare il testo e per verificare i concetti esposti. Dopo aver raccolto, catalogato ed analizzato il materiale iconografico relativo al progetto per il Foro Bonaparte si è scelto di basare il ridisegno sulla raccolta di disegni conservati presso la Biblioteca Nazionale di Francia.
Resumo:
The relatively young discipline of astronautics represents one of the scientifically most fascinating and technologically advanced achievements of our time. The human exploration in space does not offer only extraordinary research possibilities but also demands high requirements from man and technology. The space environment provides a lot of attractive experimental tools towards the understanding of fundamental mechanism in natural sciences. It has been shown that especially reduced gravity and elevated radiation, two distinctive factors in space, influence the behavior of biological systems significantly. For this reason one of the key objectives on board of an earth orbiting laboratory is the research in the field of life sciences, covering the broad range from botany, human physiology and crew health up to biotechnology. The Columbus Module is the only European low gravity platform that allows researchers to perform ambitious experiments in a continuous time frame up to several months. Biolab is part of the initial outfitting of the Columbus Laboratory; it is a multi-user facility supporting research in the field of biology, e.g. effect of microgravity and space radiation on cell cultures, micro-organisms, small plants and small invertebrates. The Biolab IEC are projects designed to work in the automatic part of Biolab. In this moment in the TO-53 department of Airbus Defence & Space (formerly Astrium) there are two experiments that are in phase C/D of the development and they are the subject of this thesis: CELLRAD and CYTOSKELETON. They will be launched in soft configuration, that means packed inside a block of foam that has the task to reduce the launch loads on the payload. Until 10 years ago the payloads which were launched in soft configuration were supposed to be structural safe by themselves and a specific structural analysis could be waived on them; with the opening of the launchers market to private companies (that are not under the direct control of the international space agencies), the requirements on the verifications of payloads are changed and they have become much more conservative. In 2012 a new random environment has been introduced due to the new Space-X launch specification that results to be particularly challenging for the soft launched payloads. The last ESA specification requires to perform structural analysis on the payload for combined loads (random vibration, quasi-steady acceleration and pressure). The aim of this thesis is to create FEM models able to reproduce the launch configuration and to verify that all the margins of safety are positive and to show how they change because of the new Space-X random environment. In case the results are negative, improved design solution are implemented. Based on the FEM result a study of the joins has been carried out and, when needed, a crack growth analysis has been performed.
Resumo:
With this dissertation research we investigate intersections between design and marketing and in this respect, which factors do contribute that a product design becomes brand formative. We have developed a Brand Formative Design (BFD) framework, which investigates individual design features in a holistic, comparable, brand relevant, and consumer specific context. We discuss what kinds of characteristics contribute to BFD but also illuminate how they should be applied and examine: rnA holistic framework leading to Brand Formative Design. Identification and assessment of BFD Drivers. The dissection of products into three Distinctive Design Levels. The detection of surprising design preferences. The appropriate degree of scheme deviation with evolutionary design. Simulated BFD development processes with three different products and the integration of consumers. Future oriented objectification, comparability and assessment of design. Recommendations for the management of design in a brand specific context. Design is a product feature, which contributes significantly to the success of products. However, the development of new design contains challenges. Design can hardly be objectified; many people have an opinion concerning the attractiveness of new products but cannot formulate their future preferences. Product design is widely developed based on intuition, which can be difficult for the management of design. Here the concept of Brand Formative Design can provide a framework which contributes to structure, objectify, develop and assess new evolutionary design in brand and future relevant contexts, but also integrates consumers and their preferences without restricting creativity too much.
Resumo:
Solid oral dosage form disintegration in the human stomach is a highly complex process dependent on physicochemical properties of the stomach contents as well as on physical variables such as hydrodynamics and mechanical stress. Understanding the role of hydrodynamics and forces in disintegration of oral solid dosage forms can help to improve in vitro disintegration testing and the predictive power of the in vitro test. The aim of this work was to obtain a deep understanding of the influence of changing hydrodynamic conditions on solid oral dosage form performance. Therefore, the hydrodynamic conditions and forces present in the compendial PhEur/USP disintegration test device were characterized using a computational fluid dynamics (CFD) approach. Furthermore, a modified device was developed and the hydrodynamic conditions present were simulated using CFD. This modified device was applied in two case studies comprising immediate release (IR) tablets and gastroretentive drug delivery systems (GRDDS). Due to the description of movement provided in the PhEur, the movement velocity of the basket-rack assembly follows a sinusoidal profile. Therefore, hydrodynamic conditions are changing continually throughout the movement cycle. CFD simulations revealed that the dosage form is exposed to a wide range of fluid velocities and shear forces during the test. The hydrodynamic conditions in the compendial device are highly variable and cannot be controlled. A new, modified disintegration test device based on computerized numerical control (CNC) technique was developed. The modified device can be moved in all three dimensions and radial movement is also possible. Simple and complex moving profiles can be developed and the influence of the hydrodynamic conditions on oral solid dosage form performance can be evaluated. Furthermore, a modified basket was designed that allows two-sided fluid flow. CFD simulations of the hydrodynamics and forces in the modified device revealed significant differences in the fluid flow field and forces when compared to the compendial device. Due to the CNC technique moving velocity and direction are arbitrary and hydrodynamics become controllable. The modified disintegration test device was utilized to examine the influence of moving velocity on disintegration times of IR tablets. Insights into the influence of moving speed, medium viscosity and basket design on disintegration times were obtained. An exponential relationship between moving velocity of the modified basket and disintegration times was established in simulated gastric fluid. The same relationship was found between the disintegration times and the CFD predicted average shear stress on the tablet surface. Furthermore, a GRDDS was developed based on the approach of an in situ polyelectrolyte complex (PEC). Different complexes composed of different grades of chitosan and carrageenan and different ratios of those were investigated for their swelling behavior, mechanical stability, and in vitro drug release. With an optimized formulation the influence of changing hydrodynamic conditions on the swelling behavior and the drug release profile was demonstrated using the modified disintegration test device. Both, swelling behavior and drug release, were largely dependent on the hydrodynamic conditions. Concluding, it has been shown within this thesis that the application of the modified disintegration test device allows for detailed insights into the influence of hydrodynamic conditions on solid oral dosage form disintegration and dissolution. By the application of appropriate test conditions, the predictive power of in vitro disintegration testing can be improved using the modified disintegration test device. Furthermore, CFD has proven a powerful tool to examine the hydrodynamics and forces in the compendial as well as in the modified disintegration test device. rn
Resumo:
Learning by reinforcement is important in shaping animal behavior, and in particular in behavioral decision making. Such decision making is likely to involve the integration of many synaptic events in space and time. However, using a single reinforcement signal to modulate synaptic plasticity, as suggested in classical reinforcement learning algorithms, a twofold problem arises. Different synapses will have contributed differently to the behavioral decision, and even for one and the same synapse, releases at different times may have had different effects. Here we present a plasticity rule which solves this spatio-temporal credit assignment problem in a population of spiking neurons. The learning rule is spike-time dependent and maximizes the expected reward by following its stochastic gradient. Synaptic plasticity is modulated not only by the reward, but also by a population feedback signal. While this additional signal solves the spatial component of the problem, the temporal one is solved by means of synaptic eligibility traces. In contrast to temporal difference (TD) based approaches to reinforcement learning, our rule is explicit with regard to the assumed biophysical mechanisms. Neurotransmitter concentrations determine plasticity and learning occurs fully online. Further, it works even if the task to be learned is non-Markovian, i.e. when reinforcement is not determined by the current state of the system but may also depend on past events. The performance of the model is assessed by studying three non-Markovian tasks. In the first task, the reward is delayed beyond the last action with non-related stimuli and actions appearing in between. The second task involves an action sequence which is itself extended in time and reward is only delivered at the last action, as it is the case in any type of board-game. The third task is the inspection game that has been studied in neuroeconomics, where an inspector tries to prevent a worker from shirking. Applying our algorithm to this game yields a learning behavior which is consistent with behavioral data from humans and monkeys, revealing themselves properties of a mixed Nash equilibrium. The examples show that our neuronal implementation of reward based learning copes with delayed and stochastic reward delivery, and also with the learning of mixed strategies in two-opponent games.