893 resultados para Model methodology of empirical research in communication
Resumo:
La cuestión principal abordada en esta tesis doctoral es la mejora de los sistemas biométricos de reconocimiento de personas a partir de la voz, proponiendo el uso de una nueva parametrización, que hemos denominado parametrización biométrica extendida dependiente de género (GDEBP en sus siglas en inglés). No se propone una ruptura completa respecto a los parámetros clásicos sino una nueva forma de utilizarlos y complementarlos. En concreto, proponemos el uso de parámetros diferentes dependiendo del género del locutor, ya que como es bien sabido, la voz masculina y femenina presentan características diferentes que deberán modelarse, por tanto, de diferente manera. Además complementamos los parámetros clásicos utilizados (MFFC extraídos de la señal de voz), con un nuevo conjunto de parámetros extraídos a partir de la deconstrucción de la señal de voz en sus componentes de fuente glótica (más relacionada con el proceso y órganos de fonación y por tanto con características físicas del locutor) y de tracto vocal (más relacionada con la articulación acústica y por tanto con el mensaje emitido). Para verificar la validez de esta propuesta se plantean diversos escenarios, utilizando diferentes bases de datos, para validar que la GDEBP permite generar una descripción más precisa de los locutores que los parámetros MFCC clásicos independientes del género. En concreto se plantean diferentes escenarios de identificación sobre texto restringido y texto independiente utilizando las bases de datos de HESPERIA y ALBAYZIN. El trabajo también se completa con la participación en dos competiciones internacionales de reconocimiento de locutor, NIST SRE (2010 y 2012) y MOBIO 2013. En el primer caso debido a la naturaleza de las bases de datos utilizadas se obtuvieron resultados cercanos al estado del arte, mientras que en el segundo de los casos el sistema presentado obtuvo la mejor tasa de reconocimiento para locutores femeninos. A pesar de que el objetivo principal de esta tesis no es el estudio de sistemas de clasificación, sí ha sido necesario analizar el rendimiento de diferentes sistemas de clasificación, para ver el rendimiento de la parametrización propuesta. En concreto, se ha abordado el uso de sistemas de reconocimiento basados en el paradigma GMM-UBM, supervectores e i-vectors. Los resultados que se presentan confirman que la utilización de características que permitan describir los locutores de manera más precisa es en cierto modo más importante que la elección del sistema de clasificación utilizado por el sistema. En este sentido la parametrización propuesta supone un paso adelante en la mejora de los sistemas de reconocimiento biométrico de personas por la voz, ya que incluso con sistemas de clasificación relativamente simples se consiguen tasas de reconocimiento realmente competitivas. ABSTRACT The main question addressed in this thesis is the improvement of automatic speaker recognition systems, by the introduction of a new front-end module that we have called Gender Dependent Extended Biometric Parameterisation (GDEBP). This front-end do not constitute a complete break with respect to classical parameterisation techniques used in speaker recognition but a new way to obtain these parameters while introducing some complementary ones. Specifically, we propose a gender-dependent parameterisation, since as it is well known male and female voices have different characteristic, and therefore the use of different parameters to model these distinguishing characteristics should provide a better characterisation of speakers. Additionally, we propose the introduction of a new set of biometric parameters extracted from the components which result from the deconstruction of the voice into its glottal source estimate (close related to the phonation process and the involved organs, and therefore the physical characteristics of the speaker) and vocal tract estimate (close related to acoustic articulation and therefore to the spoken message). These biometric parameters constitute a complement to the classical MFCC extracted from the power spectral density of speech as a whole. In order to check the validity of this proposal we establish different practical scenarios, using different databases, so we can conclude that a GDEBP generates a more accurate description of speakers than classical approaches based on gender-independent MFCC. Specifically, we propose scenarios based on text-constrain and text-independent test using HESPERIA and ALBAYZIN databases. This work is also completed with the participation in two international speaker recognition evaluations: NIST SRE (2010 and 2012) and MOBIO 2013, with diverse results. In the first case, due to the nature of the NIST databases, we obtain results closed to state-of-the-art although confirming our hypothesis, whereas in the MOBIO SRE we obtain the best simple system performance for female speakers. Although the study of classification systems is beyond the scope of this thesis, we found it necessary to analise the performance of different classification systems, in order to verify the effect of them on the propose parameterisation. In particular, we have addressed the use of speaker recognition systems based on the GMM-UBM paradigm, supervectors and i-vectors. The presented results confirm that the selection of a set of parameters that allows for a more accurate description of the speakers is as important as the selection of the classification method used by the biometric system. In this sense, the proposed parameterisation constitutes a step forward in improving speaker recognition systems, since even when using relatively simple classification systems, really competitive recognition rates are achieved.
Resumo:
In this study we are proposing a Bayesian model selection methodology, where the best model from the list of candidate structural explanatory models is selected. The model structure is based on the Zellner's (1971)explanatory model with autoregressive errors. For the selection technique we are using a parsimonious model, where the model variables are transformed using Box and Cox (1964) class of transformations.
Resumo:
Vernacular architecture has demonstrated its perfect environmental adaptation through its empirical development and improvement by generations of user-builders. Nowadays, the sustainability of vernacular architecture is the aim of some research projects in which the same method should be applied in order to be comparable. Hence, we propose a research method putting together various steps. Through the analysis of geographical, lithology, economic, cultural and social influence as well as materials and constructive systems, vernacular architecture is analyzed. But, all this information is put together with the natural landscape (topography and vegetation) and the climatic data (temperature, winds, rain and sun exposure). In addition, the use of bioclimatic charts, such as Olgyay or Givoni’s, revealed the necessities and strategies in urban and building design. They are satisfied in the vernacular architecture by the application of different energy conservation mechanisms, some of them are shown by different examples in this paper.
Resumo:
The paper identifies the potential spatial and social impacts of a proposed road-pricing scheme for different social groups in the Madrid Metropolitan Area (MMA). We appraise the accessibility of different districts within the MMA in terms of the actual and perceived cost of using the road infrastructure ‘before’ and ‘after’ implementation of the scheme. The appraisal framework was developed using quantitative survey data and qualitative focus group discussions with residents. We then simulated user behaviours (mode and route choice) based on the empirical evidence from a travel demand model for the MMA. The results from our simulation model demonstrated that implementation of the toll on the orbital metropolitan motorways (M40, M30, for example) decreases accessibility mostly in the districts where there are no viable public transport alternatives. Our specific study finding is that the economic burden of the road-pricing scheme particularly affects unskilled and lower income individuals living in the south of the MMA. The focus groups confirmed that low income drivers in the south part of the MMA would reduce their use of tolled roads and have to find new arrangements for these trips: i.e. switch to public transport, spend double the time travelling or stay at home. More generally, our research finds that European transport planners are still a long way from recognising the social equity implications of their policy decisions and that more thorough social appraisals are needed to avoid the social exclusion of low income populations when road tolling is proposed.
Resumo:
The paper explores the spatial and social impacts arising from implementation of a road-pricing scheme in the Madrid Metropolitan Area (MMA). Our analytical focus is on understanding the effects of the scheme on the transport accessibility of different social groups within the MMA. We define an evaluation framework to appraise the accessibility of different districts within the MMA in terms of the actual and perceived cost of using the road infrastructure "before" and "after" the implementation of the scheme. The framework was developed using quantitative survey data and qualitative data from focus group discussions with residents. We then simulated user behaviors (mode and route choice) based on the empirical evidence from a travel demand model for the MMA. The results from our simulation model demonstrated that implementation of the toll on the orbital metropolitan motorways (M40, M30, for example) decreases accessibility, mostly in the districts where there are no viable public transport alternatives. Our key finding is that the economic burden of the road-pricing scheme particularly affects unskilled and lower income individuals living in the south of the MMA. Consequently lower income people reduce their use of tolled roads and have to find new arrangements for these trips: i.e. switch to the public transport, spend double the time for their commuter trips or stay at home. The results of our research could be applicable more widely for anyone wishing to better understand the important relationship between increased transport cost and social equity, especially where there is an intention to introduce similar road-pricing schemes within the urban context.
Resumo:
Los sistemas transaccionales tales como los programas informáticos para la planificación de recursos empresariales (ERP software) se han implementado ampliamente mientras que los sistemas analíticos para la gestión de la cadena de suministro (SCM software) no han tenido el éxito deseado por la industria de tecnología de información (TI). Aunque se documentan beneficios importantes derivados de las implantaciones de SCM software, las empresas industriales son reacias a invertir en este tipo de sistemas. Por una parte esto es debido a la falta de métodos que son capaces de detectar los beneficios por emplear esos sistemas, y por otra parte porque el coste asociado no está identificado, detallado y cuantificado suficientemente. Los esquemas de coordinación basados únicamente en sistemas ERP son alternativas válidas en la práctica industrial siempre que la relación coste-beneficio esta favorable. Por lo tanto, la evaluación de formas organizativas teniendo en cuenta explícitamente el coste debido a procesos administrativos, en particular por ciclos iterativos, es de gran interés para la toma de decisiones en el ámbito de inversiones en TI. Con el fin de cerrar la brecha, el propósito de esta investigación es proporcionar métodos de evaluación que permitan la comparación de diferentes formas de organización y niveles de soporte por sistemas informáticos. La tesis proporciona una amplia introducción, analizando los retos a los que se enfrenta la industria. Concluye con las necesidades de la industria de SCM software: unas herramientas que facilitan la evaluación integral de diferentes propuestas de organización. A continuación, la terminología clave se detalla centrándose en la teoría de la organización, las peculiaridades de inversión en TI y la tipología de software de gestión de la cadena de suministro. La revisión de la literatura clasifica las contribuciones recientes sobre la gestión de la cadena de suministro, tratando ambos conceptos, el diseño de la organización y su soporte por las TI. La clasificación incluye criterios relacionados con la metodología de la investigación y su contenido. Los estudios empíricos en el ámbito de la administración de empresas se centran en tipologías de redes industriales. Nuevos algoritmos de planificación y esquemas de coordinación innovadoras se desarrollan principalmente en el campo de la investigación de operaciones con el fin de proponer nuevas funciones de software. Artículos procedentes del área de la gestión de la producción se centran en el análisis de coste y beneficio de las implantaciones de sistemas. La revisión de la literatura revela que el éxito de las TI para la coordinación de redes industriales depende en gran medida de características de tres dimensiones: la configuración de la red industrial, los esquemas de coordinación y las funcionalidades del software. La literatura disponible está enfocada sobre todo en los beneficios de las implantaciones de SCM software. Sin embargo, la coordinación de la cadena de suministro, basándose en el sistema ERP, sigue siendo la práctica industrial generalizada, pero el coste de coordinación asociado no ha sido abordado por los investigadores. Los fundamentos de diseño organizativo eficiente se explican en detalle en la medida necesaria para la comprensión de la síntesis de las diferentes formas de organización. Se han generado varios esquemas de coordinación variando los siguientes parámetros de diseño: la estructura organizativa, los mecanismos de coordinación y el soporte por TI. Las diferentes propuestas de organización desarrolladas son evaluadas por un método heurístico y otro basado en la simulación por eventos discretos. Para ambos métodos, se tienen en cuenta los principios de la teoría de la organización. La falta de rendimiento empresarial se debe a las dependencias entre actividades que no se gestionan adecuadamente. Dentro del método heurístico, se clasifican las dependencias y se mide su intensidad basándose en factores contextuales. A continuación, se valora la idoneidad de cada elemento de diseño organizativo para cada dependencia específica. Por último, cada forma de organización se evalúa basándose en la contribución de los elementos de diseño tanto al beneficio como al coste. El beneficio de coordinación se refiere a la mejora en el rendimiento logístico - este concepto es el objeto central en la mayoría de modelos de evaluación de la gestión de la cadena de suministro. Por el contrario, el coste de coordinación que se debe incurrir para lograr beneficios no se suele considerar en detalle. Procesos iterativos son costosos si se ejecutan manualmente. Este es el caso cuando SCM software no está implementada y el sistema ERP es el único instrumento de coordinación disponible. El modelo heurístico proporciona un procedimiento simplificado para la clasificación sistemática de las dependencias, la cuantificación de los factores de influencia y la identificación de configuraciones que indican el uso de formas organizativas y de soporte de TI más o menos complejas. La simulación de eventos discretos se aplica en el segundo modelo de evaluación utilizando el paquete de software ‘Plant Simulation’. Con respecto al rendimiento logístico, por un lado se mide el coste de fabricación, de inventario y de transporte y las penalizaciones por pérdida de ventas. Por otro lado, se cuantifica explícitamente el coste de la coordinación teniendo en cuenta los ciclos de coordinación iterativos. El método se aplica a una configuración de cadena de suministro ejemplar considerando diversos parámetros. Los resultados de la simulación confirman que, en la mayoría de los casos, el beneficio aumenta cuando se intensifica la coordinación. Sin embargo, en ciertas situaciones en las que se aplican ciclos de planificación manuales e iterativos el coste de coordinación adicional no siempre conduce a mejor rendimiento logístico. Estos resultados inesperados no se pueden atribuir a ningún parámetro particular. La investigación confirma la gran importancia de nuevas dimensiones hasta ahora ignoradas en la evaluación de propuestas organizativas y herramientas de TI. A través del método heurístico se puede comparar de forma rápida, pero sólo aproximada, la eficiencia de diferentes formas de organización. Por el contrario, el método de simulación es más complejo pero da resultados más detallados, teniendo en cuenta parámetros específicos del contexto del caso concreto y del diseño organizativo. ABSTRACT Transactional systems such as Enterprise Resource Planning (ERP) systems have been implemented widely while analytical software like Supply Chain Management (SCM) add-ons are adopted less by manufacturing companies. Although significant benefits are reported stemming from SCM software implementations, companies are reluctant to invest in such systems. On the one hand this is due to the lack of methods that are able to detect benefits from the use of SCM software and on the other hand associated costs are not identified, detailed and quantified sufficiently. Coordination schemes based only on ERP systems are valid alternatives in industrial practice because significant investment in IT can be avoided. Therefore, the evaluation of these coordination procedures, in particular the cost due to iterations, is of high managerial interest and corresponding methods are comprehensive tools for strategic IT decision making. The purpose of this research is to provide evaluation methods that allow the comparison of different organizational forms and software support levels. The research begins with a comprehensive introduction dealing with the business environment that industrial networks are facing and concludes highlighting the challenges for the supply chain software industry. Afterwards, the central terminology is addressed, focusing on organization theory, IT investment peculiarities and supply chain management software typology. The literature review classifies recent supply chain management research referring to organizational design and its software support. The classification encompasses criteria related to research methodology and content. Empirical studies from management science focus on network types and organizational fit. Novel planning algorithms and innovative coordination schemes are developed mostly in the field of operations research in order to propose new software features. Operations and production management researchers realize cost-benefit analysis of IT software implementations. The literature review reveals that the success of software solutions for network coordination depends strongly on the fit of three dimensions: network configuration, coordination scheme and software functionality. Reviewed literature is mostly centered on the benefits of SCM software implementations. However, ERP system based supply chain coordination is still widespread industrial practice but the associated coordination cost has not been addressed by researchers. Fundamentals of efficient organizational design are explained in detail as far as required for the understanding of the synthesis of different organizational forms. Several coordination schemes have been shaped through the variation of the following design parameters: organizational structuring, coordination mechanisms and software support. The different organizational proposals are evaluated using a heuristic approach and a simulation-based method. For both cases, the principles of organization theory are respected. A lack of performance is due to dependencies between activities which are not managed properly. Therefore, within the heuristic method, dependencies are classified and their intensity is measured based on contextual factors. Afterwards the suitability of each organizational design element for the management of a specific dependency is determined. Finally, each organizational form is evaluated based on the contribution of the sum of design elements to coordination benefit and to coordination cost. Coordination benefit refers to improvement in logistic performance – this is the core concept of most supply chain evaluation models. Unfortunately, coordination cost which must be incurred to achieve benefits is usually not considered in detail. Iterative processes are costly when manually executed. This is the case when SCM software is not implemented and the ERP system is the only available coordination instrument. The heuristic model provides a simplified procedure for the classification of dependencies, quantification of influence factors and systematic search for adequate organizational forms and IT support. Discrete event simulation is applied in the second evaluation model using the software package ‘Plant Simulation’. On the one hand logistic performance is measured by manufacturing, inventory and transportation cost and penalties for lost sales. On the other hand coordination cost is explicitly considered taking into account iterative coordination cycles. The method is applied to an exemplary supply chain configuration considering various parameter settings. The simulation results confirm that, in most cases, benefit increases when coordination is intensified. However, in some situations when manual, iterative planning cycles are applied, additional coordination cost does not always lead to improved logistic performance. These unexpected results cannot be attributed to any particular parameter. The research confirms the great importance of up to now disregarded dimensions when evaluating SCM concepts and IT tools. The heuristic method provides a quick, but only approximate comparison of coordination efficiency for different organizational forms. In contrast, the more complex simulation method delivers detailed results taking into consideration specific parameter settings of network context and organizational design.
Resumo:
Recent research has shown large differences between the expected and the actual energy consumption in buildings. The differences have been attributed partially, to the assumptions made during the design phase of buildings when simulation methods are employed. More accurate occupancy profiles on building operation could help to carry out more precise building performance calculations. This study focuses on the post-occupancy evaluation of two apartments, one renovated and one non renovated, in Madrid within the same building complex. The aim of this paper is to present an application of the mixed-methods methodology (Creswell, 2007) to assess thermal comfort and occupancy practices used in the case studies, and to discuss the shortcomings and opportunities associated with it. The mixed-methods methodology offers strategies for integrating qualitative and quantitative methods to investigate complex phenomena. This approach is expected to contribute to the growing knowledge of occupants’ behaviour and building performance by explaining the differences observed between energy consumption and thermal comfort in relation to people’s saving and comfort practices and the related experiences, preferences and values.
Resumo:
This paper presents a work whose objective is, first, to quantify the potential of the triticale biomass existing in each of the agricultural regions in the Madrid Community through a crop simulation model based on regression techniques and multiple correlation. Second, a methodology for defining which area has the best conditions for the installation of electricity plants from biomass has been described and applied. The study used a methodology based on compromise programming in a discrete multicriteria decision method (MDM) context. To make a ranking, the following criteria were taken into account: biomass potential, electric power infrastructure, road networks, protected spaces, and urban nuclei surfaces. The results indicate that, in the case of the Madrid Community, the Campiña region is the most suitable for setting up plants powered by biomass. A minimum of 17,339.9 tons of triticale will be needed to satisfy the requirements of a 2.2 MW power plant. The minimum range of action for obtaining the biomass necessary in Campiña region would be 6.6 km around the municipality of Algete, based on Geographic Information Systems. The total biomass which could be made available in considering this range in this region would be 18,430.68 t.
Resumo:
Polyethylene chains in the amorphous region between two crystalline lamellae M unit apart are modeled as random walks with one-step memory on a cubic lattice between two absorbing boundaries. These walks avoid the two preceding steps, though they are not true self-avoiding walks. Systems of difference equations are introduced to calculate the statistics of the restricted random walks. They yield that the fraction of loops is (2M - 2)/(2M + 1), the fraction of ties 3/(2M + 1), the average length of loops 2M - 0.5, the average length of ties 2/3M2 + 2/3M - 4/3, the average length of walks equals 3M - 3, the variance of the loop length 16/15M3 + O(M2), the variance of the tie length 28/45M4 + O(M3), and the variance of the walk length 2M3 + O(M2).
Resumo:
Overexpression of the c-myc oncogene is associated with a variety of both human and experimental tumors, and cooperation of other oncogenes and growth factors with the myc family are critical in the evolution of the malignant phenotype. The interaction of hepatocyte growth factor (HGF) with c-myc during hepatocarcinogenesis in a transgenic mouse model has been analyzed. While sustained overexpression of c-myc in the liver leads to cancer, coexpression of HGF and c-myc in the liver delayed the appearance of preneoplastic lesions and prevented malignant conversion. Furthermore, tumor promotion by phenobarbital was completely inhibited in the c-myc/HGF double transgenic mice, whereas phenobarbital was an effective tumor promoter in the c-myc single transgenic mice. The results indicate that HGF may function as a tumor suppressor during early stages of liver carcinogenesis, and suggest the possibility of therapeutic application for this cytokine.
Resumo:
The genes of the homeotic complex (HOX) encode DNA binding homeodomain proteins that control developmental fates by differentially regulating the transcription of downstream target genes. Despite their unique in vivo functions, disparate HOX proteins often bind to very similar DNA sequences in vitro. Thus, a critical question is how HOX proteins select the correct sets of target genes in vivo. The homeodomain proteins encoded by the Drosophila extradenticle gene and its mammalian homologues, the pbx genes, contribute to HOX specificity by cooperatively binding to DNA with HOX proteins. For example, the HOX protein labial cooperatively binds with extradenticle protein to a 20-bp oligonucleotide that is sufficient to direct a labial-like expression pattern in Drosophila embryos. Here we have analyzed the protein-DNA interactions that are important for forming the labial-extradenticle-DNA complex. The data suggest a model in which labial and extradenticle, separated by only 4 bp, bind this DNA as a heterodimer in a head-to-tail orientation. We have confirmed several aspects of this model by characterizing extradenticle-HOX binding to mutant oligonucleotides. Most importantly, mutations in base pairs predicted to contact the HOX N-terminal arm resulted in a change in HOX preference in the heterodimer, from labial to Ultrabithorax. These results demonstrate that extradenticle prefers to bind cooperatively with different HOX proteins depending on subtle differences in the heterodimer binding site.
Resumo:
The gastric mucosa of mammalian stomach contains several differentiated cell types specialized for the secretion of acid, digestive enzymes, mucus, and hormones. Understanding whether each of these cell lineages is derived from a common stem cell has been a challenging problem. We have used a genetic approach to analyze the ontogeny of progenitor cells within mouse stomach. Herpes simplex virus 1 thymidine kinase was targeted to parietal cells within the gastric mucosa of transgenic mice, and parietal cells were ablated by treatment of animals with the antiherpetic drug ganciclovir. Ganciclovir treatment produced complete ablation of parietal cells, dissolution of gastric glands, and loss of chief and mucus-producing cells. Termination of drug treatment led to the reemergence of all major gastric epithelial cell types and restoration of glandular architecture. Our results imply the existence of a pluripotent stem cell for the gastric mucosa. Parietal cell ablation should provide a model for analyzing cell lineage relationships within the stomach as well as mechanisms underlying gastric injury and repair.
Resumo:
When NMR hydrogen exchange was used previously to monitor the kinetics of RNase A unfolding, some peptide NH protons were found to show EX2 exchange (detected by base catalysis) in addition to the expected EX1 exchange, whose rate is limited by the kinetic unfolding process. In earlier work, two groups showed independently that a restricted two-process model successfully fits published hydrogen exchange rates of native RNase A in the range 0-0.7 M guanidinium chloride. We find that this model predicts properties that are very different from the observed properties of the EX2 exchange reactions of RNase A in conditions where guanidine-induced unfolding takes place. The model predicts that EX2 exchange should be too fast to measure by the technique used, whereas it is readily measurable. Possible explanations for the contradiction are considered here, and we show that removing the restriction from the earlier two-process model is sufficient to resolve the contradiction; instead of specifying that exchange caused by global unfolding occurs by the EX2 mechanism, we allow it to occur by the general mechanism, which includes both the EX1 and EX2 cases. It is logical to remove this restriction because global unfolding of RNase A is known to give rise to EX1 exchange in these unfolding conditions. Resolving the contradiction makes it possible to determine whether populated unfolding intermediates contribute to the EX2 exchange, and this question is considered elsewhere. The results and simulations indicate that moderate or high denaturant concentrations readily give rise to EX1 exchange in native proteins. Earlier studies showed that hydrogen exchange in native proteins typically occurs by the EX2 mechanism but that high temperatures or pH values above 7 may give rise to EX1 exchange. High denaturant concentrations should be added to the list of variables likely to cause EX1 exchange.
Resumo:
Advances in digital speech processing are now supporting application and deployment of a variety of speech technologies for human/machine communication. In fact, new businesses are rapidly forming about these technologies. But these capabilities are of little use unless society can afford them. Happily, explosive advances in microelectronics over the past two decades have assured affordable access to this sophistication as well as to the underlying computing technology. The research challenges in speech processing remain in the traditionally identified areas of recognition, synthesis, and coding. These three areas have typically been addressed individually, often with significant isolation among the efforts. But they are all facets of the same fundamental issue--how to represent and quantify the information in the speech signal. This implies deeper understanding of the physics of speech production, the constraints that the conventions of language impose, and the mechanism for information processing in the auditory system. In ongoing research, therefore, we seek more accurate models of speech generation, better computational formulations of language, and realistic perceptual guides for speech processing--along with ways to coalesce the fundamental issues of recognition, synthesis, and coding. Successful solution will yield the long-sought dictation machine, high-quality synthesis from text, and the ultimate in low bit-rate transmission of speech. It will also open the door to language-translating telephony, where the synthetic foreign translation can be in the voice of the originating talker.
Resumo:
The subject of Construction of Structures I studies, from a constructive point of view and taking into account current legislation, reinforced concrete structures used in buildings, through the acquisition of knowledge and construction criteria required in the profession of a Technical Architect. The contents acquired in this course are essential for further professional development of technicians and are closely related to many of the subjects taught in the same or other courses of the Degree in Technical Architecture at the University of Alicante. The aim of this paper is to present, analyze and discuss the development of a new methodology proposed in the mentioned subject, as it supposed an important change in the traditional way of teaching Construction and Structures I. In order to incorporate new teaching tools in 2013-2014, the course has been implemented by using a Moodle software tool to promote blended learning with online exercises. Our Moodle community allows collaborative work within an open-source platform where teachers and students share a new and personalized learning environment. Students are easily used to the interface and the platform, value the constant connection with teachers or other fellows and completely agree with the possibility of making questions or share documents 24 hours a day. The proposed methodology consists of lectures and practical classes. In the lectures, the basics of each topic are discussed; class attendance, daily study and conducting scheduled exercises are indispensable. Practical classes allow to consolidate the knowledge gained in theory classes by solving professional exercises and actual construction problems related to structures, that shall be compulsorily delivered online. So, after the correction of the teacher and the subsequent feedback of students, practical exercises ensure lifelong learning of the student, who can download any kind of material at any time (constructive details, practical exercises and even corrected exams). Regarding the general evaluation system, goals achievement is assessed on an ongoing basis (65% of the final mark) along the course through written and graphic evidences in person and online, as well as a individual development of a workbook. In all cases, the acquisition of skills, the ability to synthesize, the capacity of logical and critical thinking are assessed. The other 35 % of the mark is evaluated by a complementary graphic exam. Participation in the computing platform is essential and the student is required to do and present, at least 90% of the practices proposed. Those who do not comply with the practices in each specific date could not be assessed continuously and may only choose the final exam. In conclusion, the subject of Construction of Structures I is essential in the development of the regulated profession of Technical Architect as they are considered, among other professional profiles, as specialists in construction of building structures. The use of a new communication platform and online teaching allows the acquisition of knowledge and constructive approaches in a continuous way, with a more direct and personal monitoring by the teacher that has been highly appreciated by almost 100% of the students. Ultimately, it is important to say that the use of Moodle in this subject is a very interesting tool, which was really well welcome by students in one of the densest and important subjects of the Degree of Technical Architecture.