14 resultados para Translating and interpreting -- TFC
em AMS Tesi di Dottorato - Alm@DL - Università di Bologna
Resumo:
This volume is a collection of the work done in a three years-lasting PhD, focused in the analysis of Central and Southern Adriatic marine sediments, deriving from the collection of a borehole and many cores, achieved thanks to the good seismic-stratigraphic knowledge of the study area. The work was made out within European projects EC-EURODELTA (coordinated by Fabio Trincardi, ISMAR-CNR), EC-EUROSTRATAFORM (coordinated by Phil P. E. Weaver, NOC, UK), and PROMESS1 (coordinated by Serge Bernè, IFREMER, France). The analysed sedimentary successions presented highly expanded stratigraphic intervals, particularly for the last 400 kyr, 60 kyr and 6 kyr BP. These three different time-intervals resulted in a tri-partition of the PhD thesis. The study consisted of the analysis of planktic and benthic foraminifers’ assemblages (more than 560 samples analysed), as well as in preparing the material for oxygen and carbon stable isotope analyses, and interpreting and discussing the obtained dataset. The chronologic framework of the last 400 kyr was achieved for borehole PRAD1-2 (within the work-package WP6 of PROMESS1 project), collected in 186.5 m water depth. The proposed chronology derives from a multi-disciplinary approach, consisting of the integration of numerous and independent proxies, some of which analysed by other specialists within the project. The final framework based on: micropaleontology (calcareous nannofossils and foraminifers’ bioevents), climatic cyclicity (foraminifers’ assemblages), geochemistry (oxygen stable isotope, made out on planktic and benthic records), paleomagnetism, radiometric ages (14C AMS), teprhochronology, identification of sapropel-equivalent levels (Se). It’s worth to note the good consistency between the oxygen stable isotope curve obtained for borehole PRAD1-2 and other deeper Mediterranean records. The studied proxies allowed the recognition of all the isotopic intervals from MIS10 to MIS1 in PRAD1-2 record, and the base of the borehole has been ascribed to the early MIS11. Glacial and interglacial intervals identified in the Central Adriatic record have been analysed in detail for the paleo-environmental reconstruction, as well. For instance, glacial stages MIS6, MIS8 and MIS10 present peculiar foraminifers’ assemblages, composed by benthic species typical of polar regions and no longer living in the Central Adriatic nowadays. Moreover, a deepening trend in the paleo-bathymetry during glacial intervals was observed, from MIS10 (inner-shelf environment) to MIS4 (mid-shelf environment).Ten sapropel-equivalent levels have been recognised in PRAD1-2 Central Adriatic record. They showed different planktic foraminifers’ assemblages, which allowed the first distinction of events occurred during warm-climate (Se5, Se7), cold-climate (Se4, Se6 and Se8) and temperate-intermediate-climate (Se1, Se3, Se9, Se’, Se10) conditions, consistently with literature. Cold-climate sapropel equivalents are characterised by the absence of an oligotrophic phase, whereas warm-temeprate-climate sapropel equivalents present both the oligotrophic and the eutrophic phases (except for Se1). Sea floor conditions vary, according to benthic foraminifers’ assemblages, from relatively well oxygenated (Se1, Se3), to dysoxic (Se9, Se’, Se10), to highly dysoxic (Se4, Se6, Se8) to events during which benthic foraminifers are absent (Se5, Se7). These two latter levels are also characterised by the lamination of the sediment, feature never observed in literature in such shallow records. The enhanced stratification of the water column during the events Se8, Se7, Se6, Se5, Se4, and the concurring strong dilution of shallow water, pointed out by the isotope record, lead to the hypothesis of a period of intense precipitation in the Central Adriatic region, possibly due to a northward shift of the African Monsoon. Finally, the expression of Central Adriatic PRAD1-2 Se5 equivalent was compared with the same event, as registered in other Eastern Mediterranean areas. The sequence of substantially the same planktic foraminifers’ bioevents has been consistently recognised, indicating a similar evolution of the water column all over the Eastern Mediterranean; yet, the synchronism of these events cannot be demonstrated. A high resolution analysis of late Holocene (last 6000 years BP) climate change was carried out for the Adriatic area, through the recognition of planktic and benthic foraminifers’ bioevents. In particular, peaks of planktic Globigerinoides sacculifer (four during the last 5500 years BP in the most expanded core) have been interpreted, based on the ecological requirements of this species, as warm-climate, arid intervals, correspondent to periods of relative climatic optimum, such as, for instance, the Medieval Warm Period, the Roman Age, the Late Bronze Age and the Copper Age. Consequently, the minima in the abundance of this biomarker could correspond to relatively cooler and more rainy periods. These conclusions are in good agreement with the isotopic and the pollen data. The Last Occurrence (LO) of G. sacculifer has been dated in this work at an average age of 550 years BP, and it is the best bioevent approximating the base of the Little Ice Age in the Adriatic. Recent literature reports the same bioevent in the Levantine Basin, showing a rather consistent age. Therefore, the LO of G. sacculifer has the potential to be extended to all the Eastern Mediterranean. Within the Little Ice Age, benthic foraminifer V. complanata shows two distinct peaks in the shallower Adriatic cores analysed, collected hundred kilometres apart, inside the mud belt environment. Based on the ecological requirements of this species, these two peaks have been interpreted as the more intense (cold and rainy) oscillations inside the LIA. The chronologic framework of the analysed cores is robust, being based on several range-finding 14C AMS ages, on estimates of the secular variation of the magnetic field, on geochemical estimates of the activity depth of 210Pb short-lived radionuclide (for the core-top ages), and is in good agreement with tephrochronologic, pollen and foraminiferal data. The intra-holocenic climate oscillations find out in the Adriatic have been compared with those pointed out in literature from other records of the Northern Hemisphere, and the chronologic constraint seems quite good. Finally, the sedimentary successions analysed allowed the review and the update of the foraminifers’ ecobiostratigraphy available from literature for the Adriatic region, thanks to the achievement of 16 ecobiozones for the last 60 kyr BP. Some bioevents are restricted to the Central Adriatic (for instance the LO of benthic Hyalinea balthica , approximating the MIS3/MIS2 boundary), others occur all over the Adriatic basin (for instance the LO of planktic Globorotalia inflata during MIS3, individuating Dansgaard-Oeschger cycle 8 (Denekamp)).
Resumo:
Clusters have increasingly become an essential part of policy discourses at all levels, EU, national, regional, dealing with regional development, competitiveness, innovation, entrepreneurship, SMEs. These impressive efforts in promoting the concept of clusters on the policy-making arena have been accompanied by much less academic and scientific research work investigating the actual economic performance of firms in clusters, the design and execution of cluster policies and going beyond singular case studies to a more methodologically integrated and comparative approach to the study of clusters and their real-world impact. The theoretical background is far from being consolidated and there is a variety of methodologies and approaches for studying and interpreting this phenomenon while at the same time little comparability among studies on actual cluster performances. The conceptual framework of clustering suggests that they affect performance but theory makes little prediction as to the ultimate distribution of the value being created by clusters. This thesis takes the case of Eastern European countries for two reasons. One is that clusters, as coopetitive environments, are a new phenomenon as the previous centrally-based system did not allow for such types of firm organizations. The other is that, as new EU member states, they have been subject to the increased popularization of the cluster policy approach by the European Commission, especially in the framework of the National Reform Programmes related to the Lisbon objectives. The originality of the work lays in the fact that starting from an overview of theoretical contributions on clustering, it offers a comparative empirical study of clusters in transition countries. There have been very few examples in the literature that attempt to examine cluster performance in a comparative cross-country perspective. It adds to this an analysis of cluster policies and their implementation or lack of such as a way to analyse the way the cluster concept has been introduced to transition economies. Our findings show that the implementation of cluster policies does vary across countries with some countries which have embraced it more than others. The specific modes of implementation, however, are very similar, based mostly on soft measures such as funding for cluster initiatives, usually directed towards the creation of cluster management structures or cluster facilitators. They are essentially founded on a common assumption that the added values of clusters is in the creation of linkages among firms, human capital, skills and knowledge at the local level, most often perceived as the regional level. Often times geographical proximity is not a necessary element in the application process and cluster application are very similar to network membership. Cluster mapping is rarely a factor in the selection of cluster initiatives for funding and the relative question about critical mass and expected outcomes is not considered. In fact, monitoring and evaluation are not elements of the cluster policy cycle which have received a lot of attention. Bulgaria and the Czech Republic are the countries which have implemented cluster policies most decisively, Hungary and Poland have made significant efforts, while Slovakia and Romania have only sporadically and not systematically used cluster initiatives. When examining whether, in fact, firms located within regional clusters perform better and are more efficient than similar firms outside clusters, we do find positive results across countries and across sectors. The only country with negative impact from being located in a cluster is the Czech Republic.
Resumo:
A 30 anni dalla Dichiarazione di Alma Ata, l'Organizzazione Mondiale della Sanità, sia nei lavori della Commissione sui Determinanti Sociali della Salute che nel corso della sua 62^ Assemblea (2009) ha posto nuovamente la sua attenzione al tema dei determinanti sociali della salute e allo sviluppo di una sanità secondo un approccio "Primary Health Care", in cui la partecipazione ai processi decisionali è uno dei fattori che possono incidere sull'equità in salute tra e nelle nazioni. Dopo una presentazione dei principali elementi e concetti teorici di riferimento della tesi: Determinanti Sociali della Salute, partecipazione ed empowerment partecipativo (Cap. 1 e 2), il lavoro di tesi, a seguito dell'attività di ricerca di campo svolta in Zambia (Lusaka, Kitwe e Ndola) e presso EuropeAid (Bruxelles), si concentra sui processi di sviluppo e riforma del settore sanitario (Cap. 3), sulle politiche di cooperazione internazionale (Cap.4) e sull'azione (spesso sperimentale) della società civile in Zambia, considerando (Cap. 5): le principali criticità e limiti della/alla partecipazione, la presenza di strumenti e strategie specifiche di empowerment partecipativo, le politiche di decentramento e accountability, le buone prassi e proposte emergenti dalla società civile, le linee e i ruoli assunti dai donatori internazionali e dal Governo dello Zambia. Con questa tesi di dottorato si è voluto evidenziare e interpretare sia il dibattito recente rispetto alla partecipazione nel settore sanitario che i diversi e contraddittori gradi di attenzione alla partecipazione delle politiche di sviluppo del settore sanitario e l'emergere delle istanze e pratiche della società civile. Tutto questo incide su spazi e forme di partecipazione alla governance e ai processi decisionali nel settore sanitario, che influenzano a loro volta le politiche e condizioni di equità in salute. La metodologia adottata è stata di tipo qualitativo articolata in osservazione, interviste, analisi bibliografica e documentale.
Resumo:
Il progetto di dottorato IMITES (Interpretación de la Metáfora entre ITaliano y ESpañol) si pone come obiettivo quello di analizzare l’interpretazione simultanea del linguaggio figurato nelle combinazioni italiano-spagnolo e spagnolo-italiano. Prevede l’analisi di una serie di dati estratti da discorsi pronunciati in italiano e spagnolo in occasione di conferenze tenutesi presso la Commissione europea, e le loro versioni interpretate in spagnolo e italiano rispettivamente. Le espressioni figurate contenute nei discorsi originali sono state allineate e messe a confronto con le versioni fornite dagli interpreti, con il duplice obiettivo di a) capire quali causano maggiori problemi agli interpreti e b) analizzare le strategie di interpretazione applicate da professionisti quali quelli della Direzione Generale Interpretazione (DG SCIC) della Commissione europea nell’interpretare metafore. Il progetto prevede anche la somministrazione di un questionario agli interpreti delle cabine spagnola e italiana del DG SCIC, con l’obiettivo di sondare la loro percezione delle difficoltà che sottendono all’interpretazione del linguaggio figurato, le indicazioni metodologiche ricevute (se del caso) dai loro docenti a tale riguardo e le strategie applicate nella pratica professionale. Infine, l’ultima fase del progetto di ricerca prevede la sperimentazione di una proposta didattica attraverso uno studio caso-controllo svolto su studenti del secondo anno della Laurea Magistrale in Interpretazione delle Scuole Interpreti di Forlì e Trieste. Il gruppo-caso ha ricevuto una formazione specifica sull'interpretazione delle metafore, mentre gruppo-controllo è stato monitorato nella sua evoluzione. L’obiettivo di questa ultima fase di ricerca è quello di valutare, da una parte, l’ “insegnabilità” di strategia per affrontare il linguaggio figurato in interpretazione simultanea, e, dall’altra, l’efficacia dell’unità didattica proposta, sviluppata in base all’analisi svolta su IMITES.
Resumo:
The goal of the present research is to define a Semantic Web framework for precedent modelling, by using knowledge extracted from text, metadata, and rules, while maintaining a strong text-to-knowledge morphism between legal text and legal concepts, in order to fill the gap between legal document and its semantics. The framework is composed of four different models that make use of standard languages from the Semantic Web stack of technologies: a document metadata structure, modelling the main parts of a judgement, and creating a bridge between a text and its semantic annotations of legal concepts; a legal core ontology, modelling abstract legal concepts and institutions contained in a rule of law; a legal domain ontology, modelling the main legal concepts in a specific domain concerned by case-law; an argumentation system, modelling the structure of argumentation. The input to the framework includes metadata associated with judicial concepts, and an ontology library representing the structure of case-law. The research relies on the previous efforts of the community in the field of legal knowledge representation and rule interchange for applications in the legal domain, in order to apply the theory to a set of real legal documents, stressing the OWL axioms definitions as much as possible in order to enable them to provide a semantically powerful representation of the legal document and a solid ground for an argumentation system using a defeasible subset of predicate logics. It appears that some new features of OWL2 unlock useful reasoning features for legal knowledge, especially if combined with defeasible rules and argumentation schemes. The main task is thus to formalize legal concepts and argumentation patterns contained in a judgement, with the following requirement: to check, validate and reuse the discourse of a judge - and the argumentation he produces - as expressed by the judicial text.
Resumo:
Regulatory T cells (Treg) actively regulate alloimmune responses and promote transplantation tolerance. Polyclonal anti-thymocyte globulin (ATG), a widely used induction therapy in clinical organ transplantation, depletes peripheral T cells. However, resistance to tolerance induction is seen with certain T cell depleting strategies and is attributed to alterations in the balance of naïve, memory and regulatory T cells. Here we report a novel reagent, murine ATG (mATG), depletes T cells but preferentially spares CD25+ natural Tregs which limit skewing of T cell repertoire toward T-effector-memory (Tem) phenotype among the recovering T cells. T-cell depletion with mATG combined with CTLA4Ig and Sirolimus synergize to prolong graft survival by tipping the Treg/Tem balance further in favor of Tregs by preserving Tregs, facilitating generation of new Tregs by a conversion mechanism and limiting Tem expansion in response to alloantigen and homeostatic proliferation. These results provide the rationale for translating such novel combination therapies to promote tolerance in primate and human organ transplantation.
Resumo:
Forecasting the time, location, nature, and scale of volcanic eruptions is one of the most urgent aspects of modern applied volcanology. The reliability of probabilistic forecasting procedures is strongly related to the reliability of the input information provided, implying objective criteria for interpreting the historical and monitoring data. For this reason both, detailed analysis of past data and more basic research into the processes of volcanism, are fundamental tasks of a continuous information-gain process; in this way the precursor events of eruptions can be better interpreted in terms of their physical meanings with correlated uncertainties. This should lead to better predictions of the nature of eruptive events. In this work we have studied different problems associated with the long- and short-term eruption forecasting assessment. First, we discuss different approaches for the analysis of the eruptive history of a volcano, most of them generally applied for long-term eruption forecasting purposes; furthermore, we present a model based on the characteristics of a Brownian passage-time process to describe recurrent eruptive activity, and apply it for long-term, time-dependent, eruption forecasting (Chapter 1). Conversely, in an effort to define further monitoring parameters as input data for short-term eruption forecasting in probabilistic models (as for example, the Bayesian Event Tree for eruption forecasting -BET_EF-), we analyze some characteristics of typical seismic activity recorded in active volcanoes; in particular, we use some methodologies that may be applied to analyze long-period (LP) events (Chapter 2) and volcano-tectonic (VT) seismic swarms (Chapter 3); our analysis in general are oriented toward the tracking of phenomena that can provide information about magmatic processes. Finally, we discuss some possible ways to integrate the results presented in Chapters 1 (for long-term EF), 2 and 3 (for short-term EF) in the BET_EF model (Chapter 4).
Resumo:
Hydrothermal fluids are a fundamental resource for understanding and monitoring volcanic and non-volcanic systems. This thesis is focused on the study of hydrothermal system through numerical modeling with the geothermal simulator TOUGH2. Several simulations are presented, and geophysical and geochemical observables, arising from fluids circulation, are analyzed in detail throughout the thesis. In a volcanic setting, fluids feeding fumaroles and hot spring may play a key role in the hazard evaluation. The evolution of the fluids circulation is caused by a strong interaction between magmatic and hydrothermal systems. A simultaneous analysis of different geophysical and geochemical observables is a sound approach for interpreting monitored data and to infer a consistent conceptual model. Analyzed observables are ground displacement, gravity changes, electrical conductivity, amount, composition and temperature of the emitted gases at surface, and extent of degassing area. Results highlight the different temporal response of the considered observables, as well as the different radial pattern of variation. However, magnitude, temporal response and radial pattern of these signals depend not only on the evolution of fluid circulation, but a main role is played by the considered rock properties. Numerical simulations highlight differences that arise from the assumption of different permeabilities, for both homogeneous and heterogeneous systems. Rock properties affect hydrothermal fluid circulation, controlling both the range of variation and the temporal evolution of the observable signals. Low temperature fumaroles and low discharge rate may be affected by atmospheric conditions. Detailed parametric simulations were performed, aimed to understand the effects of system properties, such as permeability and gas reservoir overpressure, on diffuse degassing when air temperature and barometric pressure changes are applied to the ground surface. Hydrothermal circulation, however, is not only a characteristic of volcanic system. Hot fluids may be involved in several mankind problems, such as studies on geothermal engineering, nuclear waste propagation in porous medium, and Geological Carbon Sequestration (GCS). The current concept for large-scale GCS is the direct injection of supercritical carbon dioxide into deep geological formations which typically contain brine. Upward displacement of such brine from deep reservoirs driven by pressure increases resulting from carbon dioxide injection may occur through abandoned wells, permeable faults or permeable channels. Brine intrusion into aquifers may degrade groundwater resources. Numerical results show that pressure rise drives dense water up to the conduits, and does not necessarily result in continuous flow. Rather, overpressure leads to new hydrostatic equilibrium if fluids are initially density stratified. If warm and salty fluid does not cool passing through the conduit, an oscillatory solution is then possible. Parameter studies delineate steady-state (static) and oscillatory solutions.
Resumo:
Can space and place foster child development, and in particular social competence and ecological literacy? If yes, how can space and place do that? This study shows that the answer to the first question is positive and then tries to explain the way space and place can make a difference. The thesis begins with the review of literature from different disciplines – child development and child psychology, education, environmental psychology, architecture and landscape architecture. Some bridges among such disciplines are created and in some cases the ideas from the different areas of research merge: thus, this is an interdisciplinary study. The interdisciplinary knowledge from these disciplines is translated into a range of design suggestions that can foster the development of social competence and ecological literacy. Using scientific knowledge from different disciplines is a way of introducing forms of evidence into the development of design criteria. However, the definition of design criteria also has to pass through the study of a series of school buildings and un-built projects: case studies can give a positive contribution to the criteria because examples and good practices can help “translating” the theoretical knowledge into design ideas and illustrations. To do that, the different case studies have to be assessed in relation to the various themes that emerged in the literature review. Finally, research by design can be used to help define the illustrated design criteria: based on all the background knowledge that has been built, the role of the architect is to provide a series of different design solutions that can give answers to the different “questions” emerged in the literature review.
Resumo:
This research seeks to review the level of knowledge achieved in interpreting the relationship between the ethnic diversity at the workplace in the public sector and the organizational performance; as well as seeks to contribute in understanding the implications of this relationship. The study commenced with investigating the academic research in the relevant area addressing the following research questions: (a) How are diversity management and organizational performance conceptualized? (b) What are the existing findings of research concerning diversity at the workplace in the public organizations and organizational performance? (c) What factors intervene the relationship between the diversity and organizational performance? Based on the findings from the review of the academic research, this study seeks to contribute in understanding the ethnic diversity – performance relationship and its mplications at the local level in the Macedonian context. The reform process in Macedonia as a multicultural society, where for many years, inter-ethnic relations have been one of the most sensitive political issues, affecting both the stability of the country and the progress, focused mainly on the implementation of the decentralization and inclusion of ethnic minorities in the decision making process. With the implementation of the Ohrid Framework Agreement workforce at the units of local self-government in Republic of Macedonia is becoming more balanced with respect to ethnic minorities, with more workforce participation than ever by Albanians, Turks, Roma and other minorities. As public organizations at local level become more diverse along ethnic lines, it makes sense to pay more attention to how different ethnic groups interact with one another at work. Thus it gives additional importance on the research question addressed in the study and gives significance of the research in a broader scope.
Resumo:
Machine learning is widely adopted to decode multi-variate neural time series, including electroencephalographic (EEG) and single-cell recordings. Recent solutions based on deep learning (DL) outperformed traditional decoders by automatically extracting relevant discriminative features from raw or minimally pre-processed signals. Convolutional Neural Networks (CNNs) have been successfully applied to EEG and are the most common DL-based EEG decoders in the state-of-the-art (SOA). However, the current research is affected by some limitations. SOA CNNs for EEG decoding usually exploit deep and heavy structures with the risk of overfitting small datasets, and architectures are often defined empirically. Furthermore, CNNs are mainly validated by designing within-subject decoders. Crucially, the automatically learned features mainly remain unexplored; conversely, interpreting these features may be of great value to use decoders also as analysis tools, highlighting neural signatures underlying the different decoded brain or behavioral states in a data-driven way. Lastly, SOA DL-based algorithms used to decode single-cell recordings rely on more complex, slower to train and less interpretable networks than CNNs, and the use of CNNs with these signals has not been investigated. This PhD research addresses the previous limitations, with reference to P300 and motor decoding from EEG, and motor decoding from single-neuron activity. CNNs were designed light, compact, and interpretable. Moreover, multiple training strategies were adopted, including transfer learning, which could reduce training times promoting the application of CNNs in practice. Furthermore, CNN-based EEG analyses were proposed to study neural features in the spatial, temporal and frequency domains, and proved to better highlight and enhance relevant neural features related to P300 and motor states than canonical EEG analyses. Remarkably, these analyses could be used, in perspective, to design novel EEG biomarkers for neurological or neurodevelopmental disorders. Lastly, CNNs were developed to decode single-neuron activity, providing a better compromise between performance and model complexity.
Resumo:
The discovery of new materials and their functions has always been a fundamental component of technological progress. Nowadays, the quest for new materials is stronger than ever: sustainability, medicine, robotics and electronics are all key assets which depend on the ability to create specifically tailored materials. However, designing materials with desired properties is a difficult task, and the complexity of the discipline makes it difficult to identify general criteria. While scientists developed a set of best practices (often based on experience and expertise), this is still a trial-and-error process. This becomes even more complex when dealing with advanced functional materials. Their properties depend on structural and morphological features, which in turn depend on fabrication procedures and environment, and subtle alterations leads to dramatically different results. Because of this, materials modeling and design is one of the most prolific research fields. Many techniques and instruments are continuously developed to enable new possibilities, both in the experimental and computational realms. Scientists strive to enforce cutting-edge technologies in order to make progress. However, the field is strongly affected by unorganized file management, proliferation of custom data formats and storage procedures, both in experimental and computational research. Results are difficult to find, interpret and re-use, and a huge amount of time is spent interpreting and re-organizing data. This also strongly limit the application of data-driven and machine learning techniques. This work introduces possible solutions to the problems described above. Specifically, it talks about developing features for specific classes of advanced materials and use them to train machine learning models and accelerate computational predictions for molecular compounds; developing method for organizing non homogeneous materials data; automate the process of using devices simulations to train machine learning models; dealing with scattered experimental data and use them to discover new patterns.
Resumo:
Frame. Assessing the difficulty of source texts and parts thereof is important in CTIS, whether for research comparability, for didactic purposes or setting price differences in the market. In order to empirically measure it, Campbell & Hale (1999) and Campbell (2000) developed the Choice Network Analysis (CNA) framework. Basically, the CNA’s main hypothesis is that the more translation options (a group of) translators have to render a given source text stretch, the higher the difficulty of that text stretch will be. We will call this the CNA hypothesis. In a nutshell, this research project puts the CNA hypothesis to the test and studies whether it does actually measure difficulty. Data collection. Two groups of participants (n=29) of different profiles and from two universities in different countries had three translation tasks keylogged with Inputlog, and filled pre- and post-translation questionnaires. Participants translated from English (L2) into their L1s (Spanish or Italian), and worked—first in class and then at home—using their own computers, on texts ca. 800–1000 words long. Each text was translated in approximately equal halves in two 1-hour sessions, in three consecutive weeks. Only the parts translated at home were considered in the study. Results. A very different picture emerged from data than that which the CNA hypothesis might predict: there was no prevalence of disfluent task segments when there were many translation options, nor was a prevalence of fluent task segments associated to fewer translation options. Indeed, there was no correlation between the number of translation options (many and few) and behavioral fluency. Additionally, there was no correlation between pauses and both behavioral fluency and typing speed. The discussed theoretical flaws and the empirical evidence lead to the conclusion that the CNA framework does not and cannot measure text and translation difficulty.
Resumo:
The scope of the thesis is to broaden the knowledge about axially loaded pipe piles, that can play as foundations for offshore wind turbines based on jacket structures. The goal of the work was pursued by interpreting experimental data on large-scale model piles and by developing numerical tools for the prediction of their monotonic response to tensile and compressive loads to failure. The availability of experimental results on large scale model piles produced in two different campaigns at Fraunhofer IWES (Hannover, Germany) represented the reference for the whole work. Data from CPTs, blow counts during installation and load-displacement curves allowed to develop considerations on the experimental results and comparison with empirical methods from literature, such as CPT-based methods and Load Transfer methods. The understanding of soil-structure interaction mechanisms has been involved in the study in order to better assess the mechanical response of the sand with the scope to help in developing predictive tools of the experiments. A lack of information on the response of Rohsand 3152 when in contact with steel was highlighted, so the necessity of better assessing its response was fulfilled with a comprehensive campaign of interface shear test. It was found how the response of the sand to ultimate conditions evolve with the roughness of the steel, which is a precious information to take account of when attempting the prediction of a pile capacity. Parallel to this topic, the work has developed a numerical modelling procedure that was validated on the available large-scale model piles at IWES. The modelling strategy is intended to build a FE model whose mechanical properties of the sand come from an interpretation of commonly available geotechnical tests. The results of the FE model were compared with other predictive tools currently used in the engineering practice.