882 resultados para Conventional Figurative Language Theory
Resumo:
Cette thèse prend pour objet le nouage entre l’autoréflexivité et la mise en forme esthétique dans la poésie de Christophe Tarkos, produite dans les années 1990. Elle met en lumière les rapports entre l’élaboration d’une théorie du langage au sein de cette œuvre poétique et ses visées esthétiques totalisantes. Il s’agit d’identifier les principes générateurs de la théorie et de fournir une analyse de ses fondements qui s’ancrent dans la crise de la représentation moderne commençant dans la deuxième moitié du dix-neuvième siècle. Les motifs de la crise revisités par Tarkos inscrivent sa poésie dans une historicité, et notre thèse tente d’interpréter cette actualisation dans une œuvre qui donne forme au monde et à la mémoire individuelle. L’hypothèse qui chapeaute notre étude est que la théorie du langage favorise l’intelligibilité de l’œuvre totalisante en lui offrant un support réflexif. Notre thèse, qui privilégie une méthode fondée sur l’analyse des textes, se divise en trois parties. La première propose une recension de la réception critique de l’œuvre, dont nous retraçons les grandes lignes d’interprétation, de Christian Prigent à Jean-Michel Espitallier. Tout en plaçant Tarkos dans le champ poétique français, cette étape nous permet de positionner notre recherche par rapport à certains lieux communs de la critique. La deuxième partie vise à étudier la théorie du langage de Tarkos à partir de ses manifestes principaux (Le Signe =, Manifeste chou, Ma langue est poétique et La poésie est une intelligence) qui révèlent plusieurs principes, pouvoirs et limites de la langue et de la poésie. Afin de montrer la spécificité du concept de la « pâte-mot » de Tarkos, nous l’étudions dans un dialogue avec la figure de la « pâte » langagière chez la poète française Danielle Collobert. La troisième partie propose une étude de la volonté et de l’esthétique totalisantes dans l’œuvre de Tarkos, qui cherche à donner forme au réel. En effet, la poésie répond à l’excès du réel par diverses stratégies. Tout en voulant représenter son caractère débordant par une énonciation logorrhéique ou en usant de procédés comme celui de la répétition, elle cherche à le maîtriser dans des formes textuelles stables comme des fragments de prose « carrés » (Carrés, Caisses), dans des listes énumératives (Anachronisme) ou dans des réseaux d’images. La volonté totalisante chez Tarkos semble également prendre origine dans un sentiment d’urgence qui concerne, en dernière instance, une bataille contre la finitude.
Resumo:
The basic concepts of digital signal processing are taught to the students in engineering and science. The focus of the course is on linear, time invariant systems. The question as to what happens when the system is governed by a quadratic or cubic equation remains unanswered in the vast majority of literature on signal processing. Light has been shed on this problem when John V Mathews and Giovanni L Sicuranza published the book Polynomial Signal Processing. This book opened up an unseen vista of polynomial systems for signal and image processing. The book presented the theory and implementations of both adaptive and non-adaptive FIR and IIR quadratic systems which offer improved performance than conventional linear systems. The theory of quadratic systems presents a pristine and virgin area of research that offers computationally intensive work. Once the area of research is selected, the next issue is the choice of the software tool to carry out the work. Conventional languages like C and C++ are easily eliminated as they are not interpreted and lack good quality plotting libraries. MATLAB is proved to be very slow and so do SCILAB and Octave. The search for a language for scientific computing that was as fast as C, but with a good quality plotting library, ended up in Python, a distant relative of LISP. It proved to be ideal for scientific computing. An account of the use of Python, its scientific computing package scipy and the plotting library pylab is given in the appendix Initially, work is focused on designing predictors that exploit the polynomial nonlinearities inherent in speech generation mechanisms. Soon, the work got diverted into medical image processing which offered more potential to exploit by the use of quadratic methods. The major focus in this area is on quadratic edge detection methods for retinal images and fingerprints as well as de-noising raw MRI signals
Resumo:
Computational models are arising is which programs are constructed by specifying large networks of very simple computational devices. Although such models can potentially make use of a massive amount of concurrency, their usefulness as a programming model for the design of complex systems will ultimately be decided by the ease in which such networks can be programmed (constructed). This thesis outlines a language for specifying computational networks. The language (AFL-1) consists of a set of primitives, ad a mechanism to group these elements into higher level structures. An implementation of this language runs on the Thinking Machines Corporation, Connection machine. Two significant examples were programmed in the language, an expert system (CIS), and a planning system (AFPLAN). These systems are explained and analyzed in terms of how they compare with similar systems written in conventional languages.
Resumo:
Free-word order languages have long posed significant problems for standard parsing algorithms. This thesis presents an implemented parser, based on Government-Binding (GB) theory, for a particular free-word order language, Warlpiri, an aboriginal language of central Australia. The words in a sentence of a free-word order language may swap about relatively freely with little effect on meaning: the permutations of a sentence mean essentially the same thing. It is assumed that this similarity in meaning is directly reflected in the syntax. The parser presented here properly processes free word order because it assigns the same syntactic structure to the permutations of a single sentence. The parser also handles fixed word order, as well as other phenomena. On the view presented here, there is no such thing as a "configurational" or "non-configurational" language. Rather, there is a spectrum of languages that are more or less ordered. The operation of this parsing system is quite different in character from that of more traditional rule-based parsing systems, e.g., context-free parsers. In this system, parsing is carried out via the construction of two different structures, one encoding precedence information and one encoding hierarchical information. This bipartite representation is the key to handling both free- and fixed-order phenomena. This thesis first presents an overview of the portion of Warlpiri that can be parsed. Following this is a description of the linguistic theory on which the parser is based. The chapter after that describes the representations and algorithms of the parser. In conclusion, the parser is compared to related work. The appendix contains a substantial list of test cases ??th grammatical and ungrammatical ??at the parser has actually processed.
Resumo:
This thesis presents the ideas underlying a computer program that takes as input a schematic of a mechanical or hydraulic power transmission system, plus specifications and a utility function, and returns catalog numbers from predefined catalogs for the optimal selection of components implementing the design. Unlike programs for designing single components or systems, the program provides the designer with a high level "language" in which to compose new designs. It then performs some of the detailed design process. The process of "compilation" is based on a formalization of quantitative inferences about hierarchically organized sets of artifacts and operating conditions. This allows the design compilation without the exhaustive enumeration of alternatives.
Resumo:
The central thesis of this report is that human language is NP-complete. That is, the process of comprehending and producing utterances is bounded above by the class NP, and below by NP-hardness. This constructive complexity thesis has two empirical consequences. The first is to predict that a linguistic theory outside NP is unnaturally powerful. The second is to predict that a linguistic theory easier than NP-hard is descriptively inadequate. To prove the lower bound, I show that the following three subproblems of language comprehension are all NP-hard: decide whether a given sound is possible sound of a given language; disambiguate a sequence of words; and compute the antecedents of pronouns. The proofs are based directly on the empirical facts of the language user's knowledge, under an appropriate idealization. Therefore, they are invariant across linguistic theories. (For this reason, no knowledge of linguistic theory is needed to understand the proofs, only knowledge of English.) To illustrate the usefulness of the upper bound, I show that two widely-accepted analyses of the language user's knowledge (of syntactic ellipsis and phonological dependencies) lead to complexity outside of NP (PSPACE-hard and Undecidable, respectively). Next, guided by the complexity proofs, I construct alternate linguisitic analyses that are strictly superior on descriptive grounds, as well as being less complex computationally (in NP). The report also presents a new framework for linguistic theorizing, that resolves important puzzles in generative linguistics, and guides the mathematical investigation of human language.
Resumo:
The goal of this article is to reveal the computational structure of modern principle-and-parameter (Chomskian) linguistic theories: what computational problems do these informal theories pose, and what is the underlying structure of those computations? To do this, I analyze the computational complexity of human language comprehension: what linguistic representation is assigned to a given sound? This problem is factored into smaller, interrelated (but independently statable) problems. For example, in order to understand a given sound, the listener must assign a phonetic form to the sound; determine the morphemes that compose the words in the sound; and calculate the linguistic antecedent of every pronoun in the utterance. I prove that these and other subproblems are all NP-hard, and that language comprehension is itself PSPACE-hard.
Resumo:
Esta concebido con un planteamiento integrador, en el que se tratan temas como el bilingüismo y el desarrollo cognitivo, programas de inmersión, la potenciación del papel de los estudiantes pertenecientes a minorías, la educación multicultural y no racista y la diversidad cultural y lingüística.
Resumo:
A procedure based on quantum molecular similarity measures (QMSM) has been used to compare electron densities obtained from conventional ab initio and density functional methodologies at their respective optimized geometries. This method has been applied to a series of small molecules which have experimentally known properties and molecular bonds of diverse degrees of ionicity and covalency. Results show that in most cases the electron densities obtained from density functional methodologies are of a similar quality than post-Hartree-Fock generalized densities. For molecules where Hartree-Fock methodology yields erroneous results, the density functional methodology is shown to yield usually more accurate densities than those provided by the second order Møller-Plesset perturbation theory
Resumo:
A comparative systematic study of the CrO2F2 compound has been performed using different conventional ab initio methodologies and density functional procedures. Two points have been analyzed: first, the accuracy of results yielded by each method under study, and second, the computational cost required to reach such results. Weighing up both aspects, density functional theory has been found to be more appropriate than the Hartree-Fock (HF) and the analyzed post-HF methods. Hence, the structural characterization and spectroscopic elucidation of the full CrO2X2 series (X=F,Cl,Br,I) has been done at this level of theory. Emphasis has been given to the unknown CrO2I2 species, and specially to the UV/visible spectra of all four compounds. Furthermore, a topological analysis in terms of charge density distributions has revealed why the valence shell electron pair repulsion model fails in predicting the molecular shape of such CrO2X2 complexes
Resumo:
Firms form consortia in order to win contracts. Once a project has been awarded to a consortium each member then concentrates on his or her own contract with the client. Therefore, consortia are marketing devices, which present the impression of teamworking, but the production process is just as fragmented as under conventional procurement methods. In this way, the consortium forms a barrier between the client and the actual construction production process. Firms form consortia, not as a simple development of normal ways of working, but because the circumstances for specific projects make it a necessary vehicle. These circumstances include projects that are too large or too complex to undertake alone or projects that require on-going services which cannot be provided by the individual firms inhouse. It is not a preferred way of working, because participants carry extra risk in the form of liability for the actions of their partners in the consortium. The behaviour of members of consortia is determined by their relative power, based on several factors, including financial commitment and ease of replacement. The level of supply chain visibility to the public sector client and to the industry is reduced by the existence of a consortium because the consortium forms an additional obstacle between the client and the firms undertaking the actual construction work. Supply chain visibility matters to the client who otherwise loses control over the process of construction or service provision, while remaining accountable for cost overruns. To overcome this separation there is a convincing argument in favour of adopting the approach put forward in the Project Partnering Contract 2000 (PPC2000) Agreement. Members of consortia do not necessarily go on to work in the same consortia again because members need to respond flexibly to opportunities as and when they arise. Decision-making processes within consortia tend to be on an ad hoc basis. Construction risk is taken by the contractor and the construction supply chain but the reputational risk is carried by all the firms associated with a consortium. There is a wide variation in the manner that consortia are formed, determined by the individual circumstances of each project; its requirements, size and complexity, and the attitude of individual project leaders. However, there are a number of close working relationships based on generic models of consortia-like arrangements for the purpose of building production, such as the Housing Corporation Guidance Notes and the PPC2000.
Resumo:
This article outlines some of the key issues involved in developing a programme of strategy training for learners of French, in listening and in writing. It highlights the theoretical perspectives and research findings on listening and writing that informed the selection of strategies to teach learners and thence the development of appropriate materials. Examples of these materials are given as well as advice regarding their use. The article concludes with suggestions for how strategy training might be incorporated into teachers' own work with learners.
Resumo:
This review highlights the importance of right hemisphere language functions for successful social communication and advances the hypothesis that the core deficit in psychosis is a failure of segregation of right from left hemisphere functions. Lesion studies of stroke patients and dichotic listening and functional imaging studies of healthy people have shown that some language functions are mediated by the right hemisphere rather than the left. These functions include discourse planning/comprehension, understanding humour, sarcasm, metaphors and indirect requests, and the generation/comprehension of emotional prosody. Behavioural evidence indicates that patients with typical schizophrenic illnesses perform poorly on tests of these functions, and aspects of these functions are disturbed in schizo-affective and affective psychoses. The higher order language functions mediated by the right hemisphere are essential to an accurate understanding of someone's communicative intent, and the deficits displayed by patients with schizophrenia may make a significant contribution to their social interaction deficits. We outline a bi-hemispheric theory of the neural basis of language that emphasizes the role of the sapiens-specific cerebral torque in determining the four-chambered nature of the human brain in relation to the origins of language and the symptoms of schizophrenia. Future studies of abnormal lateralization of left hemisphere language functions need to take account of the consequences of a failure of lateralization of language functions to the right as well as the left hemisphere.
Resumo:
Event-related brain potentials (ERP) are important neural correlates of cognitive processes. In the domain of language processing, the N400 and P600 reflect lexical-semantic integration and syntactic processing problems, respectively. We suggest an interpretation of these markers in terms of dynamical system theory and present two nonlinear dynamical models for syntactic computations where different processing strategies correspond to functionally different regions in the system's phase space.
Resumo:
This chapter compares lexical diversity of French words used by Dutch-French bilinguals, English-French bilinguals and Flemish L2 learners of French.