944 resultados para Logic, symbolic and mathematical
Resumo:
Abstract not available
Resumo:
Gating of sensory information can be assessed using an auditory conditioning-test paradigm which measures the reduction in the auditory evoked response to a test stimulus following an initial conditioning stimulus. Recording brainwaves from specific areas of the brain using multiple electrodes is helpful in the study of the neurobiology of sensory gating. In this paper, we use such technology to investigate the role of cannabinoids in sensory gating in the CA3 region of the rat hippocampus. Our experimental results show that application of the exogenous cannabinoid agonist WIN55,212-2 can abolish sensory gating. We have developed a phenomenological model of cannabinoid dynamics incorporated within a spiking neural network model of CA3 with synaptically interacting pyramidal and basket cells. Direct numerical simulations of this model suggest that the basic mechanism for this effect can be traced to the suppression of inhibition of slow GABAB synapses. Furthermore, by working with a simpler mathematical firing rate model we are able to show the robustness of this mechanism for the abolition of sensory gating.
Resumo:
The "SNARC effect" refers to the finding that people respond faster to small numbers with the left hand and to large numbers with the right hand. This effect is often explained by hypothesizing that numbers are represented from left to right in ascending order (Mental Number Line). However, the SNARC effect may not depend on quantitative information, but on other factors such as the order in which numbers are often represented from left to right in our culture. Four experiments were performed to test this hypothesis. In the first experiment, the concept of spatial association was extended to nonnumeric mathematical symbols: the minus and plus symbols. These symbols were presented as fixation points in a spatial compatibility paradigm. The results demonstrated an opposite influence of the two symbols on the target stimulus: the minus symbol tends to favor the target presented on the left, while the plus symbol the target presented on the right, demonstrating that spatial association can emerge in the absence of a numerical context. In the last three experiments, the relationship between quantity and order was evaluated using normal numbers and mirror numbers. Although mirror numbers denote quantity, they are not encountered in a left-to-right spatial organization. In Experiments 1 and 2, participants performed a magnitude classification task with mirror and normal numbers presented together (Experiment 1) or separately (Experiment 2). In Experiment 3, participants performed a new task in which quantity information processing was not required: the mirror judgment task. The results show that participants access the quantity of both normal and mirror numbers, but only the normal numbers are spatially organized from left to right. In addition, the physical similarity between the numbers, used as a predictor variable in the last three experiments, showed that the physical characteristics of numbers influenced participants' reaction times.
Resumo:
I will argue that the doctrine of eternal recurrence of the same no better interprets cosmology than pink elephants interpret zoology. I will also argue that the eternal-reiurn-of-the-same doctrine as what Magnus calls "existential imperative" is without possibility of application and thus futile. To facilitate those arguments, the validity of the doctrine of the eternal recurrence of the same will be tested under distinct rubrics. Although each rubric will stand alone, one per chapter, as an evaluation of some specific aspect of eternal recurrence, the rubric sequence has been selected to accommodate the identification of what I shall be calling logic abridgments. The conclusions to be extracted from each rubric are grouped under the heading CONCLUSION and appear immediately following rubric ten. Then, or if, at the end of a rubric a reader is inclined to wonder which rubric or topic is next, and why, the answer can be found at the top of the following page. The question is usually answered in the very first sentence, but always answered in the first paragraph. The first rubric has been placed in order by chronological entitlement in that it deals with the evolution of the idea of eternal recurrence from the time of the ancient Greeks to Nietzsche's August, 1881 inspiration. This much-recommended technique is also known as starting at the beginning. Rubric 1 also deals with 20th. Century philosophers' assessments of the relationship between Nietzsche and ancient Greek thought. The only experience of E-R, Zarathustra's mountain vision, is second only because it sets the scene alluded to in following rubrics. The third rubric explores .ii?.ih T jc,i -I'w Nietzsche's evaluation of rationality so that his thought processes will be understood appropriately. The actual mechanism of E-R is tested in rubric four...The scientific proof Nietzsche assembled in support of E-R is assessed by contemporary philosophers in rubric five. E-R's function as an ethical imperative is debated in rubrics six and seven.. .The extent to which E-R fulfills its purpose in overcoming nihilism is measured against the comfort assured by major world religions in rubric eight. Whether E-R also serves as a redemption for revenge is questioned in rubric nine. Rubric ten assures that E-R refers to return of the identically same and not merely the similar. In addition to assemblage and evaluation of all ten rubrics, at the end of each rubric a brief recapitulation of its principal points concludes the chapter. In this essay I will assess the theoretical conditions under which the doctrine cannot be applicable and will show what contradictions and inconsistencies follow if the doctrine is taken to be operable. Harold Alderman in his book Nietzsche's Gift wrote, the "doctrine of eternal recurrence gives us a problem not in Platonic cosmology, but in Socratic selfreflection." ^ I will illustrate that the recurrence doctrine's cosmogony is unworkable and that if it were workable, it would negate self-reflection on the grounds that selfreflection cannot find its cause in eternal recurrence of the same. Thus, when the cosmology is shown to be impossible, any expected ensuing results or benefits will be rendered also impossible. The so-called "heaviest burden" will be exposed as complex, engrossing "what if speculations deserving no linkings to reality. To identify ^Alderman p. 84 abridgments of logic, contradictions and inconsistencies in Nietzsche's doctrine of eternal recurrence of the same, I. will examine the subject under the following schedule. In Chapter 1 the ancient origins of recurrence theories will be introduced. ..This chapter is intended to establish the boundaries within which the subsequent chapters, except Chapter 10, will be confined. Chapter 2, Zarathustra's vision of E-R, assesses the sections of Thus Spoke Zarathustra in which the phenomenon of recurrence of the same is reported. ..Nihilism as a psychological difficulty is introduced in this rubric, but that subject will be studied in detail in Chapter 8. In Chapter 2 the symbols of eternal recurrence of the same will be considered. Whether the recurrence image should be of a closed ring or as a coil will be of significance in many sections of my essay. I will argue that neither symbolic configuration can accommodate Nietzsche's supposed intention. Chapter 3 defends the description of E-R given by Zarathustra. Chapter 4, the cosmological mechanics of E-R, speculates on the seriousness with which Nietzsche might have intended the doctrine of eternal recurrence to be taken. My essay reports, and then assesses, the argument of those who suppose the doctrine to have been merely exploratory musings by Nietzsche on cosmological hypotheses...The cosmogony of E-R is examined. In Chapter 5, cosmological proofs tested, the proofs for Nietzsche's doctrine of return of the same are evaluated. This chapter features the position taken by Martin ' Heidegger. My essay suggests that while Heidegger's argument that recurrence of the same is a genuine cosmic agenda is admirable, it is not at all persuasive. Chapter 6, E-R is an ethical imperative, is in essence the reporting of a debate between two scholars regarding the possibility of an imperative in the doctrine of recurrence. Their debate polarizes the arguments I intend to develop. Chapter 7, does E-R of the same preclude alteration of attitudes, is a continuation of the debate presented in Chapter 6 with the focus shifted to the psychological from the cosmological aspects of eternal recurrence of the same. Chapter 8, Can E-R Overcome Nihilism?, is divided into two parts. In the first, nihilism as it applies to Nietzsche's theory is discussed. ..In part 2, the broader consequences, sources and definitions of nihilism are outlined. My essay argues that Nietzsche's doctrine is more nihilistic than are the world's major religions. Chapter 9, Is E-R a redemption for revenge?, examines the suggestion extracted from Thus Spoke Zarathustra that the doctrine of eternal recurrence is intended, among other purposes, as a redemption for mankind from the destructiveness of revenge. Chapter 10, E-R of the similar refuted, analyses a position that an element of chance can influence the doctrine of recurrence. This view appears to allow, not for recurrence of the same, but recurrence of the similar. A summary will recount briefly the various significant logic abridgments, contradictions, and inconsistencies associated with Nietzsche's doctrine of eternal recurrence of the same. In the 'conclusion' section of my essay my own opinions and observations will be assembled from the body of the essay.
Resumo:
Irregular computations pose sorne of the most interesting and challenging problems in automatic parallelization. Irregularity appears in certain kinds of numerical problems and is pervasive in symbolic applications. Such computations often use dynamic data structures, which make heavy use of pointers. This complicates all the steps of a parallelizing compiler, from independence detection to task partitioning and placement. Starting in the mid 80s there has been significant progress in the development of parallelizing compilers for logic programming (and more recently, constraint programming) resulting in quite capable parallelizers. The typical applications of these paradigms frequently involve irregular computations, and make heavy use of dynamic data structures with pointers, since logical variables represent in practice a well-behaved form of pointers. This arguably makes the techniques used in these compilers potentially interesting. In this paper, we introduce in a tutoríal way, sorne of the problems faced by parallelizing compilers for logic and constraint programs and provide pointers to sorne of the significant progress made in the area. In particular, this work has resulted in a series of achievements in the areas of inter-procedural pointer aliasing analysis for independence detection, cost models and cost analysis, cactus-stack memory management, techniques for managing speculative and irregular computations through task granularity control and dynamic task allocation such as work-stealing schedulers), etc.
Resumo:
Irregular computations pose some of the most interesting and challenging problems in automatic parallelization. Irregularity appears in certain kinds of numerical problems and is pervasive in symbolic applications. Such computations often use dynamic data structures which make heavy use of pointers. This complicates all the steps of a parallelizing compiler, from independence detection to task partitioning and placement. In the past decade there has been significant progress in the development of parallelizing compilers for logic programming and, more recently, constraint programming. The typical applications of these paradigms frequently involve irregular computations, which arguably makes the techniques used in these compilers potentially interesting. In this paper we introduce in a tutorial way some of the problems faced by parallelizing compilers for logic and constraint programs. These include the need for inter-procedural pointer aliasing analysis for independence detection and having to manage speculative and irregular computations through task granularity control and dynamic task allocation. We also provide pointers to some of the progress made in these áreas. In the associated talk we demónstrate representatives of several generations of these parallelizing compilers.
Resumo:
In this thesis we discuss in what ways computational logic (CL) and data science (DS) can jointly contribute to the management of knowledge within the scope of modern and future artificial intelligence (AI), and how technically-sound software technologies can be realised along the path. An agent-oriented mindset permeates the whole discussion, by stressing pivotal role of autonomous agents in exploiting both means to reach higher degrees of intelligence. Accordingly, the goals of this thesis are manifold. First, we elicit the analogies and differences among CL and DS, hence looking for possible synergies and complementarities along 4 major knowledge-related dimensions, namely representation, acquisition (a.k.a. learning), inference (a.k.a. reasoning), and explanation. In this regard, we propose a conceptual framework through which bridges these disciplines can be described and designed. We then survey the current state of the art of AI technologies, w.r.t. their capability to support bridging CL and DS in practice. After detecting lacks and opportunities, we propose the notion of logic ecosystem as the new conceptual, architectural, and technological solution supporting the incremental integration of symbolic and sub-symbolic AI. Finally, we discuss how our notion of logic ecosys- tem can be reified into actual software technology and extended towards many DS-related directions.
Resumo:
This paper presents the multi-threading and internet message communication capabilities of Qu-Prolog. Message addresses are symbolic and the communications package provides high-level support that completely hides details of IP addresses and port numbers as well as the underlying TCP/IP transport layer. The combination of the multi-threads and the high level inter-thread message communications provide simple, powerful support for implementing internet distributed intelligent applications.
Resumo:
Programming and mathematics are core areas of computer science (CS) and consequently also important parts of CS education. Introductory instruction in these two topics is, however, not without problems. Studies show that CS students find programming difficult to learn and that teaching mathematical topics to CS novices is challenging. One reason for the latter is the disconnection between mathematics and programming found in many CS curricula, which results in students not seeing the relevance of the subject for their studies. In addition, reports indicate that students' mathematical capability and maturity levels are dropping. The challenges faced when teaching mathematics and programming at CS departments can also be traced back to gaps in students' prior education. In Finland the high school curriculum does not include CS as a subject; instead, focus is on learning to use the computer and its applications as tools. Similarly, many of the mathematics courses emphasize application of formulas, while logic, formalisms and proofs, which are important in CS, are avoided. Consequently, high school graduates are not well prepared for studies in CS. Motivated by these challenges, the goal of the present work is to describe new approaches to teaching mathematics and programming aimed at addressing these issues: Structured derivations is a logic-based approach to teaching mathematics, where formalisms and justifications are made explicit. The aim is to help students become better at communicating their reasoning using mathematical language and logical notation at the same time as they become more confident with formalisms. The Python programming language was originally designed with education in mind, and has a simple syntax compared to many other popular languages. The aim of using it in instruction is to address algorithms and their implementation in a way that allows focus to be put on learning algorithmic thinking and programming instead of on learning a complex syntax. Invariant based programming is a diagrammatic approach to developing programs that are correct by construction. The approach is based on elementary propositional and predicate logic, and makes explicit the underlying mathematical foundations of programming. The aim is also to show how mathematics in general, and logic in particular, can be used to create better programs.
Resumo:
Genetic algorithms (GAs) have been introduced into site layout planning as reported in a number of studies. In these studies, the objective functions were defined so as to employ the GAs in searching for the optimal site layout. However, few studies have been carried out to investigate the actual closeness of relationships between site facilities; it is these relationships that ultimately govern the site layout. This study has determined that the underlying factors of site layout planning for medium-size projects include work flow, personnel flow, safety and environment, and personal preferences. By finding the weightings on these factors and the corresponding closeness indices between each facility, a closeness relationship has been deduced. Two contemporary mathematical approaches - fuzzy logic theory and an entropy measure - were adopted in finding these results in order to minimize the uncertainty and vagueness of the collected data and improve the quality of the information. GAs were then applied to searching for the optimal site layout in a medium-size government project using the GeneHunter software. The objective function involved minimizing the total travel distance. An optimal layout was obtained within a short time. This reveals that the application of GA to site layout planning is highly promising and efficient.
Resumo:
In this work, we present an implementation of quantum logic gates and algorithms in a three effective qubits system, represented by a (I = 7/2) NMR quadrupolar nuclei. To implement these protocols we have used the strong modulating pulses (SMP) and the various stages of each implementation were verified by quantum state tomography (QST). The results for the computational base states, Toffolli logic gates, and Deutsch-Jozsa and Grover algorithms are presented here. Also, we discuss the difficulties and advantages of implementing such protocols using the SMP technique in quadrupolar systems.
Resumo:
Interaction protocols establish how different computational entities can interact with each other. The interaction can be finalized to the exchange of data, as in 'communication protocols', or can be oriented to achieve some result, as in 'application protocols'. Moreover, with the increasing complexity of modern distributed systems, protocols are used also to control such a complexity, and to ensure that the system as a whole evolves with certain features. However, the extensive use of protocols has raised some issues, from the language for specifying them to the several verification aspects. Computational Logic provides models, languages and tools that can be effectively adopted to address such issues: its declarative nature can be exploited for a protocol specification language, while its operational counterpart can be used to reason upon such specifications. In this thesis we propose a proof-theoretic framework, called SCIFF, together with its extensions. SCIFF is based on Abductive Logic Programming, and provides a formal specification language with a clear declarative semantics (based on abduction). The operational counterpart is given by a proof procedure, that allows to reason upon the specifications and to test the conformance of given interactions w.r.t. a defined protocol. Moreover, by suitably adapting the SCIFF Framework, we propose solutions for addressing (1) the protocol properties verification (g-SCIFF Framework), and (2) the a-priori conformance verification of peers w.r.t. the given protocol (AlLoWS Framework). We introduce also an agent based architecture, the SCIFF Agent Platform, where the same protocol specification can be used to program and to ease the implementation task of the interacting peers.
Resumo:
The means through which the nervous system perceives its environment is one of the most fascinating questions in contemporary science. Our endeavors to comprehend the principles of neural science provide an instance of how biological processes may inspire novel methods in mathematical modeling and engineering. The application ofmathematical models towards understanding neural signals and systems represents a vibrant field of research that has spanned over half a century. During this period, multiple approaches to neuronal modeling have been adopted, and each approach is adept at elucidating a specific aspect of nervous system function. Thus while bio-physical models have strived to comprehend the dynamics of actual physical processes occurring within a nerve cell, the phenomenological approach has conceived models that relate the ionic properties of nerve cells to transitions in neural activity. Further-more, the field of neural networks has endeavored to explore how distributed parallel processing systems may become capable of storing memory. Through this project, we strive to explore how some of the insights gained from biophysical neuronal modeling may be incorporated within the field of neural net-works. We specifically study the capabilities of a simple neural model, the Resonate-and-Fire (RAF) neuron, whose derivation is inspired by biophysical neural modeling. While reflecting further biological plausibility, the RAF neuron is also analytically tractable, and thus may be implemented within neural networks. In the following thesis, we provide a brief overview of the different approaches that have been adopted towards comprehending the properties of nerve cells, along with the framework under which our specific neuron model relates to the field of neuronal modeling. Subsequently, we explore some of the time-dependent neurocomputational capabilities of the RAF neuron, and we utilize the model to classify logic gates, and solve the classic XOR problem. Finally we explore how the resonate-and-fire neuron may be implemented within neural networks, and how such a network could be adapted through the temporal backpropagation algorithm.