988 resultados para LS-DYNA (Computer file)
Resumo:
Compliant motion occurs when the manipulator position is constrained by the task geometry. Compliant motion may be produced either by a passive mechanical compliance built in to the manipulator, or by an active compliance implemented in the control servo loop. The second method, called force control, is the subject of this report. In particular, this report presents a theory of force control based on formal models of the manipulator, and the task geometry. The ideal effector is used to model the manipulator, and the task geometry is modeled by the ideal surface, which is the locus of all positions accessible to the ideal effector. Models are also defined for the goal trajectory, position control, and force control.
Resumo:
This report is concerned with the problem of achieving flexibility (additivity, modularity) and efficiency (performance, expertise) simultaneously in one AI program. It deals with the domain of elementary electronic circuit design. The proposed solution is to provide a deduction-driven problem solver with built-in-control-structure concepts. This problem solver and its knowledge base in the applicaitn areas of design and electronics are descrbed. The prgram embodying it is being used to explore the solutionof some modest problems in circuit design. It is concluded that shallow reasoning about problem-solver plans is necessary for flexibility, and can be implemented with reasonable efficiency.
Resumo:
The constraint paradigm is a model of computation in which values are deduced whenever possible, under the limitation that deductions be local in a certain sense. One may visualize a constraint 'program' as a network of devices connected by wires. Data values may flow along the wires, and computation is performed by the devices. A device computes using only locally available information (with a few exceptions), and places newly derived values on other, locally attached wires. In this way computed values are propagated. An advantage of the constraint paradigm (not unique to it) is that a single relationship can be used in more than one direction. The connections to a device are not labelled as inputs and outputs; a device will compute with whatever values are available, and produce as many new values as it can. General theorem provers are capable of such behavior, but tend to suffer from combinatorial explosion; it is not usually useful to derive all the possible consequences of a set of hypotheses. The constraint paradigm places a certain kind of limitation on the deduction process. The limitations imposed by the constraint paradigm are not the only one possible. It is argued, however, that they are restrictive enough to forestall combinatorial explosion in many interesting computational situations, yet permissive enough to allow useful computations in practical situations. Moreover, the paradigm is intuitive: It is easy to visualize the computational effects of these particular limitations, and the paradigm is a natural way of expressing programs for certain applications, in particular relationships arising in computer-aided design. A number of implementations of constraint-based programming languages are presented. A progression of ever more powerful languages is described, complete implementations are presented and design difficulties and alternatives are discussed. The goal approached, though not quite reached, is a complete programming system which will implicitly support the constraint paradigm to the same extent that LISP, say, supports automatic storage management.
Resumo:
"The Structure and Interpretation of Computer Programs" is the entry-level subject in Computer Science at the Massachusetts Institute of Technology. It is required of all students at MIT who major in Electrical Engineering or in Computer Science, as one fourth of the "common core curriculum," which also includes two subjects on circuits and linear systems and a subject on the design of digital systems. We have been involved in the development of this subject since 1978, and we have taught this material in its present form since the fall of 1980 to approximately 600 students each year. Most of these students have had little or no prior formal training in computation, although most have played with computers a bit and a few have had extensive programming or hardware design experience. Our design of this introductory Computer Science subject reflects two major concerns. First we want to establish the idea that a computer language is not just a way of getting a computer to perform operations, but rather that it is a novel formal medium for expressing ideas about methodology. Thus, programs must be written for people to read, and only incidentally for machines to execute. Secondly, we believe that the essential material to be addressed by a subject at this level, is not the syntax of particular programming language constructs, nor clever algorithms for computing particular functions of efficiently, not even the mathematical analysis of algorithms and the foundations of computing, but rather the techniques used to control the intellectual complexity of large software systems.
Resumo:
Web threats are becoming a major issue for both governments and companies. Generally, web threats increased as much as 600% during last year (WebSense, 2013). This appears to be a significant issue, since many major businesses seem to provide these services. Denial of Service (DoS) attacks are one of the most significant web threats and generally their aim is to waste the resources of the target machine (Mirkovic & Reiher, 2004). Dis-tributed Denial of Service (DDoS) attacks are typically executed from many sources and can result in large traf-fic flows. During last year 11% of DDoS attacks were over 60 Gbps (Prolexic, 2013a). The DDoS attacks are usually performed from the large botnets, which are networks of remotely controlled computers. There is an increasing effort by governments and companies to shut down the botnets (Dittrich, 2012), which has lead the attackers to look for alternative DDoS attack methods. One of the techniques to which attackers are returning to is DDoS amplification attacks. Amplification attacks use intermediate devices called amplifiers in order to amplify the attacker's traffic. This work outlines an evaluation tool and evaluates an amplification attack based on the Trivial File Transfer Proto-col (TFTP). This attack could have amplification factor of approximately 60, which rates highly alongside other researched amplification attacks. This could be a substantial issue globally, due to the fact this protocol is used in approximately 599,600 publicly open TFTP servers. Mitigation methods to this threat have also been consid-ered and a variety of countermeasures are proposed. Effects of this attack on both amplifier and target were analysed based on the proposed metrics. While it has been reported that the breaching of TFTP would be possible (Schultz, 2013), this paper provides a complete methodology for the setup of the attack, and its verification.
Resumo:
This document describes two sets of benchmark problem instances for the job shop scheduling problem. Each set of instances is supplied as a compressed (zipped) archive containing a single CSV file for each problem instance using the format described in http://rollproject.org/jssp/jsspGen.pdf
Resumo:
ISBN: 3-540-76198-5 (out of print)
Resumo:
Karwath, A. King, R. Homology induction: the use of machine learning to improve sequence similarity searches. BMC Bioinformatics. 23rd April 2002. 3:11 Additional File Describes the title organims species declaration in one string [http://www.biomedcentral.com/content/supplementary/1471- 2105-3-11-S1.doc] Sponsorship: Andreas Karwath and Ross D. King were supported by the EPSRC grant GR/L62849.
Resumo:
A 5-minute clip from a 1975 video on Bath University Library has been incorporated as slide 21 of this PowerPoint. The video was produced by the author to show library and information science students details of a library with a computer-based cataloguing and circulation system. To make sure that the video clip functions correctly within the presentation, please make sure that you have downloaded both the presentation (.ppt file) and the video (.wmv file) into the same directory. To activate the video move the cursor to the middle of Slide 21 (i.e. the screen that says Bath University) and then click the mouse.
Resumo:
B. Schafer, J. Keppens and Q. Shen. Thinking with and outside the Box: Developing Computer Support for Evidence Teaching. P. Robert and M. Redmayne (Eds.), Innovations in Evidence and Proof: Integrating Theory, Research and Teaching, pp. 139-158, 2007.
Resumo:
Y. Zhu, S. Williams and R. Zwiggelaar, 'Computer technology in detection and staging of prostate carcinoma: a review', Medical Image Analysis 10 (2), 178-199 (2006)
Resumo:
R. Zwiggelaar, C.R. Bull, M.J. Mooney and S. Czarnes, 'The detection of 'soft' materials by selective energy xray transmission imaging and computer tomography', Journal of Agricultural Engineering Research 66 (3), 203-212 (1997)
Resumo:
Murphy, L. and Thomas, L. 2008. Dangers of a fixed mindset: implications of self-theories research for computer science education. In Proceedings of the 13th Annual Conference on innovation and Technology in Computer Science Education (Madrid, Spain, June 30 - July 02, 2008). ITiCSE '08. ACM, New York, NY, 271-275.
Resumo:
Communities of faith have appeared online since the inception of computer - mediated communication (CMC)and are now ubiquitous. Yet the character and legitimacy of Internet communities as ecclesial bodies is often disputed by traditional churches; and the Internet's ability to host the church as church for online Christians remains a question. This dissertation carries out a practical theological conversation between three main sources: the phenomenon of the church online; ecclesiology (especially that characteristic of Reformed communities); and communication theory. After establishing the need for this study in Chapter 1, Chapter 2 investigates the online presence of Christians and trends in their Internet use, including its history and current expressions. Chapter 3 sets out an historical overview of the Reformed Tradition, focusing on the work of John Calvin and Karl Barth, as well as more contemporary theologians. With a theological context in which to consider online churches in place, Chapter 4 introduces four theological themes prominent in both ecclesiology and CMC studies: authority; community; mediation; and embodiment. These themes constitute the primary lens through which the dissertation conducts a critical-confessional interface between communication theory and ecclesiology in the examination of CMC. Chapter 5 continues the contextualization of online churches with consideration of communication theories that impact CMC, focusing on three major communication theories: Narrative Theory; Interpretive Theory; and Speech Act Theory. Chapter 6 contains the critical conversation between ecclesiology and communication theory by correlating the aforementioned communication theories with Narrative Theology, Communities of Practice, and Theo-Drama, and applying these to the four theological themes noted above. In addition, new or anticipated developments in CMC investigated in relationship to traditional ecclesiologies and the prospect of cyber-ecclesiology. Chapter 7 offers an evaluative tool consisting of a three-step hermeneutical process that examines: 1) the history, tradition, and ecclesiology of the particular community being evaluated; 2) communication theories and the process of religious-social shaping of technology; and 3) CMC criteria for establishing the presence of a stable, interactive, and relational community. As this hermeneutical process unfolds, it holds the church at the center of the process, seeking a contextual yet faithful understanding of the church.
Resumo:
We analyzed the logs of our departmental HTTP server http://cs-www.bu.edu as well as the logs of the more popular Rolling Stones HTTP server http://www.stones.com. These servers have very different purposes; the former caters primarily to local clients, whereas the latter caters exclusively to remote clients all over the world. In both cases, our analysis showed that remote HTTP accesses were confined to a very small subset of documents. Using a validated analytical model of server popularity and file access profiles, we show that by disseminating the most popular documents on servers (proxies) closer to the clients, network traffic could be reduced considerably, while server loads are balanced. We argue that this process could be generalized so as to provide for an automated demand-based duplication of documents. We believe that such server-based information dissemination protocols will be more effective at reducing both network bandwidth and document retrieval times than client-based caching protocols [2].