14 resultados para Computer-generated stimuli

em Massachusetts Institute of Technology


Relevância:

90.00% 90.00%

Publicador:

Resumo:

We discuss a variety of object recognition experiments in which human subjects were presented with realistically rendered images of computer-generated three-dimensional objects, with tight control over stimulus shape, surface properties, illumination, and viewpoint, as well as subjects' prior exposure to the stimulus objects. In all experiments recognition performance was: (1) consistently viewpoint dependent; (2) only partially aided by binocular stereo and other depth information, (3) specific to viewpoints that were familiar; (4) systematically disrupted by rotation in depth more than by deforming the two-dimensional images of the stimuli. These results are consistent with recently advanced computational theories of recognition based on view interpolation.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

By enhancing a real scene with computer generated objects, Augmented Reality (AR), has proven itself as a valuable Human-Computer Interface (HCI) in numerous application areas such as medical, military, entertainment and manufacturing. It enables higher performance of on-site tasks with seamless presentation of up-to-date, task-related information to the users during the operation. AR has potentials in design because the current interface provided by Computer-aided Design (CAD) packages is less intuitive and reports show that the presence of physical objects help design thinking and communication. This research explores the use of AR to improve the efficiency of a design process, specifically in mechanical design.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This report examines why women pursue careers in computer science and related fields far less frequently than men do. In 1990, only 13% of PhDs in computer science went to women, and only 7.8% of computer science professors were female. Causes include the different ways in which boys and girls are raised, the stereotypes of female engineers, subtle biases that females face, problems resulting from working in predominantly male environments, and sexual biases in language. A theme of the report is that women's underrepresentation is not primarily due to direct discrimination but to subconscious behavior that perpetuates the status quo.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A revolution\0\0\0 in earthmoving, a $100 billion industry, can be achieved with three components: the GPS location system, sensors and computers in bulldozers, and SITE CONTROLLER, a central computer system that maintains design data and directs operations. The first two components are widely available; I built SITE CONTROLLER to complete the triangle and describe it here. SITE CONTROLLER assists civil engineers in the design, estimation, and construction of earthworks, including hazardous waste site remediation. The core of SITE CONTROLLER is a site modelling system that represents existing and prospective terrain shapes, roads, hydrology, etc. Around this core are analysis, simulation, and vehicle control tools. Integrating these modules into one program enables civil engineers and contractors to use a single interface and database throughout the life of a project.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Segmentation of medical imagery is a challenging problem due to the complexity of the images, as well as to the absence of models of the anatomy that fully capture the possible deformations in each structure. Brain tissue is a particularly complex structure, and its segmentation is an important step for studies in temporal change detection of morphology, as well as for 3D visualization in surgical planning. In this paper, we present a method for segmentation of brain tissue from magnetic resonance images that is a combination of three existing techniques from the Computer Vision literature: EM segmentation, binary morphology, and active contour models. Each of these techniques has been customized for the problem of brain tissue segmentation in a way that the resultant method is more robust than its components. Finally, we present the results of a parallel implementation of this method on IBM's supercomputer Power Visualization System for a database of 20 brain scans each with 256x256x124 voxels and validate those against segmentations generated by neuroanatomy experts.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A method is presented for the visual analysis of objects by computer. It is particularly well suited for opaque objects with smoothly curved surfaces. The method extracts information about the object's surface properties, including measures of its specularity, texture, and regularity. It also aids in determining the object's shape. The application of this method to a simple recognition task ??e recognition of fruit ?? discussed. The results on a more complex smoothly curved object, a human face, are also considered.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper describes a system for the computer understanding of English. The system answers questions, executes commands, and accepts information in normal English dialog. It uses semantic information and context to understand discourse and to disambiguate sentences. It combines a complete syntactic analysis of each sentence with a "heuristic understander" which uses different kinds of information about a sentence, other parts of the discourse, and general information about the world in deciding what the sentence means. It is based on the belief that a computer cannot deal reasonably with language unless it can "understand" the subject it is discussing. The program is given a detailed model of the knowledge needed by a simple robot having only a hand and an eye. We can give it instructions to manipulate toy objects, interrogate it about the scene, and give it information it will use in deduction. In addition to knowing the properties of toy objects, the program has a simple model of its own mentality. It can remember and discuss its plans and actions as well as carry them out. It enters into a dialog with a person, responding to English sentences with actions and English replies, and asking for clarification when its heuristic programs cannot understand a sentence through use of context and physical knowledge.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The furious pace of Moore's Law is driving computer architecture into a realm where the the speed of light is the dominant factor in system latencies. The number of clock cycles to span a chip are increasing, while the number of bits that can be accessed within a clock cycle is decreasing. Hence, it is becoming more difficult to hide latency. One alternative solution is to reduce latency by migrating threads and data, but the overhead of existing implementations has previously made migration an unserviceable solution so far. I present an architecture, implementation, and mechanisms that reduces the overhead of migration to the point where migration is a viable supplement to other latency hiding mechanisms, such as multithreading. The architecture is abstract, and presents programmers with a simple, uniform fine-grained multithreaded parallel programming model with implicit memory management. In other words, the spatial nature and implementation details (such as the number of processors) of a parallel machine are entirely hidden from the programmer. Compiler writers are encouraged to devise programming languages for the machine that guide a programmer to express their ideas in terms of objects, since objects exhibit an inherent physical locality of data and code. The machine implementation can then leverage this locality to automatically distribute data and threads across the physical machine by using a set of high performance migration mechanisms. An implementation of this architecture could migrate a null thread in 66 cycles -- over a factor of 1000 improvement over previous work. Performance also scales well; the time required to move a typical thread is only 4 to 5 times that of a null thread. Data migration performance is similar, and scales linearly with data block size. Since the performance of the migration mechanism is on par with that of an L2 cache, the implementation simulated in my work has no data caches and relies instead on multithreading and the migration mechanism to hide and reduce access latencies.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The goal of the work reported here is to capture the commonsense knowledge of non-expert human contributors. Achieving this goal will enable more intelligent human-computer interfaces and pave the way for computers to reason about our world. In the domain of natural language processing, it will provide the world knowledge much needed for semantic processing of natural language. To acquire knowledge from contributors not trained in knowledge engineering, I take the following four steps: (i) develop a knowledge representation (KR) model for simple assertions in natural language, (ii) introduce cumulative analogy, a class of nearest-neighbor based analogical reasoning algorithms over this representation, (iii) argue that cumulative analogy is well suited for knowledge acquisition (KA) based on a theoretical analysis of effectiveness of KA with this approach, and (iv) test the KR model and the effectiveness of the cumulative analogy algorithms empirically. To investigate effectiveness of cumulative analogy for KA empirically, Learner, an open source system for KA by cumulative analogy has been implemented, deployed, and evaluated. (The site "1001 Questions," is available at http://teach-computers.org/learner.html). Learner acquires assertion-level knowledge by constructing shallow semantic analogies between a KA topic and its nearest neighbors and posing these analogies as natural language questions to human contributors. Suppose, for example, that based on the knowledge about "newspapers" already present in the knowledge base, Learner judges "newspaper" to be similar to "book" and "magazine." Further suppose that assertions "books contain information" and "magazines contain information" are also already in the knowledge base. Then Learner will use cumulative analogy from the similar topics to ask humans whether "newspapers contain information." Because similarity between topics is computed based on what is already known about them, Learner exhibits bootstrapping behavior --- the quality of its questions improves as it gathers more knowledge. By summing evidence for and against posing any given question, Learner also exhibits noise tolerance, limiting the effect of incorrect similarities. The KA power of shallow semantic analogy from nearest neighbors is one of the main findings of this thesis. I perform an analysis of commonsense knowledge collected by another research effort that did not rely on analogical reasoning and demonstrate that indeed there is sufficient amount of correlation in the knowledge base to motivate using cumulative analogy from nearest neighbors as a KA method. Empirically, evaluating the percentages of questions answered affirmatively, negatively and judged to be nonsensical in the cumulative analogy case compares favorably with the baseline, no-similarity case that relies on random objects rather than nearest neighbors. Of the questions generated by cumulative analogy, contributors answered 45% affirmatively, 28% negatively and marked 13% as nonsensical; in the control, no-similarity case 8% of questions were answered affirmatively, 60% negatively and 26% were marked as nonsensical.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Human object recognition is generally considered to tolerate changes of the stimulus position in the visual field. A number of recent studies, however, have cast doubt on the completeness of translation invariance. In a new series of experiments we tried to investigate whether positional specificity of short-term memory is a general property of visual perception. We tested same/different discrimination of computer graphics models that were displayed at the same or at different locations of the visual field, and found complete translation invariance, regardless of the similarity of the animals and irrespective of direction and size of the displacement (Exp. 1 and 2). Decisions were strongly biased towards same decisions if stimuli appeared at a constant location, while after translation subjects displayed a tendency towards different decisions. Even if the spatial order of animal limbs was randomized ("scrambled animals"), no deteriorating effect of shifts in the field of view could be detected (Exp. 3). However, if the influence of single features was reduced (Exp. 4 and 5) small but significant effects of translation could be obtained. Under conditions that do not reveal an influence of translation, rotation in depth strongly interferes with recognition (Exp. 6). Changes of stimulus size did not reduce performance (Exp. 7). Tolerance to these object transformations seems to rely on different brain mechanisms, with translation and scale invariance being achieved in principle, while rotation invariance is not.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Well-defined, water-soluble, pH and temperature stimuli-responsive [60]fullerene (C₆₀) containing ampholytic block copolymer of poly((methacrylic acid)-block-(2-(dimethylamino)ethyl methacrylate))-block–C₆₀ (P(MAA-b-DMAEMA)-b-C₆₀) was synthesized by the atom transfer radical polymerization (ATRP) technique. The self-assembly behaviour of the C₆₀ containing polyampholyte in aqueous solution was characterized by dynamic light scattering (DLS), and transmission electron microscopy. This amphiphilic mono-C₆₀ end-capped block copolymer shows enhanced solubility in aqueous medium at room and elevated temperatures and at low and high pH but phase-separates at intermediate pH of between 5.4 and 8.8. The self assembly of the copolymer is different from that of P(MAA-b-DMAEMA). Examination of the association behavior using DLS revealed the co-existence of unimers and aggregates at low pH at all temperatures studied, with the association being driven by the balance of hydrophobic and electrostatic interactions. Unimers and aggregates of different microstructures are also observed at high pH and at temperatures below the lower critical solution temperature (LCST) of PDMAEMA. At high pH and at temperatures above the LCST of PDMAEMA, the formation of micelles and aggregates co-existing in solution is driven by the combination of hydrophobic, electrostatic, and charge-transfer interactions.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Amphiphilic polymers are a class of polymers that self-assemble into different types of microstructure, depending on the solvent environment and external stimuli. Self assembly structures can exist in many different forms, such as spherical micelles, rod-like micelles, bi-layers, vesicles, bi-continuous structure etc. Most biological systems are basically comprised of many of these organised structures arranged in an intelligent manner, which impart functions and life to the system. We have adopted the atom transfer radical polymerization (ATRP) technique to synthesize various types of block copolymer systems that self-assemble into different microstructure when subject to an external stimuli, such as pH or temperature. The systems that we have studied are: (1) pH responsive fullerene (C60) containing poly(methacrylic acid) (PMAA-b-C60); (2) pH and temperature responsive fullerene containing poly[2-(dimethylamino)ethyl methacrylate] (C₆₀-b-PDMAEMA); (3) other responsive water-soluble fullerene systems. By varying temperature, pH and salt concentration, different types microstructure can be produced. In the presence of inorganic salts, fractal patterns at nano- to microscopic dimension were observed for negatively charged PMAA-b-C60, while such structure was not observed for positively charged PDMAEMA-b-C60. We demonstrated that negatively charged fullerene containing polymeric systems can serve as excellent nano-templates for the controlled growth of inorganic crystals at the nano- to micrometer length scale and the possible mechanism was proposed. The physical properties and the characteristics of their self-assembly properties will be discussed, and their implications to chemical and biomedical applications will be highlighted.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A targeted, stimuli-responsive, polymeric drug delivery vehicle is being developed in our lab to help alleviate severe side-effects caused by narrow therapeutic window drugs. Targeting specific cell types or organs via proteins, specifically, lectin-mediated targeting holds potential due to the high specificity and affinity of receptor-ligand interactions, rapid internalization, and relative ease of processing. Dextran, a commercially available, biodegradable polymer has been conjugated to doxorubicin and galactosamine to target hepatocytes in a three-step, one-pot synthesis. The loading of doxorubicin and galactose on the conjugates was determined by absorbance at 485 nm and elemental analysis, respectively. Conjugation efficiency based on the amount loaded of each reactant varies from 20% to 50% for doxorubicin and from 2% to 20% for galactosamine. Doxorubicin has also been attached to dextran through an acid-labile hydrazide bond. Doxorubicin acts by intercalating with DNA in the nuclei of cells. The fluorescence of doxorubicin is quenched when it binds to DNA. This allows a fluorescence-based cell-free assay to evaluate the efficacy of the polymer conjugates where we measure the fluorescence of doxorubicin and the conjugates in increasing concentrations of calf thymus DNA. Fluorescence quenching indicates that our conjugates can bind to DNA. The degree of binding increases with polymer molecular weight and substitution of doxorubicin. In cell culture experiments with hepatocytes, the relative uptake of polymer conjugates was evaluated using flow cytometry, and the killing efficiency was determined using the MTT cell proliferation assay. We have found that conjugate uptake is much lower than that of free doxorubicin. Lower uptake of conjugates may increase the maximum dose of drug tolerated by the body. Also, non-galactosylated conjugate uptake is lower than that of the galactosylated conjugate. Microscopy indicates that doxorubicin localizes almost exclusively at the nucleus, whereas the conjugates are present throughout the cell. Doxorubicin linked to dextran through a hydrazide bond was used to achieve improved killing efficiency. Following uptake, the doxorubicin dissociates from the polymer in an endosomal compartment and diffuses to the nucleus. The LC₅₀ of covalently linked doxorubicin is 7.4 μg/mL, whereas that of hydrazide linked doxorubicin is 4.4 μg/mL.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Memory errors are a common cause of incorrect software execution and security vulnerabilities. We have developed two new techniques that help software continue to execute successfully through memory errors: failure-oblivious computing and boundless memory blocks. The foundation of both techniques is a compiler that generates code that checks accesses via pointers to detect out of bounds accesses. Instead of terminating or throwing an exception, the generated code takes another action that keeps the program executing without memory corruption. Failure-oblivious code simply discards invalid writes and manufactures values to return for invalid reads, enabling the program to continue its normal execution path. Code that implements boundless memory blocks stores invalid writes away in a hash table to return as the values for corresponding out of bounds reads. he net effect is to (conceptually) give each allocated memory block unbounded size and to eliminate out of bounds accesses as a programming error. We have implemented both techniques and acquired several widely used open source servers (Apache, Sendmail, Pine, Mutt, and Midnight Commander).With standard compilers, all of these servers are vulnerable to buffer overflow attacks as documented at security tracking web sites. Both failure-oblivious computing and boundless memory blocks eliminate these security vulnerabilities (as well as other memory errors). Our results show that our compiler enables the servers to execute successfully through buffer overflow attacks to continue to correctly service user requests without security vulnerabilities.