982 resultados para Architecture history


Relevância:

20.00% 20.00%

Publicador:

Resumo:

A new high throughput and scalable architecture for unified transform coding in H.264/AVC is proposed in this paper. Such flexible structure is capable of computing all the 4x4 and 2x2 transforms for Ultra High Definition Video (UHDV) applications (4320x7680@ 30fps) in real-time and with low hardware cost. These significantly high performance levels were proven with the implementation of several different configurations of the proposed structure using both FPGA and ASIC 90 nm technologies. In addition, such experimental evaluation also demonstrated the high area efficiency of theproposed architecture, which in terms of Data Throughput per Unit of Area (DTUA) is at least 1.5 times more efficient than its more prominent related designs(1).

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper describes an architecture conceived to integrate Power Sys-tems tools in a Power System Control Centre, based on an Ambient Intelligent (AmI) paradigm. This architecture is an instantiation of the generic architecture proposed in [1] for developing systems that interact with AmI environments. This architecture has been proposed as a consequence of a methodology for the inclu-sion of Artificial Intelligence in AmI environments (ISyRAmI - Intelligent Sys-tems Research for Ambient Intelligence). The architecture presented in the paper will be able to integrate two applications in the control room of a power system transmission network. The first is SPARSE expert system, used to get diagnosis of incidents and to support power restoration. The second application is an Intelligent Tutoring System (ITS) incorporating two training tools. The first tutoring tool is used to train operators to get the diagnosis of incidents. The second one is another tutoring tool used to train operators to perform restoration procedures.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper presents the proposal of an architecture for developing systems that interact with Ambient Intelligence (AmI) environments. This architecture has been proposed as a consequence of a methodology for the inclusion of Artificial Intelligence in AmI environments (ISyRAmI - Intelligent Systems Research for Ambient Intelligence). The ISyRAmI architecture considers several modules. The first is related with the acquisition of data, information and even knowledge. This data/information knowledge deals with our AmI environment and can be acquired in different ways (from raw sensors, from the web, from experts). The second module is related with the storage, conversion, and handling of the data/information knowledge. It is understood that incorrectness, incompleteness, and uncertainty are present in the data/information/knowledge. The third module is related with the intelligent operation on the data/information/knowledge of our AmI environment. Here we include knowledge discovery systems, expert systems, planning, multi-agent systems, simulation, optimization, etc. The last module is related with the actuation in the AmI environment, by means of automation, robots, intelligent agents and users.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A novel high throughput and scalable unified architecture for the computation of the transform operations in video codecs for advanced standards is presented in this paper. This structure can be used as a hardware accelerator in modern embedded systems to efficiently compute all the two-dimensional 4 x 4 and 2 x 2 transforms of the H.264/AVC standard. Moreover, its highly flexible design and hardware efficiency allows it to be easily scaled in terms of performance and hardware cost to meet the specific requirements of any given video coding application. Experimental results obtained using a Xilinx Virtex-5 FPGA demonstrated the superior performance and hardware efficiency levels provided by the proposed structure, which presents a throughput per unit of area relatively higher than other similar recently published designs targeting the H.264/AVC standard. Such results also showed that, when integrated in a multi-core embedded system, this architecture provides speedup factors of about 120x concerning pure software implementations of the transform algorithms, therefore allowing the computation, in real-time, of all the above mentioned transforms for Ultra High Definition Video (UHDV) sequences (4,320 x 7,680 @ 30 fps).

Relevância:

20.00% 20.00%

Publicador:

Resumo:

A new high performance architecture for the computation of all the DCT operations adopted in the H.264/AVC and HEVC standards is proposed in this paper. Contrasting to other dedicated transform cores, the presented multi-standard transform architecture is supported on a completely configurable, scalable and unified structure, that is able to compute not only the forward and the inverse 8×8 and 4×4 integer DCTs and the 4×4 and 2×2 Hadamard transforms defined in the H.264/AVC standard, but also the 4×4, 8×8, 16×16 and 32×32 integer transforms adopted in HEVC. Experimental results obtained using a Xilinx Virtex-7 FPGA demonstrated the superior performance and hardware efficiency levels provided by the proposed structure, which outperforms its more prominent related designs by at least 1.8 times. When integrated in a multi-core embedded system, this architecture allows the computation, in real-time, of all the transforms mentioned above for resolutions as high as the 8k Ultra High Definition Television (UHDTV) (7680×4320 @ 30fps).

Relevância:

20.00% 20.00%

Publicador:

Resumo:

For a long time the allegorical activity was considered dogmatic and equated with artistic fossilization, archaic religious propensity and lack of creativity. However, Walter Benjamin (1928) and Paul De Man (1969), among other illustrious thinkers, came to its defense, exalting, instead, its cryptic, hybrid and abstract nature, which, incidentally, are the main characteristics of modern art. “Twin Peaks – Fire Walk with Me” (David Lynch, 1992) is a wonderful object of analysis, despite being one of the most misunderstood films in the history of cinema. The fact that its narrative is a prequel to the cult television series “Twin Peaks” and incorporates many of the characters of that show, explicitly denigrating the moral image of the protagonist, Laura Palmer, brought about an intense rejection by the fans of the series, as well as the indifference of the cinephilic community in general. However, one must go deeper, in order to understand Lynch’s brave accomplishment and its artfulness. Indeed, the opus is a powerful cinematic allegory because it contains a double layer of metaphorical meaning, one of them being explicitly metacinematic. Thus, besides assuming itself as a filmic daimonic allegory, occurring in a spiritual universe of Good versus Evil, the film is also an authorial discourse on cinema itself. More specifically, it is an allegory of spectatorship, according to Robert Stam’s definition, where the existence and crossing over to “another side” duplicates the architecture of movie theatres and the psychic processes involved in film viewing.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Conferência: IEEE 24th International Conference on Application-Specific Systems, Architectures and Processors (ASAP)- Jun 05-07, 2013

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The legacy of nineteenth century social theory followed a “nationalist” model of society, assuming that analysis of social realities depends upon national boundaries, taking the nation-state as the primary unit of analysis, and developing the concept of methodological nationalism. This perspective regarded the nation-state as the natural - and even necessary - form of society in modernity. Thus, the constitution of large cities, at the end of the 19th century, through the intense flows of immigrants coming from diverse political and linguistic communities posed an enormous challenge to all social research. One of the most significant studies responding to this set of issues was The Immigrant Press and its Control, by Robert E. Park, one of the most prominent American sociologists of the first half of the 20th century. The Immigrant Press and its Control was part of a larger project entitled Americanization Studies: The Acculturation of Immigrant Group into American Society, funded by the Carnagie Corporation following World War I, taking as its goal to study the so-called “Americanization methods” during the 1920s. This paper revisits that particular work by Park to reveal how his detailed analysis of the role of the immigrant press overcame the limitations of methodological nationalism. By granting importance to language as a tool uniting each community and by showing how the strength of foreign languages expressed itself through the immigrant press, Park demonstrated that the latter produces a more ambivalent phenomenon than simply the assimilation of immigrants. On the one hand, the immigrant press served as a connecting force, driven by the desire to preserve the mother tongue and culture while at the same time awakening national sentiments that had, until then, remained diffuse. Yet, on the other hand, it facilitated the adjustment of immigrants to the American context. As a result, Park’s work contributes to our understanding of a particular liminal moment inherent within many intercultural contexts, the space between emigrant identity (emphasizing the country of origin) and immigrant identity (emphasizing the newly adopted country). His focus on the role played by media in the socialization of immigrant groups presaged later work on this subject by communication scholars. Focusing attention on Park’s research leads to other studies of the immigrant experience from the same period (e.g., Thomas & Znaniecki, The Polish Peasant in Europe and America), and also to insights on multi-presence and interculturality as significant but often overlooked phenomena in the study of immigrant socialization.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The Robuter is a robotic mobile platform that is located in the “Hands-On” Laboratory of the IPP-Hurray! Research Group, at the School of Engineering of the Polytechnic Institute of Porto. Recently, the Robuter was subject of an upgrading process addressing two essential areas: the Hardware Architecture and the Software Architecture. This upgrade in process was triggered due to technical problems on-board of the robot and also to the fact that the hardware/software architecture has become obsolete. This Technical Report overviews the most important aspects of the new Hardware and Software Architectures of the Robuter. This document also presents a first approach on the first steps towards the use of the Robuter platform, and provides some hints on future work that may be carried out using this mobile platform.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In Distributed Computer-Controlled Systems (DCCS), both real-time and reliability requirements are of major concern. Architectures for DCCS must be designed considering the integration of processing nodes and the underlying communication infrastructure. Such integration must be provided by appropriate software support services. In this paper, an architecture for DCCS is presented, its structure is outlined, and the services provided by the support software are presented. These are considered in order to guarantee the real-time and reliability requirements placed by current and future systems.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper presents an architecture (Multi-μ) being implemented to study and develop software based fault tolerant mechanisms for Real-Time Systems, using the Ada language (Ada 95) and Commercial Off-The-Shelf (COTS) components. Several issues regarding fault tolerance are presented and mechanisms to achieve fault tolerance by software active replication in Ada 95 are discussed. The Multi-μ architecture, based on a specifically proposed Fault Tolerance Manager (FTManager), is then described. Finally, some considerations are made about the work being done and essential future developments.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In the past years, Software Architecture has attracted increased attention by academia and industry as the unifying concept to structure the design of complex systems. One particular research area deals with the possibility of reconfiguring architectures to adapt the systems they describe to new requirements. Reconfiguration amounts to adding and removing components and connections, and may have to occur without stopping the execution of the system being reconfigured. This work contributes to the formal description of such a process. Taking as a premise that a single formalism hardly ever satisfies all requirements in every situation, we present three approaches, each one with its own assumptions about the systems it can be applied to and with different advantages and disadvantages. Each approach is based on work of other researchers and has the aesthetic concern of changing as little as possible the original formalism, keeping its spirit. The first approach shows how a given reconfiguration can be specified in the same manner as the system it is applied to and in a way to be efficiently executed. The second approach explores the Chemical Abstract Machine, a formalism for rewriting multisets of terms, to describe architectures, computations, and reconfigurations in a uniform way. The last approach uses a UNITY-like parallel programming design language to describe computations, represents architectures by diagrams in the sense of Category Theory, and specifies reconfigurations by graph transformation rules.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Para obtenção do grau de Doutor pela Universidade de Vigo com menção internacional Departamento de Informática

Relevância:

20.00% 20.00%

Publicador:

Resumo:

In spite of the significant amount of scientific work in Wireless Sensor Networks (WSNs), there is a clear lack of effective, feasible and usable WSN system architectures that address both functional and non-functional requirements in an integrated fashion. This poster abstract outlines the EMMON system architecture for large-scale, dense, real-time embedded monitoring. EMMON relies on a hierarchical network architecture together with integrated middleware and command&control mechanisms. It has been designed to use standard commercially– available technologies, while maintaining as much flexibility as possible to meet specific applications’ requirements. The EMMON WSN architecture has been validated through extensive simulation and experimental evaluation, including through a 300+ node test-bed, the largest WSN test-bed in Europe to date