911 resultados para architectural computation
Resumo:
There is great demand for easily-accessible, user-friendly dietary self-management applications. Yet accurate, fully-automatic estimation of nutritional intake using computer vision methods remains an open research problem. One key element of this problem is the volume estimation, which can be computed from 3D models obtained using multi-view geometry. The paper presents a computational system for volume estimation based on the processing of two meal images. A 3D model of the served meal is reconstructed using the acquired images and the volume is computed from the shape. The algorithm was tested on food models (dummy foods) with known volume and on real served food. Volume accuracy was in the order of 90 %, while the total execution time was below 15 seconds per image pair. The proposed system combines simple and computational affordable methods for 3D reconstruction, remained stable throughout the experiments, operates in near real time, and places minimum constraints on users.
Resumo:
Well-known data mining algorithms rely on inputs in the form of pairwise similarities between objects. For large datasets it is computationally impossible to perform all pairwise comparisons. We therefore propose a novel approach that uses approximate Principal Component Analysis to efficiently identify groups of similar objects. The effectiveness of the approach is demonstrated in the context of binary classification using the supervised normalized cut as a classifier. For large datasets from the UCI repository, the approach significantly improves run times with minimal loss in accuracy.
Resumo:
Pre-combined SLR-GNSS solutions are studied and the impact of different types of datum definition on the estimated parameters is assessed. It is found that the origin is realized best by using only the SLR core network for defining the geodetic datum and the inclusion of the GNSS core sites degrades the origin. The orientation, however, requires a dense and continuous network, thus, the inclusion of the GNSS core network is absolutely needed.
Resumo:
Software architecture consists of a set of design choices that can be partially expressed in form of rules that the implementation must conform to. Architectural rules are intended to ensure properties that fulfill fundamental non-functional requirements. Verifying architectural rules is often a non- trivial activity: available tools are often not very usable and support only a narrow subset of the rules that are commonly specified by practitioners. In this paper we present a new highly-readable declarative language for specifying architectural rules. With our approach, users can specify a wide variety of rules using a single uniform notation. Rules can get tested by third-party tools by conforming to pre-defined specification templates. Practitioners can take advantage of the capabilities of a growing number of testing tools without dealing with them directly.
Resumo:
Transcription factors often determine cell fate and tissue development. Chondrogenesis is the developmental process by which cartilages form. Recently, gene targeting studies have shown that two transcription factors, L-Sox5 and Sox6, play essential and redundant roles in chondrogenesis in vivo by converting precartilaginous cell condensations into cartilages. Both are highly similar High-Mobility-Group (HMG)-domain proteins that bind and subsequently bend DNA containing the 7bp HMG site (A/T)(A/T)CAA(A/T)G. They have no transactivation domain, but homo- and hetero-dimerize and preferentially bind DNA containing two HMG sites. They are thought to play an architectural role in transactivation by facilitating long-range DNA and protein interactions. To understand their molecular mechanism of action, we investigated how phasing, orientation, and spacing between HMG sites affect L-Sox5 and Sox6 DNA-binding. We determined that L-Sox5 and Sox6 dimers bind with high affinity to paired HMG sites in DNA rather than a single HMG site. Binding of paired sites is independent of DNA helical phasing, orientation of paired HMG sites and independent of distance up to 255 base pairs between sites. Mutational analysis demonstrated that binding of L-Sox5 and Sox6, independent of orientation of the sites, is critically dependent on the presence of paired HMG sites rather than one HMG site alone. Our data support a unique and novel model whereby L-Sox5 and Sox6 dimerize and bind DNA with pronounced spatial flexibility, possibly by a flexible hinge, and act as architectural transcription factors that bring distant DNA sites and proteins together to form higher order transcriptional complexes that are essential for the activation of their target genes in chondrogenesis. ^
Resumo:
The literature on the use of free trade agreements (FTAs) has recently been growing because it is becoming more important to encourage the use of current FTAs than to increase the number of FTAs. In this paper, we discuss some practical issues in the computation of FTA utilization rates, which provide a useful measure to discover how much FTA schemes are used in trade. For example, compared with the use of customs data on FTA utilization in imports, when using certificates of origin data on FTA utilization in exports, there are several points about which we should be careful. Our practical guidance on the computation of FTA utilization rates will be helpful when computing such rates and in examining the determinants of those rates empirically.
Resumo:
Nowadays computing platforms consist of a very large number of components that require to be supplied with diferent voltage levels and power requirements. Even a very small platform, like a handheld computer, may contain more than twenty diferent loads and voltage regulators. The power delivery designers of these systems are required to provide, in a very short time, the right power architecture that optimizes the performance, meets electrical specifications plus cost and size targets. The appropriate selection of the architecture and converters directly defines the performance of a given solution. Therefore, the designer needs to be able to evaluate a significant number of options in order to know with good certainty whether the selected solutions meet the size, energy eficiency and cost targets. The design dificulties of selecting the right solution arise due to the wide range of power conversion products provided by diferent manufacturers. These products range from discrete components (to build converters) to complete power conversion modules that employ diferent manufacturing technologies. Consequently, in most cases it is not possible to analyze all the alternatives (combinations of power architectures and converters) that can be built. The designer has to select a limited number of converters in order to simplify the analysis. In this thesis, in order to overcome the mentioned dificulties, a new design methodology for power supply systems is proposed. This methodology integrates evolutionary computation techniques in order to make possible analyzing a large number of possibilities. This exhaustive analysis helps the designer to quickly define a set of feasible solutions and select the best trade-off in performance according to each application. The proposed approach consists of two key steps, one for the automatic generation of architectures and other for the optimized selection of components. In this thesis are detailed the implementation of these two steps. The usefulness of the methodology is corroborated by contrasting the results using real problems and experiments designed to test the limits of the algorithms.
Resumo:
Software architectural evaluation is a key discipline used to identify, at early stages of a real-time system (RTS) development, the problems that may arise during its operation. Typical mechanisms supporting concurrency, such as semaphores, mutexes or monitors, usually lead to concurrency problems in execution time that are difficult to be identified, reproduced and solved. For this reason, it is crucial to understand the root causes of these problems and to provide support to identify and mitigate them at early stages of the system lifecycle. This paper aims to present the results of a research work oriented to the development of the tool called ‘Deadlock Risk Evaluation of Architectural Models’ (DREAM) to assess deadlock risk in architectural models of an RTS. A particular architectural style, Pipelines of Processes in Object-Oriented Architectures–UML (PPOOA) was used to represent platform-independent models of an RTS architecture supported by the PPOOA –Visio tool. We validated the technique presented here by using several case studies related to RTS development and comparing our results with those from other deadlock detection approaches, supported by different tools. Here we present two of these case studies, one related to avionics and the other to planetary exploration robotics. Copyright © 2011 John Wiley & Sons, Ltd.
Resumo:
This paper is presented in CIB: Management and Innovation Sustainable Built Environment 2011, as the study and analysis of the residential model of a rural area from the Iberian Peninsula, specifically applied to the case of the province of Cáceres, in the autonomous region of Extremadura, in Spain. To this end, from a database made up of building projects whose real costs are known, it is intended to establish the links of the different parameters studied through the corresponding functions of statistical analysis. One of the main objectives of this process is constituted by the possibility of establishing those design variables of higher economic importance, so as to keep an economic control of these parameters, generally geometrical and typological, from the very start of the project. And, in general, a higher optimization of resources in the construction of dwellings in the rural environment from their design is intended.
Resumo:
Membrane systems are computational equivalent to Turing machines. However, their distributed and massively parallel nature obtains polynomial solutions opposite to traditional non-polynomial ones. At this point, it is very important to develop dedicated hardware and software implementations exploiting those two membrane systems features. Dealing with distributed implementations of P systems, the bottleneck communication problem has arisen. When the number of membranes grows up, the network gets congested. The purpose of distributed architectures is to reach a compromise between the massively parallel character of the system and the needed evolution step time to transit from one configuration of the system to the next one, solving the bottleneck communication problem. The goal of this paper is twofold. Firstly, to survey in a systematic and uniform way the main results regarding the way membranes can be placed on processors in order to get a software/hardware simulation of P-Systems in a distributed environment. Secondly, we improve some results about the membrane dissolution problem, prove that it is connected, and discuss the possibility of simulating this property in the distributed model. All this yields an improvement in the system parallelism implementation since it gets an increment of the parallelism of the external communication among processors. Proposed ideas improve previous architectures to tackle the communication bottleneck problem, such as reduction of the total time of an evolution step, increase of the number of membranes that could run on a processor and reduction of the number of processors.
Resumo:
Dendritic computation is a term that has been in neuro physiological research for a long time [1]. It is still controversial and far for been clarified within the concepts of both computation and neurophysiology [2], [3]. In any case, it hasnot been integrated neither in a formal computational scheme or structure, nor into formulations of artificial neural nets. Our objective here is to formulate a type of distributed computation that resembles dendritic trees, in such a way that it shows the advantages of neural network distributed computation, mostly the reliability that is shown under the existence of holes (scotomas) in the computing net, without ?blind spots?.