949 resultados para Unified Formulation
Resumo:
Accurate assessments of fish populations are often limited by re-observation or recapture events. Since the early 1990s, passive integrated transponders (PIT tags) have been used to understand the biology of many fish species. Until recently, PIT applications in small streams have been limited to physical recapture events. To maximize recapture probability, we constructed PIT antenna arrays in small streams to remotely detect individual fish. Experiences from two different laboratories (three case studies) allowed us to develop a unified approach to applying PIT technology for enhancing data assessments. Information on equipment, its installation, tag considerations, and array construction is provided. Theoretical and practical definitions are introduced to standardize metrics for assessing detection efficiency. We demonstrate how certain conditions (stream discharge, vibration, and ambient radio frequency noise) affect the detection efficiency and suggest that by monitoring these conditions, expectations of efficiency can be modified. We emphasize the importance of consistently estimating detection efficiency for fisheries applications.
Resumo:
Economic models of crime have focused primarily on the goal of deterrence; the goal of incapacitation has received much less attention. This paper adapts the standard deterrence model to incorporate incapacitation. When prison only is used, incapacitation can result in a longer or a shorter optimal prison term compared to the deterrence-only model. It is longer if there is underdeterrence, and shorter if there is overdeterrence. In contrast, when a fine is available and it is not constrained by the offender's wealth, the optimal prison term is zero. Since the fine achieves first-best deterrence, only efficient crimes are committed and hence, there is no gain from incapacitation.
Resumo:
With the globalization of economic activity, the relative weight of foreign trade in national economic activities has increased, and the question of how to measure trends in the value and quantity of international trade has become an important issue for policy-makers and economists. This paper compares the chain-linked indices formulated by Masato Kuroko, based on HS this fiscal year for individual industry categories and countries with chain-linked indices based on SITC-R1 codes, in order to study how changes in the quality composition of the same products, which cannot be considered using unit value indices based on SITC-R1 codes, can be considered using unit value indices based on the more detailed HS product classifications.
Resumo:
Koopman et al. (2014) developed a method to consistently decompose gross exports in value-added terms that accommodate infinite repercussions of international and inter-sector transactions. This provides a better understanding of trade in value added in global value chains than does the conventional gross exports method, which is affected by double-counting problems. However, the new framework is based on monetary input--output (IO) tables and cannot distinguish prices from quantities; thus, it is unable to consider financial adjustments through the exchange market. In this paper, we propose a framework based on a physical IO system, characterized by its linear programming equivalent that can clarify the various complexities relevant to the existing indicators and is proved to be consistent with Koopman's results when the physical decompositions are evaluated in monetary terms. While international monetary tables are typically described in current U.S. dollars, the physical framework can elucidate the impact of price adjustments through the exchange market. An iterative procedure to calculate the exchange rates is proposed, and we also show that the physical framework is also convenient for considering indicators associated with greenhouse gas (GHG) emissions.
Resumo:
Metodología para integrar numéricamente de forma termodinámicamente consistente las ecuaciones del problema termomecánico acoplado de un elemento discreto viscoso.
Resumo:
Identification and tracking of objects in specific environments such as harbors or security areas is a matter of great importance nowadays. With this purpose, numerous systems based on different technologies have been developed, resulting in a great amount of gathered data displayed through a variety of interfaces. Such amount of information has to be evaluated by human operators in order to take the correct decisions, sometimes under highly critical situations demanding both speed and accuracy. In order to face this problem we describe IDT-3D, a platform for identification and tracking of vessels in a harbour environment able to represent fused information in real time using a Virtual Reality application. The effectiveness of using IDT-3D as an integrated surveillance system is currently under evaluation. Preliminary results point to a significant decrease in the times of reaction and decision making of operators facing up a critical situation. Although the current application focus of IDT-3D is quite specific, the results of this research could be extended to the identification and tracking of targets in other controlled environments of interest as coastlines, borders or even urban areas.
Resumo:
OntoTag - A Linguistic and Ontological Annotation Model Suitable for the Semantic Web
1. INTRODUCTION. LINGUISTIC TOOLS AND ANNOTATIONS: THEIR LIGHTS AND SHADOWS
Computational Linguistics is already a consolidated research area. It builds upon the results of other two major ones, namely Linguistics and Computer Science and Engineering, and it aims at developing computational models of human language (or natural language, as it is termed in this area). Possibly, its most well-known applications are the different tools developed so far for processing human language, such as machine translation systems and speech recognizers or dictation programs.
These tools for processing human language are commonly referred to as linguistic tools. Apart from the examples mentioned above, there are also other types of linguistic tools that perhaps are not so well-known, but on which most of the other applications of Computational Linguistics are built. These other types of linguistic tools comprise POS taggers, natural language parsers and semantic taggers, amongst others. All of them can be termed linguistic annotation tools.
Linguistic annotation tools are important assets. In fact, POS and semantic taggers (and, to a lesser extent, also natural language parsers) have become critical resources for the computer applications that process natural language. Hence, any computer application that has to analyse a text automatically and ‘intelligently’ will include at least a module for POS tagging. The more an application needs to ‘understand’ the meaning of the text it processes, the more linguistic tools and/or modules it will incorporate and integrate.
However, linguistic annotation tools have still some limitations, which can be summarised as follows:
1. Normally, they perform annotations only at a certain linguistic level (that is, Morphology, Syntax, Semantics, etc.).
2. They usually introduce a certain rate of errors and ambiguities when tagging. This error rate ranges from 10 percent up to 50 percent of the units annotated for unrestricted, general texts.
3. Their annotations are most frequently formulated in terms of an annotation schema designed and implemented ad hoc.
A priori, it seems that the interoperation and the integration of several linguistic tools into an appropriate software architecture could most likely solve the limitations stated in (1). Besides, integrating several linguistic annotation tools and making them interoperate could also minimise the limitation stated in (2). Nevertheless, in the latter case, all these tools should produce annotations for a common level, which would have to be combined in order to correct their corresponding errors and inaccuracies. Yet, the limitation stated in (3) prevents both types of integration and interoperation from being easily achieved.
In addition, most high-level annotation tools rely on other lower-level annotation tools and their outputs to generate their own ones. For example, sense-tagging tools (operating at the semantic level) often use POS taggers (operating at a lower level, i.e., the morphosyntactic) to identify the grammatical category of the word or lexical unit they are annotating. Accordingly, if a faulty or inaccurate low-level annotation tool is to be used by other higher-level one in its process, the errors and inaccuracies of the former should be minimised in advance. Otherwise, these errors and inaccuracies would be transferred to (and even magnified in) the annotations of the high-level annotation tool.
Therefore, it would be quite useful to find a way to
(i) correct or, at least, reduce the errors and the inaccuracies of lower-level linguistic tools;
(ii) unify the annotation schemas of different linguistic annotation tools or, more generally speaking, make these tools (as well as their annotations) interoperate.
Clearly, solving (i) and (ii) should ease the automatic annotation of web pages by means of linguistic tools, and their transformation into Semantic Web pages (Berners-Lee, Hendler and Lassila, 2001). Yet, as stated above, (ii) is a type of interoperability problem. There again, ontologies (Gruber, 1993; Borst, 1997) have been successfully applied thus far to solve several interoperability problems. Hence, ontologies should help solve also the problems and limitations of linguistic annotation tools aforementioned.
Thus, to summarise, the main aim of the present work was to combine somehow these separated approaches, mechanisms and tools for annotation from Linguistics and Ontological Engineering (and the Semantic Web) in a sort of hybrid (linguistic and ontological) annotation model, suitable for both areas. This hybrid (semantic) annotation model should (a) benefit from the advances, models, techniques, mechanisms and tools of these two areas; (b) minimise (and even solve, when possible) some of the problems found in each of them; and (c) be suitable for the Semantic Web. The concrete goals that helped attain this aim are presented in the following section.
2. GOALS OF THE PRESENT WORK
As mentioned above, the main goal of this work was to specify a hybrid (that is, linguistically-motivated and ontology-based) model of annotation suitable for the Semantic Web (i.e. it had to produce a semantic annotation of web page contents). This entailed that the tags included in the annotations of the model had to (1) represent linguistic concepts (or linguistic categories, as they are termed in ISO/DCR (2008)), in order for this model to be linguistically-motivated; (2) be ontological terms (i.e., use an ontological vocabulary), in order for the model to be ontology-based; and (3) be structured (linked) as a collection of ontology-based
Resumo:
Modeling the evolution of the state of program memory during program execution is critical to many parallehzation techniques. Current memory analysis techniques either provide very accurate information but run prohibitively slowly or produce very conservative results. An approach based on abstract interpretation is presented for analyzing programs at compile time, which can accurately determine many important program properties such as aliasing, logical data structures and shape. These properties are known to be critical for transforming a single threaded program into a versión that can be run on múltiple execution units in parallel. The analysis is shown to be of polynomial complexity in the size of the memory heap. Experimental results for benchmarks in the Jolden suite are given. These results show that in practice the analysis method is efflcient and is capable of accurately determining shape information in programs that créate and manipúlate complex data structures.
Resumo:
The complete formulation of B.E.M. applied to the analysis of axisymmetric bodies acting in the plastic range is presented in this paper. The concept of derivative of a singular integral given by Mikhlin has been used in order to calculate the stresses in internal points. Also a semianalytical approach is proposed to compute the matrix coefficients, presenting the way in which it can be done and the results obtained.
Resumo:
A mathematical formulation for finite strain elasto plastic consolidation of fully saturated soil media is presented. Strong and weak forms of the boundary-value problem are derived using both the material and spatial descriptions. The algorithmic treatment of finite strain elastoplasticity for the solid phase is based on multiplicative decomposition and is coupled with the algorithm for fluid flow via the Kirchhoff pore water pressure. Balance laws are written for the soil-water mixture following the motion of the soil matrix alone. It is shown that the motion of the fluid phase only affects the Jacobian of the solid phase motion, and therefore can be characterized completely by the motion of the soil matrix. Furthermore, it is shown from energy balance consideration that the effective, or intergranular, stress is the appropriate measure of stress for describing the constitutive response of the soil skeleton since it absorbs all the strain energy generated in the saturated soil-water mixture. Finally, it is shown that the mathematical model is amenable to consistent linearization, and that explicit expressions for the consistent tangent operators can be derived for use in numerical solutions such as those based on the finite element method.
Resumo:
The goal of the W3C's Media Annotation Working Group (MAWG) is to promote interoperability between multimedia metadata formats on the Web. As experienced by everybody, audiovisual data is omnipresent on today's Web. However, different interaction interfaces and especially diverse metadata formats prevent unified search, access, and navigation. MAWG has addressed this issue by developing an interlingua ontology and an associated API. This article discusses the rationale and core concepts of the ontology and API for media resources. The specifications developed by MAWG enable interoperable contextualized and semantic annotation and search, independent of the source metadata format, and connecting multimedia data to the Linked Data cloud. Some demonstrators of such applications are also presented in this article.
Resumo:
Computation of Independent Sensitivities Using Maggi’s Formulation
Resumo:
The performance of a CATR relies on the planarity of the synthesized test wave, which is generated within a bounded volume for which specifications are drawn. Millimetre-wave facilities deal with the classical limitations of this frequency band, among which two become critical in our analysis: time-extensive acquisition campaigns and impact of environmental variables. Both features become more evident when increasing the frequency of operation. The variation in atmospheric variables, such as humidity, temperature and pressure has an influence over the performance of all the elements of the facility. The instrumentation behavior is influenced both by the warming up process, and the ambience conditions that surround the equipment. On the changes of the atmosphere itself, they affect the electromagnetic wave propagation, given the physical link between the conditions of the atmosphere and its electric properties as an electromagnetic waves propagation medium
Resumo:
Automated and semi-automated accessibility evaluation tools are key to streamline the process of accessibility assessment, and ultimately ensure that software products, contents, and services meet accessibility requirements. Different evaluation tools may better fit different needs and concerns, accounting for a variety of corporate and external policies, content types, invocation methods, deployment contexts, exploitation models, intended audiences and goals; and the specific overall process where they are introduced. This has led to the proliferation of many evaluation tools tailored to specific contexts. However, tool creators, who may be not familiar with the realm of accessibility and may be part of a larger project, lack any systematic guidance when facing the implementation of accessibility evaluation functionalities. Herein we present a systematic approach to the development of accessibility evaluation tools, leveraging the different artifacts and activities of a standardized development process model (the Unified Software Development Process), and providing templates of these artifacts tailored to accessibility evaluation tools. The work presented specially considers the work in progress in this area by the W3C/WAI Evaluation and Report Working Group (ERT WG)
Resumo:
Due to the high dependence of photovoltaic energy efficiency on environmental conditions (temperature, irradiation...), it is quite important to perform some analysis focusing on the characteristics of photovoltaic devices in order to optimize energy production, even for small-scale users. The use of equivalent circuits is the preferred option to analyze solar cells/panels performance. However, the aforementioned small-scale users rarely have the equipment or expertise to perform large testing/calculation campaigns, the only information available for them being the manufacturer datasheet. The solution to this problem is the development of new and simple methods to define equivalent circuits able to reproduce the behavior of the panel for any working condition, from a very small amount of information. In the present work a direct and completely explicit method to extract solar cell parameters from the manufacturer datasheet is presented and tested. This method is based on analytical formulation which includes the use of the Lambert W-function to turn the series resistor equation explicit. The presented method is used to analyze commercial solar panel performance (i.e., the current-voltage–I-V–curve) at different levels of irradiation and temperature. The analysis performed is based only on the information included in the manufacturer’s datasheet.