820 resultados para efficient causation
Resumo:
Feature selection is a central problem in machine learning and pattern recognition. On large datasets (in terms of dimension and/or number of instances), using search-based or wrapper techniques can be cornputationally prohibitive. Moreover, many filter methods based on relevance/redundancy assessment also take a prohibitively long time on high-dimensional. datasets. In this paper, we propose efficient unsupervised and supervised feature selection/ranking filters for high-dimensional datasets. These methods use low-complexity relevance and redundancy criteria, applicable to supervised, semi-supervised, and unsupervised learning, being able to act as pre-processors for computationally intensive methods to focus their attention on smaller subsets of promising features. The experimental results, with up to 10(5) features, show the time efficiency of our methods, with lower generalization error than state-of-the-art techniques, while being dramatically simpler and faster.
Resumo:
The aim of the present work is to provide insight into the mechanism of laccase reactions using syringyl-type mediators. We studied the pH dependence and the kinetics of oxidation of syringyl-type phenolics using the low CotA and the high redox potential TvL laccases. Additionally, the efficiency of these compounds as redox mediators for the oxidation of non-phenolic lignin units was tested at different pH values and increasing mediator/non-phenolic ratios. Finally, the intermediates and products of reactions were identified by LC-MS and H-1 NMR. These approaches allow concluding on the (1) mechanism involved in the oxidation of phenolics by bacterial laccases, (2) importance of the chemical nature and properties of phenolic mediators, (3) apparent independence of the enzyme's properties on the yields of non-phenolics conversion, (4) competitive routes involved in the catalytic cycle of the laccase-mediator system with several new C-O coupling type structures being proposed.
Resumo:
Empowered by virtualisation technology, cloud infrastructures enable the construction of flexi- ble and elastic computing environments, providing an opportunity for energy and resource cost optimisation while enhancing system availability and achieving high performance. A crucial re- quirement for effective consolidation is the ability to efficiently utilise system resources for high- availability computing and energy-efficiency optimisation to reduce operational costs and carbon footprints in the environment. Additionally, failures in highly networked computing systems can negatively impact system performance substantially, prohibiting the system from achieving its initial objectives. In this paper, we propose algorithms to dynamically construct and readjust vir- tual clusters to enable the execution of users’ jobs. Allied with an energy optimising mechanism to detect and mitigate energy inefficiencies, our decision-making algorithms leverage virtuali- sation tools to provide proactive fault-tolerance and energy-efficiency to virtual clusters. We conducted simulations by injecting random synthetic jobs and jobs using the latest version of the Google cloud tracelogs. The results indicate that our strategy improves the work per Joule ratio by approximately 12.9% and the working efficiency by almost 15.9% compared with other state-of-the-art algorithms.
Resumo:
IEEE International Symposium on Circuits and Systems, pp. 220 – 223, Seattle, EUA
Resumo:
In search of an efficient but simple, low cost procedure for the serodiagnosis of Toxoplasmosis, especially suited for routine laboratories facing technical and budget limitations as in less developed countries, the diagnostic capability of Hematoxo® , an hemagglutination test for toxoplasmosis, was evaluated in relation to a battery of tests including IgG- and IgM-immunofluorescence tests, hemagglutination and an IgM-capture enzymatic assay. Detecting a little as 5 I.U. of IgG antitoxoplasma antibodies, Hematoxo® showed a straight agreement as to reactivity and non-reactivity for the 443 non-reactive and the 387 reactive serum samples, included in this study. In 23 cases presenting a serological pattern of acute toxoplasmosis and showing IgM antibodies, Hematoxo® could detect IgM antibodies in 18, indicated by negativation or a significant decrease in titers as a result of treating samples with 2-mercapto-ethanol. However, a neat increase in sensitivity for IgM specific antibodies could be achieved by previously removing IgG from the sample, as demonstrated in a series of acute toxoplasmosis sera. A simple procedure was developed for this purpose, by reconstituting a lyophilized suspension of Protein A - rich Staphylococcus with the lowest serum dilution to be tested. Of low cost and easy to perform, Hematoxo® affords not only a practical qualitative procedure for screening reactors and non-reactors, as in prenatal services, but also quantitative assays that permit to titrate antibodies as well as to identify IgM antibodies.
Resumo:
Proceedings of IEEE, ISCAS 2003, Vol.I, pp. 877-880
Resumo:
Recent integrated circuit technologies have opened the possibility to design parallel architectures with hundreds of cores on a single chip. The design space of these parallel architectures is huge with many architectural options. Exploring the design space gets even more difficult if, beyond performance and area, we also consider extra metrics like performance and area efficiency, where the designer tries to design the architecture with the best performance per chip area and the best sustainable performance. In this paper we present an algorithm-oriented approach to design a many-core architecture. Instead of doing the design space exploration of the many core architecture based on the experimental execution results of a particular benchmark of algorithms, our approach is to make a formal analysis of the algorithms considering the main architectural aspects and to determine how each particular architectural aspect is related to the performance of the architecture when running an algorithm or set of algorithms. The architectural aspects considered include the number of cores, the local memory available in each core, the communication bandwidth between the many-core architecture and the external memory and the memory hierarchy. To exemplify the approach we did a theoretical analysis of a dense matrix multiplication algorithm and determined an equation that relates the number of execution cycles with the architectural parameters. Based on this equation a many-core architecture has been designed. The results obtained indicate that a 100 mm(2) integrated circuit design of the proposed architecture, using a 65 nm technology, is able to achieve 464 GFLOPs (double precision floating-point) for a memory bandwidth of 16 GB/s. This corresponds to a performance efficiency of 71 %. Considering a 45 nm technology, a 100 mm(2) chip attains 833 GFLOPs which corresponds to 84 % of peak performance These figures are better than those obtained by previous many-core architectures, except for the area efficiency which is limited by the lower memory bandwidth considered. The results achieved are also better than those of previous state-of-the-art many-cores architectures designed specifically to achieve high performance for matrix multiplication.
Resumo:
The need to increase agricultural yield led, among others, to an increase in the consumption of nitrogen based fertilizers. As a consequence, there are excessive concentrations of nitrates, the most abundant of the reactive nitrogen (Nr) species, in several areas of the world. The demographic changes and projected population growth for the next decades, and the economic shifts which are already shaping the near future are powerful drivers for a further intensification in the use of fertilizers, with a predicted increase of the nitrogen loads in soils. Nitrate easily diffuses in the subsurface environments, portraying high mobility in soils. Moreover, the presence of high nitrate loads in water has the potential to cause an array of health dysfunctions, such as methemoglobinemia and several cancers. Permeable Reactive Barriers (PRB) placed strategically relatively to the nitrate source constitute an effective technology to tackle nitrate pollution. Ergo, PRB avoid various adverse impacts resulting from the displacement of reactive nitrogen downstream along water bodies. A four stages literature review was carried out in 34 databases. Initially, a set of pertinent key words were identified to perform the initial databases searches. Then, the synonyms of those initial key words were used to carry out a second set of databases searches. The third stage comprised the identification of other additional relevant terms from the research papers identified in the previous two stages. Again, databases searches were performed with this third set of key words. The final step consisted of the identification of relevant papers from the bibliography of the relevant papers identified in the previous three stages of the literature review process. The set of papers identified as relevant for in-depth analysis were assessed considering a set of relevant characterization variables.
Resumo:
Task scheduling is one of the key mechanisms to ensure timeliness in embedded real-time systems. Such systems have often the need to execute not only application tasks but also some urgent routines (e.g. error-detection actions, consistency checkers, interrupt handlers) with minimum latency. Although fixed-priority schedulers such as Rate-Monotonic (RM) are in line with this need, they usually make a low processor utilization available to the system. Moreover, this availability usually decreases with the number of considered tasks. If dynamic-priority schedulers such as Earliest Deadline First (EDF) are applied instead, high system utilization can be guaranteed but the minimum latency for executing urgent routines may not be ensured. In this paper we describe a scheduling model according to which urgent routines are executed at the highest priority level and all other system tasks are scheduled by EDF. We show that the guaranteed processor utilization for the assumed scheduling model is at least as high as the one provided by RM for two tasks, namely 2(2√−1). Seven polynomial time tests for checking the system timeliness are derived and proved correct. The proposed tests are compared against each other and to an exact but exponential running time test.
Resumo:
The reaction between 2-aminobenzenesulfonic acid and 2-hydroxy-3-methoxybenzaldehyde produces the acyclic Schiff base 2-[(2-hydroxy-3-methoxyphenyl) methylideneamino] benzenesulfonic acid (H2L center dot 3H(2)O) (1). In situ reactions of this compound with Cu(II) salts and, eventually, in the presence of pyridine (py) or 2,2'-bipyridine (2,2'-bipy) lead to the formation of the mononuclear complexes [CuL(H2O)(2)] (2) and [CuL(2,2'-bipy)]center dot DMF center dot H2O (3) and the diphenoxo-bridged dicopper compounds [CuL(py)](2) (4) and [CuL(EtOH)](2)center dot 2H(2)O (5). In 2-5 the L-2-ligand acts as a tridentate chelating species by means of one of the O-sulfonate atoms, the O-phenoxo and the N-atoms. The remaining coordination sites are then occupied by H2O (in 2), 2,2'-bipyridine (in 3), pyridine (in 4) or EtOH (in 5). Hydrogen bond interactions resulted in R-2(2) (14) and in R-4(4)(12) graph sets leading to dimeric species (in 2 and 3, respectively), 1D chain associations (in 2 and 5) or a 2D network (1). Complexes 2-5 are applied as selective catalysts for the homogeneous peroxidative (with tert-butylhydroperoxide, TBHP) oxidation of primary and secondary alcohols, under solvent-and additive-free conditions and under low power microwave (MW) irradiation. A quantitative yield of acetophenone was obtained by oxidation of 1-phenylethanol with compound 4 [TOFs up to 7.6 x 10(3) h(-1)] after 20 min of MW irradiation, whereas the oxidation of benzyl alcohol to benzaldehyde is less effective (TOF 992 h(-1)). The selectivity of 4 to oxidize the alcohol relative to the ene function is demonstrated when using cinnamyl alcohol as substrate.
Resumo:
In this work I articulate...
Resumo:
Coarse Grained Reconfigurable Architectures (CGRAs) are emerging as enabling platforms to meet the high performance demanded by modern applications (e.g. 4G, CDMA, etc.). Recently proposed CGRAs offer time-multiplexing and dynamic applications parallelism to enhance device utilization and reduce energy consumption at the cost of additional memory (up to 50% area of the overall platform). To reduce the memory overheads, novel CGRAs employ either statistical compression, intermediate compact representation, or multicasting. Each compaction technique has different properties (i.e. compression ratio, decompression time and decompression energy) and is best suited for a particular class of applications. However, existing research only deals with these methods separately. Moreover, they only analyze the compaction ratio and do not evaluate the associated energy overheads. To tackle these issues, we propose a polymorphic compression architecture that interleaves these techniques in a unique platform. The proposed architecture allows each application to take advantage of a separate compression/decompression hierarchy (consisting of various types and implementations of hardware/software decoders) tailored to its needs. Simulation results, using different applications (FFT, Matrix multiplication, and WLAN), reveal that the choice of compression hierarchy has a significant impact on compression ratio (up to 52%), decompression energy (up to 4 orders of magnitude), and configuration time (from 33 n to 1.5 s) for the tested applications. Synthesis results reveal that introducing adaptivity incurs negligible additional overheads (1%) compared to the overall platform area.
Resumo:
20th International Conference on Reliable Software Technologies - Ada-Europe 2015 (Ada-Europe 2015), Madrid, Spain.
Resumo:
In this paper we present the operational matrices of the left Caputo fractional derivative, right Caputo fractional derivative and Riemann–Liouville fractional integral for shifted Legendre polynomials. We develop an accurate numerical algorithm to solve the two-sided space–time fractional advection–dispersion equation (FADE) based on a spectral shifted Legendre tau (SLT) method in combination with the derived shifted Legendre operational matrices. The fractional derivatives are described in the Caputo sense. We propose a spectral SLT method, both in temporal and spatial discretizations for the two-sided space–time FADE. This technique reduces the two-sided space–time FADE to a system of algebraic equations that simplifies the problem. Numerical results carried out to confirm the spectral accuracy and efficiency of the proposed algorithm. By selecting relatively few Legendre polynomial degrees, we are able to get very accurate approximations, demonstrating the utility of the new approach over other numerical methods.
Resumo:
The shifted Legendre orthogonal polynomials are used for the numerical solution of a new formulation for the multi-dimensional fractional optimal control problem (M-DFOCP) with a quadratic performance index. The fractional derivatives are described in the Caputo sense. The Lagrange multiplier method for the constrained extremum and the operational matrix of fractional integrals are used together with the help of the properties of the shifted Legendre orthonormal polynomials. The method reduces the M-DFOCP to a simpler problem that consists of solving a system of algebraic equations. For confirming the efficiency and accuracy of the proposed scheme, some test problems are implemented with their approximate solutions.