156 resultados para Log conformance


Relevância:

10.00% 10.00%

Publicador:

Resumo:

Gelatin graft copolymers of different compositions were tested for microbial susceptibility in a synthetic medium with pure cultures of Pseudomonas aeruginosa, Bacillus subtilis, and Serratia marcescens. The percent weight losses were recorded over 6 weeks of incubation period in nitrogen-free and nitrogen-rich media. The relationship between [log(rate)] during the first week of the test period and composition of the grafted samples showed a linear behavior. There was no difference in the aggressivity of these bacterial strains. Nitrogen analysis data and pH measurements of the media seem to reinforce our earlier observations. Soil burial tests also indicate degradation of polymer samples under natural weathering conditions. This article also summarizes the salient features of our series of investigations.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Overconsolidated soils exhibit a bilinear e-log p relationship. During virgin compression, microstructural units form larger stable groups, thereby reducing the operating specific surface and, in turn, net osmotic repulsive forces in the soil. The rebound portion of the e-log p curve is consequently flatter. The generalized relationship for compressibility of uncemented soils in the overconsolidated state has been developed in the form of e/eL = 1.122 = 0.188 log pc - 0.0463 log p in which e/eL is the generalized soil state parameter, pc is the preconsolidation pressure in kPa, p is the effective overburden pressure in kPa, e is the in situ void ratio, and eL is the void ratio corresponding to the liquid limit water content (wLG). This relationship can be usefully employed to predict both the preconsolidation pressure and compressibility responses of overconsolidated uncemented soils.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

A channel router is an important design aid in the design automation of VLSI circuit layout. Many algorithms have been developed based on various wiring models with routing done on two layers. With the recent advances in VLSI process technology, it is possible to have three independent layers for interconnection. In this paper two algorithms are presented for three-layer channel routing. The first assumes a very simple wiring model. This enables the routing problem to be solved optimally in a time of O(n log n). The second algorithm is for a different wiring model and has an upper bound of O(n2) for its execution time. It uses fewer horizontal tracks than the first algorithm. For the second model the channel width is not bounded by the channel density.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The unconfined aquifer of the Continental Terminal in Niger was investigated by magnetic resonance sounding (MRS) and by 14 pumping tests in order to improve calibration of MRS outputs at field scale. The reliability of the standard relationship used for estimating aquifer transmissivity by MRS was checked; it was found that the parametric factor can be estimated with an uncertainty a parts per thousand currency sign150% by a single point of calibration. The MRS water content (theta (MRS)) was shown to be positively correlated with the specific yield (Sy), and theta (MRS) always displayed higher values than Sy. A conceptual model was subsequently developed, based on estimated changes of the total porosity, Sy, and the specific retention Sr as a function of the median grain size. The resulting relationship between theta (MRS) and Sy showed a reasonably good fit with the experimental dataset, considering the inherent heterogeneity of the aquifer matrix (residual error is similar to 60%). Interpreted in terms of aquifer parameters, MRS data suggest a log-normal distribution of the permeability and a one-sided Gaussian distribution of Sy. These results demonstrate the efficiency of the MRS method for fast and low-cost prospection of hydraulic parameters for large unconfined aquifers.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Emerging embedded applications are based on evolving standards (e.g., MPEG2/4, H.264/265, IEEE802.11a/b/g/n). Since most of these applications run on handheld devices, there is an increasing need for a single chip solution that can dynamically interoperate between different standards and their derivatives. In order to achieve high resource utilization and low power dissipation, we propose REDEFINE, a polymorphic ASIC in which specialized hardware units are replaced with basic hardware units that can create the same functionality by runtime re-composition. It is a ``future-proof'' custom hardware solution for multiple applications and their derivatives in a domain. In this article, we describe a compiler framework and supporting hardware comprising compute, storage, and communication resources. Applications described in high-level language (e.g., C) are compiled into application substructures. For each application substructure, a set of compute elements on the hardware are interconnected during runtime to form a pattern that closely matches the communication pattern of that particular application. The advantage is that the bounded CEs are neither processor cores nor logic elements as in FPGAs. Hence, REDEFINE offers the power and performance advantage of an ASIC and the hardware reconfigurability and programmability of that of an FPGA/instruction set processor. In addition, the hardware supports custom instruction pipelining. Existing instruction-set extensible processors determine a sequence of instructions that repeatedly occur within the application to create custom instructions at design time to speed up the execution of this sequence. We extend this scheme further, where a kernel is compiled into custom instructions that bear strong producer-consumer relationship (and not limited to frequently occurring sequences of instructions). Custom instructions, realized as hardware compositions effected at runtime, allow several instances of the same to be active in parallel. A key distinguishing factor in majority of the emerging embedded applications is stream processing. To reduce the overheads of data transfer between custom instructions, direct communication paths are employed among custom instructions. In this article, we present the overview of the hardware-aware compiler framework, which determines the NoC-aware schedule of transports of the data exchanged between the custom instructions on the interconnect. The results for the FFT kernel indicate a 25% reduction in the number of loads/stores, and throughput improves by log(n) for n-point FFT when compared to sequential implementation. Overall, REDEFINE offers flexibility and a runtime reconfigurability at the expense of 1.16x in power and 8x in area when compared to an ASIC. REDEFINE implementation consumes 0.1x the power of an FPGA implementation. In addition, the configuration overhead of the FPGA implementation is 1,000x more than that of REDEFINE.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The conventional Clauser-chart method for determination of local skin friction in zero or weak pressure-gradient turbulent boundary layer flows fails entirely in strong pressure-gradient situations. This failure occurs due to the large departure of the mean velocity profile from the universal logarithmic law upon which the conventional Clauser-chart method is based. It is possible to extend this method,even for strong pressure-gradient situations involving equilibrium or near-equilibrium turbulent boundary layers by making use of the so-called non-universal logarithmic laws. These non-universal log laws depend on the local strength of the pressure gradient and may be regarded as perturbations of the universal log law.The present paper shows that the modified Clauser-chart method, so developed, yields quit satisfactory results in terms of estimation of local skin friction in strongly accelerated or retarded equilibrium and near-equilibrium turbulent boundary layers that are not very close to relaminarization or separation.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

An axis-parallel k-dimensional box is a Cartesian product R-1 x R-2 x...x R-k where R-i (for 1 <= i <= k) is a closed interval of the form [a(i), b(i)] on the real line. For a graph G, its boxicity box(G) is the minimum dimension k, such that G is representable as the intersection graph of (axis-parallel) boxes in k-dimensional space. The concept of boxicity finds applications in various areas such as ecology, operations research etc. A number of NP-hard problems are either polynomial time solvable or have much better approximation ratio on low boxicity graphs. For example, the max-clique problem is polynomial time solvable on bounded boxicity graphs and the maximum independent set problem for boxicity d graphs, given a box representation, has a left perpendicular1 + 1/c log n right perpendicular(d-1) approximation ratio for any constant c >= 1 when d >= 2. In most cases, the first step usually is computing a low dimensional box representation of the given graph. Deciding whether the boxicity of a graph is at most 2 itself is NP-hard. We give an efficient randomized algorithm to construct a box representation of any graph G on n vertices in left perpendicular(Delta + 2) ln nright perpendicular dimensions, where Delta is the maximum degree of G. This algorithm implies that box(G) <= left perpendicular(Delta + 2) ln nright perpendicular for any graph G. Our bound is tight up to a factor of ln n. We also show that our randomized algorithm can be derandomized to get a polynomial time deterministic algorithm. Though our general upper bound is in terms of maximum degree Delta, we show that for almost all graphs on n vertices, their boxicity is O(d(av) ln n) where d(av) is the average degree.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

Microbiological quality of the treated wastewater is an important parameter for its reuse. The data oil the Fecal Coliform (FC) and Fecal Streptococcus (FS) at different stages of treatment in the Sewage Treatment Plants (STPs) in Delhi watershed is not available, therefore in the present study microbial profiling of STPs was carried out to assess the effluent quality for present and future reuse options. This Study further evaluates the water quality profiles at different stages of treatment for 16 STPs in Delhi city. These STPs are based on conventional Activated Sludge Process (ASP), extended aeration, physical, chemical and biological treatment (BIOFORE), Trickling Filter and Oxidation Pond. The primary effluent quality produced from most of the STPs was suitable for Soil Aquifer Treatment (SAT). Extended Hydraulic Retention Time (HRT) as a result Of low inflow to the STPS Was responsible for high turbidity, COD and BODs removal. Conventional ASP based STPs achieved 1.66 log FC and 1.06 log FS removal. STPs with extended aeration treatment process produced better quality effluent with maximum 4 log order reduction in FC and FS levels. ``Kondli'' and ``Nilothi'' STPs employing ASP, produced better quality secondary effluent as compared to other STPs based oil similar treatment process. Oxidation Pond based STPs showed better FC and FS removals, whereas good physiochemical quality was achieved during the first half of the treatment. Based upon physical, chemical and microbiological removal efficiencies, actual integrated efficiency (IEa) of each STP was determined to evaluate its Suitability for reuse for irrigation purposes. Except Mehrauli'' and ``Oxidation Pond'', effluents from all other STPs require tertiary treatment for further reuse. Possible reuse options, depending Upon the geographical location, proximity of facilities of potential users based oil the beneficial uses, and sub-soil types, etc. for the Delhi city have been investigated, which include artificial groundwater recharge, aquaculture, horticulture and industrial uses Such as floor washing, boiler feed, and cooling towers, etc. (C) 2009 Elsevier B.V. All rights reserved.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

In recent years, spatial variability modeling of soil parameters using random field theory has gained distinct importance in geotechnical analysis. In the present Study, commercially available finite difference numerical code FLAC 5.0 is used for modeling the permeability parameter as spatially correlated log-normally distributed random variable and its influence on the steady state seepage flow and on the slope stability analysis are studied. Considering the case of a 5.0 m high cohesive-frictional soil slope of 30 degrees, a range of coefficients of variation (CoV%) from 60 to 90% in the permeability Values, and taking different values of correlation distance in the range of 0.5-15 m, parametric studies, using Monte Carlo simulations, are performed to study the following three aspects, i.e., (i) effect ostochastic soil permeability on the statistics of seepage flow in comparison to the analytic (Dupuit's) solution available for the uniformly constant permeability property; (ii) strain and deformation pattern, and (iii) stability of the given slope assessed in terms of factor of safety (FS). The results obtained in this study are useful to understand the role of permeability variations in slope stability analysis under different slope conditions and material properties. (C) 2009 Elsevier B.V. All rights reserved.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

A common trick for designing faster quantum adiabatic algorithms is to apply the adiabaticity condition locally at every instant. However it is often difficult to determine the instantaneous gap between the lowest two eigenvalues, which is an essential ingredient in the adiabaticity condition. In this paper we present a simple linear algebraic technique for obtaining a lower bound on the instantaneous gap even in such a situation. As an illustration, we investigate the adiabatic un-ordered search of van Dam et al. [17] and Roland and Cerf [15] when the non-zero entries of the diagonal final Hamiltonian are perturbed by a polynomial (in log N, where N is the length of the unordered list) amount. We use our technique to derive a bound on the running time of a local adiabatic schedule in terms of the minimum gap between the lowest two eigenvalues.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We consider the problem of determining if two finite groups are isomorphic. The groups are assumed to be represented by their multiplication tables. We present an O(n) algorithm that determines if two Abelian groups with n elements each are isomorphic. This improves upon the previous upper bound of O(n log n) [Narayan Vikas, An O(n) algorithm for Abelian p-group isomorphism and an O(n log n) algorithm for Abelian group isomorphism, J. Comput. System Sci. 53 (1996) 1-9] known for this problem. We solve a more general problem of computing the orders of all the elements of any group (not necessarily Abelian) of size n in O(n) time. Our algorithm for isomorphism testing of Abelian groups follows from this result. We use the property that our order finding algorithm works for any group to design a simple O(n) algorithm for testing whether a group of size n, described by its multiplication table, is nilpotent. We also give an O(n) algorithm for determining if a group of size n, described by its multiplication table, is Abelian. (C) 2007 Elsevier Inc. All rights reserved.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

The Hadwiger number eta(G) of a graph G is the largest integer n for which the complete graph K-n on n vertices is a minor of G. Hadwiger conjectured that for every graph G, eta(G) >= chi(G), where chi(G) is the chromatic number of G. In this paper, we study the Hadwiger number of the Cartesian product G square H of graphs. As the main result of this paper, we prove that eta(G(1) square G(2)) >= h root 1 (1 - o(1)) for any two graphs G(1) and G(2) with eta(G(1)) = h and eta(G(2)) = l. We show that the above lower bound is asymptotically best possible when h >= l. This asymptotically settles a question of Z. Miller (1978). As consequences of our main result, we show the following: 1. Let G be a connected graph. Let G = G(1) square G(2) square ... square G(k) be the ( unique) prime factorization of G. Then G satisfies Hadwiger's conjecture if k >= 2 log log chi(G) + c', where c' is a constant. This improves the 2 log chi(G) + 3 bound in [2] 2. Let G(1) and G(2) be two graphs such that chi(G1) >= chi(G2) >= clog(1.5)(chi(G(1))), where c is a constant. Then G1 square G2 satisfies Hadwiger's conjecture. 3. Hadwiger's conjecture is true for G(d) (Cartesian product of G taken d times) for every graph G and every d >= 2. This settles a question by Chandran and Sivadasan [2]. ( They had shown that the Hadiwger's conjecture is true for G(d) if d >= 3).

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We consider the problem of computing an approximate minimum cycle basis of an undirected edge-weighted graph G with m edges and n vertices; the extension to directed graphs is also discussed. In this problem, a {0,1} incidence vector is associated with each cycle and the vector space over F-2 generated by these vectors is the cycle space of G. A set of cycles is called a cycle basis of G if it forms a basis for its cycle space. A cycle basis where the sum of the weights of the cycles is minimum is called a minimum cycle basis of G. Cycle bases of low weight are useful in a number of contexts, e.g. the analysis of electrical networks, structural engineering, chemistry, and surface reconstruction. We present two new algorithms to compute an approximate minimum cycle basis. For any integer k >= 1, we give (2k - 1)-approximation algorithms with expected running time 0(kmn(1+2/k) + mn((1+1/k)(omega-1))) and deterministic running time 0(n(3+2/k)), respectively. Here omega is the best exponent of matrix multiplication. It is presently known that omega < 2.376. Both algorithms are o(m(omega)) for dense graphs. This is the first time that any algorithm which computes sparse cycle bases with a guarantee drops below the Theta(m(omega)) bound. We also present a 2-approximation algorithm with O(m(omega) root n log n) expected running time, a linear time 2-approximation algorithm for planar graphs and an O(n(3)) time 2.42-approximation algorithm for the complete Euclidean graph in the plane.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

We consider a scenario in which a wireless sensor network is formed by randomly deploying n sensors to measure some spatial function over a field, with the objective of computing a function of the measurements and communicating it to an operator station. We restrict ourselves to the class of type-threshold functions (as defined in the work of Giridhar and Kumar, 2005), of which max, min, and indicator functions are important examples: our discussions are couched in terms of the max function. We view the problem as one of message-passing distributed computation over a geometric random graph. The network is assumed to be synchronous, and the sensors synchronously measure values and then collaborate to compute and deliver the function computed with these values to the operator station. Computation algorithms differ in (1) the communication topology assumed and (2) the messages that the nodes need to exchange in order to carry out the computation. The focus of our paper is to establish (in probability) scaling laws for the time and energy complexity of the distributed function computation over random wireless networks, under the assumption of centralized contention-free scheduling of packet transmissions. First, without any constraint on the computation algorithm, we establish scaling laws for the computation time and energy expenditure for one-time maximum computation. We show that for an optimal algorithm, the computation time and energy expenditure scale, respectively, as Theta(radicn/log n) and Theta(n) asymptotically as the number of sensors n rarr infin. Second, we analyze the performance of three specific computation algorithms that may be used in specific practical situations, namely, the tree algorithm, multihop transmission, and the Ripple algorithm (a type of gossip algorithm), and obtain scaling laws for the computation time and energy expenditure as n rarr infin. In particular, we show that the computation time for these algorithms scales as Theta(radicn/lo- g n), Theta(n), and Theta(radicn log n), respectively, whereas the energy expended scales as , Theta(n), Theta(radicn/log n), and Theta(radicn log n), respectively. Finally, simulation results are provided to show that our analysis indeed captures the correct scaling. The simulations also yield estimates of the constant multipliers in the scaling laws. Our analyses throughout assume a centralized optimal scheduler, and hence, our results can be viewed as providing bounds for the performance with practical distributed schedulers.

Relevância:

10.00% 10.00%

Publicador:

Resumo:

A new approach for unwrapping phase maps, obtained during the measurement of 3-D surfaces using sinusoidal structured light projection technique, is proposed. "Takeda's method" is used to obtain the wrapped phase map. Proposed method of unwrapping makes use of an additional image of the object captured under the illumination of a specifically designed color-coded pattern. The new approach demonstrates, for the first time, a method of producing reliable unwrapping of objects even with surface discontinuities from a single-phase map. It is shown to be significantly faster and reliable than temporal phase unwrapping procedure that uses a complete exponential sequence. For example, if a measurement with the accuracy obtained by interrogating the object with S fringes in the projected pattern is carried out with both the methods, new method requires only 2 frames as compared to (log(2)S +1) frames required by the later method.