991 resultados para Combining method


Relevância:

100.00% 100.00%

Publicador:

Resumo:

The performance of the maximum ratio combining method for the combining of antenna-diversity signals in correlated Rician-fading channels is rigorously studied. The distribution function of the normalized signal-to-noise ratio (SNR) is expanded in terms of a power series and calculated numerically. This power series can easily take into account the signal correlations and antenna gains and can be applied to any number of receiving antennas. An application of the method to dual-antenna diversity systems produces useful distribution curves for the normalized SNR which can be used to find the diversity gain. It is revealed that signal correlation in Rician-fading channels helps to increase the diversity gain rather than to decrease it as in the Rayleigh fading channels. It is also shown that with a relative strong direct signal component, the diversity gain can be much higher than that without a direct signal component.

Relevância:

70.00% 70.00%

Publicador:

Resumo:

A power combining strategy for Class-E and inverse Class-E amplifiers operating at high frequencies such that they can operate into unbalanced loads is proposed. This power combining method is particularly important for the inverse Class-E amplifier configuration whose single-stage topology is naturally limited for small-to-medium power applications. Design examples for the power combining synthesis of classical Class-E and then inverse Class-E amplifiers with specification 3 V-1.5 W-2.5 GHz are given. For this specification, it is shown that a three-branch combiner has a natural 50 V output impedance. The resulting circuits are simulated within Agilent Advanced Design Systems environment with good agreement to theoretical prediction. Further the performance of the proposed circuits when operated in a Linear amplification using Nonlinear Components transmitter configuration whereby two-branch amplifiers are driven with constant amplitude conjugate input phase signals is investigated.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

The Optimum-Path Forest (OPF) classifier is a recent and promising method for pattern recognition, with a fast training algorithm and good accuracy results. Therefore, the investigation of a combining method for this kind of classifier can be important for many applications. In this paper we report a fast method to combine OPF-based classifiers trained with disjoint training subsets. Given a fixed number of subsets, the algorithm chooses random samples, without replacement, from the original training set. Each subset accuracy is improved by a learning procedure. The final decision is given by majority vote. Experiments with simulated and real data sets showed that the proposed combining method is more efficient and effective than naive approach provided some conditions. It was also showed that OPF training step runs faster for a series of small subsets than for the whole training set. The combining scheme was also designed to support parallel or distributed processing, speeding up the procedure even more. © 2011 Springer-Verlag.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Purpose: To evaluate and compare the performance of Ripplet Type-1 transform and directional discrete cosine transform (DDCT) and their combinations for improved representation of MRI images while preserving its fine features such as edges along the smooth curves and textures. Methods: In a novel image representation method based on fusion of Ripplet type-1 and conventional/directional DCT transforms, source images were enhanced in terms of visual quality using Ripplet and DDCT and their various combinations. The enhancement achieved was quantified on the basis of peak signal to noise ratio (PSNR), mean square error (MSE), structural content (SC), average difference (AD), maximum difference (MD), normalized cross correlation (NCC), and normalized absolute error (NAE). To determine the attributes of both transforms, these transforms were combined to represent the entire image as well. All the possible combinations were tested to present a complete study of combinations of the transforms and the contrasts were evaluated amongst all the combinations. Results: While using the direct combining method (DDCT) first and then the Ripplet method, a PSNR value of 32.3512 was obtained which is comparatively higher than the PSNR values of the other combinations. This novel designed technique gives PSNR value approximately equal to the PSNR’s of parent techniques. Along with this, it was able to preserve edge information, texture information and various other directional image features. The fusion of DDCT followed by the Ripplet reproduced the best images. Conclusion: The transformation of images using Ripplet followed by DDCT ensures a more efficient method for the representation of images with preservation of its fine details like edges and textures.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

A computational impact analysis methodology has been developed, based on modal analysis and a local contact force-deflection model. The contact law is based on Hertz contact theory while contact stresses are elastic, defines a modified contact theory to take account of local permanent indentation, and considers elastic recovery during unloading. The model was validated experimentally through impact testing of glass-carbon hybrid braided composite panels. Specimens were mounted in a support frame and the contact force was inferred from the deceleration of the impactor, measured by high-speed photography. A Finite Element analysis of the panel and support frame assembly was performed to compute the modal responses. The new contact model performed well in predicting the peak forces and impact durations for moderate energy impacts (15 J), where contact stresses locally exceed the linear elastic limit and damage may be deemed to have occurred. C-scan measurements revealed substantial damage for impact energies in the range of 30-50 J. For this regime the new model predictions might be improved by characterisation of the contact law hysteresis during the unloading phase, and a modification of the elastic vibration response in line with damage levels acquired during the impact. © 2011 Elsevier Ltd. All rights reserved.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

The assessment of building energy efficiency is one of the most effective measures for reducing building energy consumption. This paper proposes a holistic method (HMEEB) for assessing and certifying building energy efficiency based on the D-S (Dempster-Shafer) theory of evidence and the Evidential Reasoning (ER) approach. HMEEB has three main features: (i) it provides both a method to assess and certify building energy efficiency, and exists as an analytical tool to identify improvement opportunities; (ii) it combines a wealth of information on building energy efficiency assessment, including identification of indicators and a weighting mechanism; and (iii) it provides a method to identify and deal with inherent uncertainties within the assessment procedure. This paper demonstrates the robustness, flexibility and effectiveness of the proposed method, using two examples to assess the energy efficiency of two residential buildings, both located in the ‘Hot Summer and Cold Winter’ zone in China. The proposed certification method provides detailed recommendations for policymakers in the context of carbon emission reduction targets and promoting energy efficiency in the built environment. The method is transferable to other countries and regions, using an indicator weighting system to modify local climatic, economic and social factors.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Particulate antigen assemblies in the nanometer range and DNA plasmids are particularly interesting for designing vaccines. We hypothesised that a combination of these approaches could result in a new delivery method of gp160 envelope HIV-1 vaccine which could combine the potency of virus-like particles (VLPs) and the simplicity of use of DNA vaccines. Characterisation of lentivirus-like particles (lentiVLPs) by western blot, dynamic light scattering and electron microscopy revealed that their protein pattern, size and structure make them promising candidates for HIV-1 vaccines. Although all particles were similar with regard to size and distribution, they clearly differed in p24 capsid protein content suggesting that Rev may be required for particle maturation and Gag processing. In vivo, lentiVLP pseudotyping with the gp160 envelope or with a combination of gp160 and VSV-G envelopes did not influence the magnitude of the immune response but the combination of lentiVLPs with Alum adjuvant resulted in a more potent response. Interestingly, the strongest immune response was obtained when plasmids encoding lentiVLPs were co-delivered to mice muscles by electrotransfer, suggesting that lentiVLPs were efficiently produced in vivo or the packaging genes mediate an adjuvant effect. DNA electrotransfer of plasmids encoding lentivirus-like particles offers many advantages and appears therefore as a promising delivery method of HIV-1 vaccines. Keywords:VLP, Electroporation, Electrotransfer, HIV vaccine, DNA vaccine

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Jakarta is vulnerable to flooding mainly caused by prolonged and heavy rainfall and thus a robust hydrological modeling is called for. A good quality of spatial precipitation data is therefore desired so that a good hydrological model could be achieved. Two types of rainfall sources are available: satellite and gauge station observations. At-site rainfall is considered to be a reliable and accurate source of rainfall. However, the limited number of stations makes the spatial interpolation not very much appealing. On the other hand, the gridded rainfall nowadays has high spatial resolution and improved accuracy, but still, relatively less accurate than its counterpart. To achieve a better precipitation data set, the study proposes cokriging method, a blending algorithm, to yield the blended satellite-gauge gridded rainfall at approximately 10-km resolution. The Global Satellite Mapping of Precipitation (GSMaP, 0.1⁰×0.1⁰) and daily rainfall observations from gauge stations are used. The blended product is compared with satellite data by cross-validation method. The newly-yield blended product is then utilized to re-calibrate the hydrological model. Several scenarios are simulated by the hydrological models calibrated by gauge observations alone and blended product. The performance of two calibrated hydrological models is then assessed and compared based on simulated and observed runoff.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

This paper reports on the development and optimization of a modified Quick, Easy, Cheap Effective, Rugged and Safe (QuEChERS) based extraction technique coupled with a clean-up dispersive-solid phase extraction (dSPE) as a new, reliable and powerful strategy to enhance the extraction efficiency of free low molecular-weight polyphenols in selected species of dietary vegetables. The process involves two simple steps. First, the homogenized samples are extracted and partitioned using an organic solvent and salt solution. Then, the supernatant is further extracted and cleaned using a dSPE technique. Final clear extracts of vegetables were concentrated under vacuum to near dryness and taken up into initial mobile phase (0.1% formic acid and 20% methanol). The separation and quantification of free low molecular weight polyphenols from the vegetable extracts was achieved by ultrahigh pressure liquid chromatography (UHPLC) equipped with a phodiode array (PDA) detection system and a Trifunctional High Strength Silica capillary analytical column (HSS T3), specially designed for polar compounds. The performance of the method was assessed by studying the selectivity, linear dynamic range, the limit of detection (LOD) and limit of quantification (LOQ), precision, trueness, and matrix effects. The validation parameters of the method showed satisfactory figures of merit. Good linearity (View the MathML sourceRvalues2>0.954; (+)-catechin in carrot samples) was achieved at the studied concentration range. Reproducibility was better than 3%. Consistent recoveries of polyphenols ranging from 78.4 to 99.9% were observed when all target vegetable samples were spiked at two concentration levels, with relative standard deviations (RSDs, n = 5) lower than 2.9%. The LODs and the LOQs ranged from 0.005 μg mL−1 (trans-resveratrol, carrot) to 0.62 μg mL−1 (syringic acid, garlic) and from 0.016 μg mL−1 (trans-resveratrol, carrot) to 0.87 μg mL−1 ((+)-catechin, carrot) depending on the compound. The method was applied for studying the occurrence of free low molecular weight polyphenols in eight selected dietary vegetables (broccoli, tomato, carrot, garlic, onion, red pepper, green pepper and beetroot), providing a valuable and promising tool for food quality evaluation.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Fundação de Amparo à Pesquisa do Estado de São Paulo (FAPESP)

Relevância:

40.00% 40.00%

Publicador:

Resumo:

In the biomedical studies, the general data structures have been the matched (paired) and unmatched designs. Recently, many researchers are interested in Meta-Analysis to obtain a better understanding from several clinical data of a medical treatment. The hybrid design, which is combined two data structures, may create the fundamental question for statistical methods and the challenges for statistical inferences. The applied methods are depending on the underlying distribution. If the outcomes are normally distributed, we would use the classic paired and two independent sample T-tests on the matched and unmatched cases. If not, we can apply Wilcoxon signed rank and rank sum test on each case. ^ To assess an overall treatment effect on a hybrid design, we can apply the inverse variance weight method used in Meta-Analysis. On the nonparametric case, we can use a test statistic which is combined on two Wilcoxon test statistics. However, these two test statistics are not in same scale. We propose the Hybrid Test Statistic based on the Hodges-Lehmann estimates of the treatment effects, which are medians in the same scale.^ To compare the proposed method, we use the classic meta-analysis T-test statistic on the combined the estimates of the treatment effects from two T-test statistics. Theoretically, the efficiency of two unbiased estimators of a parameter is the ratio of their variances. With the concept of Asymptotic Relative Efficiency (ARE) developed by Pitman, we show ARE of the hybrid test statistic relative to classic meta-analysis T-test statistic using the Hodges-Lemann estimators associated with two test statistics.^ From several simulation studies, we calculate the empirical type I error rate and power of the test statistics. The proposed statistic would provide effective tool to evaluate and understand the treatment effect in various public health studies as well as clinical trials.^

Relevância:

40.00% 40.00%

Publicador:

Resumo:

The boundary element method (BEM) has been applied successfully to many engineering problems during the last decades. Compared with domain type methods like the finite element method (FEM) or the finite difference method (FDM) the BEM can handle problems where the medium extends to infinity much easier than domain type methods as there is no need to develop special boundary conditions (quiet or absorbing boundaries) or infinite elements at the boundaries introduced to limit the domain studied. The determination of the dynamic stiffness of arbitrarily shaped footings is just one of these fields where the BEM has been the method of choice, especially in the 1980s. With the continuous development of computer technology and the available hardware equipment the size of the problems under study grew and, as the flop count for solving the resulting linear system of equations grows with the third power of the number of equations, there was a need for the development of iterative methods with better performance. In [1] the GMRES algorithm was presented which is now widely used for implementations of the collocation BEM. While the FEM results in sparsely populated coefficient matrices, the BEM leads, in general, to fully or densely populated ones, depending on the number of subregions, posing a serious memory problem even for todays computers. If the geometry of the problem permits the surface of the domain to be meshed with equally shaped elements a lot of the resulting coefficients will be calculated and stored repeatedly. The present paper shows how these unnecessary operations can be avoided reducing the calculation time as well as the storage requirement. To this end a similar coefficient identification algorithm (SCIA), has been developed and implemented in a program written in Fortran 90. The vertical dynamic stiffness of a single pile in layered soil has been chosen to test the performance of the implementation. The results obtained with the 3-d model may be compared with those obtained with an axisymmetric formulation which are considered to be the reference values as the mesh quality is much better. The entire 3D model comprises more than 35000 dofs being a soil region with 21168 dofs the biggest single region. Note that the memory necessary to store all coefficients of this single region is about 6.8 GB, an amount which is usually not available with personal computers. In the problem under study the interface zone between the two adjacent soil regions as well as the surface of the top layer may be meshed with equally sized elements. In this case the application of the SCIA leads to an important reduction in memory requirements. The maximum memory used during the calculation has been reduced to 1.2 GB. The application of the SCIA thus permits problems to be solved on personal computers which otherwise would require much more powerful hardware.