95 resultados para CONVEX
Resumo:
The purpose of this article is to consider two themes, both of which emanate from and involve the Kobayashi and the Carath,odory metric. First, we study the biholomorphic invariant introduced by B. Fridman on strongly pseudoconvex domains, on weakly pseudoconvex domains of finite type in C (2), and on convex finite type domains in C (n) using the scaling method. Applications include an alternate proof of the Wong-Rosay theorem, a characterization of analytic polyhedra with noncompact automorphism group when the orbit accumulates at a singular boundary point, and a description of the Kobayashi balls on weakly pseudoconvex domains of finite type in C (2) and convex finite type domains in C (n) in terms of Euclidean parameters. Second, a version of Vitushkin's theorem about the uniform extendability of a compact subgroup of automorphisms of a real analytic strongly pseudoconvex domain is proved for C (1)-isometries of the Kobayashi and Carath,odory metrics on a smoothly bounded strongly pseudoconvex domain.
Resumo:
We consider functions that map the open unit disc conformally onto the complement of an unbounded convex set with opening angle pa, a ? (1, 2], at infinity. In this paper, we show that every such function is close-to-convex of order (a - 1) and is included in the set of univalent functions of bounded boundary rotation. Many interesting consequences of this result are obtained. We also determine the extreme points of the set of concave functions with respect to the linear structure of the Hornich space.
Resumo:
Let be a smooth real surface in and let be a point at which the tangent plane is a complex line. How does one determine whether or not is locally polynomially convex at such a p-i.e. at a CR singularity? Even when the order of contact of with at p equals 2, no clean characterisation exists; difficulties are posed by parabolic points. Hence, we study non-parabolic CR singularities. We show that the presence or absence of Bishop discs around certain non-parabolic CR singularities is completely determined by a Maslov-type index. This result subsumes all known facts about Bishop discs around order-two, non-parabolic CR singularities. Sufficient conditions for Bishop discs have earlier been investigated at CR singularities having high order of contact with . These results relied upon a subharmonicity condition, which fails in many simple cases. Hence, we look beyond potential theory and refine certain ideas going back to Bishop.
Resumo:
In this paper we study the problem of designing SVM classifiers when the kernel matrix, K, is affected by uncertainty. Specifically K is modeled as a positive affine combination of given positive semi definite kernels, with the coefficients ranging in a norm-bounded uncertainty set. We treat the problem using the Robust Optimization methodology. This reduces the uncertain SVM problem into a deterministic conic quadratic problem which can be solved in principle by a polynomial time Interior Point (IP) algorithm. However, for large-scale classification problems, IP methods become intractable and one has to resort to first-order gradient type methods. The strategy we use here is to reformulate the robust counterpart of the uncertain SVM problem as a saddle point problem and employ a special gradient scheme which works directly on the convex-concave saddle function. The algorithm is a simplified version of a general scheme due to Juditski and Nemirovski (2011). It achieves an O(1/T-2) reduction of the initial error after T iterations. A comprehensive empirical study on both synthetic data and real-world protein structure data sets show that the proposed formulations achieve the desired robustness, and the saddle point based algorithm outperforms the IP method significantly.
Resumo:
We address the problem of identifying the constituent sources in a single-sensor mixture signal consisting of contributions from multiple simultaneously active sources. We propose a generic framework for mixture signal analysis based on a latent variable approach. The basic idea of the approach is to detect known sources represented as stochastic models, in a single-channel mixture signal without performing signal separation. A given mixture signal is modeled as a convex combination of known source models and the weights of the models are estimated using the mixture signal. We show experimentally that these weights indicate the presence/absence of the respective sources. The performance of the proposed approach is illustrated through mixture speech data in a reverberant enclosure. For the task of identifying the constituent speakers using data from a single microphone, the proposed approach is able to identify the dominant source with up to 8 simultaneously active background sources in a room with RT60 = 250 ms, using models obtained from clean speech data for a Source to Interference Ratio (SIR) greater than 2 dB.
Resumo:
The n-interior point variant of the Erdos-Szekeres problem is to show the following: For any n, n-1, every point set in the plane with sufficient number of interior points contains a convex polygon containing exactly n-interior points. This has been proved only for n-3. In this paper, we prove it for pointsets having atmost logarithmic number of convex layers. We also show that any pointset containing atleast n interior points, there exists a 2-convex polygon that contains exactly n-interior points.
Resumo:
High-level loop transformations are a key instrument in mapping computational kernels to effectively exploit the resources in modern processor architectures. Nevertheless, selecting required compositions of loop transformations to achieve this remains a significantly challenging task; current compilers may be off by orders of magnitude in performance compared to hand-optimized programs. To address this fundamental challenge, we first present a convex characterization of all distinct, semantics-preserving, multidimensional affine transformations. We then bring together algebraic, algorithmic, and performance analysis results to design a tractable optimization algorithm over this highly expressive space. Our framework has been implemented and validated experimentally on a representative set of benchmarks running on state-of-the-art multi-core platforms.
Resumo:
In many real world prediction problems the output is a structured object like a sequence or a tree or a graph. Such problems range from natural language processing to compu- tational biology or computer vision and have been tackled using algorithms, referred to as structured output learning algorithms. We consider the problem of structured classifi- cation. In the last few years, large margin classifiers like sup-port vector machines (SVMs) have shown much promise for structured output learning. The related optimization prob -lem is a convex quadratic program (QP) with a large num-ber of constraints, which makes the problem intractable for large data sets. This paper proposes a fast sequential dual method (SDM) for structural SVMs. The method makes re-peated passes over the training set and optimizes the dual variables associated with one example at a time. The use of additional heuristics makes the proposed method more efficient. We present an extensive empirical evaluation of the proposed method on several sequence learning problems.Our experiments on large data sets demonstrate that the proposed method is an order of magnitude faster than state of the art methods like cutting-plane method and stochastic gradient descent method (SGD). Further, SDM reaches steady state generalization performance faster than the SGD method. The proposed SDM is thus a useful alternative for large scale structured output learning.
Resumo:
This paper extends some geometric properties of a one-parameter family of relative entropies. These arise as redundancies when cumulants of compressed lengths are considered instead of expected compressed lengths. These parametric relative entropies are a generalization of the Kullback-Leibler divergence. They satisfy the Pythagorean property and behave like squared distances. This property, which was known for finite alphabet spaces, is now extended for general measure spaces. Existence of projections onto convex and certain closed sets is also established. Our results may have applications in the Rényi entropy maximization rule of statistical physics.
Resumo:
The interaction between the digital human model (DHM) and environment typically occurs in two distinct modes; one, when the DHM maintains contacts with the environment using its self weight, wherein associated reaction forces at the interface due to gravity are unidirectional; two, when the DHM applies both tension and compression on the environment through anchoring. For static balancing in first mode of interaction, it is sufficient to maintain the projection of the centre of mass (COM) inside the convex region induced by the weight supporting segments of the body on a horizontal plane. In DHM, static balancing is required while performing specified tasks such as reach, manipulation and locomotion; otherwise the simulations would not be realistic. This paper establishes the geometric relationships that must be satisfied for maintaining static balance while altering the support configurations for a given posture and altering the posture for a given support condition. For a given location of the COM for a system supported by multiple point contacts, the conditions for simultaneous withdrawal of a specified set of contacts have been determined in terms of the convex hulls of the subsets of the points of contact. When the projection of COM must move beyond the existing support for performing some task, new supports must be enabled for maintaining static balance. This support seeking behavior could also manifest while planning for reduction of support stresses. Feasibility of such a support depends upon the availability of necessary features in the environment. Geometric conditions necessary for selection of new support on horizontal,inclined and vertical surfaces within the workspace of the DHM for such dynamic scenario have been derived. The concepts developed are demonstrated using the cases of sit-to-stand posture transition for manipulation of COM within the convex supporting polygon, and statically stable walking gaits for support seeking within the kinematic capabilities of the DHM. The theory developed helps in making the DHM realize appropriate behaviors in diverse scenarios autonomously.
Resumo:
We address the problem of mining targeted association rules over multidimensional market-basket data. Here, each transaction has, in addition to the set of purchased items, ancillary dimension attributes associated with it. Based on these dimensions, transactions can be visualized as distributed over cells of an n-dimensional cube. In this framework, a targeted association rule is of the form {X -> Y} R, where R is a convex region in the cube and X. Y is a traditional association rule within region R. We first describe the TOARM algorithm, based on classical techniques, for identifying targeted association rules. Then, we discuss the concepts of bottom-up aggregation and cubing, leading to the CellUnion technique. This approach is further extended, using notions of cube-count interleaving and credit-based pruning, to derive the IceCube algorithm. Our experiments demonstrate that IceCube consistently provides the best execution time performance, especially for large and complex data cubes.
Resumo:
Chebyshev-inequality-based convex relaxations of Chance-Constrained Programs (CCPs) are shown to be useful for learning classifiers on massive datasets. In particular, an algorithm that integrates efficient clustering procedures and CCP approaches for computing classifiers on large datasets is proposed. The key idea is to identify high density regions or clusters from individual class conditional densities and then use a CCP formulation to learn a classifier on the clusters. The CCP formulation ensures that most of the data points in a cluster are correctly classified by employing a Chebyshev-inequality-based convex relaxation. This relaxation is heavily dependent on the second-order statistics. However, this formulation and in general such relaxations that depend on the second-order moments are susceptible to moment estimation errors. One of the contributions of the paper is to propose several formulations that are robust to such errors. In particular a generic way of making such formulations robust to moment estimation errors is illustrated using two novel confidence sets. An important contribution is to show that when either of the confidence sets is employed, for the special case of a spherical normal distribution of clusters, the robust variant of the formulation can be posed as a second-order cone program. Empirical results show that the robust formulations achieve accuracies comparable to that with true moments, even when moment estimates are erroneous. Results also illustrate the benefits of employing the proposed methodology for robust classification of large-scale datasets.
Resumo:
The paper reports exchange-spring soft and hard ferrite nanocomposites synthesized by chemical co-precipitation with or without the application of ultrasonic vibration. The composites contained BaFe12O19 as the hard phase and CoFe2O4/MgFe2O4 as the soft phase. X-ray diffraction patterns of the samples in the optimum calcined condition indicated the presence of soft ferrites as face-centred cubic (fcc) and hard ferrites as hexagonal close packed (hcp) structure respectively. Temperature dependence of magnetization in the range of 20-700 degrees C demonstrated distinct presence of soft and hard ferrites as magnetic phases which are characterized by wide difference in magnetic anisotropy and coercivity. Exchange-spring mechanism led these nanocomposite systems to exchange-coupled, which ultimately produced convex hysteresis loops characteristic of a single-phase permanent magnet. Fairly high value of coercivity and maximum energy product were observed for the samples in the optimum calcined conditions with a maximum applied field of 1600 kA/m (2 T).
Resumo:
We consider the MIMO X channel (XC), a system consisting of two transmit-receive pairs, where each transmitter communicates with both the receivers. Both the transmitters and receivers are equipped with multiple antennas. First, we derive an upper bound on the sum-rate capacity of the MIMO XC under individual power constraint at each transmitter. The sum-rate capacity of the two-user multiple access channel (MAC) that results when receiver cooperation is assumed forms an upper bound on the sum-rate capacity of the MIMO XC. We tighten this bound by considering noise correlation between the receivers and deriving the worst noise covariance matrix. It is shown that the worst noise covariance matrix is a saddle-point of a zero-sum, two-player convex-concave game, which is solved through a primal-dual interior point method that solves the maximization and the minimization parts of the problem simultaneously. Next, we propose an achievable scheme which employs dirty paper coding at the transmitters and successive decoding at the receivers. We show that the derived upper bound is close to the achievable region of the proposed scheme at low to medium SNRs.
Resumo:
In this paper, we evaluate secrecy rates in cooperative relay beamforming in the presence of imperfect channel state information (CSI) and multiple eavesdroppers. A source-destination pair aided by.. out of.. relays, 1 <= k <= M, using decode-and-forward relay beamforming is considered. We compute the worst case secrecy rate with imperfect CSI in the presence of multiple eavesdroppers, where the number of eavesdroppers can be more than the number of relays. We solve the optimization problem for all possible relay combinations to find the secrecy rate and optimum source and relay weights subject to a total power constraint. We relax the rank-1 constraint on the complex semi-definite relay weight matrix and use S-procedure to reformulate the optimization problem that can be solved using convex semi-definite programming.