901 resultados para weights of ideals
Resumo:
Las recientes crisis económicas que ha vivido el mundo nos llevan a hacer una reflexión en torno a las responsabilidades que tienen los dirigentes empresariales en los cambios económicos e incluso sociales que se están viviendo. Es por tanto el momento para hacer un análisis más profundo del aporte que la administración, como ciencia, ha hecho a la sociedad. Se ha identificado que la administración permite el desarrollo y crecimiento de un individuo, al tiempo que es un medio de crecimiento y desarrollo para una comunidad, una región y una cultura, lo que se convierte en una realidad indiscutible que nos lleva a reflexionar acerca del porqué, ocurren problemas como las quiebras, ausencia de recursos, y conflictos internos en las organizaciones. En esa medida, una reflexión puede ser que los ideales que ha venido liderando la administración han sido desvirtuados y apartados de una moral y ética empresarial que permita al directivo pensar en algo más que utilidades. Podría parecer extraño decir que la empresa y la acción gerencial deben ir más allá. Podríamos iniciar con el análisis de la obsolescencia de los ideales administrativos o incluso en la ausencia de ideales en la administración. Sobre si se está teniendo en cuenta que la administración que conocemos y aprendemos sigue vigente en un mundo que ha evolucionado y cambiado, el mundo de hoy es distinto al que conocimos incluso en nuestra época de formación profesional, pues se nos enseñó a pensar linealmente, causa y efecto, todo cumple un orden y un plan predeterminado, hoy la visión de certidumbre y control del hombre sobre las decisiones son guiadas por el azar, y por más que el hombre trabaja y diseña estrategias, éstas serán transformadas de acuerdo a las emergencias. Por lo anterior, es importante entender que para avanzar y dejar la obsolescencia en los ideales que ha definido la administración, los directivos deben cambiar su visón y entender la dinámica de interrelación de sus organizaciones en un mundo en permanente caos. Debemos estudiar desde dinámicas diferentes a la de funcionalidad, y saber que es momento para que la administración tenga una nueva visión de la gestión administrativa, por lo que debemos ampliar nuestro horizonte y ver a través de la complejidad. Las decisiones que toman los directivos, las estrategias que definen y la forma de relacionarse con la competencia en mercados altamente competitivos requiere de un cambio en la administración y de un marco ético muy distinto, debe concebirse la gestión directiva y administrativa para mantener y conservar la vida del sector, de la empresa y, por ende, mantener el medio que nos rodea. Por ello se propone ejercer la administración desde la bioética.
Resumo:
Simple Adaptive Momentum [1] was introduced as a simple means of speeding the training of multi-layer perceptrons (MLPs) by changing the momentum term depending on the angle between the current and previous changes in the weights of the MLP. In the original paper. the weight changes of the whole network are used in determining this angle. This paper considers adapting the momentum term using certain subsets of these weights. This idea was inspired by the author's object oriented approach to programming MLPs. successfully used in teaching students: this approach is also described. It is concluded that the angle is best determined using the weight changes in each layer separately.
Resumo:
A novel sparse kernel density estimator is derived based on a regression approach, which selects a very small subset of significant kernels by means of the D-optimality experimental design criterion using an orthogonal forward selection procedure. The weights of the resulting sparse kernel model are calculated using the multiplicative nonnegative quadratic programming algorithm. The proposed method is computationally attractive, in comparison with many existing kernel density estimation algorithms. Our numerical results also show that the proposed method compares favourably with other existing methods, in terms of both test accuracy and model sparsity, for constructing kernel density estimates.
Resumo:
Using the classical Parzen window (PW) estimate as the desired response, the kernel density estimation is formulated as a regression problem and the orthogonal forward regression technique is adopted to construct sparse kernel density (SKD) estimates. The proposed algorithm incrementally minimises a leave-one-out test score to select a sparse kernel model, and a local regularisation method is incorporated into the density construction process to further enforce sparsity. The kernel weights of the selected sparse model are finally updated using the multiplicative nonnegative quadratic programming algorithm, which ensures the nonnegative and unity constraints for the kernel weights and has the desired ability to reduce the model size further. Except for the kernel width, the proposed method has no other parameters that need tuning, and the user is not required to specify any additional criterion to terminate the density construction procedure. Several examples demonstrate the ability of this simple regression-based approach to effectively construct a SKID estimate with comparable accuracy to that of the full-sample optimised PW density estimate. (c) 2007 Elsevier B.V. All rights reserved.
Resumo:
A new probabilistic neural network (PNN) learning algorithm based on forward constrained selection (PNN-FCS) is proposed. An incremental learning scheme is adopted such that at each step, new neurons, one for each class, are selected from the training samples arid the weights of the neurons are estimated so as to minimize the overall misclassification error rate. In this manner, only the most significant training samples are used as the neurons. It is shown by simulation that the resultant networks of PNN-FCS have good classification performance compared to other types of classifiers, but much smaller model sizes than conventional PNN.
Resumo:
A sparse kernel density estimator is derived based on the zero-norm constraint, in which the zero-norm of the kernel weights is incorporated to enhance model sparsity. The classical Parzen window estimate is adopted as the desired response for density estimation, and an approximate function of the zero-norm is used for achieving mathemtical tractability and algorithmic efficiency. Under the mild condition of the positive definite design matrix, the kernel weights of the proposed density estimator based on the zero-norm approximation can be obtained using the multiplicative nonnegative quadratic programming algorithm. Using the -optimality based selection algorithm as the preprocessing to select a small significant subset design matrix, the proposed zero-norm based approach offers an effective means for constructing very sparse kernel density estimates with excellent generalisation performance.
Resumo:
A generalized or tunable-kernel model is proposed for probability density function estimation based on an orthogonal forward regression procedure. Each stage of the density estimation process determines a tunable kernel, namely, its center vector and diagonal covariance matrix, by minimizing a leave-one-out test criterion. The kernel mixing weights of the constructed sparse density estimate are finally updated using the multiplicative nonnegative quadratic programming algorithm to ensure the nonnegative and unity constraints, and this weight-updating process additionally has the desired ability to further reduce the model size. The proposed tunable-kernel model has advantages, in terms of model generalization capability and model sparsity, over the standard fixed-kernel model that restricts kernel centers to the training data points and employs a single common kernel variance for every kernel. On the other hand, it does not optimize all the model parameters together and thus avoids the problems of high-dimensional ill-conditioned nonlinear optimization associated with the conventional finite mixture model. Several examples are included to demonstrate the ability of the proposed novel tunable-kernel model to effectively construct a very compact density estimate accurately.
Resumo:
This paper derives an efficient algorithm for constructing sparse kernel density (SKD) estimates. The algorithm first selects a very small subset of significant kernels using an orthogonal forward regression (OFR) procedure based on the D-optimality experimental design criterion. The weights of the resulting sparse kernel model are then calculated using a modified multiplicative nonnegative quadratic programming algorithm. Unlike most of the SKD estimators, the proposed D-optimality regression approach is an unsupervised construction algorithm and it does not require an empirical desired response for the kernel selection task. The strength of the D-optimality OFR is owing to the fact that the algorithm automatically selects a small subset of the most significant kernels related to the largest eigenvalues of the kernel design matrix, which counts for the most energy of the kernel training data, and this also guarantees the most accurate kernel weight estimate. The proposed method is also computationally attractive, in comparison with many existing SKD construction algorithms. Extensive numerical investigation demonstrates the ability of this regression-based approach to efficiently construct a very sparse kernel density estimate with excellent test accuracy, and our results show that the proposed method compares favourably with other existing sparse methods, in terms of test accuracy, model sparsity and complexity, for constructing kernel density estimates.
Resumo:
In financial decision-making processes, the adopted weights of the objective functions have significant impacts on the final decision outcome. However, conventional rating and weighting methods exhibit difficulty in deriving appropriate weights for complex decision-making problems with imprecise information. Entropy is a quantitative measure of uncertainty and has been useful in exploring weights of attributes in decision making. A fuzzy and entropy-based mathematical approach is employed to solve the weighting problem of the objective functions in an overall cash-flow model. The multiproject being undertaken by a medium-size construction firm in Hong Kong was used as a real case study to demonstrate the application of entropy. Its application in multiproject cash flow situations is demonstrated. The results indicate that the overall before-tax profit was HK$ 0.11 millions lower after the introduction of appropriate weights. In addition, the best time to invest in new projects arising from positive cash flow was identified to be two working months earlier than the nonweight system.
Resumo:
A new structure of Radial Basis Function (RBF) neural network called the Dual-orthogonal RBF Network (DRBF) is introduced for nonlinear time series prediction. The hidden nodes of a conventional RBF network compare the Euclidean distance between the network input vector and the centres, and the node responses are radially symmetrical. But in time series prediction where the system input vectors are lagged system outputs, which are usually highly correlated, the Euclidean distance measure may not be appropriate. The DRBF network modifies the distance metric by introducing a classification function which is based on the estimation data set. Training the DRBF networks consists of two stages. Learning the classification related basis functions and the important input nodes, followed by selecting the regressors and learning the weights of the hidden nodes. In both cases, a forward Orthogonal Least Squares (OLS) selection procedure is applied, initially to select the important input nodes and then to select the important centres. Simulation results of single-step and multi-step ahead predictions over a test data set are included to demonstrate the effectiveness of the new approach.
Resumo:
A new sparse kernel probability density function (pdf) estimator based on zero-norm constraint is constructed using the classical Parzen window (PW) estimate as the target function. The so-called zero-norm of the parameters is used in order to achieve enhanced model sparsity, and it is suggested to minimize an approximate function of the zero-norm. It is shown that under certain condition, the kernel weights of the proposed pdf estimator based on the zero-norm approximation can be updated using the multiplicative nonnegative quadratic programming algorithm. Numerical examples are employed to demonstrate the efficacy of the proposed approach.
Resumo:
This study was undertaken to explore gel permeation chromatography (GPC) for estimating molecular weights of proanthocyanidin fractions isolated from sainfoin (Onobrychis viciifolia). The results were compared with data obtained by thiolytic degradation of the same fractions. Polystyrene, polyethylene glycol and polymethyl methacrylate standards were not suitable for estimating the molecular weights of underivatized proanthocyanidins. Therefore, a novel HPLC-GPC method was developed based on two serially connected PolarGel-L columns using DMF that contained 5% water, 1% acetic acid and 0.15 M LiBr at 0.7 ml/min and 50 degrees C. This yielded a single calibration curve for galloyl glucoses (trigalloyl glucose, pentagalloyl glucose), ellagitannins (pedunculagin, vescalagin, punicalagin, oenothein B, gemin A), proanthocyanidins (procyanidin B2, cinnamtannin B1), and several other polyphenols (catechin, epicatechin gallate, epicallocatechin gallate, amentoflavone). These GPC predicted molecular weights represented a considerable advance over previously reported HPLC-GPC methods for underivatized proanthocyanidins. (C) 2011 Elsevier B.V. All rights reserved.
Resumo:
Abstract. We prove that the vast majority of JC∗-triples satisfy the condition of universal reversibility. Our characterisation is that a JC∗-triple is universally reversible if and only if it has no triple homomorphisms onto Hilbert spaces of dimension greater than two nor onto spin factors of dimension greater than four. We establish corresponding characterisations in the cases of JW∗-triples and of TROs (regarded as JC∗-triples). We show that the distinct natural operator space structures on a universally reversible JC∗-triple E are in bijective correspondence with a distinguished class of ideals in its universal TRO, identify the Shilov boundaries of these operator spaces and prove that E has a unique natural operator space structure precisely when E contains no ideal isometric to a nonabelian TRO. We deduce some decomposition and completely contractive properties of triple homomorphisms on TROs.
Resumo:
In this paper, we propose a novel online modeling algorithm for nonlinear and nonstationary systems using a radial basis function (RBF) neural network with a fixed number of hidden nodes. Each of the RBF basis functions has a tunable center vector and an adjustable diagonal covariance matrix. A multi-innovation recursive least square (MRLS) algorithm is applied to update the weights of RBF online, while the modeling performance is monitored. When the modeling residual of the RBF network becomes large in spite of the weight adaptation, a node identified as insignificant is replaced with a new node, for which the tunable center vector and diagonal covariance matrix are optimized using the quantum particle swarm optimization (QPSO) algorithm. The major contribution is to combine the MRLS weight adaptation and QPSO node structure optimization in an innovative way so that it can track well the local characteristic in the nonstationary system with a very sparse model. Simulation results show that the proposed algorithm has significantly better performance than existing approaches.
Resumo:
Document design and typeface design: A typographic specification for a new Intermediate Greek-English Lexicon by CUP, accompanied by typefaces modified for the specific typographic requirements of the text. The Lexicon is a substantial (over 1400 pages) publication for HE students and academics intended to complement Liddell-Scott (the standard reference for classical Greek since the 1850s), and has been in preparation for over a decade. The typographic appearance of such works has changed very little since the original editions, largely to the lack of suitable typefaces: early digital proofs of the Lexicon utilised directly digitised versions of historical typefaces, making the entries difficult to navigate, and the document uneven in typographic texture. Close collaboration with the editors of the Lexicon, and discussion of the historical precedents for such documents informed the design at all typographic levels to achieve a highly reader-friendly results that propose a model for this kind of typography. Uniquely for a work of this kind, typeface design decisions were integrated into the wider document design specification. A rethinking of the complex typography for Greek and English based on historical editions as well as equivalent bilingual reference works at this level (from OUP, CUP, Brill, Mondadori, and other publishers) led a redefinition of multi-script typeface pairing for the specific context, taking into account recent developments in typeface design. Specifically, the relevant weighting of elements within each entry were redefined, as well as the typographic texture of type styles across the two scripts. In details, Greek typefaces were modified to emphasise clarity and readability, particularly of diacritics, at very small sizes. The relative weights of typefaces typeset side-by-side were fine-tuned so that the visual hierarchy of the entires was unambiguous despite the dense typesetting.