912 resultados para Piecewise constant argument
Resumo:
Im Mittelpunkt dieser Arbeit steht Beweis der Existenz- und Eindeutigkeit von Quadraturformeln, die für das Qualokationsverfahren geeignet sind. Letzteres ist ein von Sloan, Wendland und Chandler entwickeltes Verfahren zur numerischen Behandlung von Randintegralgleichungen auf glatten Kurven (allgemeiner: periodische Pseudodifferentialgleichungen). Es erreicht die gleichen Konvergenzordnungen wie das Petrov-Galerkin-Verfahren, wenn man durch den Operator bestimmte Quadraturformeln verwendet. Zunächst werden die hier behandelten Pseudodifferentialoperatoren und das Qualokationsverfahren vorgestellt. Anschließend wird eine Theorie zur Existenz und Eindeutigkeit von Quadraturformeln entwickelt. Ein wesentliches Hilfsmittel hierzu ist die hier bewiesene Verallgemeinerung eines Satzes von Nürnberger über die Existenz und Eindeutigkeit von Quadraturformeln mit positiven Gewichten, die exakt für Tschebyscheff-Räume sind. Es wird schließlich gezeigt, dass es stets eindeutig bestimmte Quadraturformeln gibt, welche die in den Arbeiten von Sloan und Wendland formulierten Bedingungen erfüllen. Desweiteren werden 2-Punkt-Quadraturformeln für so genannte einfache Operatoren bestimmt, mit welchen das Qualokationsverfahren mit einem Testraum von stückweise konstanten Funktionen eine höhere Konvergenzordnung hat. Außerdem wird gezeigt, dass es für nicht-einfache Operatoren im Allgemeinen keine Quadraturformel gibt, mit der die Konvergenzordnung höher als beim Petrov-Galerkin-Verfahren ist. Das letzte Kapitel beinhaltet schließlich numerische Tests mit Operatoren mit konstanten und variablen Koeffizienten, welche die theoretischen Ergebnisse der vorangehenden Kapitel bestätigen.
Resumo:
This paper proposes Poisson log-linear multilevel models to investigate population variability in sleep state transition rates. We specifically propose a Bayesian Poisson regression model that is more flexible, scalable to larger studies, and easily fit than other attempts in the literature. We further use hierarchical random effects to account for pairings of individuals and repeated measures within those individuals, as comparing diseased to non-diseased subjects while minimizing bias is of epidemiologic importance. We estimate essentially non-parametric piecewise constant hazards and smooth them, and allow for time varying covariates and segment of the night comparisons. The Bayesian Poisson regression is justified through a re-derivation of a classical algebraic likelihood equivalence of Poisson regression with a log(time) offset and survival regression assuming piecewise constant hazards. This relationship allows us to synthesize two methods currently used to analyze sleep transition phenomena: stratified multi-state proportional hazards models and log-linear models with GEE for transition counts. An example data set from the Sleep Heart Health Study is analyzed.
Resumo:
Since 2010, the client base of online-trading service providers has grown significantly. Such companies enable small investors to access the stock market at advantageous rates. Because small investors buy and sell stocks in moderate amounts, they should consider fixed transaction costs, integral transaction units, and dividends when selecting their portfolio. In this paper, we consider the small investor’s problem of investing capital in stocks in a way that maximizes the expected portfolio return and guarantees that the portfolio risk does not exceed a prescribed risk level. Portfolio-optimization models known from the literature are in general designed for institutional investors and do not consider the specific constraints of small investors. We therefore extend four well-known portfolio-optimization models to make them applicable for small investors. We consider one nonlinear model that uses variance as a risk measure and three linear models that use the mean absolute deviation from the portfolio return, the maximum loss, and the conditional value-at-risk as risk measures. We extend all models to consider piecewise-constant transaction costs, integral transaction units, and dividends. In an out-of-sample experiment based on Swiss stock-market data and the cost structure of the online-trading service provider Swissquote, we apply both the basic models and the extended models; the former represent the perspective of an institutional investor, and the latter the perspective of a small investor. The basic models compute portfolios that yield on average a slightly higher return than the portfolios computed with the extended models. However, all generated portfolios yield on average a higher return than the Swiss performance index. There are considerable differences between the four risk measures with respect to the mean realized portfolio return and the standard deviation of the realized portfolio return.
Resumo:
Since 2010, the client base of online-trading service providers has grown significantly. Such companies enable small investors to access the stock market at advantageous rates. Because small investors buy and sell stocks in moderate amounts, they should consider fixed transaction costs, integral transaction units, and dividends when selecting their portfolio. In this paper, we consider the small investor’s problem of investing capital in stocks in a way that maximizes the expected portfolio return and guarantees that the portfolio risk does not exceed a prescribed risk level. Portfolio-optimization models known from the literature are in general designed for institutional investors and do not consider the specific constraints of small investors. We therefore extend four well-known portfolio-optimization models to make them applicable for small investors. We consider one nonlinear model that uses variance as a risk measure and three linear models that use the mean absolute deviation from the portfolio return, the maximum loss, and the conditional value-at-risk as risk measures. We extend all models to consider piecewise-constant transaction costs, integral transaction units, and dividends. In an out-of-sample experiment based on Swiss stock-market data and the cost structure of the online-trading service provider Swissquote, we apply both the basic models and the extended models; the former represent the perspective of an institutional investor, and the latter the perspective of a small investor. The basic models compute portfolios that yield on average a slightly higher return than the portfolios computed with the extended models. However, all generated portfolios yield on average a higher return than the Swiss performance index. There are considerable differences between the four risk measures with respect to the mean realized portfolio return and the standard deviation of the realized portfolio return.
Resumo:
En esta Tesis Doctoral se aborda la utilización de filtros de difusión no lineal para obtener imágenes constantes a trozos como paso previo al proceso de segmentación. En una primera parte se propone un formulación intrínseca para la ecuación de difusión no lineal que proporcione las condiciones de diseño necesarias sobre los filtros de difusión. A partir del marco teórico propuesto, se proporciona una nueva familia de difusividades; éstas son obtenidas a partir de técnicas de difusión no lineal relacionadas con los procesos de difusión regresivos. El objetivo es descomponer la imagen en regiones cerradas que sean homogéneas en sus niveles de grises sin contornos difusos. Asimismo, se prueba que la función de difusividad propuesta satisface las condiciones de un correcto planteamiento semi-discreto. Esto muestra que mediante el esquema semi-implícito habitualmente utilizado, realmente se hace un proceso de difusión no lineal directa, en lugar de difusión inversa, conectando con proceso de preservación de bordes. Bajo estas condiciones establecidas, se plantea un criterio de parada para el proceso de difusión, para obtener imágenes constantes a trozos con un bajo coste computacional. Una vez aplicado todo el proceso al caso unidimensional, se extienden los resultados teóricos, al caso de imágenes en 2D y 3D. Para el caso en 3D, se detalla el esquema numérico para el problema evolutivo no lineal, con condiciones de contorno Neumann homogéneas. Finalmente, se prueba el filtro propuesto para imágenes reales en 2D y 3D y se ilustran los resultados de la difusividad propuesta como método para obtener imágenes constantes a trozos. En el caso de imágenes 3D, se aborda la problemática del proceso previo a la segmentación del hígado, mediante imágenes reales provenientes de Tomografías Axiales Computarizadas (TAC). En ese caso, se obtienen resultados sobre la estimación de los parámetros de la función de difusividad propuesta. This Ph.D. Thesis deals with the case of using nonlinear diffusion filters to obtain piecewise constant images as a previous process for segmentation techniques. I have first shown an intrinsic formulation for the nonlinear diffusion equation to provide some design conditions on the diffusion filters. According to this theoretical framework, I have proposed a new family of diffusivities; they are obtained from nonlinear diffusion techniques and are related with backward diffusion. Their goal is to split the image in closed contours with a homogenized grey intensity inside and with no blurred edges. It has also proved that the proposed filters satisfy the well-posedness semi-discrete and full discrete scale-space requirements. This shows that by using semi-implicit schemes, a forward nonlinear diffusion equation is solved, instead of a backward nonlinear diffusion equation, connecting with an edgepreserving process. Under the conditions established for the diffusivity and using a stopping criterion I for the diffusion time, I have obtained piecewise constant images with a low computational effort. The whole process in the one-dimensional case is extended to the case where 2D and 3D theoretical results are applied to real images. For 3D, develops in detail the numerical scheme for nonlinear evolutionary problem with homogeneous Neumann boundary conditions. Finally, I have tested the proposed filter with real images for 2D and 3D and I have illustrated the effects of the proposed diffusivity function as a method to get piecewise constant images. For 3D I have developed a preprocess for liver segmentation with real images from CT (Computerized Tomography). In this case, I have obtained results on the estimation of the parameters of the given diffusivity function.
Resumo:
En este trabajo se expone la formulación del B.I.E.M. (en problemas de potencial)con elementos mixtos, que representan una interpolación lineal en la función de campo y constante en su derivada. El objetivo primordial de dicha formulación, es el soslayar los problemas que se presentan con los planteamientos anteriores, cuando existen puntos singulares. Los ejemplos que se incluyen, resueltos mediante un programa desarrollado en un miniordenador, confirman que el método propuesto consigue mejores resultados, sin ninguna complicación adicional sobre la formulación general, y con un ahorro significativo en cuanto al número de incógnitas y ecuaciones que se han de resolver = In this paper the mixed B.I.E.M. for bidimensional potential problems is presented. The interpolation of the field variable is done by piecewise constant one. The main idea is to swep out the solution the parasitic disturbances introduced near singular points by a finite value although an important byproduct is the possibility of conexion with domain methods, as F.E.M., in which the same kind of interpolation is worked. Results are very good, as shown by the exemples, and also interesting is the reduction in computer time comparatively to the classical B.l.E.M approach.
Resumo:
A method has been constructed for the solution of a wide range of chemical plant simulation models including differential equations and optimization. Double orthogonal collocation on finite elements is applied to convert the model into an NLP problem that is solved either by the VF 13AD package based on successive quadratic programming, or by the GRG2 package, based on the generalized reduced gradient method. This approach is termed simultaneous optimization and solution strategy. The objective functional can contain integral terms. The state and control variables can have time delays. Equalities and inequalities containing state and control variables can be included into the model as well as algebraic equations and inequalities. The maximum number of independent variables is 2. Problems containing 3 independent variables can be transformed into problems having 2 independent variables using finite differencing. The maximum number of NLP variables and constraints is 1500. The method is also suitable for solving ordinary and partial differential equations. The state functions are approximated by a linear combination of Lagrange interpolation polynomials. The control function can either be approximated by a linear combination of Lagrange interpolation polynomials or by a piecewise constant function over finite elements. The number of internal collocation points can vary by finite elements. The residual error is evaluated at arbitrarily chosen equidistant grid-points, thus enabling the user to check the accuracy of the solution between collocation points, where the solution is exact. The solution functions can be tabulated. There is an option to use control vector parameterization to solve optimization problems containing initial value ordinary differential equations. When there are many differential equations or the upper integration limit should be selected optimally then this approach should be used. The portability of the package has been addressed converting the package from V AX FORTRAN 77 into IBM PC FORTRAN 77 and into SUN SPARC 2000 FORTRAN 77. Computer runs have shown that the method can reproduce optimization problems published in the literature. The GRG2 and the VF I 3AD packages, integrated into the optimization package, proved to be robust and reliable. The package contains an executive module, a module performing control vector parameterization and 2 nonlinear problem solver modules, GRG2 and VF I 3AD. There is a stand-alone module that converts the differential-algebraic optimization problem into a nonlinear programming problem.
Resumo:
We investigate an application of the method of fundamental solutions (MFS) to heat conduction in two-dimensional bodies, where the thermal diffusivity is piecewise constant. We extend the MFS proposed in Johansson and Lesnic [A method of fundamental solutions for transient heat conduction, Eng. Anal. Bound. Elem. 32 (2008), pp. 697–703] for one-dimensional heat conduction with the sources placed outside the space domain of interest, to the two-dimensional setting. Theoretical properties of the method, as well as numerical investigations, are included, showing that accurate results can be obtained efficiently with small computational cost.
Resumo:
In this paper we investigate an application of the method of fundamental solutions (MFS) to transient heat conduction in layered materials, where the thermal diffusivity is piecewise constant. Recently, in Johansson and Lesnic [A method of fundamental solutions for transient heat conduction. Eng Anal Boundary Elem 2008;32:697–703], a MFS was proposed with the sources placed outside the space domain of interest, and we extend that technique to numerically approximate the heat flow in layered materials. Theoretical properties of the method, as well as numerical investigations are included.
Resumo:
Regions containing internal boundaries such as composite materials arise in many applications.We consider a situation of a layered domain in IR3 containing a nite number of bounded cavities. The model is stationary heat transfer given by the Laplace equation with piecewise constant conductivity. The heat ux (a Neumann condition) is imposed on the bottom of the layered region and various boundary conditions are imposed on the cavities. The usual transmission (interface) conditions are satised at the interface layer, that is continuity of the solution and its normal derivative. To eciently calculate the stationary temperature eld in the semi-innite region, we employ a Green's matrix technique and reduce the problem to boundary integral equations (weakly singular) over the bounded surfaces of the cavities. For the numerical solution of these integral equations, we use Wienert's approach [20]. Assuming that each cavity is homeomorphic with the unit sphere, a fully discrete projection method with super-algebraic convergence order is proposed. A proof of an error estimate for the approximation is given as well. Numerical examples are presented that further highlights the eciency and accuracy of the proposed method.
Resumo:
Providing transportation system operators and travelers with accurate travel time information allows them to make more informed decisions, yielding benefits for individual travelers and for the entire transportation system. Most existing advanced traveler information systems (ATIS) and advanced traffic management systems (ATMS) use instantaneous travel time values estimated based on the current measurements, assuming that traffic conditions remain constant in the near future. For more effective applications, it has been proposed that ATIS and ATMS should use travel times predicted for short-term future conditions rather than instantaneous travel times measured or estimated for current conditions. ^ This dissertation research investigates short-term freeway travel time prediction using Dynamic Neural Networks (DNN) based on traffic detector data collected by radar traffic detectors installed along a freeway corridor. DNN comprises a class of neural networks that are particularly suitable for predicting variables like travel time, but has not been adequately investigated for this purpose. Before this investigation, it was necessary to identifying methods for data imputation to account for missing data usually encountered when collecting data using traffic detectors. It was also necessary to identify a method to estimate the travel time on the freeway corridor based on data collected using point traffic detectors. A new travel time estimation method referred to as the Piecewise Constant Acceleration Based (PCAB) method was developed and compared with other methods reported in the literatures. The results show that one of the simple travel time estimation methods (the average speed method) can work as well as the PCAB method, and both of them out-perform other methods. This study also compared the travel time prediction performance of three different DNN topologies with different memory setups. The results show that one DNN topology (the time-delay neural networks) out-performs the other two DNN topologies for the investigated prediction problem. This topology also performs slightly better than the simple multilayer perceptron (MLP) neural network topology that has been used in a number of previous studies for travel time prediction.^
Resumo:
Providing transportation system operators and travelers with accurate travel time information allows them to make more informed decisions, yielding benefits for individual travelers and for the entire transportation system. Most existing advanced traveler information systems (ATIS) and advanced traffic management systems (ATMS) use instantaneous travel time values estimated based on the current measurements, assuming that traffic conditions remain constant in the near future. For more effective applications, it has been proposed that ATIS and ATMS should use travel times predicted for short-term future conditions rather than instantaneous travel times measured or estimated for current conditions. This dissertation research investigates short-term freeway travel time prediction using Dynamic Neural Networks (DNN) based on traffic detector data collected by radar traffic detectors installed along a freeway corridor. DNN comprises a class of neural networks that are particularly suitable for predicting variables like travel time, but has not been adequately investigated for this purpose. Before this investigation, it was necessary to identifying methods for data imputation to account for missing data usually encountered when collecting data using traffic detectors. It was also necessary to identify a method to estimate the travel time on the freeway corridor based on data collected using point traffic detectors. A new travel time estimation method referred to as the Piecewise Constant Acceleration Based (PCAB) method was developed and compared with other methods reported in the literatures. The results show that one of the simple travel time estimation methods (the average speed method) can work as well as the PCAB method, and both of them out-perform other methods. This study also compared the travel time prediction performance of three different DNN topologies with different memory setups. The results show that one DNN topology (the time-delay neural networks) out-performs the other two DNN topologies for the investigated prediction problem. This topology also performs slightly better than the simple multilayer perceptron (MLP) neural network topology that has been used in a number of previous studies for travel time prediction.
Resumo:
Thesis (Ph.D.)--University of Washington, 2016-08
Resumo:
A high-frequency time domain finite element scattering code using a combination of edge and piecewise constant elements on unstructured tetrahedral meshes is described. A comparison of computation with theory is given for scattering from a sphere. A parallel implementation making use of the bulk synchronous parallel (BSP) programming model is described in detail; a BSP performance model of the parallelized field calculation is derived and compared to timing measurements on up to 128 processors on a Cray T3D.
Resumo:
In this paper we show how to construct the Evans function for traveling wave solutions of integral neural field equations when the firing rate function is a Heaviside. This allows a discussion of wave stability and bifurcation as a function of system parameters, including the speed and strength of synaptic coupling and the speed of axonal signals. The theory is illustrated with the construction and stability analysis of front solutions to a scalar neural field model and a limiting case is shown to recover recent results of L. Zhang [On stability of traveling wave solutions in synaptically coupled neuronal networks, Differential and Integral Equations, 16, (2003), pp.513-536.]. Traveling fronts and pulses are considered in more general models possessing either a linear or piecewise constant recovery variable. We establish the stability of coexisting traveling fronts beyond a front bifurcation and consider parameter regimes that support two stable traveling fronts of different speed. Such fronts may be connected and depending on their relative speed the resulting region of activity can widen or contract. The conditions for the contracting case to lead to a pulse solution are established. The stability of pulses is obtained for a variety of examples, in each case confirming a previously conjectured stability result. Finally we show how this theory may be used to describe the dynamic instability of a standing pulse that arises in a model with slow recovery. Numerical simulations show that such an instability can lead to the shedding of a pair of traveling pulses.