24 resultados para Mixed methodology
Resumo:
[ES]Este Trabajo de Fin de Grado (TFG) está relacionado con las prácticas efectuadas en la Guardia Municipal de San Sebastián (País Vasco) y pretende analizar las órdenes de protección concedidas a víctimas de violencia de género inmigrantes y nacionales para comprobar si existe alguna diferencia en cuanto a su aplicación. Asimismo, se pretende analizar el procedimiento efectuado por este cuerpo policial y servicios sociales en estos casos, así como el perfil de la víctima y el agresor. Por otra parte, se procura analizar aspectos controvertidos de la Ley Orgánica 1/2004 de Medidas de Protección Integral de la Violencia de Género. Para ello, se utilizara una metodología mixta. Por una parte de corte cualitativo realizando cuatro entrevistas (a dos víctimas, una agente de policía y una trabajadora social) para ahondar más respecto a este tema. En segundo lugar, desde una perspectiva cuantitativa se explorará la base de datos de la Guardia Municipal en materia de violencia de género para realizar un análisis estadístico. Finalmente, se abordaran las conclusiones a las que se ha llegado con este trabajo y se propondrán mejoras de cara a futuras investigaciones y a la operatividad de la Guardia Municipal.
Resumo:
Impreso por la Diputación Foral de Álava, D.L. VI-430/99.
Resumo:
The aim of this technical report is to present some detailed explanations in order to help to understand and use the Message Passing Interface (MPI) parallel programming for solving several mixed integer optimization problems. We have developed a C++ experimental code that uses the IBM ILOG CPLEX optimizer within the COmputational INfrastructure for Operations Research (COIN-OR) and MPI parallel computing for solving the optimization models under UNIX-like systems. The computational experience illustrates how can we solve 44 optimization problems which are asymmetric with respect to the number of integer and continuous variables and the number of constraints. We also report a comparative with the speedup and efficiency of several strategies implemented for some available number of threads.
Resumo:
In this paper we introduce four scenario Cluster based Lagrangian Decomposition (CLD) procedures for obtaining strong lower bounds to the (optimal) solution value of two-stage stochastic mixed 0-1 problems. At each iteration of the Lagrangian based procedures, the traditional aim consists of obtaining the solution value of the corresponding Lagrangian dual via solving scenario submodels once the nonanticipativity constraints have been dualized. Instead of considering a splitting variable representation over the set of scenarios, we propose to decompose the model into a set of scenario clusters. We compare the computational performance of the four Lagrange multiplier updating procedures, namely the Subgradient Method, the Volume Algorithm, the Progressive Hedging Algorithm and the Dynamic Constrained Cutting Plane scheme for different numbers of scenario clusters and different dimensions of the original problem. Our computational experience shows that the CLD bound and its computational effort depend on the number of scenario clusters to consider. In any case, our results show that the CLD procedures outperform the traditional LD scheme for single scenarios both in the quality of the bounds and computational effort. All the procedures have been implemented in a C++ experimental code. A broad computational experience is reported on a test of randomly generated instances by using the MIP solvers COIN-OR and CPLEX for the auxiliary mixed 0-1 cluster submodels, this last solver within the open source engine COIN-OR. We also give computational evidence of the model tightening effect that the preprocessing techniques, cut generation and appending and parallel computing tools have in stochastic integer optimization. Finally, we have observed that the plain use of both solvers does not provide the optimal solution of the instances included in the testbed with which we have experimented but for two toy instances in affordable elapsed time. On the other hand the proposed procedures provide strong lower bounds (or the same solution value) in a considerably shorter elapsed time for the quasi-optimal solution obtained by other means for the original stochastic problem.
Resumo:
Single-species management objectives may not be consistent within mixed fisheries. They may lead species to unsafe situations, promote discarding of over-quota and/or misreporting of catches. We provide an algorithm for characterising bio-economic reference points for a mixed fishery as the steady-state solution of a dynamic optimal management problem. The optimisation problem takes into account: i) that species are fishing simultaneously in unselective fishing operations and ii)intertemporal discounting and fleet costs to relate reference points to discounted economic profits along optimal trajectories. We illustrate how the algorithm can be implemented by applying it to the European Northern Stock of Hake (Merluccius merluccius), where fleets also capture Northern megrim (Lepidorhombus whiffiagonis) and Northern anglerfish (Lophius piscatorius and Lophius budegassa). We find that optimal mixed management leads to a target reference point that is quite similar to the 2/3 of the Fmsy single-species (hake) target. Mixed management is superior to singlespecies management because it leads the fishery to higher discounted profits with higher long-term SSB for all species. We calculate that the losses due to the use of the Fmsy single-species (hake) target in this mixed fishery account for 11.4% of total discounted profits.
Resumo:
In this article we describe the methodology developed for the semiautomatic annotation of EPEC-RolSem, a Basque corpus labeled at predicate level following the PropBank-VerbNet model. The methodology presented is the product of detailed theoretical study of the semantic nature of verbs in Basque and of their similarities and differences with verbs in other languages. As part of the proposed methodology, we are creating a Basque lexicon on the PropBank-VerbNet model that we have named the Basque Verb Index (BVI). Our work thus dovetails the general trend toward building lexicons from tagged corpora that is clear in work conducted for other languages. EPEC-RolSem and BVI are two important resources for the computational semantic processing of Basque; as far as the authors are aware, they are also the first resources of their kind developed for Basque. In addition, each entry in BVI is linked to the corresponding verb-entry in well-known resources like PropBank, VerbNet, WordNet, Levin’s Classification and FrameNet. We have also implemented several automatic processes to aid in creating and annotating the BVI, including processes designed to facilitate the task of manual annotation.
Resumo:
We present a scheme to generate clusters submodels with stage ordering from a (symmetric or a nonsymmetric one) multistage stochastic mixed integer optimization model using break stage. We consider a stochastic model in compact representation and MPS format with a known scenario tree. The cluster submodels are built by storing first the 0-1 the variables, stage by stage, and then the continuous ones, also stage by stage. A C++ experimental code has been implemented for reordering the stochastic model as well as the cluster decomposition after the relaxation of the non-anticipativiy constraints until the so-called breakstage. The computational experience shows better performance of the stage ordering in terms of elapsed time in a randomly generated testbed of multistage stochastic mixed integer problems.
Resumo:
Background: Patients with chronic obstructive pulmonary disease (COPD) often experience exacerbations of the disease that require hospitalization. Current guidelines offer little guidance for identifying patients whose clinical situation is appropriate for admission to the hospital, and properly developed and validated severity scores for COPD exacerbations are lacking. To address these important gaps in clinical care, we created the IRYSS-COPD Appropriateness Study. Methods/Design: The RAND/UCLA Appropriateness Methodology was used to identify appropriate and inappropriate scenarios for hospital admission for patients experiencing COPD exacerbations. These scenarios were then applied to a prospective cohort of patients attending the emergency departments (ED) of 16 participating hospitals. Information was recorded during the time the patient was evaluated in the ED, at the time a decision was made to admit the patient to the hospital or discharge home, and during follow-up after admission or discharge home. While complete data were generally available at the time of ED admission, data were often missing at the time of decision making. Predefined assumptions were used to impute much of the missing data. Discussion: The IRYSS-COPD Appropriateness Study will validate the appropriateness criteria developed by the RAND/UCLA Appropriateness Methodology and thus better delineate the requirements for admission or discharge of patients experiencing exacerbations of COPD. The study will also provide a better understanding of the determinants of outcomes of COPD exacerbations, and evaluate the equity and variability in access and outcomes in these patients.
Resumo:
34 p.
Resumo:
47 p.
Resumo:
[ES]Este trabajo tiene como objetivo analizar las soluciones existentes para el diseño de las líneas mixtas. Se trata de un tipo de líneas bastante complejas de diseñar y que por lo general han sido poco estudiadas por la sociedad académica, a pesar de las numerosas ventajas que ofrece en comparación a las líneas simples, mucho más investigadas. Para facilitar la comprensión de su diseño, y antes de su estudio, se realiza un análisis de las características de las líneas mixtas. Posteriormente se introduce una metodología desarrollada por Lluis Cuatrecasas (Presidente del Instituto Lean Management de España), que permite salvar la brecha que existe a su juicio entre los estudios sobre los diseños de este tipo de líneas y su aplicación a la empresa. Finalmente, se ilustra dicha metodología mediante un caso práctico y se evalúan sus fortalezas y debilidades.
Resumo:
The evaluation and comparison of internal cluster validity indices is a critical problem in the clustering area. The methodology used in most of the evaluations assumes that the clustering algorithms work correctly. We propose an alternative methodology that does not make this often false assumption. We compared 7 internal cluster validity indices with both methodologies and concluded that the results obtained with the proposed methodology are more representative of the actual capabilities of the compared indices.
Resumo:
La actividad aseguradora supone la transferencia de riesgos del asegurado al asegurador. El asegurador se compromete al pago de una prestación si el riesgo se realiza. Se produce un cambio en el ciclo productivo. El asegurador vende una cobertura sin conocer el momento y el coste exacto de dicha cobertura. Esta particularidad de la actividad aseguradora explica la necesidad para una entidad aseguradora de ser solvente en cada momento y ante cualquier imprevisto. Por ello, la solvencia de las entidades aseguradoras es un aspecto que se ha ido recogiendo en las distintas normativas que han regulado la actividad aseguradora y al que se ha ido dando cada vez más importancia. Actualmente la legislación vigente en materia de solvencia de las aseguradoras esta regulada por la directiva europea Solvencia I. Esta directiva establece dos conceptos para garantizar la solvencia: las provisiones técnicas y el margen de solvencia. Las provisiones técnicas son las calculadas para garantizar la solvencia estática de la compañía, es decir aquella que hace frente, en un instante temporal determinado, a los compromisos asumidos por la entidad. El margen de solvencia se destina a cubrir la solvencia dinámica, aquella que hace referencia a eventos futuros que puedan afectar la capacidad del asegurador. Sin embargo en una corriente de gestión global del riesgo en la que el sector bancario ya se había adelantado al sector asegurador con la normativa Basilea II, se decidió iniciar un proyecto europeo de reforma de Solvencia I y en noviembre del 2009 se adoptó la directiva 2009/138/CE del parlamento europeo y del consejo, sobre el seguro de vida, el acceso a la actividad de seguro y de reaseguro y su ejercicio mas conocida como Solvencia II. Esta directiva supone un profundo cambio en las reglas actuales de solvencia para las entidades aseguradoras. Este cambio persigue el objetivo de establecer un marco regulador común a nivel europeo que sea más adaptado al perfil de riesgo de cada entidad aseguradora. Esta nueva directiva define dos niveles distintos de capital: el SCR (requerimiento estándar de capital de solvencia) y el MCR (requerimiento mínimo de capital). Para el calculo del SCR se ha establecido que el asegurador tendrá la libertad de elegir entre dos modelos. Un modelo estándar propuesto por la Autoridad Europea de Seguros y Pensiones de Jubilación (EIOPA por sus siglas en inglés), que permitirá un calculo simple, y un modelo interno desarrollado por la propia entidad que deberá ser aprobado por las autoridades competentes. También se contempla la posibilidad de utilizar un modelo mixto que combine ambos, el estándar y el interno. Para el desarrollo del modelo estándar se han realizado una serie de estudios de impacto cuantitativos (QIS). El último estudio (QIS 5) ha sido el que ha planteado de forma más precisa el cálculo del SCR. Plantea unos shocks que se deberán de aplicar al balance de la entidad con el objetivo de estresarlo, y en base a los resultados obtenidos constituir el SCR. El objetivo de este trabajo es realizar una síntesis de las especificaciones técnicas del QIS5 para los seguros de vida y realizar una aplicación práctica para un seguro de vida mixto puro. En la aplicación práctica se determinarán los flujos de caja asociados a este producto para calcular su mejor estimación (Best estimate). Posteriormente se determinará el SCR aplicando los shocks para los riesgos de mortalidad, rescates y gastos. Por último, calcularemos el margen de riesgo asociado al SCR. Terminaremos el presente TFG con unas conclusiones, la bibliografía empleada así como un anexo con las tablas empleadas.
Resumo:
In this work we extend to the multistage case two recent risk averse measures for two-stage stochastic programs based on first- and second-order stochastic dominance constraints induced by mixed-integer linear recourse. Additionally, we consider Time Stochastic Dominance (TSD) along a given horizon. Given the dimensions of medium-sized problems augmented by the new variables and constraints required by those risk measures, it is unrealistic to solve the problem up to optimality by plain use of MIP solvers in a reasonable computing time, at least. Instead of it, decomposition algorithms of some type should be used. We present an extension of our Branch-and-Fix Coordination algorithm, so named BFC-TSD, where a special treatment is given to cross scenario group constraints that link variables from different scenario groups. A broad computational experience is presented by comparing the risk neutral approach and the tested risk averse strategies. The performance of the new version of the BFC algorithm versus the plain use of a state-of-the-artMIP solver is also reported.
Resumo:
Documentos de Trabajo