10 resultados para New classification

em AMS Tesi di Dottorato - Alm@DL - Università di Bologna


Relevância:

70.00% 70.00%

Publicador:

Resumo:

Introduction: Recently, the American Association of Gynecologic Laparoscopists proposed a new classification and scoring system with the specific aim to assess surgical complexity. This study sought to assess if a higher AAGL score correlates with an increased risk of peri-operative complications in women submitted to surgery for endometriosis. Methods: This is a retrospective cohort study conducted in a third level referral center. We collected data from women with endometriosis submitted to complete surgical removal of endometriosis from January 2019 to December 2021. ENZIAN, r-ASRM classifications and AAGL total score was calculated for each patient. Population was divided in two groups according to the occurrence or not of at least one peri-operative complication. Our primary outcome was to evaluate the correlation between AAGL score and occurrence of complications. Results: During the study period we analyzed data from 282 eligible patients. Among them, 80 (28.4%) experienced peri-operative complications. No statistically significant difference was found between the two groups in terms of baseline characteristics, except for pre-operative hemoglobin (Hb), which was lower in patients with complications (p=0.001). Surgical variables associated with the occurrence of complications were recto-sigmoid surgery (p=0.003), ileocecal resection (0.034), and longer operative time (p=0.007). Furthermore, a higher ENZIAN B score (p=0.006), AAGL score (p=0.045) and stage (p=0.022) were found in the group of patients with complications. The multivariate analysis only confirmed the significant association between the occurrence of peri-operative complications and lower pre-operative Hb level (OR 0.74; 95% CI, 0.59 - 0.94; p=0.014), longer operative time (OR 1.00; 95% CI, 1.00 – 1.01; p=0.013), recto-sigmoid surgery - especially discoid resection (OR 8.73; 95% CI, 2.18 – 35; p=0.016) and ENZIAN B3 (OR 3.62; 95% CI, 1.46 – 8.99; p= 0.006). Conclusion: According to our findings, high AAGL scores or stages do not seem to increase the risk of peri-operative complications.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Background. A new classification system of human breast tumours based on the immunohistochemical characterization has been applied to mammary tumours of the female dog with the aim to verify its association with invasion and grade, and prognostic aid in veterinary medicine. Methods. Forty-five canine mammary carcinomas with a two-year post-mastectomy follow-up were selected from our database, and the following antibodies were applied: anti-cytokeratines 14, 5/6, oestrogen receptor (ER), progesterone receptor (PR), and ERB-B2. . The tumours were grouped for phenotype as: luminal-like (ER+ and/or PR+, CK14-, CK5/6-) type A (ERB-B2-), and B (ERB-B2+); basal-like (ER-, PR-, CK14+ and/or CK5/6+, ERB-B2-); ERB-B2 (ER-, PR-, CK14-, CK5/6-, ERB-B2+). Association with invasion, grade and histotypes were evaluated and Kaplan-Meier survival curves estimated, then compared by survival analysis. Results. Thirty-five cases with luminal pattern (ER+ and PR+) were subgrouped into 13 A type and 22 B type, if ERB-B2 positive or negative . Most luminal-like A and basal-like cases were grade 1 carcinomas, while the percentage of luminal B cases was higher in grade 2 and 3 (Pearson Chi-square P=0.009). No difference in the percentage of molecular subtypes was evidenced between simple and complex/mixed carcinomas (Pearson Chi-square P=0.47). No significant results were obtained by survival analysis, even if basal-like had a more favourable prognosis than luminal-like. Conclusion. The panel of antibodies identified only 3 groups (luminal-like A and B, and basal-like) in the dog. Even though canine mammary tumours may be a model of human breast cancer, the existence of the same types of carcinoma as in the woman need to be confirmed. Canine mammary carcinomas show high molecular heterogeneity, which would benefit from a classification based on molecular differences. However, by multivariate analysis, the molecular classification appears a variable with a dependent value if compared to invasion and grade that are independent, suggesting that, at present, caution should be used in the application of such a classification to the dog, in which invasion and grade supply the most important prognostic information.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

The work presented in this thesis is focused on the open-ended coaxial-probe frequency-domain reflectometry technique for complex permittivity measurement at microwave frequencies of dispersive dielectric multilayer materials. An effective dielectric model is introduced and validated to extend the applicability of this technique to multilayer materials in on-line system context. In addition, the thesis presents: 1) a numerical study regarding the imperfectness of the contact at the probe-material interface, 2) a review of the available models and techniques, 3) a new classification of the extraction schemes with guidelines on how they can be used to improve the overall performance of the probe according to the problem requirements.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

The purpose of this Thesis is to develop a robust and powerful method to classify galaxies from large surveys, in order to establish and confirm the connections between the principal observational parameters of the galaxies (spectral features, colours, morphological indices), and help unveil the evolution of these parameters from $z \sim 1$ to the local Universe. Within the framework of zCOSMOS-bright survey, and making use of its large database of objects ($\sim 10\,000$ galaxies in the redshift range $0 < z \lesssim 1.2$) and its great reliability in redshift and spectral properties determinations, first we adopt and extend the \emph{classification cube method}, as developed by Mignoli et al. (2009), to exploit the bimodal properties of galaxies (spectral, photometric and morphologic) separately, and then combining together these three subclassifications. We use this classification method as a test for a newly devised statistical classification, based on Principal Component Analysis and Unsupervised Fuzzy Partition clustering method (PCA+UFP), which is able to define the galaxy population exploiting their natural global bimodality, considering simultaneously up to 8 different properties. The PCA+UFP analysis is a very powerful and robust tool to probe the nature and the evolution of galaxies in a survey. It allows to define with less uncertainties the classification of galaxies, adding the flexibility to be adapted to different parameters: being a fuzzy classification it avoids the problems due to a hard classification, such as the classification cube presented in the first part of the article. The PCA+UFP method can be easily applied to different datasets: it does not rely on the nature of the data and for this reason it can be successfully employed with others observables (magnitudes, colours) or derived properties (masses, luminosities, SFRs, etc.). The agreement between the two classification cluster definitions is very high. ``Early'' and ``late'' type galaxies are well defined by the spectral, photometric and morphological properties, both considering them in a separate way and then combining the classifications (classification cube) and treating them as a whole (PCA+UFP cluster analysis). Differences arise in the definition of outliers: the classification cube is much more sensitive to single measurement errors or misclassifications in one property than the PCA+UFP cluster analysis, in which errors are ``averaged out'' during the process. This method allowed us to behold the \emph{downsizing} effect taking place in the PC spaces: the migration between the blue cloud towards the red clump happens at higher redshifts for galaxies of larger mass. The determination of $M_{\mathrm{cross}}$ the transition mass is in significant agreement with others values in literature.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In the past decade, the advent of efficient genome sequencing tools and high-throughput experimental biotechnology has lead to enormous progress in the life science. Among the most important innovations is the microarray tecnology. It allows to quantify the expression for thousands of genes simultaneously by measurin the hybridization from a tissue of interest to probes on a small glass or plastic slide. The characteristics of these data include a fair amount of random noise, a predictor dimension in the thousand, and a sample noise in the dozens. One of the most exciting areas to which microarray technology has been applied is the challenge of deciphering complex disease such as cancer. In these studies, samples are taken from two or more groups of individuals with heterogeneous phenotypes, pathologies, or clinical outcomes. these samples are hybridized to microarrays in an effort to find a small number of genes which are strongly correlated with the group of individuals. Eventhough today methods to analyse the data are welle developed and close to reach a standard organization (through the effort of preposed International project like Microarray Gene Expression Data -MGED- Society [1]) it is not unfrequant to stumble in a clinician's question that do not have a compelling statistical method that could permit to answer it.The contribution of this dissertation in deciphering disease regards the development of new approaches aiming at handle open problems posed by clinicians in handle specific experimental designs. In Chapter 1 starting from a biological necessary introduction, we revise the microarray tecnologies and all the important steps that involve an experiment from the production of the array, to the quality controls ending with preprocessing steps that will be used into the data analysis in the rest of the dissertation. While in Chapter 2 a critical review of standard analysis methods are provided stressing most of problems that In Chapter 3 is introduced a method to adress the issue of unbalanced design of miacroarray experiments. In microarray experiments, experimental design is a crucial starting-point for obtaining reasonable results. In a two-class problem, an equal or similar number of samples it should be collected between the two classes. However in some cases, e.g. rare pathologies, the approach to be taken is less evident. We propose to address this issue by applying a modified version of SAM [2]. MultiSAM consists in a reiterated application of a SAM analysis, comparing the less populated class (LPC) with 1,000 random samplings of the same size from the more populated class (MPC) A list of the differentially expressed genes is generated for each SAM application. After 1,000 reiterations, each single probe given a "score" ranging from 0 to 1,000 based on its recurrence in the 1,000 lists as differentially expressed. The performance of MultiSAM was compared to the performance of SAM and LIMMA [3] over two simulated data sets via beta and exponential distribution. The results of all three algorithms over low- noise data sets seems acceptable However, on a real unbalanced two-channel data set reagardin Chronic Lymphocitic Leukemia, LIMMA finds no significant probe, SAM finds 23 significantly changed probes but cannot separate the two classes, while MultiSAM finds 122 probes with score >300 and separates the data into two clusters by hierarchical clustering. We also report extra-assay validation in terms of differentially expressed genes Although standard algorithms perform well over low-noise simulated data sets, multi-SAM seems to be the only one able to reveal subtle differences in gene expression profiles on real unbalanced data. In Chapter 4 a method to adress similarities evaluation in a three-class prblem by means of Relevance Vector Machine [4] is described. In fact, looking at microarray data in a prognostic and diagnostic clinical framework, not only differences could have a crucial role. In some cases similarities can give useful and, sometimes even more, important information. The goal, given three classes, could be to establish, with a certain level of confidence, if the third one is similar to the first or the second one. In this work we show that Relevance Vector Machine (RVM) [2] could be a possible solutions to the limitation of standard supervised classification. In fact, RVM offers many advantages compared, for example, with his well-known precursor (Support Vector Machine - SVM [3]). Among these advantages, the estimate of posterior probability of class membership represents a key feature to address the similarity issue. This is a highly important, but often overlooked, option of any practical pattern recognition system. We focused on Tumor-Grade-three-class problem, so we have 67 samples of grade I (G1), 54 samples of grade 3 (G3) and 100 samples of grade 2 (G2). The goal is to find a model able to separate G1 from G3, then evaluate the third class G2 as test-set to obtain the probability for samples of G2 to be member of class G1 or class G3. The analysis showed that breast cancer samples of grade II have a molecular profile more similar to breast cancer samples of grade I. Looking at the literature this result have been guessed, but no measure of significance was gived before.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this thesis we made the first steps towards the systematic application of a methodology for automatically building formal models of complex biological systems. Such a methodology could be useful also to design artificial systems possessing desirable properties such as robustness and evolvability. The approach we follow in this thesis is to manipulate formal models by means of adaptive search methods called metaheuristics. In the first part of the thesis we develop state-of-the-art hybrid metaheuristic algorithms to tackle two important problems in genomics, namely, the Haplotype Inference by parsimony and the Founder Sequence Reconstruction Problem. We compare our algorithms with other effective techniques in the literature, we show strength and limitations of our approaches to various problem formulations and, finally, we propose further enhancements that could possibly improve the performance of our algorithms and widen their applicability. In the second part, we concentrate on Boolean network (BN) models of gene regulatory networks (GRNs). We detail our automatic design methodology and apply it to four use cases which correspond to different design criteria and address some limitations of GRN modeling by BNs. Finally, we tackle the Density Classification Problem with the aim of showing the learning capabilities of BNs. Experimental evaluation of this methodology shows its efficacy in producing network that meet our design criteria. Our results, coherently to what has been found in other works, also suggest that networks manipulated by a search process exhibit a mixture of characteristics typical of different dynamical regimes.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Information is nowadays a key resource: machine learning and data mining techniques have been developed to extract high-level information from great amounts of data. As most data comes in form of unstructured text in natural languages, research on text mining is currently very active and dealing with practical problems. Among these, text categorization deals with the automatic organization of large quantities of documents in priorly defined taxonomies of topic categories, possibly arranged in large hierarchies. In commonly proposed machine learning approaches, classifiers are automatically trained from pre-labeled documents: they can perform very accurate classification, but often require a consistent training set and notable computational effort. Methods for cross-domain text categorization have been proposed, allowing to leverage a set of labeled documents of one domain to classify those of another one. Most methods use advanced statistical techniques, usually involving tuning of parameters. A first contribution presented here is a method based on nearest centroid classification, where profiles of categories are generated from the known domain and then iteratively adapted to the unknown one. Despite being conceptually simple and having easily tuned parameters, this method achieves state-of-the-art accuracy in most benchmark datasets with fast running times. A second, deeper contribution involves the design of a domain-independent model to distinguish the degree and type of relatedness between arbitrary documents and topics, inferred from the different types of semantic relationships between respective representative words, identified by specific search algorithms. The application of this model is tested on both flat and hierarchical text categorization, where it potentially allows the efficient addition of new categories during classification. Results show that classification accuracy still requires improvements, but models generated from one domain are shown to be effectively able to be reused in a different one.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Intelligent Transport Systems (ITS) consists in the application of ICT to transport to offer new and improved services to the mobility of people and freights. While using ITS, travellers produce large quantities of data that can be collected and analysed to study their behaviour and to provide information to decision makers and planners. The thesis proposes innovative deployments of classification algorithms for Intelligent Transport System with the aim to support the decisions on traffic rerouting, bus transport demand and behaviour of two wheelers vehicles. The first part of this work provides an overview and a classification of a selection of clustering algorithms that can be implemented for the analysis of ITS data. The first contribution of this thesis is an innovative use of the agglomerative hierarchical clustering algorithm to classify similar travels in terms of their origin and destination, together with the proposal for a methodology to analyse drivers’ route choice behaviour using GPS coordinates and optimal alternatives. The clusters of repetitive travels made by a sample of drivers are then analysed to compare observed route choices to the modelled alternatives. The results of the analysis show that drivers select routes that are more reliable but that are more expensive in terms of travel time. Successively, different types of users of a service that provides information on the real time arrivals of bus at stop are classified using Support Vector Machines. The results shows that the results of the classification of different types of bus transport users can be used to update or complement the census on bus transport flows. Finally, the problem of the classification of accidents made by two wheelers vehicles is presented together with possible future application of clustering methodologies aimed at identifying and classifying the different types of accidents.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In these last years a great effort has been put in the development of new techniques for automatic object classification, also due to the consequences in many applications such as medical imaging or driverless cars. To this end, several mathematical models have been developed from logistic regression to neural networks. A crucial aspect of these so called classification algorithms is the use of algebraic tools to represent and approximate the input data. In this thesis, we examine two different models for image classification based on a particular tensor decomposition named Tensor-Train (TT) decomposition. The use of tensor approaches preserves the multidimensional structure of the data and the neighboring relations among pixels. Furthermore the Tensor-Train, differently from other tensor decompositions, does not suffer from the curse of dimensionality making it an extremely powerful strategy when dealing with high-dimensional data. It also allows data compression when combined with truncation strategies that reduce memory requirements without spoiling classification performance. The first model we propose is based on a direct decomposition of the database by means of the TT decomposition to find basis vectors used to classify a new object. The second model is a tensor dictionary learning model, based on the TT decomposition where the terms of the decomposition are estimated using a proximal alternating linearized minimization algorithm with a spectral stepsize.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This thesis presents a study of globular clusters (GCs), based on analysis of Monte Carlo simulations of globular clusters (GCs) with the aim to define new empirical parameters measurable from observations and able to trace the different phases of their dynamical evolution history. During their long term dynamical evolution, due to mass segregation and and dynamical friction, massive stars transfer kinetic energy to lower-mass objects, causing them to sink toward the cluster center. This continuous transfer of kinetic energy from the core to the outskirts triggers the runaway contraction of the core, known as "core collapse" (CC), followed by episodes of expansion and contraction called gravothermal oscillations. Clearly, such an internal dynamical evolution corresponds to significant variations also of the structure of the system. Determining the dynamical age of a cluster can be challenging as it depends on various internal and external properties. The traditional classification of GCs as CC or post-CC systems relies on detecting a steep power-law cusp in the central density profile, which may not always be reliable due to post-CC oscillations or other processes. In this thesis, based on the normalized cumulative radial distribution (nCRD) within a fraction of the half-mass radius is analyzed, and three diagnostics (A5, P5, and S2.5) are defined. These diagnostics show sensitivity to dynamical evolution and can distinguish pre-CC clusters from post-CC clusters.The analysis performed using multiple simulations with different initial conditions, including varying binary fractions and the presence of dark remnants showed the time variations of the diagnostics follow distinct patterns depending on the binary fraction and the retention or ejection of black holes. This analysis is extended to a larger set of simulations matching the observed properties of Galactic GCs, and the parameters show a potential to distinguish the dynamical stages of the observed clusters as well.