942 resultados para Free Boundary Value Problem
Resumo:
We establish existence and non-existence results to the Brezis-Nirenberg type problem involving the square root of the Laplacian in a bounded domain with zero Dirichlet boundary condition.
Resumo:
BACKGROUND: The mutation status of the BRAF and KRAS genes has been proposed as prognostic biomarker in colorectal cancer. Of them, only the BRAF V600E mutation has been validated independently as prognostic for overall survival and survival after relapse, while the prognostic value of KRAS mutation is still unclear. We investigated the prognostic value of BRAF and KRAS mutations in various contexts defined by stratifications of the patient population. METHODS: We retrospectively analyzed a cohort of patients with stage II and III colorectal cancer from the PETACC-3 clinical trial (N = 1,423), by assessing the prognostic value of the BRAF and KRAS mutations in subpopulations defined by all possible combinations of the following clinico-pathological variables: T stage, N stage, tumor site, tumor grade and microsatellite instability status. In each such subpopulation, the prognostic value was assessed by log rank test for three endpoints: overall survival, relapse-free survival, and survival after relapse. The significance level was set to 0.01 for Bonferroni-adjusted p-values, and a second threshold for a trend towards statistical significance was set at 0.05 for unadjusted p-values. The significance of the interactions was tested by Wald test, with significance level of 0.05. RESULTS: In stage II-III colorectal cancer, BRAF mutation was confirmed a marker of poor survival only in subpopulations involving microsatellite stable and left-sided tumors, with higher effects than in the whole population. There was no evidence for prognostic value in microsatellite instable or right-sided tumor groups. We found that BRAF was also prognostic for relapse-free survival in some subpopulations. We found no evidence that KRAS mutations had prognostic value, although a trend was observed in some stratifications. We also show evidence of heterogeneity in survival of patients with BRAF V600E mutation. CONCLUSIONS: The BRAF mutation represents an additional risk factor only in some subpopulations of colorectal cancers, in others having limited prognostic value. However, in the subpopulations where it is prognostic, it represents a marker of much higher risk than previously considered. KRAS mutation status does not seem to represent a strong prognostic variable.
Resumo:
Sentinel lymph node dissection (SLND) identifies melanoma patients with metastatic disease who would benefit from radical lymph node dissection (RLND). Rarely, patients with melanoma have an underlying lymphoproliferative disease, and melanoma metastases might develop as collision tumours in the sentinel lymph node (SLN). The aim of this study was to measure the incidence and examine the effect of collision tumours on the accuracy of SLND and on the validity of staging in this setting. Between 1998 and 2012, 750 consecutive SLNDs were performed in melanoma patients using the triple technique (lymphoscintigraphy, gamma probe and blue dye). The validity of SLND in collision tumours was analysed. False negativity was reflected by the disease-free survival. The literature was reviewed on collision tumours in melanoma. Collision tumours of melanoma and chronic lymphocytic leukaemia (CLL) were found in two SLN and in one RLND (0.4%). Subsequent RLNDs of SLND-positive cases were negative for melanoma. The patient with negative SLND developed relapse after 28 months with an inguinal lymph node metastasis of melanoma; RLND showed collision tumours. The literature review identified 12 cases of collision tumours. CLL was associated with increased melanoma incidence and reduced overall survival. This is, to our knowledge, the first assessment of the clinical value of SLND when collision tumours of melanoma and CLL are found. In this small series of three patients with both malignancies present in the same lymph node basin, lymphocytic infiltration of the CLL did not alter radioisotope uptake into the SLN. No false-negative result was observed. Our data suggest the validity of SLND in collision tumours, but given the rarity of the problem, further studies are necessary to confirm this reliability.
Resumo:
BACKGROUND: Allogeneic stem cell transplantation is usually considered the only curative treatment option for patients with advanced or transformed myelodysplastic syndromes in complete remission, but post-remission chemotherapy and autologous stem cell transplantation are potential alternatives, especially in patients over 45 years old. DESIGN AND METHODS: We evaluated, after intensive anti-leukemic remission-induction chemotherapy, the impact of the availability of an HLA-identical sibling donor on an intention-to treat basis. Additionally, all patients without a sibling donor in complete remission after the first consolidation course were randomized to either autologous peripheral blood stem cell transplantation or a second consolidation course consisting of high-dose cytarabine. RESULTS: The 4-year survival of the 341 evaluable patients was 28%. After achieving complete remission, the 4-year survival rates of patients under 55 years old with or without a donor were 54% and 41%, respectively, with an adjusted hazard ratio of 0.81 (95% confidence interval [95% CI], 0.49-1.35) for survival and of 0.67 (95% CI, 0.42-1.06) for disease-free survival. In patients with intermediate/high risk cytogenetic abnormalities the hazard ratio in multivariate analysis was 0.58 (99% CI, 0.22-1.50) (P=0.14) for survival and 0.46 (99% CI, 0.22-1.50) for disease-free survival (P=0.03). In contrast, in patients with low risk cytogenetic characteristics the hazard ratio for survival was 1.17 (99% CI, 0.40-3.42) and that for disease-free survival was 1.02 (99% CI, 0.40-2.56). The 4-year survival of the 65 patients randomized to autologous peripheral blood stem cell transplantation or a second consolidation course of high-dose cytarabine was 37% and 27%, respectively. The hazard ratio in multivariate analysis was 1.22 (95% CI, 0.65-2.27) for survival and 1.02 (95% CI, 0.56-1.85) for disease-free survival. CONCLUSIONS: Patients with a donor and candidates for allogeneic stem cell transplantation in first complete remission may have a better disease-free survival than those without a donor in case of myelodysplastic syndromes with intermediate/high-risk cytogenetics. Autologous peripheral blood stem cell transplantation does not provide longer survival than intensive chemotherapy.
Resumo:
In this paper the two main drawbacks of the heat balance integral methods are examined. Firstly we investigate the choice of approximating function. For a standard polynomial form it is shown that combining the Heat Balance and Refined Integral methods to determine the power of the highest order term will either lead to the same, or more often, greatly improved accuracy on standard methods. Secondly we examine thermal problems with a time-dependent boundary condition. In doing so we develop a logarithmic approximating function. This new function allows us to model moving peaks in the temperature profile, a feature that previous heat balance methods cannot capture. If the boundary temperature varies so that at some time t & 0 it equals the far-field temperature, then standard methods predict that the temperature is everywhere at this constant value. The new method predicts the correct behaviour. It is also shown that this function provides even more accurate results, when coupled with the new CIM, than the polynomial profile. Analysis primarily focuses on a specified constant boundary temperature and is then extended to constant flux, Newton cooling and time dependent boundary conditions.
Resumo:
In this study I try to explain the systemic problem of the low economic competitiveness of nuclear energy for the production of electricity by carrying out a biophysical analysis of its production process. Given the fact that neither econometric approaches nor onedimensional methods of energy analyses are effective, I introduce the concept of biophysical explanation as a quantitative analysis capable of handling the inherent ambiguity associated with the concept of energy. In particular, the quantities of energy, considered as relevant for the assessment, can only be measured and aggregated after having agreed on a pre-analytical definition of a grammar characterizing a given set of finite transformations. Using this grammar it becomes possible to provide a biophysical explanation for the low economic competitiveness of nuclear energy in the production of electricity. When comparing the various unit operations of the process of production of electricity with nuclear energy to the analogous unit operations of the process of production of fossil energy, we see that the various phases of the process are the same. The only difference is related to characteristics of the process associated with the generation of heat which are completely different in the two systems. Since the cost of production of fossil energy provides the base line of economic competitiveness of electricity, the (lack of) economic competitiveness of the production of electricity from nuclear energy can be studied, by comparing the biophysical costs associated with the different unit operations taking place in nuclear and fossil power plants when generating process heat or net electricity. In particular, the analysis focuses on fossil-fuel requirements and labor requirements for those phases that both nuclear plants and fossil energy plants have in common: (i) mining; (ii) refining/enriching; (iii) generating heat/electricity; (iv) handling the pollution/radioactive wastes. By adopting this approach, it becomes possible to explain the systemic low economic competitiveness of nuclear energy in the production of electricity, because of: (i) its dependence on oil, limiting its possible role as a carbon-free alternative; (ii) the choices made in relation to its fuel cycle, especially whether it includes reprocessing operations or not; (iii) the unavoidable uncertainty in the definition of the characteristics of its process; (iv) its large inertia (lack of flexibility) due to issues of time scale; and (v) its low power level.
Resumo:
The work in this paper deals with the development of momentum and thermal boundary layers when a power law fluid flows over a flat plate. At the plate we impose either constant temperature, constant flux or a Newton cooling condition. The problem is analysed using similarity solutions, integral momentum and energy equations and an approximation technique which is a form of the Heat Balance Integral Method. The fluid properties are assumed to be independent of temperature, hence the momentum equation uncouples from the thermal problem. We first derive the similarity equations for the velocity and present exact solutions for the case where the power law index n = 2. The similarity solutions are used to validate the new approximation method. This new technique is then applied to the thermal boundary layer, where a similarity solution can only be obtained for the case n = 1.
Resumo:
PURPOSE: To determine the diagnostic value of the intravascular contrast agent gadocoletic acid (B-22956) in three-dimensional, free breathing coronary magnetic resonance angiography (MRA) for stenosis detection in patients with suspected or known coronary artery disease. METHODS: Eighteen patients underwent three-dimensional, free breathing coronary MRA of the left and right coronary system before and after intravenous application of a single dose of gadocoletic acid (B-22956) using three different dose regimens (group A 0.050 mmol/kg; group B 0.075 mmol/kg; group C 0.100 mmol/kg). Precontrast scanning followed a coronary MRA standard non-contrast T2 preparation/turbo-gradient echo sequence (T2Prep); for postcontrast scanning an inversion-recovery gradient echo sequence was used (real-time navigator correction for both scans). In pre- and postcontrast scans quantitative analysis of coronary MRA data was performed to determine the number of visible side branches, vessel length and vessel sharpness of each of the three coronary arteries (LAD, LCX, RCA). The number of assessable coronary artery segments was determined to calculate sensitivity and specificity for detection of stenosis > or = 50% on a segment-to-segment basis (16-segment-model) in pre- and postcontrast scans with x-ray coronary angiography as the standard of reference. RESULTS: Dose group B (0.075 mmol/kg) was preferable with regard to improvement of MR angiographic parameters: in postcontrast scans all MR angiographic parameters increased significantly except for the number of visible side branches of the left circumflex artery. In addition, assessability of coronary artery segments significantly improved postcontrast in this dose group (67 versus 88%, p < 0.01). Diagnostic performance (sensitivity, specificity, accuracy) was 83, 77 and 78% for precontrast and 86, 95 and 94% for postcontrast scans. CONCLUSIONS: The use of gadocoletic acid (B-22956) results in an improvement of MR angiographic parameters, asssessability of coronary segments and detection of coronary stenoses > or = 50%.
Resumo:
Thank you Chairman I would like to extend a warm welcome to our keynote speakers, David Byrne of the European Commission, Derek Yach from the World Health Organisation, and Paul Quinn representing Congressman Marty Meehan who sends his apologies. When we include the speakers who will address later sessions, this is, undoubtedly, one of the strongest teams that have been assembled on tobacco control in Europe. The very strength of the team underlines what I see as a shift – a very necessary shift – in the way we perceive the tobacco issue. For the last twenty years, we have lived out a paradox. It isn´t a social side issue. I make no apology for the bluntness of what I´m saying, and will come back, a little later, to the radicalism I believe we need to bring – nationally – to this issue. For starters, though, I want to lay it on the line that what we´re talking about is an epidemic as deadly as any suffered by human kind throughout the centuries. Slower than some of those epidemics in its lethal action, perhaps. But an epidemic, nonetheless. According to the World Health Organisation tobacco accounted for just over 3 million annual deaths in 1990, rising to 4.023 million annual deaths in 1998. The numbers of deaths due to tobacco will rise to 8.4 million in 2020 and reach roughly 10 million annually by 2030. This is quite simply ghastly. Tobacco kills. It kills in many different ways. It kills increasing numbers of women. It does its damage directly and indirectly. For children, much of the damage comes from smoking by adults where children live, study, play and work. The very least we should be able to offer every child is breathable air. Air that doesn´t do them damage. We´re now seeing a global public health response to the tobacco epidemic. The Tobacco Free Initiative launched by the World Health Organisation was matched by significant tobacco control initiatives throughout the world. During this conference we will hear about the experiences our speakers had in driving these initiatives. This Tobacco Free Initiative poses unique challenges to our legal frameworks at both national and international levels; in particular it raises challenges about the legal context in which tobacco products are traded and asks questions about the impact of commercial speech especially on children, and the extent of the limitations that should be imposed on it. Politicians, supported by economists and lawyers as well as the medical profession, must continue to explore and develop this context to find innovative ways to wrap public health considerations around the trade in tobacco products – very tightly. We also have the right to demand a totally new paradigm from the tobacco industry. Bluntly, the tobacco industry plays the PR game at its cynical worst. The industry sells its products without regard to the harm these products cause. At the same time, to gain social acceptance, it gives donations, endowments and patronage to high profile events and people. Not good enough. This model of behaviour is no longer acceptable in a modern society. We need one where the industry integrates social responsibility and accountability into its day-to-day activities. We have waited for this change in behaviour from the tobacco industry for many decades. Unfortunately the documents disclosed during litigation in the USA and from other sources make very depressing reading; it is clear from them that any trust society placed in the tobacco industry in the past to address the health problems associated with its products was misplaced. This industry appears to lack the necessary leadership to guide it towards just and responsible action. Instead, it chooses evasion, deception and at times illegal activity to protect its profits at any price and to avoid its responsibilities to society and its customers. It has engaged in elaborate ´spin´ to generate political tolerance, scientific uncertainty and public acceptance of its products. Legislators must act now. I see no reason why the global community should continue to wait. Effective legal controls must be laid on this errant industry. We should also keep these controls under review at regular intervals and if they are failing to achieve the desired outcomes we should be prepared to amend them. In Ireland, as Minister for Health and Children, I launched a comprehensive tobacco control policy entitled “Towards a Tobacco Free Society“. OTT?Excessive?Unrealistic? On the contrary – I believe it to be imperative and inevitable. I honestly hold that, given the range of fatal diseases caused by tobacco use we have little alternative but to pursue the clear objective of creating a tobacco free society. Aiming at a tobacco free society means ensuring public and political opinion are properly informed. It requires help to be given to smokers to break the addiction. It demands that people are protected against environmental tobacco smoke and children are protected from any inducement to experiment with this product. Over the past year we have implemented a number of measures which will support these objectives; we have established an independent Office of Tobacco Control, we have introduced free nicotine replacement therapy for low-income earners, we have extended our existing prohibitions on tobacco advertising to the print media with some minor derogations for international publications. We have raised the legal age at which a person can be sold tobacco products to eighteen years. We have invested substantially more funds in health promotion activities and we have mounted sustained information campaigns. We have engaged in sponsorship arrangements, which are new and innovative for public bodies. I have provided health boards with additional resources to let them mount a sustained inspection and enforcement service. Health boards will engage new Directors of Tobacco Control responsible for coordinating each health board´s response and for liasing with the Tobacco Control Agency I set up earlier this year. Most recently, I have published a comprehensive Bill – The Public Health (Tobacco) Bill, 2001. This Bill will, among other things, end all forms of product display and in-store advertising and will require all retailers to register with the new Tobacco Control Agency. Ten packs of cigarettes will be banned and transparent and independent testing procedures of tobacco products will be introduced. Enforcement officers will be given all the necessary powers to ensure there is full compliance with the law. On smoking in public places we will extend the existing areas covered and it is proposed that I, as Minister for Health and Children, will have the powers to introduce further prohibitions in public places such as pubs and the work place. I will also provide for the establishment of a Tobacco Free Council to advise and assist on an ongoing basis. I believe the measures already introduced and those additional ones proposed in the Bill have widespread community support. In fact, you´re going to hear a detailed presentation from the MRBI which will amply illustrate the extent of this support. The great thing is that the support comes from smokers and non-smokers alike. Bottom line, Ladies and Gentlemen, is that we are at a watershed. As a society (if you´ll allow me to play with a popular phrase) we´ve realised it´s time to ´wake up and smell the cigarettes.´ Smell them. See them for what they are. And get real about destroying their hold on our people. The MRBI survey makes it clear that the single strongest weapon we have when it comes to preventing the habit among young people is price. Simple as that. Price. Up to now, the fear of inflation has been a real impediment to increasing taxes on tobacco. It sounds a serious, logical argument. Until you take it out and look at it a little more closely. Weigh it, as it were, in two hands. I believe – and I believe this with a great passion – that we must take cigarettes out of the equation we use when awarding wage increases. I am calling on IBEC and ICTU, on employers and trade unions alike, to move away from any kind of tolerance of a trade that is killing our citizens. At one point in industrial history, cigarettes were a staple of the workingman´s life. So it was legitimate to include them in the ´basket´ of goods that goes to make up the Consumer Price Index. It isn´t legitimate to include them any more. Today, I´m saying that society collectively must take the step to remove cigarettes from the basket of normality, from the list of elements which constitute necessary consumer spending. I´m saying: “We can no longer delude ourselves. We must exclude cigarettes from the considerations we address in central wage bargaining. We must price cigarettes out of the reach of the children those cigarettes will kill.” Right now, in the monthly Central Statistics Office reports on consumer spending, the figures include cigarettes. But – right down at the bottom of the page – there´s another figure. Calculated without including cigarettes. I believe that if we continue to use the first figure as our constant measure, it will be an indictment of us as legislators, as advocates for working people, as public health professionals. If, on the other hand, we move to the use of the second figure, we will be sending out a message of startling clarity to the nation. We will be saying “We don´t count an addictive, killer drug as part of normal consumer spending.” Taking cigarettes out of the basket used to determine the Consumer Price Index will take away the inflation argument. It will not be easy, in its implications for the social partners. But it is morally inescapable. We must do it. Because it will help us stop the killer that is tobacco. If we can do it, we will give so much extra strength to health educators and the new Tobacco Control Association. This new organisation of young people who already have branches in over fifteen counties, is represented here today. The young adults who make up its membership are well placed to advise children of the dangers of tobacco addiction in a way that older generations cannot. It would strengthen their hand if cigarettes move – in price terms – out of the easy reach of our children Finally, I would like to commend so many public health advocates who have shown professional and indeed personal courage in their commitment to this critical public health issue down through the years. We need you to continue to challenge and confront this grave public health problem and to repudiate the questionable science of the tobacco industry. The Research Institute for a Tobacco Free Society represents a new and dynamic form of partnership between government and civil society. It will provide an effective platform to engage and mobilise the many different professional and academic skills necessary to guide and challenge us. I wish the conference every success.
Resumo:
The generator problem was posed by Kadison in 1967, and it remains open until today. We provide a solution for the class of C*-algebras absorbing the Jiang-Su algebra Z tensorially. More precisely, we show that every unital, separable, Z-stable C*-algebra A is singly generated, which means that there exists an element x є A that is not contained in any proper sub-C*- algebra of A. To give applications of our result, we observe that Z can be embedded into the reduced group C*-algebra of a discrete group that contains a non-cyclic, free subgroup. It follows that certain tensor products with reduced group C*-algebras are singly generated. In particular, C*r (F ∞) ⨂ C*r (F ∞) is singly generated.
Resumo:
BACKGROUND: Plasma free and urinary metanephrines are recognized biomarkers for the assessment of pheochromocytoma. Plasma total metanephrines with a long half-life may represent another useful biomarker. OBJECTIVE: The aim of this study is to evaluate the diagnostic performances of plasma total metanephrines alone or combined with free metanephrines and fractionated 24-h urinary metanephrines. METHODS: A retrospective, case-control diagnostic test study was conducted between 1999 and 2007 in two university hospitals in Switzerland and two institutions in France. The patients included 46 cases with histologically proven pheochromocytoma, and 181 controls suspected of tumor with negative investigations and 3-year follow-up. None had renal dysfunction. Sensitivity and specificity were compared after expressing each measurement result as a ratio over its upper reference limit, adding the ratios of normetanephrine and metanephrine, and defining cut-off values of 1 or 2 for this sum. RESULTS: Applying a cut-off value of 1, plasma free and total metanephrines and urinary fractionated metanephrines had similar sensitivities of 96% (95% confidence interval, 86-99%), 95% (85-99%), and 95% (84-99%) along with similar specificities of 89% (83-94%), 91% (84-95%), and 86% (80-91%). A cut-off of 2 for the sum of ratios over reference limit improves the specificity, and it can be used for a confirmation test based on another biomarker taken among the three biomarkers. CONCLUSION: All three metanephrine-based tests perform equivalently for diagnosing pheochromocytoma in the absence of renal insufficiency, and can be conveniently associated two by two for confirming/excluding tumor.
Resumo:
OBJECTIVES: This study sought to establish an accurate and reproducible T(2)-mapping cardiac magnetic resonance (CMR) methodology at 3 T and to evaluate it in healthy volunteers and patients with myocardial infarct. BACKGROUND: Myocardial edema affects the T(2) relaxation time on CMR. Therefore, T(2)-mapping has been established to characterize edema at 1.5 T. A 3 T implementation designed for longitudinal studies and aimed at guiding and monitoring therapy remains to be implemented, thoroughly characterized, and evaluated in vivo. METHODS: A free-breathing navigator-gated radial CMR pulse sequence with an adiabatic T(2) preparation module and an empirical fitting equation for T(2) quantification was optimized using numerical simulations and was validated at 3 T in a phantom study. Its reproducibility for myocardial T(2) quantification was then ascertained in healthy volunteers and improved using an external reference phantom with known T(2). In a small cohort of patients with established myocardial infarction, the local T(2) value and extent of the edematous region were determined and compared with conventional T(2)-weighted CMR and x-ray coronary angiography, where available. RESULTS: The numerical simulations and phantom study demonstrated that the empirical fitting equation is significantly more accurate for T(2) quantification than that for the more conventional exponential decay. The volunteer study consistently demonstrated a reproducibility error as low as 2 ± 1% using the external reference phantom and an average myocardial T(2) of 38.5 ± 4.5 ms. Intraobserver and interobserver variability in the volunteers were -0.04 ± 0.89 ms (p = 0.86) and -0.23 ± 0.91 ms (p = 0.87), respectively. In the infarction patients, the T(2) in edema was 62.4 ± 9.2 ms and was consistent with the x-ray angiographic findings. Simultaneously, the extent of the edematous region by T(2)-mapping correlated well with that from the T(2)-weighted images (r = 0.91). CONCLUSIONS: The new, well-characterized 3 T methodology enables robust and accurate cardiac T(2)-mapping at 3 T with high spatial resolution, while the addition of a reference phantom improves reproducibility. This technique may be well suited for longitudinal studies in patients with suspected or established heart disease.
Resumo:
Background The 'database search problem', that is, the strengthening of a case - in terms of probative value - against an individual who is found as a result of a database search, has been approached during the last two decades with substantial mathematical analyses, accompanied by lively debate and centrally opposing conclusions. This represents a challenging obstacle in teaching but also hinders a balanced and coherent discussion of the topic within the wider scientific and legal community. This paper revisits and tracks the associated mathematical analyses in terms of Bayesian networks. Their derivation and discussion for capturing probabilistic arguments that explain the database search problem are outlined in detail. The resulting Bayesian networks offer a distinct view on the main debated issues, along with further clarity. Methods As a general framework for representing and analyzing formal arguments in probabilistic reasoning about uncertain target propositions (that is, whether or not a given individual is the source of a crime stain), this paper relies on graphical probability models, in particular, Bayesian networks. This graphical probability modeling approach is used to capture, within a single model, a series of key variables, such as the number of individuals in a database, the size of the population of potential crime stain sources, and the rarity of the corresponding analytical characteristics in a relevant population. Results This paper demonstrates the feasibility of deriving Bayesian network structures for analyzing, representing, and tracking the database search problem. The output of the proposed models can be shown to agree with existing but exclusively formulaic approaches. Conclusions The proposed Bayesian networks allow one to capture and analyze the currently most well-supported but reputedly counter-intuitive and difficult solution to the database search problem in a way that goes beyond the traditional, purely formulaic expressions. The method's graphical environment, along with its computational and probabilistic architectures, represents a rich package that offers analysts and discussants with additional modes of interaction, concise representation, and coherent communication.
Resumo:
The statistical analysis of literary style is the part of stylometry that compares measurable characteristicsin a text that are rarely controlled by the author, with those in other texts. When thegoal is to settle authorship questions, these characteristics should relate to the author’s style andnot to the genre, epoch or editor, and they should be such that their variation between authors islarger than the variation within comparable texts from the same author.For an overview of the literature on stylometry and some of the techniques involved, see for exampleMosteller and Wallace (1964, 82), Herdan (1964), Morton (1978), Holmes (1985), Oakes (1998) orLebart, Salem and Berry (1998).Tirant lo Blanc, a chivalry book, is the main work in catalan literature and it was hailed to be“the best book of its kind in the world” by Cervantes in Don Quixote. Considered by writterslike Vargas Llosa or Damaso Alonso to be the first modern novel in Europe, it has been translatedseveral times into Spanish, Italian and French, with modern English translations by Rosenthal(1996) and La Fontaine (1993). The main body of this book was written between 1460 and 1465,but it was not printed until 1490.There is an intense and long lasting debate around its authorship sprouting from its first edition,where its introduction states that the whole book is the work of Martorell (1413?-1468), while atthe end it is stated that the last one fourth of the book is by Galba (?-1490), after the death ofMartorell. Some of the authors that support the theory of single authorship are Riquer (1990),Chiner (1993) and Badia (1993), while some of those supporting the double authorship are Riquer(1947), Coromines (1956) and Ferrando (1995). For an overview of this debate, see Riquer (1990).Neither of the two candidate authors left any text comparable to the one under study, and thereforediscriminant analysis can not be used to help classify chapters by author. By using sample textsencompassing about ten percent of the book, and looking at word length and at the use of 44conjunctions, prepositions and articles, Ginebra and Cabos (1998) detect heterogeneities that mightindicate the existence of two authors. By analyzing the diversity of the vocabulary, Riba andGinebra (2000) estimates that stylistic boundary to be near chapter 383.Following the lead of the extensive literature, this paper looks into word length, the use of the mostfrequent words and into the use of vowels in each chapter of the book. Given that the featuresselected are categorical, that leads to three contingency tables of ordered rows and therefore tothree sequences of multinomial observations.Section 2 explores these sequences graphically, observing a clear shift in their distribution. Section 3describes the problem of the estimation of a suden change-point in those sequences, in the followingsections we propose various ways to estimate change-points in multinomial sequences; the methodin section 4 involves fitting models for polytomous data, the one in Section 5 fits gamma modelsonto the sequence of Chi-square distances between each row profiles and the average profile, theone in Section 6 fits models onto the sequence of values taken by the first component of thecorrespondence analysis as well as onto sequences of other summary measures like the averageword length. In Section 7 we fit models onto the marginal binomial sequences to identify thefeatures that distinguish the chapters before and after that boundary. Most methods rely heavilyon the use of generalized linear models
Resumo:
BACKGROUND Ovarian carcinoma is the most important cause of gynecological cancer-related mortality in Western societies. Despite the improved median overall survival in patients receiving chemotherapy regimens such as paclitaxel and carboplatin combination, relapse still occurs in most advanced diseased patients. Increased angiogenesis is associated with rapid recurrence and decreased survival in ovarian cancer. This study was planned to identify an angiogenesis-related gene expression profile with prognostic value in advanced ovarian carcinoma patients. METHODOLOGY/PRINCIPAL FINDINGS RNAs were collected from formalin-fixed paraffin-embedded samples of 61 patients with III/IV FIGO stage ovarian cancer who underwent surgical cytoreduction and received a carboplatin plus paclitaxel regimen. Expression levels of 82 angiogenesis related genes were measured by quantitative real-time polymerase chain reaction using TaqMan low-density arrays. A 34-gene-profile which was able to predict the overall survival of ovarian carcinoma patients was identified. After a leave-one-out cross validation, the profile distinguished two groups of patients with different outcomes. Median overall survival and progression-free survival for the high risk group was 28.3 and 15.0 months, respectively, and was not reached by patients in the low risk group at the end of follow-up. Moreover, the profile maintained an independent prognostic value in the multivariate analysis. The hazard ratio for death was 2.3 (95% CI, 1.5 to 3.2; p<0.001). CONCLUSIONS/SIGNIFICANCE It is possible to generate a prognostic model for advanced ovarian carcinoma based on angiogenesis-related genes using formalin-fixed paraffin-embedded samples. The present results are consistent with the increasing weight of angiogenesis genes in the prognosis of ovarian carcinoma.