852 resultados para Initial data problem
Resumo:
A detailed investigation has been conducted on core samples taken from 17 portland cement concrete pavements located in Iowa. The goal of the investigation was to help to clarify the root cause of the premature deterioration problem that has become evident since the early 1990s. Laboratory experiments were also conducted to evaluate how cement composition, mixing time, and admixtures could have influenced the occurrence of premature deterioration. The cements used in this study were selected in an attempt to cover the main compositional parameters pertinent to the construction industry in Iowa. The hardened air content determinations conducted during this study indicated that the pavements that exhibited premature deterioration often contained poor to marginal entrained-air void systems. In addition, petrographic studies indicated that sometimes the entrained-air void system had been marginal after mixing and placement of the pavement slab, while in other instances a marginal to adequate entrained-air void system had been filled with ettringite. The filling was most probably accelerated because of shrinkage cracking at the surface of the concrete pavements. The results of this study suggest that the durability—more sciecifically, the frost resistance—of the concrete pavements should be less than anticipated during the design stage of the pavements. Construction practices played a significant role in the premature deterioration problem. The pavements that exhibited premature distress also exhibited features that suggested poor mixing and poor control of aggregate grading. Segregation was very common in the cores extracted from the pavements that exhibited premature distress. This suggests that the vibrators on the paver were used to overcome a workability problem. Entrained-air voids formed in concrete mixtures experiencing these types of problems normally tend to be extremely coarse, and hence they can easily be lost during the paving process. This tends to leave the pavement with a low air content and a poor distribution of air voids. All of these features were consistent with a premature stiffening problem that drastically influenced the ability of the contractor to place the concrete mixture. Laboratory studies conducted during this project indicated that most premature stiffening problems can be directly attributed to the portland cement used on the project. The admixtures (class C fly ash and water reducer) tended to have only a minor influence on the premature stiffening problem when they were used at the dosage rates described in this study.
Resumo:
A number of experimental methods have been reported for estimating the number of genes in a genome, or the closely related coding density of a genome, defined as the fraction of base pairs in codons. Recently, DNA sequence data representative of the genome as a whole have become available for several organisms, making the problem of estimating coding density amenable to sequence analytic methods. Estimates of coding density for a single genome vary widely, so that methods with characterized error bounds have become increasingly desirable. We present a method to estimate the protein coding density in a corpus of DNA sequence data, in which a ‘coding statistic’ is calculated for a large number of windows of the sequence under study, and the distribution of the statistic is decomposed into two normal distributions, assumed to be the distributions of the coding statistic in the coding and noncoding fractions of the sequence windows. The accuracy of the method is evaluated using known data and application is made to the yeast chromosome III sequence and to C.elegans cosmid sequences. It can also be applied to fragmentary data, for example a collection of short sequences determined in the course of STS mapping.
Resumo:
Acquiring lexical information is a complex problem, typically approached by relying on a number of contexts to contribute information for classification. One of the first issues to address in this domain is the determination of such contexts. The work presented here proposes the use of automatically obtained FORMAL role descriptors as features used to draw nouns from the same lexical semantic class together in an unsupervised clustering task. We have dealt with three lexical semantic classes (HUMAN, LOCATION and EVENT) in English. The results obtained show that it is possible to discriminate between elements from different lexical semantic classes using only FORMAL role information, hence validating our initial hypothesis. Also, iterating our method accurately accounts for fine-grained distinctions within lexical classes, namely distinctions involving ambiguous expressions. Moreover, a filtering and bootstrapping strategy employed in extracting FORMAL role descriptors proved to minimize effects of sparse data and noise in our task.
Resumo:
AbstractFor a wide range of environmental, hydrological, and engineering applications there is a fast growing need for high-resolution imaging. In this context, waveform tomographic imaging of crosshole georadar data is a powerful method able to provide images of pertinent electrical properties in near-surface environments with unprecedented spatial resolution. In contrast, conventional ray-based tomographic methods, which consider only a very limited part of the recorded signal (first-arrival traveltimes and maximum first-cycle amplitudes), suffer from inherent limitations in resolution and may prove to be inadequate in complex environments. For a typical crosshole georadar survey the potential improvement in resolution when using waveform-based approaches instead of ray-based approaches is in the range of one order-of- magnitude. Moreover, the spatial resolution of waveform-based inversions is comparable to that of common logging methods. While in exploration seismology waveform tomographic imaging has become well established over the past two decades, it is comparably still underdeveloped in the georadar domain despite corresponding needs. Recently, different groups have presented finite-difference time-domain waveform inversion schemes for crosshole georadar data, which are adaptations and extensions of Tarantola's seminal nonlinear generalized least-squares approach developed for the seismic case. First applications of these new crosshole georadar waveform inversion schemes on synthetic and field data have shown promising results. However, there is little known about the limits and performance of such schemes in complex environments. To this end, the general motivation of my thesis is the evaluation of the robustness and limitations of waveform inversion algorithms for crosshole georadar data in order to apply such schemes to a wide range of real world problems.One crucial issue to making applicable and effective any waveform scheme to real-world crosshole georadar problems is the accurate estimation of the source wavelet, which is unknown in reality. Waveform inversion schemes for crosshole georadar data require forward simulations of the wavefield in order to iteratively solve the inverse problem. Therefore, accurate knowledge of the source wavelet is critically important for successful application of such schemes. Relatively small differences in the estimated source wavelet shape can lead to large differences in the resulting tomograms. In the first part of my thesis, I explore the viability and robustness of a relatively simple iterative deconvolution technique that incorporates the estimation of the source wavelet into the waveform inversion procedure rather than adding additional model parameters into the inversion problem. Extensive tests indicate that this source wavelet estimation technique is simple yet effective, and is able to provide remarkably accurate and robust estimates of the source wavelet in the presence of strong heterogeneity in both the dielectric permittivity and electrical conductivity as well as significant ambient noise in the recorded data. Furthermore, our tests also indicate that the approach is insensitive to the phase characteristics of the starting wavelet, which is not the case when directly incorporating the wavelet estimation into the inverse problem.Another critical issue with crosshole georadar waveform inversion schemes which clearly needs to be investigated is the consequence of the common assumption of frequency- independent electromagnetic constitutive parameters. This is crucial since in reality, these parameters are known to be frequency-dependent and complex and thus recorded georadar data may show significant dispersive behaviour. In particular, in the presence of water, there is a wide body of evidence showing that the dielectric permittivity can be significantly frequency dependent over the GPR frequency range, due to a variety of relaxation processes. The second part of my thesis is therefore dedicated to the evaluation of the reconstruction limits of a non-dispersive crosshole georadar waveform inversion scheme in the presence of varying degrees of dielectric dispersion. I show that the inversion algorithm, combined with the iterative deconvolution-based source wavelet estimation procedure that is partially able to account for the frequency-dependent effects through an "effective" wavelet, performs remarkably well in weakly to moderately dispersive environments and has the ability to provide adequate tomographic reconstructions.
Resumo:
One of the most important issues in portland cement concrete pavement research today is surface characteristics. The issue is one of balancing surface texture construction with the need for durability, skid resistance, and noise reduction. The National Concrete Pavement Technology Center at Iowa State University, in conjunction with the Federal Highway Administration, American Concrete Pavement Association, International Grinding and Grooving Association, Iowa Highway Research Board, and other states, have entered into a three-part National Surface Characteristics Program to resolve the balancing problem. As a portion of Part 2, this report documents the construction of 18 separate pavement surfaces for use in the first level of testing for the national project. It identifies the testing to be done and the limitations observed in the construction process. The results of the actual tests will be included in the subsequent national study reports.
Resumo:
From a managerial point of view, the more effcient, simple, and parameter-free (ESP) an algorithm is, the more likely it will be used in practice for solving real-life problems. Following this principle, an ESP algorithm for solving the Permutation Flowshop Sequencing Problem (PFSP) is proposed in this article. Using an Iterated Local Search (ILS) framework, the so-called ILS-ESP algorithm is able to compete in performance with other well-known ILS-based approaches, which are considered among the most effcient algorithms for the PFSP. However, while other similar approaches still employ several parameters that can affect their performance if not properly chosen, our algorithm does not require any particular fine-tuning process since it uses basic "common sense" rules for the local search, perturbation, and acceptance criterion stages of the ILS metaheuristic. Our approach defines a new operator for the ILS perturbation process, a new acceptance criterion based on extremely simple and transparent rules, and a biased randomization process of the initial solution to randomly generate different alternative initial solutions of similar quality -which is attained by applying a biased randomization to a classical PFSP heuristic. This diversification of the initial solution aims at avoiding poorly designed starting points and, thus, allows the methodology to take advantage of current trends in parallel and distributed computing. A set of extensive tests, based on literature benchmarks, has been carried out in order to validate our algorithm and compare it against other approaches. These tests show that our parameter-free algorithm is able to compete with state-of-the-art metaheuristics for the PFSP. Also, the experiments show that, when using parallel computing, it is possible to improve the top ILS-based metaheuristic by just incorporating to it our biased randomization process with a high-quality pseudo-random number generator.
Resumo:
When continuous data are coded to categorical variables, two types of coding are possible: crisp coding in the form of indicator, or dummy, variables with values either 0 or 1; or fuzzy coding where each observation is transformed to a set of "degrees of membership" between 0 and 1, using co-called membership functions. It is well known that the correspondence analysis of crisp coded data, namely multiple correspondence analysis, yields principal inertias (eigenvalues) that considerably underestimate the quality of the solution in a low-dimensional space. Since the crisp data only code the categories to which each individual case belongs, an alternative measure of fit is simply to count how well these categories are predicted by the solution. Another approach is to consider multiple correspondence analysis equivalently as the analysis of the Burt matrix (i.e., the matrix of all two-way cross-tabulations of the categorical variables), and then perform a joint correspondence analysis to fit just the off-diagonal tables of the Burt matrix - the measure of fit is then computed as the quality of explaining these tables only. The correspondence analysis of fuzzy coded data, called "fuzzy multiple correspondence analysis", suffers from the same problem, albeit attenuated. Again, one can count how many correct predictions are made of the categories which have highest degree of membership. But here one can also defuzzify the results of the analysis to obtain estimated values of the original data, and then calculate a measure of fit in the familiar percentage form, thanks to the resultant orthogonal decomposition of variance. Furthermore, if one thinks of fuzzy multiple correspondence analysis as explaining the two-way associations between variables, a fuzzy Burt matrix can be computed and the same strategy as in the crisp case can be applied to analyse the off-diagonal part of this matrix. In this paper these alternative measures of fit are defined and applied to a data set of continuous meteorological variables, which are coded crisply and fuzzily into three categories. Measuring the fit is further discussed when the data set consists of a mixture of discrete and continuous variables.
Resumo:
Sampling of an industrial drill string from the northeastern Paris Basin (Montcornet, France) provides early Jurassic magnetostratigraphic data coupled with biochronological control. About 375 paleomagnetic samples were obtained from a 145 m thick series of Pliensbachian rocks. A composite demagnetization thermal up to 300 C and an alternating field up to 80 mT were used to separate the magnetic components. A low unblocking temperature component (<250degreesC) with an inclination of about 64 is interpreted as a present-day field overprint. The characteristic remanent component with both normal and reversed antipodal directions was isolated between 5 and 50 mT. Twenty-nine polarity intervals were recognized. Correlation of these new results from the Paris Basin with data from the Breggia Gorge section (Ticino, southern Alps, Switzerland), which is generally considered as the reference section for Pliensbachian magnetostratigraphy, reveals almost identical patterns of magnetic polarity reversals. However, the correlation implies significant paleontological age discrepancies. Revised age assignments of biostratigraphic data of Breggia as well as an objective evaluation of the uncertainties on zonal boundaries in both Breggia and Moncornet resolve the initial discrepancies between magnetostratigraphic correlations and biostratigraphic ages. Hence, the sequence of magnetic reversals is significantly strengthened and the age calibration is notably improved for the Pliensbachian, a stage for which sections combining adequate magnetic signal and biostratigraphic constraints are still very few. (C) 2002 Elsevier Science B.V. All rights reserved.
Resumo:
The passage of the Workforce Investment Act (WIA) of 1998 [Public Law 105-220] by the 105th Congress has ushered in a new era of collaboration, coordination, cooperation and accountability. The overall goal of the Act is “to increase the employability, retention, and earnings of participants, and increase occupational skill attainment by participants, and, as a result improve the quality of the workforce, reduce welfare dependency, and enhance the productivity and competitiveness of the Nation.” The key principles inculcated in the Act are: • Streamlining services; • Empowering individuals; • Universal access; • Increased accountability; • New roles for local boards; • State and local flexibility; • Improved youth programs. The purpose of Title II, The Adult Education and Family Literacy Act (AEFLA), of the Workforce Investment Act of 1998 is to create a partnership among the federal government, states, and localities to provide, on a voluntary basis, adult education and literacy services in order to: • Assist adults become literate and obtain the knowledge and skills necessary for employment and self-sufficiency; • Assist adults who are parents obtain the educational skills necessary to become full partners in the educational development of their children; • Assist adults in the completion of a secondary school education. Adult education is an important part of the workforce investment system. Title II restructures and improves programs previously authorized by the Adult Education Act. AEFLA focuses on strengthening program quality by requiring States to give priority in awarding funds to local programs that are based on a solid foundation of research, address the diverse needs of adult learners, and utilize other effective practices and strategies. To promote continuous program involvement and to ensure optimal return on the Federal investment, AEFLA also establishes a State performance accountability system. Under this system, the Secretary and each State must reach agreement on annual levels of performance for a number of “core indicators” specified in the law: • Demonstrated improvements in literacy skill levels in reading, writing, and speaking the English language, numeracy, problem solving, English language acquisition, and other literacy skills. • Placement in, retention in, or completion of postsecondary education, training, unsubsidized employment or career advancement. • Receipt of a secondary school diploma or its recognized equivalent. Iowa’s community college based adult basic education program has implemented a series of proactive strategies in order to effectively and systematically meet the challenges posed by WIA. The Iowa TOPSpro Data Dictionary is a direct result of Iowa’s pro-active efforts in this educational arena.
Resumo:
II Resumo Cabo Verde é um país que tem acompanhado as grandes mudanças políticas, sociais, económicas e educativas, atribuindo um papel preponderante aos professores na educação/formação dos indivíduos. Neste sentido a formação inicial não pode constituir um acto “acabado” que não tem em linha de conta as mudanças que decorrem no seu seio. A formação contínua aparece com um processo de construção da mudança, apoiado no desenvolvimento profissional dos professores. A avaliação de necessidade de formação em professores do ensino secundário em Cabo Verde é o tema do trabalho de investigação, realizado nas escolas secundárias da ilha de Santo Antão, em Cabo Verde, no sentido de identificar questões problemáticas na formação inicial e contínua de professores e as suas implicações na qualidade educativa. Neste sentido, ao longo de 5 capítulos apresentamos os pontos desta investigação, estruturados de forma a se poder acompanhar a sua evolução. No 1º capítulo, apresenta-se a Problemática de Investigação, que constitui a parte inicial da dissertação, pela exposição do quadro conceptual do ensino secundário em Cabo Verde através da contextualização e identificação do problema, da formulação dos objectivos e questões de investigação. No 2º capítulo faz-se o percurso histórico/educativo de Cabo Verde, desde a época colonial, passando pela educação após a independência em 1975, à constituição da Lei de Bases do Sistema Educativo e a Reforma Educativa dos anos 90. No 3º capítulo, intitulado “A Formação Inicial e Continua de Professores em Cabo Verde”, é abordado o enquadramento teórico desta investigação, com referência à contextualização geral da formação e modelos de formação de professores, evoluindo para a realidade Cabo-verdiana, tendo em consideração as instituições de formação de professores e o contexto actual do ensino secundário em Cabo Verde. No 4º capítulo faz-se a apresentação da Metodologia de Investigação, adoptada num estudo extensivo, do qual abordamos a natureza da investigação e caracterizamos a amostra e respondentes. Como técnicas de recolha de dados, são utilizados: o inquérito por questionário (n=77), a professores do ensino secundário e o inquérito por entrevista (n=9), aos directores das escolas secundárias da ilha de Santo Antão e a formadores de professores em São Vicente. Como técnica de análise de dados, são utilizados diversos procedimentos estatísticos e a análise de conteúdo. O 5º capítulo corresponde a apresentação e discussão dos resultados da investigação desses três públicos analisados. Para finalizar, apresenta-se uma conclusão que corresponde a síntese dos resultados obtidos, propostas de sugestões de melhoria e limitações do estudo. Cabo Verde is a country willing to follow the great political, social, economic and educational changes. This important role have attributed to teachers to educate and train individuals. In this sense the initial training cannot be considered as a “finished” task without taking into consideration those changes. The in service training appears as a scientific and pedagogic autonomization towards the growing change. The Evaluation of Secondary School teacher training Needs in Cabo Verde constitutes an investigation work, carried out in Santo Antão's island Secondary schools, aiming to identify problems in initial and continuous teachers' training and their implications in the educational quality. In this sense, along five chapters we will present the points of this investigation, structured to facilitate the understanding of its evolution. In the first chapter we present the Problem of Investigation which is an introductory part of the dissertation, the conceptual theory of the Cabo Verde secondary education through the contextualization and identification of the problem, the formulation of the objectives and investigation issues. In the second chapter we present the Cabo Verde historical/educational background, from the colonial time, going through after independence in 1975, to the constitution of the Basic Educational Law of the System and the nineties Educational Reform. In the third chapter entitled “The Initial and in-service Teacher´s Training in Cabo Verde” brings the theoretical framework of this investigation, where we present a general contextualization of teachers' training and models, developed from the reality of Cabo Verde. In this regard we will consider the teachers' training institutions and the current secondary education context in Cabo Verde. In the fourth chapter we present the Research Methodology, adopted in an extensive study, of which we draw the nature of the investigation and we characterized the sample. We carried out questionnaire (n=77) to teachers, interviews (n=9) to headmasters from Santo Antão's island secondary schools and trainers from São Vicente. We used statistical analyses and content analyses as data analyse techniques. The fifth chapter deals with presentation the discussion of results of those three analyzed publics' investigation. Finally, we present a summary of the obtained results, suggestions for improvement and limitations of study.
Resumo:
One of the disadvantages of old age is that there is more past than future: this,however, may be turned into an advantage if the wealth of experience and, hopefully,wisdom gained in the past can be reflected upon and throw some light on possiblefuture trends. To an extent, then, this talk is necessarily personal, certainly nostalgic,but also self critical and inquisitive about our understanding of the discipline ofstatistics. A number of almost philosophical themes will run through the talk: searchfor appropriate modelling in relation to the real problem envisaged, emphasis onsensible balances between simplicity and complexity, the relative roles of theory andpractice, the nature of communication of inferential ideas to the statistical layman, theinter-related roles of teaching, consultation and research. A list of keywords might be:identification of sample space and its mathematical structure, choices betweentransform and stay, the role of parametric modelling, the role of a sample spacemetric, the underused hypothesis lattice, the nature of compositional change,particularly in relation to the modelling of processes. While the main theme will berelevance to compositional data analysis we shall point to substantial implications forgeneral multivariate analysis arising from experience of the development ofcompositional data analysis…
Resumo:
It is common in econometric applications that several hypothesis tests arecarried out at the same time. The problem then becomes how to decide whichhypotheses to reject, accounting for the multitude of tests. In this paper,we suggest a stepwise multiple testing procedure which asymptoticallycontrols the familywise error rate at a desired level. Compared to relatedsingle-step methods, our procedure is more powerful in the sense that itoften will reject more false hypotheses. In addition, we advocate the useof studentization when it is feasible. Unlike some stepwise methods, ourmethod implicitly captures the joint dependence structure of the teststatistics, which results in increased ability to detect alternativehypotheses. We prove our method asymptotically controls the familywise errorrate under minimal assumptions. We present our methodology in the context ofcomparing several strategies to a common benchmark and deciding whichstrategies actually beat the benchmark. However, our ideas can easily beextended and/or modied to other contexts, such as making inference for theindividual regression coecients in a multiple regression framework. Somesimulation studies show the improvements of our methods over previous proposals. We also provide an application to a set of real data.
Resumo:
In this paper we study the disability transition probabilities (as well as the mortalityprobabilities) due to concurrent factors to age such as income, gender and education. Althoughit is well known that ageing and socioeconomic status influence the probability ofcausing functional disorders, surprisingly little attention has been paid to the combined effectof those factors along the individuals' life and how this affects the transition from one degreeof disability to another. The assumption that tomorrow's disability state is only a functionof the today's state is very strong, since disability is a complex variable that depends onseveral other elements than time. This paper contributes into the field in two ways: (1) byattending the distinction between the initial disability level and the process that leads tohis course (2) by addressing whether and how education, age and income differentially affectthe disability transitions. Using a Markov chain discrete model and a survival analysis, weestimate the probability by year and individual characteristics that changes the state of disabilityand the duration that it takes its progression in each case. We find that people withan initial state of disability have a higher propensity to change and take less time to transitfrom different stages. Men do that more frequently than women. Education and incomehave negative effects on transition. Moreover, we consider the disability benefits associatedto those changes along different stages of disability and therefore we offer some clues onthe potential savings of preventive actions that may delay or avoid those transitions. Onpure cost considerations, preventive programs for improvement show higher benefits thanthose for preventing deterioration, and in general terms, those focussing individuals below65 should go first. Finally the trend of disability in Spain seems not to change among yearsand regional differences are not found.
Resumo:
BACKGROUND: Socioeconomic status is thought to have a significant influence on stroke incidence, risk factors and outcome. Its influence on acute stroke severity, stroke mechanisms, and acute recanalisation treatment is less known. METHODS: Over a 4-year period, all ischaemic stroke patients admitted within 24 h were entered prospectively in a stroke registry. Data included insurance status, demographics, risk factors, time to hospital arrival, initial stroke severity (NIHSS), etiology, use of acute treatments, short-term outcome (modified Rankin Scale, mRS). Private insured patients (PI) were compared with basic insured patients (BI). RESULTS: Of 1062 consecutive acute ischaemic stroke patients, 203 had PI and 859 had BI. They were 585 men and 477 women. Both populations were similar in age, cardiovascular risk factors and preventive medications. The onset to admission time, thrombolysis rate, and stroke etiology according to TOAST classification were not different between PI and BI. Mean NIHSS at admission was significantly higher for BI. Good outcome (mRS ≤ 2) at 7 days and 3 months was more frequent in PI than in BI. CONCLUSION: We found better outcome and lesser stroke severity on admission in patients with higher socioeconomic status in an acute stroke population. The reason for milder strokes in patients with better socioeconomic status in a universal health care system needs to be explained.
Resumo:
The Drivers Scheduling Problem (DSP) consists of selecting a set of duties for vehicle drivers, for example buses, trains, plane or boat drivers or pilots, for the transportation of passengers or goods. This is a complex problem because it involves several constraints related to labour and company rules and can also present different evaluation criteria and objectives. Being able to develop an adequate model for this problem that can represent the real problem as close as possible is an important research area.The main objective of this research work is to present new mathematical models to the DSP problem that represent all the complexity of the drivers scheduling problem, and also demonstrate that the solutions of these models can be easily implemented in real situations. This issue has been recognized by several authors and as important problem in Public Transportation. The most well-known and general formulation for the DSP is a Set Partition/Set Covering Model (SPP/SCP). However, to a large extend these models simplify some of the specific business aspects and issues of real problems. This makes it difficult to use these models as automatic planning systems because the schedules obtained must be modified manually to be implemented in real situations. Based on extensive passenger transportation experience in bus companies in Portugal, we propose new alternative models to formulate the DSP problem. These models are also based on Set Partitioning/Covering Models; however, they take into account the bus operator issues and the perspective opinions and environment of the user.We follow the steps of the Operations Research Methodology which consist of: Identify the Problem; Understand the System; Formulate a Mathematical Model; Verify the Model; Select the Best Alternative; Present the Results of theAnalysis and Implement and Evaluate. All the processes are done with close participation and involvement of the final users from different transportation companies. The planner s opinion and main criticisms are used to improve the proposed model in a continuous enrichment process. The final objective is to have a model that can be incorporated into an information system to be used as an automatic tool to produce driver schedules. Therefore, the criteria for evaluating the models is the capacity to generate real and useful schedules that can be implemented without many manual adjustments or modifications. We have considered the following as measures of the quality of the model: simplicity, solution quality and applicability. We tested the alternative models with a set of real data obtained from several different transportation companies and analyzed the optimal schedules obtained with respect to the applicability of the solution to the real situation. To do this, the schedules were analyzed by the planners to determine their quality and applicability. The main result of this work is the proposition of new mathematical models for the DSP that better represent the realities of the passenger transportation operators and lead to better schedules that can be implemented directly in real situations.