896 resultados para Software testing. Test generation. Grammars


Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper discusses the role of deterministic components in the DGP and in the auxiliaryregression model which underlies the implementation of the Fractional Dickey-Fuller (FDF) test for I(1) against I(d) processes with d [0, 1). This is an important test in many economic applications because I(d) processess with d < 1 are mean-reverting although, when 0.5 = d < 1, like I(1) processes, they are nonstationary. We show how simple is the implementation of the FDF in these situations, and argue that it has better properties than LM tests. A simple testing strategy entailing only asymptotically normally distributedtests is also proposed. Finally, an empirical application is provided where the FDF test allowing for deterministic components is used to test for long-memory in the per capita GDP of several OECD countries, an issue that has important consequences to discriminate between growth theories, and on which there is some controversy.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Small sample properties are of fundamental interest when only limited data is avail-able. Exact inference is limited by constraints imposed by speci.c nonrandomizedtests and of course also by lack of more data. These e¤ects can be separated as we propose to evaluate a test by comparing its type II error to the minimal type II error among all tests for the given sample. Game theory is used to establish this minimal type II error, the associated randomized test is characterized as part of a Nash equilibrium of a .ctitious game against nature.We use this method to investigate sequential tests for the di¤erence between twomeans when outcomes are constrained to belong to a given bounded set. Tests ofinequality and of noninferiority are included. We .nd that inference in terms oftype II error based on a balanced sample cannot be improved by sequential sampling or even by observing counter factual evidence providing there is a reasonable gap between the hypotheses.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The paper proposes a technique to jointly test for groupings of unknown size in the cross sectional dimension of a panel and estimates the parameters of each group, and applies it to identifying convergence clubs in income per-capita. The approach uses the predictive density of the data, conditional on the parameters of the model. The steady state distribution of European regional data clusters around four poles of attraction with different economic features. The distribution of incomeper-capita of OECD countries has two poles of attraction and each grouphas clearly identifiable economic characteristics.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Expected utility theory (EUT) has been challenged as a descriptive theoryin many contexts. The medical decision analysis context is not an exception.Several researchers have suggested that rank dependent utility theory (RDUT)may accurately describe how people evaluate alternative medical treatments.Recent research in this domain has addressed a relevant feature of RDU models-probability weighting-but to date no direct test of this theoryhas been made. This paper provides a test of the main axiomatic differencebetween EUT and RDUT when health profiles are used as outcomes of riskytreatments. Overall, EU best described the data. However, evidence on theediting and cancellation operation hypothesized in Prospect Theory andCumulative Prospect Theory was apparent in our study. we found that RDUoutperformed EU in the presentation of the risky treatment pairs in whichthe common outcome was not obvious. The influence of framing effects onthe performance of RDU and their importance as a topic for future researchis discussed.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Consider the problem of testing k hypotheses simultaneously. In this paper,we discuss finite and large sample theory of stepdown methods that providecontrol of the familywise error rate (FWE). In order to improve upon theBonferroni method or Holm's (1979) stepdown method, Westfall and Young(1993) make eective use of resampling to construct stepdown methods thatimplicitly estimate the dependence structure of the test statistics. However,their methods depend on an assumption called subset pivotality. The goalof this paper is to construct general stepdown methods that do not requiresuch an assumption. In order to accomplish this, we take a close look atwhat makes stepdown procedures work, and a key component is a monotonicityrequirement of critical values. By imposing such monotonicity on estimatedcritical values (which is not an assumption on the model but an assumptionon the method), it is demonstrated that the problem of constructing a validmultiple test procedure which controls the FWE can be reduced to the problemof contructing a single test which controls the usual probability of a Type 1error. This reduction allows us to draw upon an enormous resamplingliterature as a general means of test contruction.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

We consider a dynamic multifactor model of investment with financing imperfections,adjustment costs and fixed and variable capital. We use the model to derive a test offinancing constraints based on a reduced form variable capital equation. Simulation resultsshow that this test correctly identifies financially constrained firms even when the estimationof firms investment opportunities is very noisy. In addition, the test is well specified inthe presence of both concave and convex adjustment costs of fixed capital. We confirmempirically the validity of this test on a sample of small Italian manufacturing companies.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This paper illustrates the philosophy which forms the basis of calibrationexercises in general equilibrium macroeconomic models and the details of theprocedure, the advantages and the disadvantages of the approach, with particularreference to the issue of testing ``false'' economic models. We provide anoverview of the most recent simulation--based approaches to the testing problemand compare them to standard econometric methods used to test the fit of non--lineardynamic general equilibrium models. We illustrate how simulation--based techniques can be used to formally evaluate the fit of a calibrated modelto the data and obtain ideas on how to improve the model design using a standardproblem in the international real business cycle literature, i.e. whether amodel with complete financial markets and no restrictions to capital mobility is able to reproduce the second order properties of aggregate savingand aggregate investment in an open economy.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Studies assessing skin irritation to chemicals have traditionally used laboratory animals; however, such methods are questionable regarding their relevance for humans. New in vitro methods have been validated, such as the reconstructed human epidermis (RHE) model (Episkin®, Epiderm®). The comparison (accuracy) with in vivo results such as the 4-h human patch test (HPT) is 76% at best (Epiderm®). There is a need to develop an in vitro method that better simulates the anatomo-pathological changes encountered in vivo. To develop an in vitro method to determine skin irritation using human viable skin through histopathology, and compare the results of 4 tested substances to the main in vitro methods and in vivo animal method (Draize test). Human skin removed during surgery was dermatomed and mounted on an in vitro flow-through diffusion cell system. Ten chemicals with known non-irritant (heptylbutyrate, hexylsalicylate, butylmethacrylate, isoproturon, bentazon, DEHP and methylisothiazolinone (MI)) and irritant properties (folpet, 1-bromohexane and methylchloroisothiazolinone (MCI/MI)), a negative control (sodiumchloride) and a positive control (sodiumlaurylsulphate) were applied. The skin was exposed at least for 4h. Histopathology was performed to investigate irritation signs (spongiosis, necrosis, vacuolization). We obtained 100% accuracy with the HPT model; 75% with the RHE models and 50% with the Draize test for 4 tested substances. The coefficients of variation (CV) between our three test batches were <0.1, showing good reproducibility. Furthermore, we reported objectively histopathological irritation signs (irritation scale): strong (folpet), significant (1-bromohexane), slight (MCI/MI at 750/250ppm) and none (isoproturon, bentazon, DEHP and MI). This new in vitro test method presented effective results for the tested chemicals. It should be further validated using a greater number of substances; and tested in different laboratories in order to suitably evaluate reproducibility.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Revenue management (RM) is a complicated business process that can best be described ascontrol of sales (using prices, restrictions, or capacity), usually using software as a tool to aiddecisions. RM software can play a mere informative role, supplying analysts with formatted andsummarized data who use it to make control decisions (setting a price or allocating capacity fora price point), or, play a deeper role, automating the decisions process completely, at the otherextreme. The RM models and algorithms in the academic literature by and large concentrateon the latter, completely automated, level of functionality.A firm considering using a new RM model or RM system needs to evaluate its performance.Academic papers justify the performance of their models using simulations, where customerbooking requests are simulated according to some process and model, and the revenue perfor-mance of the algorithm compared to an alternate set of algorithms. Such simulations, whilean accepted part of the academic literature, and indeed providing research insight, often lackcredibility with management. Even methodologically, they are usually awed, as the simula-tions only test \within-model" performance, and say nothing as to the appropriateness of themodel in the first place. Even simulations that test against alternate models or competition arelimited by their inherent necessity on fixing some model as the universe for their testing. Theseproblems are exacerbated with RM models that attempt to model customer purchase behav-ior or competition, as the right models for competitive actions or customer purchases remainsomewhat of a mystery, or at least with no consensus on their validity.How then to validate a model? Putting it another way, we want to show that a particularmodel or algorithm is the cause of a certain improvement to the RM process compared to theexisting process. We take care to emphasize that we want to prove the said model as the causeof performance, and to compare against a (incumbent) process rather than against an alternatemodel.In this paper we describe a \live" testing experiment that we conducted at Iberia Airlineson a set of flights. A set of competing algorithms control a set of flights during adjacentweeks, and their behavior and results are observed over a relatively long period of time (9months). In parallel, a group of control flights were managed using the traditional mix of manualand algorithmic control (incumbent system). Such \sandbox" testing, while common at manylarge internet search and e-commerce companies is relatively rare in the revenue managementarea. Sandbox testing has an undisputable model of customer behavior but the experimentaldesign and analysis of results is less clear. In this paper we describe the philosophy behind theexperiment, the organizational challenges, the design and setup of the experiment, and outlinethe analysis of the results. This paper is a complement to a (more technical) related paper thatdescribes the econometrics and statistical analysis of the results.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Purpose: To evaluate whether parametric imaging with contrast material-enhanced ultrasonography (US) is superior to visual assessment for the differential diagnosis of focal liver lesions (FLLs). Materials and Methods: This study had institutional review board approval, and verbal patient informed consent was obtained. Between August 2005 and October 2008, 146 FLLs in 145 patients (63 women, 82 men; mean age, 62.5 years; age range, 22-89 years) were imaged with real-time low-mechanical-index contrast-enhanced US after a bolus injection of 2.4 mL of a second-generation contrast agent. Clips showing contrast agent uptake kinetics (including arterial, portal, and late phases) were recorded and subsequently analyzed off-line with dedicated image processing software. Analysis of the dynamic vascular patterns (DVPs) of lesions with respect to adjacent parenchyma allowed mapping DVP signatures on a single parametric image. Cine loops of contrast-enhanced US and results from parametric imaging of DVP were assessed separately by three independent off-site readers who classified each lesion as benign, malignant, or indeterminate. Sensitivity, specificity, accuracy, and positive and negative predictive values were calculated for both techniques. Interobserver agreement (κ statistics) was determined. Results: Sensitivities for visual interpretation of cine loops for the three readers were 85.0%, 77.9%, and 87.6%, which improved significantly to 96.5%, 97.3%, and 96.5% for parametric imaging, respectively (P < .05, McNemar test), while retaining high specificity (90.9% for all three readers). Accuracy scores of parametric imaging were higher than those of conventional contrast-enhanced US for all three readers (P < .001, McNemar test). Interobserver agreement increased with DVP parametric imaging compared with conventional contrast-enhanced US (change of κ from 0.54 to 0.99). Conclusion: Parametric imaging of DVP improves diagnostic performance of contrast-enhanced US in the differentiation between malignant and benign FLLs; it also provides excellent interobserver agreement.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

BACKGROUND: Since the advent of combined antiretroviral therapy (ART), the incidence of non-AIDS-defining cancers (non-ADCs) among HIV-positive patients is rising. We previously described HIV testing rates of <5% in our oncology centre, against a local HIV prevalence of 0.4% (1). We have since worked with the Service of Oncology to identify, how HIV testing can be optimized, we have conducted a study on investigating barriers in HIV-testing oncology patients (IBITOP) among treating oncologists and their patients. METHODS: After an initial two-month pilot study to examine feasibility (2), we conducted the first phase of the IBITOP study between 1st July and 31st October 2013. Patients of unknown HIV status, newly diagnosed with solid-organ non-AIDS-defining cancer, and treated at Lausanne University Hospital were invited to participate. Patients were offered HIV testing as a part of their initial oncology work-up. Oncologist testing proposals and patient acceptance were the primary endpoints. RESULTS: Of 235 patients with a new oncology diagnosis, 10 were excluded (7 with ADCs and 3 of known HIV-positive status). Mean age was 62 years; 48% were men and 71% were Swiss. Of 225 patients, 75 (33%) were offered HIV testing. Of these, 56 (75%) accepted, of whom 52 (93%) were tested. A further ten patients were tested (without documentation of being offered a test), which gave a total testing rate of 28% (62/225). Among the 19 patients who declined testing, reasons cited included self-perceived absence of HIV risk, previous testing and palliative care. Of the 140 patients not offered HIV testing and not tested, reasons were documented for 35 (25%), the most common being previous testing and follow-up elsewhere. None of the 62 patients HIV tested had a reactive test. CONCLUSIONS: In this study, one third of patients seen were offered testing and the HIV testing rate was fivefold higher than that of previously observed in this service. Most patients accepted testing when offered. As HIV-positive status impacts on the medical management of cancer patients, we recommend that HIV screening should be performed in settings, where HIV prevalence is >0.1%. Phase II of the IBITOP study is now underway to explore barriers to HIV screening among oncologists and patients following the updated national HIV testing guidelines which recommend testing in non-ADC patients undergoing chemotherapy.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

A family of scaling corrections aimed to improve the chi-square approximation of goodness-of-fit test statistics in small samples, large models, and nonnormal data was proposed in Satorra and Bentler (1994). For structural equations models, Satorra-Bentler's (SB) scaling corrections are available in standard computer software. Often, however, the interest is not on the overall fit of a model, but on a test of the restrictions that a null model say ${\cal M}_0$ implies on a less restricted one ${\cal M}_1$. If $T_0$ and $T_1$ denote the goodness-of-fit test statistics associated to ${\cal M}_0$ and ${\cal M}_1$, respectively, then typically the difference $T_d = T_0 - T_1$ is used as a chi-square test statistic with degrees of freedom equal to the difference on the number of independent parameters estimated under the models ${\cal M}_0$ and ${\cal M}_1$. As in the case of the goodness-of-fit test, it is of interest to scale the statistic $T_d$ in order to improve its chi-square approximation in realistic, i.e., nonasymptotic and nonnormal, applications. In a recent paper, Satorra (1999) shows that the difference between two Satorra-Bentler scaled test statistics for overall model fit does not yield the correct SB scaled difference test statistic. Satorra developed an expression that permits scaling the difference test statistic, but his formula has some practical limitations, since it requires heavy computations that are notavailable in standard computer software. The purpose of the present paper is to provide an easy way to compute the scaled difference chi-square statistic from the scaled goodness-of-fit test statistics of models ${\cal M}_0$ and ${\cal M}_1$. A Monte Carlo study is provided to illustrate the performance of the competing statistics.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Although it is commonly accepted that most macroeconomic variables are nonstationary, it is often difficult to identify the source of the non-stationarity. In particular, it is well-known that integrated and short memory models containing trending components that may display sudden changes in their parameters share some statistical properties that make their identification a hard task. The goal of this paper is to extend the classical testing framework for I(1) versus I(0)+ breaks by considering a a more general class of models under the null hypothesis: non-stationary fractionally integrated (FI) processes. A similar identification problem holds in this broader setting which is shown to be a relevant issue from both a statistical and an economic perspective. The proposed test is developed in the time domain and is very simple to compute. The asymptotic properties of the new technique are derived and it is shown by simulation that it is very well-behaved in finite samples. To illustrate the usefulness of the proposed technique, an application using inflation data is also provided.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The drug discovery process has been deeply transformed recently by the use of computational ligand-based or structure-based methods, helping the lead compounds identification and optimization, and finally the delivery of new drug candidates more quickly and at lower cost. Structure-based computational methods for drug discovery mainly involve ligand-protein docking and rapid binding free energy estimation, both of which require force field parameterization for many drug candidates. Here, we present a fast force field generation tool, called SwissParam, able to generate, for arbitrary small organic molecule, topologies, and parameters based on the Merck molecular force field, but in a functional form that is compatible with the CHARMM force field. Output files can be used with CHARMM or GROMACS. The topologies and parameters generated by SwissParam are used by the docking software EADock2 and EADock DSS to describe the small molecules to be docked, whereas the protein is described by the CHARMM force field, and allow them to reach success rates ranging from 56 to 78%. We have also developed a rapid binding free energy estimation approach, using SwissParam for ligands and CHARMM22/27 for proteins, which requires only a short minimization to reproduce the experimental binding free energy of 214 ligand-protein complexes involving 62 different proteins, with a standard error of 2.0 kcal mol(-1), and a correlation coefficient of 0.74. Together, these results demonstrate the relevance of using SwissParam topologies and parameters to describe small organic molecules in computer-aided drug design applications, together with a CHARMM22/27 description of the target protein. SwissParam is available free of charge for academic users at www.swissparam.ch.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Background: Bacteria form biofilms on the surface of orthopaedic devices, causing persistent infections. Monitoring biofilm formation on bone grafts and bone substitutes is challenging due to heterogeneous surface characteristics. We analyzed various bone grafts and bone substitutes regarding their propensity for in-vitro biofilm formation caused by S. aureus and S. epidermidis. Methods: Beta-tricalciumphosphate (b-TCP, ChronOsTM), processed human spongiosa (TutoplastTM) and PMMA (PalacosTM) were investigated. PE was added as a growth control. As test strains S. aureus (ATCC 29213) and S. epidermidis RP62A (ATCC 35984) were used. Test materials were incubated with 105 cfu/ml. After 24 h, test materials were removed and washed, followed by a standardised sonication protocol. The resulting sonication fluid was plated and bacterial counts were enumerated and expressed as cfu/sample. Sonicated samples were transferred to a microcalorimeter (TA Instrument) and heat flow monitored over a 24 h period with a precision of 0.0001°C and a sensitiviy of 200 μW. Experiments were performed in triplicates to calculate the mean ± SD. One-way ANOVA analysis was used for statistical analysis. Results: Bacterial counts (log10 cfu/sample) were highest on b-TCP (S. aureus 7.67 ± 0.17; S. epidermidis 8.14 ± 0.05) while bacterial density (log10 cfu/surface) was highest on PMMA (S. aureus 6.12 ± 0.2, S. epidermidis 7.65 ± 0.13). Detection time for S. aureus biofilms was shorter for the porous materials (b-TCP and Tutoplast, p <0.001) compared to the smooth materials (PMMA and PE) with no differences between b-TCP and TutoplastTM (p >0.05) or PMMA and PE (p >0.05). In contrast, for S. epidermidis biofilms the detection time was different (p <0.001) between all materials except between Tutoplast and PE (p >0.05). Conclusion: Our results demonstrate biofilm formation with both strains on all tested materials. Microcalorimetry was able to detect quantitatively the amount of biofilm. Further studies are needed to see whether calorimetry is a suitable tool also to monitor approaches to prevent and treat infections associated with bone grafts and bone substitutes.