959 resultados para automated software testing


Relevância:

20.00% 20.00%

Publicador:

Resumo:

20th International Conference on Reliable Software Technologies - Ada-Europe 2015 (Ada-Europe 2015), Madrid, Spain.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

OCEANS, 2001. MTS/IEEE Conference and Exhibition (Volume:2 )

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Article in Press, Corrected Proof

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The paper presents a RFDSCA automated synthesis procedure. This algorithm determines several RFDSCA circuits from the top-level system specifications all with the same maximum performance. The genetic synthesis tool optimizes a fitness function proportional to the RFDSCA quality factor and uses the epsiv-concept and maximin sorting scheme to achieve a set of solutions well distributed along a non-dominated front. To confirm the results of the algorithm, three RFDSCAs were simulated in SpectreRF and one of them was implemented and tested. The design used a 0.25 mum BiCMOS process. All the results (synthesized, simulated and measured) are very close, which indicate that the genetic synthesis method is a very useful tool to design optimum performance RFDSCAs.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Consumer-electronics systems are becoming increasingly complex as the number of integrated applications is growing. Some of these applications have real-time requirements, while other non-real-time applications only require good average performance. For cost-efficient design, contemporary platforms feature an increasing number of cores that share resources, such as memories and interconnects. However, resource sharing causes contention that must be resolved by a resource arbiter, such as Time-Division Multiplexing. A key challenge is to configure this arbiter to satisfy the bandwidth and latency requirements of the real-time applications, while maximizing the slack capacity to improve performance of their non-real-time counterparts. As this configuration problem is NP-hard, a sophisticated automated configuration method is required to avoid negatively impacting design time. The main contributions of this article are: 1) An optimal approach that takes an existing integer linear programming (ILP) model addressing the problem and wraps it in a branch-and-price framework to improve scalability. 2) A faster heuristic algorithm that typically provides near-optimal solutions. 3) An experimental evaluation that quantitatively compares the branch-and-price approach to the previously formulated ILP model and the proposed heuristic. 4) A case study of an HD video and graphics processing system that demonstrates the practical applicability of the approach.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

AIMS: Protocols using sublingual nitrates have been increasingly used to improve diagnostic accuracy of head-up tilt testing (HUT). Nevertheless, exaggerated responses to nitrates have been frequently described, particularly in elderly patients. The aim of this article is to evaluate, in an elderly population with unexplained syncope, whether the impact of sublingual nitroglycerin (NTG) used as a provocative agent is dose-dependent. METHODS AND RESULTS: One hundred and twenty consecutive elderly patients submitted to HUT using NTG after an asymptomatic drug-free phase were studied. Patients were divided into three groups according to the NTG dosage: 500, 375 and 250 microg. The test was considered positive when there was reproduction of symptoms with bradycardia and/or arterial hypotension. A gradual decrease in the blood pressure after NTG was considered an exaggerated response to nitrates. There were no differences in the clinical characteristics of the different subgroups. A positive test was obtained in 50% of the patients in each group. The rate of exaggerated responses was identical in all groups and ranged between 15 and 17%. CONCLUSION: In an elderly population with syncope of unknown origin submitted to HUT, the response to NTG is not dose-dependent, and no difference was found in the rate of exaggerated responses to nitrates with different NTG dosages.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Recent embedded processor architectures containing multiple heterogeneous cores and non-coherent caches renewed attention to the use of Software Transactional Memory (STM) as a building block for developing parallel applications. STM promises to ease concurrent and parallel software development, but relies on the possibility of abort conflicting transactions to maintain data consistency, which in turns affects the execution time of tasks carrying transactions. Because of this fact the timing behaviour of the task set may not be predictable, thus it is crucial to limit the execution time overheads resulting from aborts. In this paper we formalise a FIFO-based algorithm to order the sequence of commits of concurrent transactions. Then, we propose and evaluate two non-preemptive and one SRP-based fully-preemptive scheduling strategies, in order to avoid transaction starvation.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

An "in-house" RT-PCR method was developed that allows the simultaneous detection of the RNA of the Hepatitis C Virus (HCV) and an artificial RNA employed as an external control. Samples were analyzed in pools of 6-12 donations, each donation included in two pools, one horizontal and one vertical, permitting the immediate identification of a reactive donation, obviating the need for pool dismembering. The whole process took 6-8 hours per day and results were issued in parallel to serology. The method was shown to detect all six HCV genotypes and a sensitivity of 500 IU/mL was achieved (95% hit rate). Until July 2005, 139,678 donations were tested and 315 (0.23%) were found reactive for HCV-RNA. Except for five false-positives, all 310 presented the corresponding antibody as well, so the yield of NAT-only donations was zero, presenting a specificity of 99.83%. Detection of a window period donation, in the population studied, will probably demand testing of a larger number of donations. International experience is showing a rate of 1:200,000 - 1:500,000 of isolated HCV-RNA reactive donations.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The antifungal activities of itraconazole, ketoconazole, fluconazole, terbinafine and griseofulvin were tested by broth microdilution methods against 71 isolates of dermatophytes isolated from Nigerian children. Most drugs were very active against all the dermatophytes and the MIC 90 ranged from 0.03 to 8.0 µg/mL. This appears to be the first documented data on the antifungal susceptibility testing of isolates of dermatophytes from Nigerian children.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

INTRODUCTION: A growing body of evidence shows the prognostic value of oxygen uptake efficiency slope (OUES), a cardiopulmonary exercise test (CPET) parameter derived from the logarithmic relationship between O(2) consumption (VO(2)) and minute ventilation (VE) in patients with chronic heart failure (CHF). OBJECTIVE: To evaluate the prognostic value of a new CPET parameter - peak oxygen uptake efficiency (POUE) - and to compare it with OUES in patients with CHF. METHODS: We prospectively studied 206 consecutive patients with stable CHF due to dilated cardiomyopathy - 153 male, aged 53.3±13.0 years, 35.4% of ischemic etiology, left ventricular ejection fraction 27.7±8.0%, 81.1% in sinus rhythm, 97.1% receiving ACE-Is or ARBs, 78.2% beta-blockers and 60.2% spironolactone - who performed a first maximal symptom-limited treadmill CPET, using the modified Bruce protocol. In 33% of patients an cardioverter-defibrillator (ICD) or cardiac resynchronization therapy device (CRT-D) was implanted during follow-up. Peak VO(2), percentage of predicted peak VO(2), VE/VCO(2) slope, OUES and POUE were analyzed. OUES was calculated using the formula VO(2) (l/min) = OUES (log(10)VE) + b. POUE was calculated as pVO(2) (l/min) / log(10)peakVE (l/min). Correlation coefficients between the studied parameters were obtained. The prognosis of each variable adjusted for age was evaluated through Cox proportional hazard models and R2 percent (R2%) and V index (V6) were used as measures of the predictive accuracy of events of each of these variables. Receiver operating characteristic (ROC) curves from logistic regression models were used to determine the cut-offs for OUES and POUE. RESULTS: pVO(2): 20.5±5.9; percentage of predicted peak VO(2): 68.6±18.2; VE/VCO(2) slope: 30.6±8.3; OUES: 1.85±0.61; POUE: 0.88±0.27. During a mean follow-up of 33.1±14.8 months, 45 (21.8%) patients died, 10 (4.9%) underwent urgent heart transplantation and in three patients (1.5%) a left ventricular assist device was implanted. All variables proved to be independent predictors of this combined event; however, VE/VCO2 slope was most strongly associated with events (HR 11.14). In this population, POUE was associated with a higher risk of events than OUES (HR 9.61 vs. 7.01), and was also a better predictor of events (R2: 28.91 vs. 22.37). CONCLUSION: POUE was more strongly associated with death, urgent heart transplantation and implantation of a left ventricular assist device and proved to be a better predictor of events than OUES. These results suggest that this new parameter can increase the prognostic value of CPET in patients with CHF.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

The recent technological advancements and market trends are causing an interesting phenomenon towards the convergence of High-Performance Computing (HPC) and Embedded Computing (EC) domains. On one side, new kinds of HPC applications are being required by markets needing huge amounts of information to be processed within a bounded amount of time. On the other side, EC systems are increasingly concerned with providing higher performance in real-time, challenging the performance capabilities of current architectures. The advent of next-generation many-core embedded platforms has the chance of intercepting this converging need for predictable high-performance, allowing HPC and EC applications to be executed on efficient and powerful heterogeneous architectures integrating general-purpose processors with many-core computing fabrics. To this end, it is of paramount importance to develop new techniques for exploiting the massively parallel computation capabilities of such platforms in a predictable way. P-SOCRATES will tackle this important challenge by merging leading research groups from the HPC and EC communities. The time-criticality and parallelisation challenges common to both areas will be addressed by proposing an integrated framework for executing workload-intensive applications with real-time requirements on top of next-generation commercial-off-the-shelf (COTS) platforms based on many-core accelerated architectures. The project will investigate new HPC techniques that fulfil real-time requirements. The main sources of indeterminism will be identified, proposing efficient mapping and scheduling algorithms, along with the associated timing and schedulability analysis, to guarantee the real-time and performance requirements of the applications.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Thirty Candida albicans isolated from oral candidosis patients and 30 C. albicans isolated from control individuals were studied. In vitro susceptibility tests were performed for amphotericin B, fluconazole, 5-flucytosine and itraconazole through the Clinical and Laboratorial Standards Institute (CLSI) reference method and E test system. The results obtained were analyzed and compared. MIC values were similar for the strains isolated from oral candidosis patients and control individuals. The agreement rate for the two methods was 66.67% for amphotericin B, 53.33% for fluconazole, 65% for flucytosine and 45% for itraconazole. According to our data, E test method could be an alternative to trial routine susceptibility testing due to its simplicity. However, it can not be considered a substitute for the CLSI reference method.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Trabalho de projecto apresentado como requisito parcial para a obtenção do grau de mestre em Ciência e Sistemas de Informação Geográfica

Relevância:

20.00% 20.00%

Publicador:

Resumo:

O âmbito deste trabalho envolve o teste do modelo BIM numa obra em construção pela Mota-Engil – Engenharia, na extração experimental de peças desenhadas de preparação e apoio à execução de obra. No capítulo 1 deste relatório são definidos o âmbito e os objetivos deste trabalho, é feito um enquadramento histórico do tema e abordados conceitos e atividades da preparação de obra, na sua forma tradicional. O estado do conhecimento da preparação de obras e mais em concreto da tecnologia BIM a nível nacional e internacional é abordado no capítulo 2. Nesse sentido procura-se definir os conceitos principais inerentes a esta nova metodologia, que passa por identificar e caraterizar a tecnologia envolvida e o seu nível de desenvolvimento. Com suporte em casos práticos de preparação de obra na sua forma tradicional, identificados e desenvolvidos no capítulo 3, foi compilado um processo tipo de peças desenhadas de suporte identificadas e caracterizadas no capítulo 4, frequentes e comuns à execução de diversos tipos de obras de edifícios. Assente na compilação baseada em casos práticos e no estudo do projeto de execução da empreitada que sustenta o presente trabalho, com base no qual o modelo BIM foi concebido, identificou-se um conjunto de peças desenhadas de preparação e apoio à execução dos trabalhos, em 2D, a extrair do modelo. No capítulo 5, é feita uma descrição do modo como foi estudado o projeto da obra, com evidência para os fatores mais relevantes, especificando os desenhos a extrair. Suportada pelo programa de modelação ArchiCAD, a extração do conjunto de desenhos identificados anteriormente foi conseguida com recurso às funcionalidades disponíveis no software, que permite a criação de desenhos 2D atualizáveis ou não automaticamente a partir do modelo. Qualquer alteração introduzida no modelo virtual é automaticamente atualizada nos desenhos bidimensionais, caso o utilizador assim o pretenda. Ao longo desse trabalho foram detetados e analisados os condicionalismos inerentes ao processo de extração, referidos no capítulo 6, para estabelecimento de regras de modelação padrão a adotar em futuras empreitadas, que possam simplificar a obtenção dos elementos desenhados de preparação necessários à sua execução. No ponto 6.3 são identificadas melhorias a introduzir no modelo. Em conclusão no capítulo 7 são abordadas especificidades do setor da construção que sustentam e evidenciam cada vez mais a necessidade de utilizar as novas tecnologias com vista à adoção de práticas e ferramentas padrão de apoio à execução de obras. Sendo a tecnologia BIM, transversal a todo o setor, a sua utilização com regras padrão na conceção dos modelos e na extração de dados, potencia a otimização dos custos, do tempo, dos recursos e da qualidade final de um empreendimento, ao longo de todo o seu ciclo de vida, para além de apoiar com elevada fiabilidade as tomadas de decisão ao longo desse período. A tecnologia BIM, possibilita a antevisão do edifício a construir com um elevado grau de pormenor, com todas as vantagens que daí advêm.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Este working paper centra-se no desenvolvimento das técnicas de edição de documentos ou publicações, que têm conhecido um incremento espectacular na última década. Os avanços e descobertas mais significativas entre 1450 e 1986, eclipsam-se praticamente, quando comparados com as grandes modificações dos últimos sete anos (isto é a História da auto-edição e si). Desde o aparecimento do primeiro programa de auto-edição e da primeira impressora laser, dotada de um intérprete postscript, até hoje, a evolução e aperfeiçoamento tanto do software como do hardware tem sido tanta, continuando a fazer-se a um tal ritmo, que se tornou irreversível.