989 resultados para Module level tests
Resumo:
I (Prática Pedagógica) - O estágio foi realizado na Academia de Música de Almada (AMA) com três alunos de níveis diferentes: Aluno x, Iniciação Musical; Aluna y, 2º grau; e Aluna z, 5º grau. A planificação do trabalho foi feita com a colaboração do professor Vasco Broco e sob a supervisão do professor Pedro Saglimbeni Muñoz, tendo por base o programa da AMA e procurando corresponder às necessidades individuais de cada aluno, bem como respeitar os diferentes ritmos de aprendizagem. Dadas as diferenças de idade e de nível, os objectivos que procurei atingir com cada aluno durante o ano lectivo são naturalmente díspares, bem como a perspectiva de cada um em relação ao próprio instrumento e ao ensino da música. Enquanto que os mais novos encaram a aprendizagem musical como uma actividade meramente recreativa, pelo menos por enquanto, a Aluna z ambiciona fazer uma carreira como instrumentista. Mesmo entre os mais novos há diferenças, pois apesar de o Aluno x ainda estar a frequentar a Iniciação Musical, já iniciou a sua aprendizagem há mais tempo do que a Aluna y, encarando o ensino com mais seriedade. A AMA tem a vantagem de ter um pianista acompanhador disponível para cada dia da semana, o que tornou possível desenvolver um trabalho consistente com os alunos em termos de ensaios e audições. As audições e momentos de avaliação foram marcados em reuniões de Departamento Curricular, à excepção da calendarização das Provas Globais que foi feita em reunião de Conselho Pedagógico. O aproveitamento foi positivo para os três alunos no final do estágio, tendo todos transitado para o nível de ensino seguinte.
Resumo:
The principal topic of this work is the application of data mining techniques, in particular of machine learning, to the discovery of knowledge in a protein database. In the first chapter a general background is presented. Namely, in section 1.1 we overview the methodology of a Data Mining project and its main algorithms. In section 1.2 an introduction to the proteins and its supporting file formats is outlined. This chapter is concluded with section 1.3 which defines that main problem we pretend to address with this work: determine if an amino acid is exposed or buried in a protein, in a discrete way (i.e.: not continuous), for five exposition levels: 2%, 10%, 20%, 25% and 30%. In the second chapter, following closely the CRISP-DM methodology, whole the process of construction the database that supported this work is presented. Namely, it is described the process of loading data from the Protein Data Bank, DSSP and SCOP. Then an initial data exploration is performed and a simple prediction model (baseline) of the relative solvent accessibility of an amino acid is introduced. It is also introduced the Data Mining Table Creator, a program developed to produce the data mining tables required for this problem. In the third chapter the results obtained are analyzed with statistical significance tests. Initially the several used classifiers (Neural Networks, C5.0, CART and Chaid) are compared and it is concluded that C5.0 is the most suitable for the problem at stake. It is also compared the influence of parameters like the amino acid information level, the amino acid window size and the SCOP class type in the accuracy of the predictive models. The fourth chapter starts with a brief revision of the literature about amino acid relative solvent accessibility. Then, we overview the main results achieved and finally discuss about possible future work. The fifth and last chapter consists of appendices. Appendix A has the schema of the database that supported this thesis. Appendix B has a set of tables with additional information. Appendix C describes the software provided in the DVD accompanying this thesis that allows the reconstruction of the present work.
Resumo:
The physiological responses of the clam R. decussatus from the Ria Formosa, southern Portugal, were examined in relation to normoxia, hypoxia (11, 6, 3 and 1.2 kPa) and anoxia; acute elevation of temperature (at 20, 27 and 32 °C), and its effect on the resistance to air exposure (at 20, 28 and 35 °C); current velocity (0.6, 3, 8 17, 24 and 36 cm. s-1) and turbidity (10, 100 and 300 mg. l-1 dry weight of particulate matter), and the efficiency of this species in retaining particles of different size (at 10 and 100 mg. l-1); and to copper contamination considering both short-term acute exposure to high levels (0.1-10 mg Cu. l-1) and chronic environmental levels (0.01 mg Cu. l-1). Clearance rates, respiration rates, absorption efficiency and excretion rates were assessed through the physiological energetics in terms of the energy budget and scope for growth (SFG). Stress independent respiration rates (R) and clearance rates (CR) were observed in relation to hypoxia down to 12 kPa and 6 kPa, respectively. Anoxic rates were 3.6 % of normoxic rates. Scope for growth was greatly reduced under extreme hypoxia (14 % of SFG in normoxia). Respiration rate was temperature independent in the range 20-32 °C but the decline in clearance rate resulted in negative SFG at 32 °C. Gaping during air exposure and the maintenance of faster aerobic metabolism led to 100 % mortality in 20 hours at 35 °C, 4 days at 28 °C and 5 days at 20 °C. Low current velocities (≤ 8 cm. s-1) supported high clearance rates. Shear stresses ≥ 0.9 Pa induced sediment movement and disturbed the feeding processes resulting in decreased clearance rates (at 36 cm. s-1, is 10 % of maximum CR). The observed ability of jetting out depleted water at a different level than the one of the inhalant current results is an important adaptation of clams to the slow currents of sheltered environments. Ingestion at high seston concentrations (> 100 mg. l-1) is controled by reducing the amount filtered, lowering CR (to 30 % of CR at low seston loads) and producing pseudofeces. Observed efficient retention of particles (70-100 %) in the range 3 to 8 μm is beneficial when algal cells are diluted by fine silt particles as it is likely to occur in the clams natural environment. R. decussatus in the short term escaped the exposure to copper by valve closure and therefore acute tests are not applicable to adult clams of this species. At environmental levels chronic exposure to copper did not induce lethal effects during the exposure period (20 days), but scope for growth was reduced to c. 30 %, indicating sustained impairment of physiological functions. The sensitivity of the physiological energetics and the integrated scope for growth measurement in assessing stress effects caused by natural environmental factors was highlighted.
Resumo:
A new integrated mathematical model for the simulation of offshore wind energy conversion system performance is presented in this paper. The mathematical model considers an offshore variable-speed turbine in deep water equipped with a permanent magnet synchronous generator using full-power two-level converter, converting the energy of a variable frequency source in injected energy into the electric network with constant frequency, through a high voltage DC transmission submarine cable. The mathematical model for the drive train is a concentrate two mass model which incorporates the dynamic for the structure and tower due to the need to emulate the effects of the moving surface. Controller strategy considered is a proportional integral one. Also, pulse width modulation using space vector modulation supplemented with sliding mode is used for trigger the transistor of the converter. Finally, a case study is presented to access the system performance. © 2014 IEEE.
Resumo:
Consider a single processor and a software system. The software system comprises components and interfaces where each component has an associated interface and each component comprises a set of constrained-deadline sporadic tasks. A scheduling algorithm (called global scheduler) determines at each instant which component is active. The active component uses another scheduling algorithm (called local scheduler) to determine which task is selected for execution on the processor. The interface of a component makes certain information about a component visible to other components; the interfaces of all components are used for schedulability analysis. We address the problem of generating an interface for a component based on the tasks inside the component. We desire to (i) incur only a small loss in schedulability analysis due to the interface and (ii) ensure that the amount of space (counted in bits) of the interface is small; this is because such an interface hides as much details of the component as possible. We present an algorithm for generating such an interface.
Resumo:
Não existe uma definição única de processo de memória de longo prazo. Esse processo é geralmente definido como uma série que possui um correlograma decaindo lentamente ou um espectro infinito de frequência zero. Também se refere que uma série com tal propriedade é caracterizada pela dependência a longo prazo e por não periódicos ciclos longos, ou que essa característica descreve a estrutura de correlação de uma série de longos desfasamentos ou que é convencionalmente expressa em termos do declínio da lei-potência da função auto-covariância. O interesse crescente da investigação internacional no aprofundamento do tema é justificado pela procura de um melhor entendimento da natureza dinâmica das séries temporais dos preços dos ativos financeiros. Em primeiro lugar, a falta de consistência entre os resultados reclama novos estudos e a utilização de várias metodologias complementares. Em segundo lugar, a confirmação de processos de memória longa tem implicações relevantes ao nível da (1) modelação teórica e econométrica (i.e., dos modelos martingale de preços e das regras técnicas de negociação), (2) dos testes estatísticos aos modelos de equilíbrio e avaliação, (3) das decisões ótimas de consumo / poupança e de portefólio e (4) da medição de eficiência e racionalidade. Em terceiro lugar, ainda permanecem questões científicas empíricas sobre a identificação do modelo geral teórico de mercado mais adequado para modelar a difusão das séries. Em quarto lugar, aos reguladores e gestores de risco importa saber se existem mercados persistentes e, por isso, ineficientes, que, portanto, possam produzir retornos anormais. O objetivo do trabalho de investigação da dissertação é duplo. Por um lado, pretende proporcionar conhecimento adicional para o debate da memória de longo prazo, debruçando-se sobre o comportamento das séries diárias de retornos dos principais índices acionistas da EURONEXT. Por outro lado, pretende contribuir para o aperfeiçoamento do capital asset pricing model CAPM, considerando uma medida de risco alternativa capaz de ultrapassar os constrangimentos da hipótese de mercado eficiente EMH na presença de séries financeiras com processos sem incrementos independentes e identicamente distribuídos (i.i.d.). O estudo empírico indica a possibilidade de utilização alternativa das obrigações do tesouro (OT’s) com maturidade de longo prazo no cálculo dos retornos do mercado, dado que o seu comportamento nos mercados de dívida soberana reflete a confiança dos investidores nas condições financeiras dos Estados e mede a forma como avaliam as respetiva economias com base no desempenho da generalidade dos seus ativos. Embora o modelo de difusão de preços definido pelo movimento Browniano geométrico gBm alegue proporcionar um bom ajustamento das séries temporais financeiras, os seus pressupostos de normalidade, estacionariedade e independência das inovações residuais são adulterados pelos dados empíricos analisados. Por isso, na procura de evidências sobre a propriedade de memória longa nos mercados recorre-se à rescaled-range analysis R/S e à detrended fluctuation analysis DFA, sob abordagem do movimento Browniano fracionário fBm, para estimar o expoente Hurst H em relação às séries de dados completas e para calcular o expoente Hurst “local” H t em janelas móveis. Complementarmente, são realizados testes estatísticos de hipóteses através do rescaled-range tests R/S , do modified rescaled-range test M - R/S e do fractional differencing test GPH. Em termos de uma conclusão única a partir de todos os métodos sobre a natureza da dependência para o mercado acionista em geral, os resultados empíricos são inconclusivos. Isso quer dizer que o grau de memória de longo prazo e, assim, qualquer classificação, depende de cada mercado particular. No entanto, os resultados gerais maioritariamente positivos suportam a presença de memória longa, sob a forma de persistência, nos retornos acionistas da Bélgica, Holanda e Portugal. Isto sugere que estes mercados estão mais sujeitos a maior previsibilidade (“efeito José”), mas também a tendências que podem ser inesperadamente interrompidas por descontinuidades (“efeito Noé”), e, por isso, tendem a ser mais arriscados para negociar. Apesar da evidência de dinâmica fractal ter suporte estatístico fraco, em sintonia com a maior parte dos estudos internacionais, refuta a hipótese de passeio aleatório com incrementos i.i.d., que é a base da EMH na sua forma fraca. Atendendo a isso, propõem-se contributos para aperfeiçoamento do CAPM, através da proposta de uma nova fractal capital market line FCML e de uma nova fractal security market line FSML. A nova proposta sugere que o elemento de risco (para o mercado e para um ativo) seja dado pelo expoente H de Hurst para desfasamentos de longo prazo dos retornos acionistas. O expoente H mede o grau de memória de longo prazo nos índices acionistas, quer quando as séries de retornos seguem um processo i.i.d. não correlacionado, descrito pelo gBm(em que H = 0,5 , confirmando- se a EMH e adequando-se o CAPM), quer quando seguem um processo com dependência estatística, descrito pelo fBm(em que H é diferente de 0,5, rejeitando-se a EMH e desadequando-se o CAPM). A vantagem da FCML e da FSML é que a medida de memória de longo prazo, definida por H, é a referência adequada para traduzir o risco em modelos que possam ser aplicados a séries de dados que sigam processos i.i.d. e processos com dependência não linear. Então, estas formulações contemplam a EMH como um caso particular possível.
Resumo:
Introdução – O efeito de êmbolo é um dos principais problemas relacionados com a eficácia de uma prótese. Uma diminuição do mesmo pode levar a uma marcha mais natural através do aumento da propriocetividade. Objetivos – Verificar se existe diferença de valores do efeito de êmbolo entre vários sistemas de suspensão para próteses transtibiais com a utilização de Liners e testar a aplicação de testes de imagiologia na análise da melhor solução protésica para um determinado indivíduo. Metodologia – Foi obtida uma radiografia da prótese em carga na posição ortostática, mantendo o peso do indivíduo igualmente distribuído pelos dois pés. Seguidamente foi realizada outra radiografia no plano sagital com o joelho com 30° de flexão, com a prótese suspensa e um peso de 5kg aplicado na extremidade distal da mesma durante 30 seg. Através destes dois exames efetuaram-se as medições do êmbolo para cada tipo de sistema de suspensão. Resultados – Dos quatro sistemas estudados apenas três apresentam valores de êmbolo, visto que um dos sistemas não criou suspensão suficiente para suportar o peso colocado na extremidade distal da prótese. Através das medições realizadas nos exames imagiológicos dos três sistemas pudemos encontrar variações de efeito de êmbolo que vão dos 47,91mm aos 72,55mm. Conclusão – Através da realização do estudo imagiológico verificaram-se diferenças a nível do efeito de êmbolo nos vários sistemas de suspensão, provando que esta é uma ferramenta viável na avaliação do mesmo. Também através da análise dos resultados ficou notório que o sistema de suspensão Vacuum Assisted Suspention System (VASS) é o que apresenta menos êmbolo.
Resumo:
Compositional real-time scheduling clearly requires that ”normal” real-time scheduling challenges are addressed but challenges intrinsic to compositionality must be addressed as well, in particular: (i) how should interfaces be described? and (ii) how should numerical values be assigned to parameters constituting the interfaces? The real-time systems community has traditionally used narrow interfaces for describing a component (for example, a utilization/bandwidthlike metric and the distribution of this bandwidth in time). In this paper, we introduce the concept of competitive ratio of an interface and show that typical narrow interfaces cause poor performance for scheduling constrained-deadline sporadic tasks (competitive ratio is infinite). Therefore, we explore more expressive interfaces; in particular a class called medium-wide interfaces. For this class, we propose an interface type and show how the parameters of the interface should be selected. We also prove that this interface is 8-competitive.
Resumo:
Coastal low-level jets (CLLJ) are a low-tropospheric wind feature driven by the pressure gradient produced by a sharp contrast between high temperatures over land and lower temperatures over the sea. This contrast between the cold ocean and the warm land in the summer is intensified by the impact of the coastal parallel winds on the ocean generating upwelling currents, sharpening the temperature gradient close to the coast and giving rise to strong baroclinic structures at the coast. During summertime, the Iberian Peninsula is often under the effect of the Azores High and of a thermal low pressure system inland, leading to a seasonal wind, in the west coast, called the Nortada (northerly wind). This study presents a regional climatology of the CLLJ off the west coast of the Iberian Peninsula, based on a 9km resolution downscaling dataset, produced using the Weather Research and Forecasting (WRF) mesoscale model, forced by 19 years of ERA-Interim reanalysis (1989-2007). The simulation results show that the jet hourly frequency of occurrence in the summer is above 30% and decreases to about 10% during spring and autumn. The monthly frequencies of occurrence can reach higher values, around 40% in summer months, and reveal large inter-annual variability in all three seasons. In the summer, at a daily base, the CLLJ is present in almost 70% of the days. The CLLJ wind direction is mostly from north-northeasterly and occurs more persistently in three areas where the interaction of the jet flow with local capes and headlands is more pronounced. The coastal jets in this area occur at heights between 300 and 400 m, and its speed has a mean around 15 m/s, reaching maximum speeds of 25 m/s.
Resumo:
Waste oil recycling companies play a very important role in our society. Competition among companies is tough and process optimization is essential for survival. By equipping oil containers with a level monitoring system that periodically reports the level and alerts when it reaches the preset threshold, the oil recycling companies are able to streamline the oil collection process and, thus, reduce the operation costs while maintaining the quality of service. This paper describes the development of this level monitoring system by a team of four students from different engineering backgrounds and nationalities. The team conducted a study of the state of the art, draw marketing and sustainable development plans and, finally, designed and implemented a prototype that continuously measures the container content level and sends an alert message as soon as it reaches the preset capacity.
Resumo:
Real-time scheduling usually considers worst-case values for the parameters of task (or message stream) sets, in order to provide safe schedulability tests for hard real-time systems. However, worst-case conditions introduce a level of pessimism that is often inadequate for a certain class of (soft) real-time systems. In this paper we provide an approach for computing the stochastic response time of tasks where tasks have inter-arrival times described by discrete probabilistic distribution functions, instead of minimum inter-arrival (MIT) values.
Resumo:
OBJECTIVE The objective of this study was to analyze the prevalence of diabetes in older people and the adopted control measures.METHODS Data regarding older diabetic individuals who participated in the Health Surveys conducted in the Municipality of Sao Paulo, SP, ISA-Capital, in 2003 and 2008, which were cross-sectional studies, were analyzed. Prevalences and confidence intervals were compared between 2003 and 2008, according to sociodemographic variables. The combination of the databases was performed when the confidence intervals overlapped. The Chi-square (level of significance of 5%) and the Pearson’s Chi-square (Rao-Scott) tests were performed. The variables without overlap between the confidence intervals were not tested.RESULTS The age of the older adults was 60-69 years. The majority were women, Caucasian, with an income of between > 0.5 and 2.5 times the minimum salary and low levels of schooling. The prevalence of diabetes was 17.6% (95%CI 14.9;20.6) in 2003 and 20.1% (95%CI 17.3;23.1) in 2008, which indicates a growth over this period (p at the limit of significance). The most prevalent measure adopted by the older adults to control diabetes was hypoglycemic agents, followed by diet. Physical activity was not frequent, despite the significant differences observed between 2003 and 2008 results. The use of public health services to control diabetes was significantly higher in older individuals with lower income and lower levels of education.CONCLUSIONS Diabetes is a complex and challenging disease for patients and the health systems. Measures that encourage health promotion practices are necessary because they presented a smaller proportion than the use of hypoglycemic agents. Public health policies should be implemented, and aimed mainly at older individuals with low income and schooling levels. These changes are essential to improve the health condition of older diabetic patients.
Resumo:
A QoS adaptation to dynamically changing system conditions that takes into consideration the user’s constraints on the stability of service provisioning is presented. The goal is to allow the system to make QoS adaptation decisions in response to fluctuations in task traffic flow, under the control of the user. We pay special attention to the case where monitoring the stability period and resource load variation of Service Level Agreements for different types of services is used to dynamically adapt future stability periods, according to a feedback control scheme. System’s adaptation behaviour can be configured according to a desired confidence level on future resource usage. The viability of the proposed approach is validated by preliminary experiments.
Resumo:
Typically common embedded systems are designed with high resource constraints. Static designs are often chosen to address very specific use cases. On contrast, a dynamic design must be used if the system must supply a real-time service where the input may contain factors of indeterminism. Thus, adding new functionality on these systems is often accomplished by higher development time, tests and costs, since new functionality push the system complexity and dynamics to a higher level. Usually, these systems have to adapt themselves to evolving requirements and changing service requests. In this perspective, run-time monitoring of the system behaviour becomes an important requirement, allowing to dynamically capturing the actual scheduling progress and resource utilization. For this to succeed, operating systems need to expose their internal behaviour and state, making it available to the external applications, usually using a run-time monitoring mechanism. However, such mechanism can impose a burden in the system itself if not wisely used. In this paper we explore this problem and propose a framework, which is intended to provide this run-time mechanism whilst achieving code separation, run-time efficiency and flexibility for the final developer.
Resumo:
Locomotor tasks characterization plays an important role in trying to improve the quality of life of a growing elderly population. This paper focuses on this matter by trying to characterize the locomotion of two population groups with different functional fitness levels (high or low) while executing three different tasks-gait, stair ascent and stair descent. Features were extracted from gait data, and feature selection methods were used in order to get the set of features that allow differentiation between functional fitness level. Unsupervised learning was used to validate the sets obtained and, ultimately, indicated that it is possible to distinguish the two population groups. The sets of best discriminate features for each task are identified and thoroughly analysed. Copyright © 2014 SCITEPRESS - Science and Technology Publications. All rights reserved.