42 resultados para frequency analysis problem
em Instituto Politécnico do Porto, Portugal
Resumo:
Uma nova tecnologia, os EDLC, também denominados por supercondensadores, tem-se tornado numa importante e aliciante área de interesse. Estes regem-se pelos mesmos princípios fundamentais dos condensadores clássicos, no entanto possibilitam receber capacidades superiores, devido a uma maior área de superfície e a um dielétrico menos espesso. Esta particularidade permite obter uma maior densidade energética, comparativamente com os condensadores clássicos e uma maior densidade de potência, comparativamente com as baterias. Consequentemente a utilização de supercondensadores tem aumentado, representando já uma alternativa fiável, segura e amiga do ambiente, em detrimento das baterias comuns. Assim, este projeto tem como principais objetivos, identificar os diferentes tipos de supercondensadores, apresentar as vantagens de cada tipo e explorar a sua resposta, quer no domínio das frequências quer no domínio dos tempos, e por fim modelá-los recorrendo a componentes elétricos clássicos, nomeadamente resistências e condensadores. A modelação foi realizada recorrendo ao MALTAB, através da função de minimização fminunc e foram construídos quatro modelos equivalentes, com o objetivo de modelar a resposta dos vários EDLC analisados. Por escassez de tempo o principal foco de análise recaiu sobre o EDLC de 0,022 F.
Resumo:
This paper presents the dynamic analysis of robotic biped systems. The main goal is to gain insight into the phenomena of walking and to evaluate its performance. In this study, we propose three methods to quantitatively measure the dynamic efficiency of walking: energy analysis, perturbation analysis and lowpass frequency analysis. In order to accomplish this goal, the prescribed motion of the biped is completely characterised in terms of a set of locomotion variables, namely: step lenght, hip height, hip ripple, hip offset, foot clearance and link lenghts. based on these variables and their influence, the performance measures aer discussed and the results compared with those observed in nature.
Resumo:
Signal Processing: Algorithms, Architectures, Arrangements, and Applications (SPA), 2013
Resumo:
The process of resources systems selection takes an important part in Distributed/Agile/Virtual Enterprises (D/A/V Es) integration. However, the resources systems selection is still a difficult matter to solve in a D/A/VE, as it is pointed out in this paper. Globally, we can say that the selection problem has been equated from different aspects, originating different kinds of models/algorithms to solve it. In order to assist the development of a web prototype tool (broker tool), intelligent and flexible, that integrates all the selection model activities and tools, and with the capacity to adequate to each D/A/V E project or instance (this is the major goal of our final project), we intend in this paper to show: a formulation of a kind of resources selection problem and the limitations of the algorithms proposed to solve it. We formulate a particular case of the problem as an integer programming, which is solved using simplex and branch and bound algorithms, and identify their performance limitations (in terms of processing time) based on simulation results. These limitations depend on the number of processing tasks and on the number of pre-selected resources per processing tasks, defining the domain of applicability of the algorithms for the problem studied. The limitations detected open the necessity of the application of other kind of algorithms (approximate solution algorithms) outside the domain of applicability founded for the algorithms simulated. However, for a broker tool it is very important the knowledge of algorithms limitations, in order to, based on problem features, develop and select the most suitable algorithm that guarantees a good performance.
Resumo:
The internal impedance of a wire is the function of the frequency. In a conductor, where the conductivity is sufficiently high, the displacement current density can be neglected. In this case, the conduction current density is given by the product of the electric field and the conductance. One of the aspects of the high-frequency effects is the skin effect (SE). The fundamental problem with SE is it attenuates the higher frequency components of a signal.
Resumo:
The internal impedance of a wire is the function of the frequency. In a conductor, where the conductivity is sufficiently high, the displacement current density can be neglected. In this case, the conduction current density is given by the product of the electric field and the conductance. One of the aspects the high-frequency effects is the skin effect (SE). The fundamental problem with SE is it attenuates the higher frequency components of a signal. The SE was first verified by Kelvin in 1887. Since then many researchers developed work on the subject and presently a comprehensive physical model, based on the Maxwell equations, is well established. The Maxwell formalism plays a fundamental role in the electromagnetic theory. These equations lead to the derivation of mathematical descriptions useful in many applications in physics and engineering. Maxwell is generally regarded as the 19th century scientist who had the greatest influence on 20th century physics, making contributions to the fundamental models of nature. The Maxwell equations involve only the integer-order calculus and, therefore, it is natural that the resulting classical models adopted in electrical engineering reflect this perspective. Recently, a closer look of some phenomas present in electrical systems and the motivation towards the development of precise models, seem to point out the requirement for a fractional calculus approach. Bearing these ideas in mind, in this study we address the SE and we re-evaluate the results demonstrating its fractional-order nature.
Resumo:
Manipulator systems are rather complex and highly nonlinear which makes difficult their analysis and control. Classic system theory is veil known, however it is inadequate in the presence of strong nonlinear dynamics. Nonlinear controllers produce good results [1] and work has been done e. g. relating the manipulator nonlinear dynamics with frequency response [2–5]. Nevertheless, given the complexity of the problem, systematic methods which permit to draw conclusions about stability, imperfect modelling effects, compensation requirements, etc. are still lacking. In section 2 we start by analysing the variation of the poles and zeros of the descriptive transfer functions of a robot manipulator in order to motivate the development of more robust (and computationally efficient) control algorithms. Based on this analysis a new multirate controller which is an improvement of the well known “computed torque controller” [6] is announced in section 3. Some research in this area was done by Neuman [7,8] showing tbat better robustness is possible if the basic controller structure is modified. The present study stems from those ideas, and attempts to give a systematic treatment, which results in easy to use standard engineering tools. Finally, in section 4 conclusions are presented.
Resumo:
Deoxyribonucleic acid, or DNA, is the most fundamental aspect of life but present day scientific knowledge has merely scratched the surface of the problem posed by its decoding. While experimental methods provide insightful clues, the adoption of analysis tools supported by the formalism of mathematics will lead to a systematic and solid build-up of knowledge. This paper studies human DNA from the perspective of system dynamics. By associating entropy and the Fourier transform, several global properties of the code are revealed. The fractional order characteristics emerge as a natural consequence of the information content. These properties constitute a small piece of scientific knowledge that will support further efforts towards the final aim of establishing a comprehensive theory of the phenomena involved in life.
Resumo:
This paper aims to study the relationships between chromosomal DNA sequences of twenty species. We propose a methodology combining DNA-based word frequency histograms, correlation methods, and an MDS technique to visualize structural information underlying chromosomes (CRs) and species. Four statistical measures are tested (Minkowski, Cosine, Pearson product-moment, and Kendall τ rank correlations) to analyze the information content of 421 nuclear CRs from twenty species. The proposed methodology is built on mathematical tools and allows the analysis and visualization of very large amounts of stream data, like DNA sequences, with almost no assumptions other than the predefined DNA “word length.” This methodology is able to produce comprehensible three-dimensional visualizations of CR clustering and related spatial and structural patterns. The results of the four test correlation scenarios show that the high-level information clusterings produced by the MDS tool are qualitatively similar, with small variations due to each correlation method characteristics, and that the clusterings are a consequence of the input data and not method’s artifacts.
Resumo:
Purpose – The aim of this article is to present some results from research undertaken into the information behaviour of European Documentation Centre (EDC) users. It will reflect on the practices of a group of 234 users of 55 EDCs covering 21 Member States of the European Union (EU), used to access European information. Design/methodology/approach – In order to collect the data presented here, five questionnaires were sent to users in all the EDCs in Finland, Ireland, Hungary and Portugal. In the remaining EU countries, five questionnaires were sent to two EDCs chosen at random. The questionnaires were sent by post, following telephone contact with the EDC managers. Findings – Factors determining access to information on the European Union and the frequency of this access are identified. The information providers most commonly used to access European information and the information sources considered the most reliable by respondents will also be analysed. Another area of analysis concerns the factors cited by respondents as facilitating access to information on Europe or, conversely, making it more difficult to access. Parallel to this, the aspects of accessing information on EU that are valued most by users will also be assessed. Research limitations/implications – Questionnaires had to be used, as the intention was to cover a very extensive geographical area. However, in opting for closed questions, it is acknowledged that standard responses have been obtained with no scope for capturing the individual circumstances of each respondent, thus making a qualitative approach difficult. Practical implications – The results provide an overall picture of certain aspects of the information behaviour of EDC users. They may serve as a starting point for planning training sessions designed to develop the skills required to search, access, evaluate and apply European information within an academic context. From a broader perspective, they also constitute factors which the European Commission should take into consideration when formulating its information and communication policy. Originality/value – This is the first piece of academic research into the EDCs and their users, which aimed to cover all Members State of the EU.
Resumo:
Mathematical Program with Complementarity Constraints (MPCC) finds many applications in fields such as engineering design, economic equilibrium and mathematical programming theory itself. A queueing system model resulting from a single signalized intersection regulated by pre-timed control in traffic network is considered. The model is formulated as an MPCC problem. A MATLAB implementation based on an hyperbolic penalty function is used to solve this practical problem, computing the total average waiting time of the vehicles in all queues and the green split allocation. The problem was codified in AMPL.
Resumo:
Different problems are daily discuss on environmental aspects such acid rain, eutrophication, global warming and an others problems. Rarely do we find some discussions about phosphorus problematic. Through the years the phosphorus as been a real problem and must be more discussed. On this thesis was done a global material flow analysis of phosphorus, based on data from the year 2004, the production of phosphate rock in that year was 18.9 million tones, almost this amount it was used as fertilizer on the soil and the plants only can uptake, on average, 20% of the input of fertilizer to grow up, the remainder is lost for the phosphorus soil. In the phosphorus soil there is equilibrium between the phosphorus available to uptake from the plants and the phosphorus associate with other compounds, this equilibrium depends of the kind of soil and is related with the soil pH. A reserve inventory was done and we have 15,000 million tones as reserve, the amount that is economical available. The reserve base is estimated in 47,000 million tones. The major reserves can be found in Morocco and Western Sahara, United Sates, China and South Africa. The reserve estimated in 2009 was 15,000 million tone of phosphate rock or 1,963 million tone of P. If every year the mined phosphate rock is around 22 Mt/yr (phosphorus production on 2008 USGS 2009), and each year the consumption of phosphorus increases because of the food demand, the reserves of phosphate rock will be finished in about 90 years, or maybe even less. About the value/impact assessment was done a qualitative analysis, if on the future we don’t have more phosphate rock to produce fertilizers, it is expected a drop on the crops yields, each depends of the kind of the soil and the impact on the humans feed and animal production will not be a relevant problem. We can recovery phosphorus from different waste streams such as ploughing crop residues back into the soil, Food processing plants and food retailers, Human and animal excreta, Meat and bone meal, Manure fibre, Sewage sludge and wastewater. Some of these examples are developed in the paper.
Resumo:
A geração de trajectórias de robôs em tempo real é uma tarefa muito complexa, não
existindo ainda um algoritmo que a permita resolver de forma eficaz. De facto, há
controladores eficientes para trajectórias previamente definidas, todavia, a adaptação a
variações imprevisíveis, como sendo terrenos irregulares ou obstáculos, constitui ainda um
problema em aberto na geração de trajectórias em tempo real de robôs.
Neste trabalho apresentam-se modelos de geradores centrais de padrões de locomoção
(CPGs), inspirados na biologia, que geram os ritmos locomotores num robô quadrúpede.
Os CPGs são modelados matematicamente por sistemas acoplados de células (ou
neurónios), sendo a dinâmica de cada célula dada por um sistema de equações diferenciais
ordinárias não lineares. Assume-se que as trajectórias dos robôs são constituídas por esta
parte rítmica e por uma parte discreta. A parte discreta pode ser embebida na parte rítmica,
(a.1) como um offset ou (a.2) adicionada às expressões rítmicas, ou (b) pode ser calculada
independentemente e adicionada exactamente antes do envio dos sinais para as articulações
do robô. A parte discreta permite inserir no passo locomotor uma perturbação, que poderá
estar associada à locomoção em terrenos irregulares ou à existência de obstáculos na
trajectória do robô. Para se proceder á análise do sistema com parte discreta, será variado o
parâmetro g. O parâmetro g, presente nas equações da parte discreta, representa o offset do
sinal após a inclusão da parte discreta.
Revê-se a teoria de bifurcação e simetria que permite a classificação das soluções
periódicas produzidas pelos modelos de CPGs com passos locomotores quadrúpedes. Nas
simulações numéricas, usam-se as equações de Morris-Lecar e o oscilador de Hopf como
modelos da dinâmica interna de cada célula para a parte rítmica. A parte discreta é
modelada por um sistema inspirado no modelo VITE. Medem-se a amplitude e a
frequência de dois passos locomotores para variação do parâmetro g, no intervalo [-5;5].
Consideram-se duas formas distintas de incluir a parte discreta na parte rítmica: (a) como
um (a.1) offset ou (a.2) somada nas expressões que modelam a parte rítmica, e (b) somada
ao sinal da parte rítmica antes de ser enviado às articulações do robô. No caso (a.1),
considerando o oscilador de Hopf como dinâmica interna das células, verifica-se que a amplitude e frequência se mantêm constantes para -5
Resumo:
The mechanisms of speech production are complex and have been raising attention from researchers of both medical and computer vision fields. In the speech production mechanism, the articulator’s study is a complex issue, since they have a high level of freedom along this process, namely the tongue, which instigates a problem in its control and observation. In this work it is automatically characterized the tongues shape during the articulation of the oral vowels of Portuguese European by using statistical modeling on MR-images. A point distribution model is built from a set of images collected during artificially sustained articulations of Portuguese European sounds, which can extract the main characteristics of the motion of the tongue. The model built in this work allows under standing more clearly the dynamic speech events involved during sustained articulations. The tongue shape model built can also be useful for speech rehabilitation purposes, specifically to recognize the compensatory movements of the articulators during speech production.
Resumo:
In recent years several countries have set up policies that allow exchange of kidneys between two or more incompatible patient–donor pairs. These policies lead to what is commonly known as kidney exchange programs. The underlying optimization problems can be formulated as integer programming models. Previously proposed models for kidney exchange programs have exponential numbers of constraints or variables, which makes them fairly difficult to solve when the problem size is large. In this work we propose two compact formulations for the problem, explain how these formulations can be adapted to address some problem variants, and provide results on the dominance of some models over others. Finally we present a systematic comparison between our models and two previously proposed ones via thorough computational analysis. Results show that compact formulations have advantages over non-compact ones when the problem size is large.