920 resultados para Frontiers
Resumo:
The most biologically-inspired artificial neurons are those of the third generation, and are termed spiking neurons, as individual pulses or spikes are the means by which stimuli are communicated. In essence, a spike is a short-term change in electrical potential and is the basis of communication between biological neurons. Unlike previous generations of artificial neurons, spiking neurons operate in the temporal domain, and exploit time as a resource in their computation. In 1952, Alan Lloyd Hodgkin and Andrew Huxley produced the first model of a spiking neuron; their model describes the complex electro-chemical process that enables spikes to propagate through, and hence be communicated by, spiking neurons. Since this time, improvements in experimental procedures in neurobiology, particularly with in vivo experiments, have provided an increasingly more complex understanding of biological neurons. For example, it is now well-understood that the propagation of spikes between neurons requires neurotransmitter, which is typically of limited supply. When the supply is exhausted neurons become unresponsive. The morphology of neurons, number of receptor sites, amongst many other factors, means that neurons consume the supply of neurotransmitter at different rates. This in turn produces variations over time in the responsiveness of neurons, yielding various computational capabilities. Such improvements in the understanding of the biological neuron have culminated in a wide range of different neuron models, ranging from the computationally efficient to the biologically realistic. These models enable the modeling of neural circuits found in the brain.
Resumo:
The most biologically-inspired artificial neurons are those of the third generation, and are termed spiking neurons, as individual pulses or spikes are the means by which stimuli are communicated. In essence, a spike is a short-term change in electrical potential and is the basis of communication between biological neurons. Unlike previous generations of artificial neurons, spiking neurons operate in the temporal domain, and exploit time as a resource in their computation. In 1952, Alan Lloyd Hodgkin and Andrew Huxley produced the first model of a spiking neuron; their model describes the complex electro-chemical process that enables spikes to propagate through, and hence be communicated by, spiking neurons. Since this time, improvements in experimental procedures in neurobiology, particularly with in vivo experiments, have provided an increasingly more complex understanding of biological neurons. For example, it is now well understood that the propagation of spikes between neurons requires neurotransmitter, which is typically of limited supply. When the supply is exhausted neurons become unresponsive. The morphology of neurons, number of receptor sites, amongst many other factors, means that neurons consume the supply of neurotransmitter at different rates. This in turn produces variations over time in the responsiveness of neurons, yielding various computational capabilities. Such improvements in the understanding of the biological neuron have culminated in a wide range of different neuron models, ranging from the computationally efficient to the biologically realistic. These models enable the modelling of neural circuits found in the brain. In recent years, much of the focus in neuron modelling has moved to the study of the connectivity of spiking neural networks. Spiking neural networks provide a vehicle to understand from a computational perspective, aspects of the brain’s neural circuitry. This understanding can then be used to tackle some of the historically intractable issues with artificial neurons, such as scalability and lack of variable binding. Current knowledge of feed-forward, lateral, and recurrent connectivity of spiking neurons, and the interplay between excitatory and inhibitory neurons is beginning to shed light on these issues, by improved understanding of the temporal processing capabilities and synchronous behaviour of biological neurons. This research topic aims to amalgamate current research aimed at tackling these phenomena.
Resumo:
The Editorial on the Research Topic: Facing the Other: Novel Theories and Methods in Face Perception Research
Resumo:
Face detection and recognition should be complemented by recognition of facial expression, for example for social robots which must react to human emotions. Our framework is based on two multi-scale representations in cortical area V1: keypoints at eyes, nose and mouth are grouped for face detection [1]; lines and edges provide information for face recognition [2].
Resumo:
This article outlines the approaches to modeling the distribution of threatened invertebrates using data from atlases, museums and databases. Species Distribution Models (SDMs) are useful for estimating species’ ranges, identifying suitable habitats, and identifying the primary factors affecting species’ distributions. The study tackles the strategies used to obtain SDMs without reliable absence data while exploring their applications for conservation. I examine the conservation status of Copris species and Graellsia isabelae by delimiting their populations and exploring the effectiveness of protected areas. I show that the method of pseudo‐absence selection strongly determines the model obtained, generating different model predictions along the gradient between potential and realized distributions. After assessing the effects of species’ traits and data characteristics on accuracy, I found that species are modeled more accurately when sample sizes are larger, no matter the technique used.
Resumo:
Trabalho de projecto de mestrado, Ciências da Educação (Formação de Adultos), Universidade de Lisboa, Instituto de Educação, 2011
Resumo:
It is commonly accepted that the educational environment has been undergoing considerable change due to the use of the Information and Communication tools. But learning depends upon actions such as experimenting, visualizing and demonstrating through which the learner succeeds in constructing his own knowledge. Although it is not easy to achieve these actions through current ICT supported learning approaches, Role Playing Games (RPG) may well develop such capacities. The creation of an interactive computer game with RPG characteristics, about the 500th anniversary of the city of Funchal, the capital of Madeira Island, is invested with compelling educational/pedagogical implications, aiming clearly at teaching history and social relations through playing. Players interpret different characters in different settings/scenarios, experiencing adventures, meeting challenges and trying to reach multiple and simultaneous goals in the areas of education, entertainment and social integration along the first 150 years of the history of Funchal. Through this process they will live and understand all the social and historical factors of that epoch.
Resumo:
The use of remote labs in undergraduate courses has been reported in literature several times since the mid 90's. Nevertheless, very few articles present results about the correspondent learning gains obtained by students, and in what conditions those systems can be more efficient, thus suggesting a lack of data concerning their pedagogical effectiveness. This paper addresses such a gap by presenting some initial findings concerning the use of a remote lab (VISIR), in a large undergraduate course on Physics, with over 550 students enrolled.
Resumo:
Practical sessions are the backbone of qualification in engineering education. It leads to a better understanding and allows mastering scientific concepts and theories. The lack of the availability of practical sessions at many universities and institutions owing to the cost and the unavailability of instructors the most of the time caused a significant decline in experimentation in engineering education over the last decades. Recently, with the progress of computer-based learning, remote laboratories have been proven to be the best alternative to the traditional ones, regarding to its low cost and ubiquity. Some universities have already started to deploy remote labs in their practical sessions. This contribution compiles diverse experiences based on the deployment of the remote laboratory, Virtual Instrument Systems in Reality (VISIR), on the practices of undergraduate engineering grades at various universities within the VISIR community. It aims to show the impact of its usage on engineering education concerning the assessments of students and teachers as well.
Resumo:
A new method, based on linear correlation and phase diagrams was successfully developed for processes like the sedimentary process, where the deposition phase can have different time duration - represented by repeated values in a series - and where the erosion can play an important rule deleting values of a series. The sampling process itself can be the cause of repeated values - large strata twice sampled - or deleted values: tiny strata fitted between two consecutive samples. What we developed was a mathematical procedure which, based upon the depth chemical composition evolution, allows the establishment of frontiers as well as the periodicity of different sedimentary environments. The basic tool isn't more than a linear correlation analysis which allow us to detect the existence of eventual evolution rules, connected with cyclical phenomena within time series (considering the space assimilated to time), with the final objective of prevision. A very interesting discovery was the phenomenon of repeated sliding windows that represent quasi-cycles of a series of quasi-periods. An accurate forecast can be obtained if we are inside a quasi-cycle (it is possible to predict the other elements of the cycle with the probability related with the number of repeated and deleted points). We deal with an innovator methodology, reason why it's efficiency is being tested in some case studies, with remarkable results that shows it's efficacy. Keywords: sedimentary environments, sequence stratigraphy, data analysis, time-series, conditional probability.
Resumo:
Dissertação de Mestrado apresentado ao Instituto Superior de Contabilidade e Administração do Porto para obtenção de grau de Mestre em Empreendedorismo e Internacionalização, sob a orientação da Professora Doutora Manuela Maria Ribeiro da Silva Patrício
Resumo:
The last decade has witnessed a major shift towards the deployment of embedded applications on multi-core platforms. However, real-time applications have not been able to fully benefit from this transition, as the computational gains offered by multi-cores are often offset by performance degradation due to shared resources, such as main memory. To efficiently use multi-core platforms for real-time systems, it is hence essential to tightly bound the interference when accessing shared resources. Although there has been much recent work in this area, a remaining key problem is to address the diversity of memory arbiters in the analysis to make it applicable to a wide range of systems. This work handles diverse arbiters by proposing a general framework to compute the maximum interference caused by the shared memory bus and its impact on the execution time of the tasks running on the cores, considering different bus arbiters. Our novel approach clearly demarcates the arbiter-dependent and independent stages in the analysis of these upper bounds. The arbiter-dependent phase takes the arbiter and the task memory-traffic pattern as inputs and produces a model of the availability of the bus to a given task. Then, based on the availability of the bus, the arbiter-independent phase determines the worst-case request-release scenario that maximizes the interference experienced by the tasks due to the contention for the bus. We show that the framework addresses the diversity problem by applying it to a memory bus shared by a fixed-priority arbiter, a time-division multiplexing (TDM) arbiter, and an unspecified work-conserving arbiter using applications from the MediaBench test suite. We also experimentally evaluate the quality of the analysis by comparison with a state-of-the-art TDM analysis approach and consistently showing a considerable reduction in maximum interference.
Resumo:
IEEE International Conference on Communications (IEEE ICC 2015). 8 to 12, Jun, 2015, IEEE ICC 2015 - Communications QoS, Reliability and Modeling, London, United Kingdom.
Resumo:
Wireless body area networks (WBANs) are expected to play a significant role in smart healthcare systems. One of the most important attributes of WBANs is to increase network lifetime by introducing novel and low-power techniques on the energy-constrained sensor nodes. Medium access control (MAC) protocols play a significant role in determining the energy consumption in WBANs. Existing MAC protocols are unable to accommodate communication requirements in WBANs. There is a need to develop novel, scalable and reliable MAC protocols that must be able to address all these requirements in a reliable manner. In this special issue, we attracted high quality research and review papers on the recent advances in MAC protocols for WBANs.
Resumo:
This thesis provides a complete analysis of the Standard Capital Requirements given by Solvency II for a real insurance portfolio. We analyze the investment portfolio of BPI Vida e Pensões, an insurance company affiliated with a Portuguese bank BPI, both at security, sub-portfolio and asset class levels. By using the Standard Formula from EIOPA, Total SCR amounts to 239M€. This value is mostly explained by Market and Default Risk whereas the former is driven by Spread and Concentration Risks. Following the methodology of Leblanc (2011), we examine the Marginal Contribution of an asset to the SCR which allows for the evaluation of the risks of each security given its characteristics and interactions in the portfolio. The top contributors to the SCR are Corporate Bonds and Term Deposits. By exploring further the composition of the portfolio, our results show that slight changes in allocation of Term and Cash Deposits have severe impacts on the total Concentration and Default Risks, respectively. Also, diversification effects are very relevant by representing savings of 122M€. Finally, Solvency II represents an opportunity for the portfolio optimization. By constructing efficient frontiers, we find that as the target expected return increases, a shift from Term Deposits/ Commercial Papers to Eurozone/Peripheral and finally Equities occurs.