939 resultados para TOT (time of task)
Resumo:
Acoustic Emission (AE) monitoring can be used to detect the presence of damage as well as determine its location in Structural Health Monitoring (SHM) applications. Information on the time difference of the signal generated by the damage event arriving at different sensors is essential in performing localization. This makes the time of arrival (ToA) an important piece of information to retrieve from the AE signal. Generally, this is determined using statistical methods such as the Akaike Information Criterion (AIC) which is particularly prone to errors in the presence of noise. And given that the structures of interest are surrounded with harsh environments, a way to accurately estimate the arrival time in such noisy scenarios is of particular interest. In this work, two new methods are presented to estimate the arrival times of AE signals which are based on Machine Learning. Inspired by great results in the field, two models are presented which are Deep Learning models - a subset of machine learning. They are based on Convolutional Neural Network (CNN) and Capsule Neural Network (CapsNet). The primary advantage of such models is that they do not require the user to pre-define selected features but only require raw data to be given and the models establish non-linear relationships between the inputs and outputs. The performance of the models is evaluated using AE signals generated by a custom ray-tracing algorithm by propagating them on an aluminium plate and compared to AIC. It was found that the relative error in estimation on the test set was < 5% for the models compared to around 45% of AIC. The testing process was further continued by preparing an experimental setup and acquiring real AE signals to test on. Similar performances were observed where the two models not only outperform AIC by more than a magnitude in their average errors but also they were shown to be a lot more robust as compared to AIC which fails in the presence of noise.
Resumo:
I vantaggi dell’Industria 4.0 hanno stravolto il manufacturing. Ma cosa vuol dire "Industria 4.0"? Essa è la nuova frontiera del manufacturing, basata su princìpi che seguono i passi avanti dei sistemi IT e della tecnologia. Dunque, i suoi pilastri sono: integrazione, verticale e orizzontale, digitalizzazione e automazione. L’Industria 4.0 coinvolge molte aree della supply chain, dai flussi informativi alla logistica. In essa e nell’intralogistica, la priorità è sviluppare dei sistemi di material handling flessibili, automatizzati e con alta prontezza di risposta. Il modello ideale è autonomo, in cui i veicoli fanno parte di una flotta le cui decisioni sono rese decentralizzate grazie all'alta connettività e alla loro abilità di collezionare dati e scambiarli rapidamente nel cloud aziendale.Tutto ciò non sarebbe raggiungibile se ci si affidasse a un comune sistema di trasporto AGV, troppo rigido e centralizzato. La tesi si focalizza su un tipo di material handlers più flessibile e intelligente: gli Autonomous Mobile Robots. Grazie alla loro intelligenza artificiale e alla digitalizzazione degli scambi di informazioni, interagiscono con l’ambiente per evitare ostacoli e calcolare il percorso ottimale. Gli scenari dell’ambiente lavorativo determinano perdite di tempo nel tragitto dei robot e sono queste che dovremo studiare. Nella tesi, i vantaggi apportati dagli AMR, come la loro decentralizzazione delle decisioni, saranno introdotti mediante una literature review e poi l’attenzione verterà sull’analisi di ogni scenario di lavoro. Fondamentali sono state le esperienze nel Logistics 4.0 Lab di NTNU, per ricreare fisicamente alcuni scenari. Inoltre, il software AnyLogic sarà usato per riprodurre e simulare tutti gli scenari rilevanti. I risultati delle simulazioni verranno infine usati per creare un modello che associ ad ogni scenario rilevante una perdita di tempo, attraverso una funzione. Per questo saranno usati software di data analysis come Minitab e MatLab.
Resumo:
The present study investigated whether developmental changes in cognitive control may underlie improvements of time-based prospective memory. Five-, 7-, 9-, and 11-year-olds (N = 166) completed a driving simulation task (ongoing task) in which they had to refuel their vehicle at specific points in time (PM task). The availability of cognitive control resources was experimentally manipulated by imposing a secondary task that required divided attention. Children completed the driving simulation task both in a full attention condition and a divided attention condition where they had to carry out a secondary task. Results revealed that older children performed better than younger children on the ongoing task and PM task. Children performed worse on the ongoing and PM tasks in the divided attention condition compared to the full attention condition. With respect to time monitoring in the final interval prior to the PM target, divided attention interacted with age such that older children’s time monitoring was more negatively affected by the secondary task compared to younger children. Results are discussed in terms of developmental shifts from reactive to proactive monitoring strategies.
Resumo:
The present study was designed to investigate the influences of type of psychophysical task (two-alternative forced-choice [2AFC] and reminder tasks), type of interval (filled vs. empty), sensory modality (auditory vs. visual), and base duration (ranging from 100 through 1,000 ms) on performance on duration discrimination. All of these factors were systematically varied in an experiment comprising 192 participants. This approach allowed for obtaining information not only on the general (main) effect of each factor alone, but also on the functional interplay and mutual interactions of some or all of these factors combined. Temporal sensitivity was markedly higher for auditory than for visual intervals, as well as for the reminder relative to the 2AFC task. With regard to base duration, discrimination performance deteriorated with decreasing base durations for intervals below 400 ms, whereas longer intervals were not affected. No indication emerged that overall performance on duration discrimination was influenced by the type of interval, and only two significant interactions were apparent: Base Duration × Type of Interval and Base Duration × Sensory Modality. With filled intervals, the deteriorating effect of base duration was limited to very brief base durations, not exceeding 100 ms, whereas with empty intervals, temporal discriminability was also affected for the 200-ms base duration. Similarly, the performance decrement observed with visual relative to auditory intervals increased with decreasing base durations. These findings suggest that type of task, sensory modality, and base duration represent largely independent sources of variance for performance on duration discrimination that can be accounted for by distinct nontemporal mechanisms.
Resumo:
The study aimed to elucidate electrophysiological and cortical mechanisms involved in anticipatory actions when healthy subjects had to catch balls in free drop; specifically through quantitative electroencephalography (qEEG) alpha absolute power changes. Our hypothesis is that during the preparation of motor action (i.e., 2 s before ball`s drop) occurred integration among left medial frontal, left primary somatomotor and left posterior parietal cortices, showing a differentiated activity involving expectation, planning and preparedness. This hypothesis supports a lateralization of motor function. Although we contend that in right-handers the left hemisphere takes on a dominant role for the regulation of motor behavior. The sample was composed of 23 healthy subjects (13 male and 10 female), right handed, with ages varying between 25 and 40 years old (32.5 +/- 7.5), absence of mental and physical illness, right handed, and do not make use of any psychoactive or psychotropic substance at the time of the study. The experiment consisted of a task of catching balls in free drop. The three-way ANOVA analysis demonstrated an interaction between moment and position in left medial frontal cortex (F3 electrode), somatomotor cortex (C3 electrode) and posterior parietal cortex (P3 electrode: p < 0.001). Summarizing, through experimental task employed, it was possible to observe integration among frontal, central and parietal regions. This integration appears to be more predominant in expectation, planning and motor preparation. In this way, it established an absolute predominance of this mechanism under the left hemisphere. (C) 2008 Elsevier Ireland Ltd. All rights reserved.
Resumo:
This paper proposes a global multiprocessor scheduling algorithm for the Linux kernel that combines the global EDF scheduler with a priority-aware work-stealing load balancing scheme, enabling parallel real-time tasks to be executed on more than one processor at a given time instant. We state that some priority inversion may actually be acceptable, provided it helps reduce contention, communication, synchronisation and coordination between parallel threads, while still guaranteeing the expected system’s predictability. Experimental results demonstrate the low scheduling overhead of the proposed approach comparatively to an existing real-time deadline-oriented scheduling class for the Linux kernel.
Resumo:
Multicore platforms have transformed parallelism into a main concern. Parallel programming models are being put forward to provide a better approach for application programmers to expose the opportunities for parallelism by pointing out potentially parallel regions within tasks, leaving the actual and dynamic scheduling of these regions onto processors to be performed at runtime, exploiting the maximum amount of parallelism. It is in this context that this paper proposes a scheduling approach that combines the constant-bandwidth server abstraction with a priority-aware work-stealing load balancing scheme which, while ensuring isolation among tasks, enables parallel tasks to be executed on more than one processor at a given time instant.
Resumo:
The current industry trend is towards using Commercially available Off-The-Shelf (COTS) based multicores for developing real time embedded systems, as opposed to the usage of custom-made hardware. In typical implementation of such COTS-based multicores, multiple cores access the main memory via a shared bus. This often leads to contention on this shared channel, which results in an increase of the response time of the tasks. Analyzing this increased response time, considering the contention on the shared bus, is challenging on COTS-based systems mainly because bus arbitration protocols are often undocumented and the exact instants at which the shared bus is accessed by tasks are not explicitly controlled by the operating system scheduler; they are instead a result of cache misses. This paper makes three contributions towards analyzing tasks scheduled on COTS-based multicores. Firstly, we describe a method to model the memory access patterns of a task. Secondly, we apply this model to analyze the worst case response time for a set of tasks. Although the required parameters to obtain the request profile can be obtained by static analysis, we provide an alternative method to experimentally obtain them by using performance monitoring counters (PMCs). We also compare our work against an existing approach and show that our approach outperforms it by providing tighter upper-bound on the number of bus requests generated by a task.
Resumo:
Consider the problem of scheduling a set of sporadic tasks on a multiprocessor system to meet deadlines using a tasksplitting scheduling algorithm. Task-splitting (also called semipartitioning) scheduling algorithms assign most tasks to just one processor but a few tasks are assigned to two or more processors, and they are dispatched in a way that ensures that a task never executes on two or more processors simultaneously. A certain type of task-splitting algorithms, called slot-based task-splitting, is of particular interest because of its ability to schedule tasks at high processor utilizations. We present a new schedulability analysis for slot-based task-splitting scheduling algorithms that takes the overhead into account and also a new task assignment algorithm.
Resumo:
Embedded real-time systems often have to support the embedding system in very different and changing application scenarios. An aircraft taxiing, taking off and in cruise flight is one example. The different application scenarios are reflected in the software structure with a changing task set and thus different operational modes. At the same time there is a strong push for integrating previously isolated functionalities in single-chip multicore processors. On such multicores the behavior of the system during a mode change, when the systems transitions from one mode to another, is complex but crucial to get right. In the past we have investigated mode change in multiprocessor systems where a mode change requires a complete change of task set. Now, we present the first analysis which considers mode changes in multicore systems, which use global EDF to schedule a set of mode independent (MI) and mode specific (MS) tasks. In such systems, only the set of MS tasks has to be replaced during mode changes, without jeopardizing the schedulability of the MI tasks. Of prime concern is that the mode change is safe and efficient: i.e. the mode change needs to be performed in a predefined time window and no deadlines may be missed as a function of the mode change.
Resumo:
Multiprocessors, particularly in the form of multicores, are becoming standard building blocks for executing reliable software. But their use for applications with hard real-time requirements is non-trivial. Well-known realtime scheduling algorithms in the uniprocessor context (Rate-Monotonic [1] or Earliest-Deadline-First [1]) do not perform well on multiprocessors. For this reason the scientific community in the area of real-time systems has produced new algorithms specifically for multiprocessors. In the meanwhile, a proposal [2] exists for extending the Ada language with new basic constructs which can be used for implementing new algorithms for real-time scheduling; the family of task splitting algorithms is one of them which was emphasized in the proposal [2]. Consequently, assessing whether existing task splitting multiprocessor scheduling algorithms can be implemented with these constructs is paramount. In this paper we present a list of state-of-art task-splitting multiprocessor scheduling algorithms and, for each of them, we present detailed Ada code that uses the new constructs.
Resumo:
Real-time scheduling usually considers worst-case values for the parameters of task (or message stream) sets, in order to provide safe schedulability tests for hard real-time systems. However, worst-case conditions introduce a level of pessimism that is often inadequate for a certain class of (soft) real-time systems. In this paper we provide an approach for computing the stochastic response time of tasks where tasks have inter-arrival times described by discrete probabilistic distribution functions, instead of minimum inter-arrival (MIT) values.
Resumo:
Recent embedded processor architectures containing multiple heterogeneous cores and non-coherent caches renewed attention to the use of Software Transactional Memory (STM) as a building block for developing parallel applications. STM promises to ease concurrent and parallel software development, but relies on the possibility of abort conflicting transactions to maintain data consistency, which in turns affects the execution time of tasks carrying transactions. Because of this fact the timing behaviour of the task set may not be predictable, thus it is crucial to limit the execution time overheads resulting from aborts. In this paper we formalise a FIFO-based algorithm to order the sequence of commits of concurrent transactions. Then, we propose and evaluate two non-preemptive and one SRP-based fully-preemptive scheduling strategies, in order to avoid transaction starvation.
Resumo:
The Commission on Classification and Terminology and the Commission on Epidemiology of the International League Against Epilepsy (ILAE) have charged a Task Force to revise concepts, definition, and classification of status epilepticus (SE). The proposed new definition of SE is as follows: Status epilepticus is a condition resulting either from the failure of the mechanisms responsible for seizure termination or from the initiation of mechanisms, which lead to abnormally, prolonged seizures (after time point t1 ). It is a condition, which can have long-term consequences (after time point t2 ), including neuronal death, neuronal injury, and alteration of neuronal networks, depending on the type and duration of seizures. This definition is conceptual, with two operational dimensions: the first is the length of the seizure and the time point (t1 ) beyond which the seizure should be regarded as "continuous seizure activity." The second time point (t2 ) is the time of ongoing seizure activity after which there is a risk of long-term consequences. In the case of convulsive (tonic-clonic) SE, both time points (t1 at 5 min and t2 at 30 min) are based on animal experiments and clinical research. This evidence is incomplete, and there is furthermore considerable variation, so these time points should be considered as the best estimates currently available. Data are not yet available for other forms of SE, but as knowledge and understanding increase, time points can be defined for specific forms of SE based on scientific evidence and incorporated into the definition, without changing the underlying concepts. A new diagnostic classification system of SE is proposed, which will provide a framework for clinical diagnosis, investigation, and therapeutic approaches for each patient. There are four axes: (1) semiology; (2) etiology; (3) electroencephalography (EEG) correlates; and (4) age. Axis 1 (semiology) lists different forms of SE divided into those with prominent motor systems, those without prominent motor systems, and currently indeterminate conditions (such as acute confusional states with epileptiform EEG patterns). Axis 2 (etiology) is divided into subcategories of known and unknown causes. Axis 3 (EEG correlates) adopts the latest recommendations by consensus panels to use the following descriptors for the EEG: name of pattern, morphology, location, time-related features, modulation, and effect of intervention. Finally, axis 4 divides age groups into neonatal, infancy, childhood, adolescent and adulthood, and elderly.
Resumo:
Older adults represent the most sedentary segment of the adult population, and thus it is critical to investigate factors that influence exercise behaviour for this age group. The purpose of this study was to examine the influence of a general exercise program, incorporating cardiovascular, strength, flexibility, and balance components, on task selfefficacy and SPA in older adult men and women. Participants (n=114, Mage = 67 years) were recruited from the Niagara region and randomly assigned to a 12-week supervised exercise program or a wait-list control. Task self-efficacy and SPA measures were taken at baseline and program end. The present study found that task self-efficacy was a significant predictor of leisure time physical activity for older adults. In addition, change in task self-efficacy was a significant predictor of change in SPA. The findings of this study suggest that sources of task self-efficacy should be considered for exercise interventions targeting older adults.