912 resultados para Signal Processing, Computer-Assisted
Resumo:
This paper presents a new type of Flexible Macroblock Ordering (FMO) type for the H.264 Advanced Video Coding (AVC) standard, which can more efficiently flag the position and shape of regions of interest (ROIs) in each frame. In H.264/AVC, 7 types of FMO have been defined, all of which are designed for error resilience. Most previous work related to ROI processing has adopted Type-2 (foreground & background), or Type-6 (explicit), to flag the position and shape of the ROI. However, only rectangular shapes are allowed in Type-2 and for non-rectangular shapes, the non-ROI macroblocks may be wrongly flagged as being within the ROI, which could seriously affect subsequent processing of the ROI. In Type-6, each macroblock in a frame uses fixed-length bits to indicate to its slice group. In general, each ROI is assigned to one slice group identity. Although this FMO type can more accurately flag the position and shape of the ROI, it incurs a significant bitrate overhead. The proposed new FMO type uses the smallest rectangle that covers the ROI to indicate its position and a spiral binary mask is employed within the rectangle to indicate the shape of the ROI. This technique can accurately flag the ROI and provide significantly savings in the bitrate overhead. Compared with Type-6, an 80% to 90% reduction in the bitrate overhead can be obtained while achieving the same accuracy.
Resumo:
Energy in today's short-range wireless communication is mostly spent on the analog- and digital hardware rather than on radiated power. Hence,purely information-theoretic considerations fail to achieve the lowest energy per information bit and the optimization process must carefully consider the overall transceiver. In this paper, we propose to perform cross-layer optimization, based on an energy-aware rate adaptation scheme combined with a physical layer that is able to properly adjust its processing effort to the data rate and the channel conditions to minimize the energy consumption per information bit. This energy proportional behavior is enabled by extending the classical system modes with additional configuration parameters at the various layers. Fine grained models of the power consumption of the hardware are developed to provide awareness of the physical layer capabilities to the medium access control layer. The joint application of the proposed energy-aware rate adaptation and modifications to the physical layer of an IEEE802.11n system, improves energy-efficiency (averaged over many noise and channel realizations) in all considered scenarios by up to 44%.
Resumo:
Timely and individualized feedback on coursework is desirable from a student perspective as it facilitates formative development and encourages reflective learning practice. Faculty however are faced with a significant and potentially time consuming challenge when teaching larger cohorts if they are to provide feedback which is timely, individualized and detailed. Additionally, for subjects which assess non-traditional submissions, such as Computer-Aided-Design (CAD), the methods for assessment and feedback tend not to be so well developed or optimized. Issues can also arise over the consistency of the feedback provided. Evaluations of Computer-Assisted feedback in other disciplines (Denton et al, 2008), (Croft et al, 2001) have shown students prefer this method of feedback to traditional “red pen” marking and also that such methods can be more time efficient for faculty.
Herein, approaches are described which make use of technology and additional software tools to speed up, simplify and automate assessment and the provision of feedback for large cohorts of first and second year engineering students studying modules where CAD files are submitted electronically. A range of automated methods are described and compared with more “manual” approaches. Specifically one method uses an application programming interface (API) to interrogate SolidWorks models and extract information into an Excel spreadsheet, which is then used to automatically send feedback emails. Another method describes the use of audio recordings made during model interrogation which reduces the amount of time while increasing the level of detail provided as feedback.
Limitations found with these methods and problems encountered are discussed along with a quantified assessment of time saving efficiencies made.
Resumo:
This paper presents a novel method of audio-visual fusion for person identification where both the speech and facial modalities may be corrupted, and there is a lack of prior knowledge about the corruption. Furthermore, we assume there is a limited amount of training data for each modality (e.g., a short training speech segment and a single training facial image for each person). A new representation and a modified cosine similarity are introduced for combining and comparing bimodal features with limited training data as well as vastly differing data rates and feature sizes. Optimal feature selection and multicondition training are used to reduce the mismatch between training and testing, thereby making the system robust to unknown bimodal corruption. Experiments have been carried out on a bimodal data set created from the SPIDRE and AR databases with variable noise corruption of speech and occlusion in the face images. The new method has demonstrated improved recognition accuracy.
Resumo:
Due to its efficiency and simplicity, the finite-difference time-domain method is becoming a popular choice for solving wideband, transient problems in various fields of acoustics. So far, the issue of extracting a binaural response from finite difference simulations has only been discussed in the context of embedding a listener geometry in the grid. In this paper, we propose and study a method for binaural response rendering based on a spatial decomposition of the sound field. The finite difference grid is locally sampled using a volumetric array of receivers, from which a plane wave density function is computed and integrated with free-field head related transfer functions, in the spherical harmonics domain. The volumetric array is studied in terms of numerical robustness and spatial aliasing. Analytic formulas that predict the performance of the array are developed, facilitating spatial resolution analysis and numerical binaural response analysis for a number of finite difference schemes. Particular emphasis is placed on the effects of numerical dispersion on array processing and on the resulting binaural responses. Our method is compared to a binaural simulation based on the image method. Results indicate good spatial and temporal agreement between the two methods.
Resumo:
In this paper, we introduce a statistical data-correction framework that aims at improving the DSP system performance in presence of unreliable memories. The proposed signal processing framework implements best-effort error mitigation for signals that are corrupted by defects in unreliable storage arrays using a statistical correction function extracted from the signal statistics, a data-corruption model, and an application-specific cost function. An application example to communication systems demonstrates the efficacy of the proposed approach.
Resumo:
In this paper, we investigate the impact of faulty memory bit-cells on the performance of LDPC and Turbo channel decoders based on realistic memory failure models. Our study investigates the inherent error resilience of such codes to potential memory faults affecting the decoding process. We develop two mitigation mechanisms that reduce the impact of memory faults rather than correcting every single error. We show how protection of only few bit-cells is sufficient to deal with high defect rates. In addition, we show how the use of repair-iterations specifically helps mitigating the impact of faults that occur inside the decoder itself.
Resumo:
Inherently error-resilient applications in areas such as signal processing, machine learning and data analytics provide opportunities for relaxing reliability requirements, and thereby reducing the overhead incurred by conventional error correction schemes. In this paper, we exploit the tolerable imprecision of such applications by designing an energy-efficient fault-mitigation scheme for unreliable data memories to meet target yield. The proposed approach uses a bit-shuffling mechanism to isolate faults into bit locations with lower significance. This skews the bit-error distribution towards the low order bits, substantially limiting the output error magnitude. By controlling the granularity of the shuffling, the proposed technique enables trading-off quality for power, area, and timing overhead. Compared to error-correction codes, this can reduce the overhead by as much as 83% in read power, 77% in read access time, and 89% in area, when applied to various data mining applications in 28nm process technology.
Resumo:
We consider a multi-pair two-way amplify-and-forward relaying system with a massive antenna array at the relay and estimated channel state information, assuming maximum-ratio combining/transmission processing. Closed-form approximations of the sum spectral effi- ciency are developed and simple analytical power scaling laws are presented, which reveal a fundamental trade-off between the transmit powers of each user/the relay and of each pilot symbol. Finally, the optimal power allocation problem is studied.
Resumo:
Presents the introductory welcome message from the conference proceedings. May include the conference officers' congratulations to all involved with the conference event and publication of the proceedings record.
Resumo:
The increasing design complexity associated with modern Field Programmable Gate Array (FPGA) has prompted the emergence of 'soft'-programmable processors which attempt to replace at least part of the custom circuit design problem with a problem of programming parallel processors. Despite substantial advances in this technology, its performance and resource efficiency for computationally complex operations remains in doubt. In this paper we present the first recorded implementation of a softcore Fast-Fourier Transform (FFT) on Xilinx Virtex FPGA technology. By employing a streaming processing architecture, we show how it is possible to achieve architectures which offer 1.1 GSamples/s throughput and up to 19 times speed-up against the Xilinx Radix-2 FFT dedicated circuit with comparable cost.
Resumo:
The present PhD work aims the research and development of materials that exhibit multiferroic properties, in particular having a significant interaction between ferromagnetism and ferroelectricity; either directly within an intrinsic single phase or by combining extrinsic materials, achieving the coupling of properties through mechanic phenomena of the respective magnetostriction and piezoelectricity. These hybrid properties will allow the cross modification of magnetic and electric polarization states by the application of cross external magnetic and/or electric fields, giving way to a vast area for scientific investigation and potential technological applications in a new generation of electronic devices, such as computer memories, signal processing, transducers, sensors, etc. Initial experimental work consisted in chemical synthesis of nano powders oxides by urea pyrolysis method: A series of ceramic bulk composites with potential multiferroic properties comprised: of LuMnO3 with La0.7Sr0.3MnO3 and BaTiO3 with La0.7Ba0.3MnO3; and a series based on the intrinsic multiferroic LuMn1-zO3 phase modified with of Manganese vacancies. The acquisition of a new magnetron RF sputtering deposition system, in the Physics Department of Aveiro University, contributed to the proposal of an analogous experimental study in multiferroic thin films and multilayer samples. Besides the operational debut of this equipment several technical upgrades were completed like: the design and construction of the heater electrical contacts; specific shutters and supports for the magnetrons and for the substrate holder and; the addition of mass flow controllers, which allowed the introduction of N2 or O2 active atmosphere in the chamber; and the addition of a second RF generator, enabling co-deposition of different targets. Base study of the deposition conditions and resulting thin films characteristics in different substrates was made from an extensive list of targets. Particular attention was given to thin film deposition of magnetic phases La1-xSrxMnO3, La1-xBaxMnO3 and Ni2+x-yMn1-xGa1+y alloy, from the respective targets: La0.7Sr0.3MnO3, La0.7Ba0.3MnO3; and NiGa with NiMn. Main structural characterization of samples was performed by conventional and high resolution X-Ray Diffraction (XRD); chemical composition was determined by Electron Dispersion Spectroscopy (EDS); magnetization measurements recur to a Vibrating Sample Magnetometer (VSM) prototype; and surface probing (SPM) using Magnetic-Force (MFM) and Piezo-Response (PFM) Microscopy. Results clearly show that the composite bulk samples (LuM+LSM and BTO+LBM) feat the intended quality objectives in terms of phase composition and purity, having spurious contents below 0.5 %. SEM images confirm compact grain packaging and size distribution around the 50 nm scale. Electric conductivity, magnetization intensity and magneto impedance spreading response are coherent with the relative amount of magnetic phase in the sample. The existence of coupling between the functional phases is confirmed by the Magnetoelectric effect measurements of the sample “78%LuM+22%LSM” reaching 300% of electric response for 1 T at 100 kHz; while in the “78%BTO+22%LBM” sample the structural transitions of the magnetic phase at ~350 K result in a inversion of ME coefficient the behavior. A functional Magneto-Resistance measurement system was assembled from the concept stage until the, development and operational status; it enabled to test samples from 77 to 350 K, under an applied magnetic field up to 1 Tesla with 360º horizontal rotation; this system was also designed to measure Hall effect and has the potential to be further upgraded. Under collaboration protocols established with national and international institutions, complementary courses and sample characterization studies were performed using Magneto-Resistance (MR), Magneto-Impedance (MZ) and Magneto-Electric (ME) measurements; Raman and X-ray Photoelectron Spectroscopy (XPS); SQUID and VSM magnetization; Scanning Electron Microscopy (SEM) and Rutherford Back Scattering (RBS); Scan Probe Microscopy (SPM) with Band Excitation Probe Spectroscopy (BEPS); Neutron Powder Diffraction (NPD) and Perturbed Angular Correlations (PAC). Additional collaboration in research projects outside the scope of multiferroic materials provided further experience in sample preparation and characterization techniques, namely VSM and XPS measurements were performed in cubane molecular complex compounds and enable to identify the oxidation state of the integrating cluster of Ru ions; also, XRD and EDS/SEM analysis of the acquired targets and substrates implied the devolution of some items not in conformity with the specifications. Direct cooperation with parallel research projects regarding multiferroic materials, enable the assess to supplementary samples, namely a preliminary series of nanopowder Y1-x-yCaxØyMn1O3 and of Eu0.8Y0.2MnO3, a series of micropowder composites of LuMnO3 with La0.625Sr0.375MnO3 and of BaTiO3 with hexagonal ferrites; mono and polycrystalline samples of Pr1-xCaxMnO3, La1-xSrxMnO3 and La1-xCaxMnO3.
Resumo:
Therapeutic drug monitoring (TDM) aims to optimize treatments by individualizing dosage regimens based on the measurement of blood concentrations. Dosage individualization to maintain concentrations within a target range requires pharmacokinetic and clinical capabilities. Bayesian calculations currently represent the gold standard TDM approach but require computation assistance. In recent decades computer programs have been developed to assist clinicians in this assignment. The aim of this survey was to assess and compare computer tools designed to support TDM clinical activities. The literature and the Internet were searched to identify software. All programs were tested on personal computers. Each program was scored against a standardized grid covering pharmacokinetic relevance, user friendliness, computing aspects, interfacing and storage. A weighting factor was applied to each criterion of the grid to account for its relative importance. To assess the robustness of the software, six representative clinical vignettes were processed through each of them. Altogether, 12 software tools were identified, tested and ranked, representing a comprehensive review of the available software. Numbers of drugs handled by the software vary widely (from two to 180), and eight programs offer users the possibility of adding new drug models based on population pharmacokinetic analyses. Bayesian computation to predict dosage adaptation from blood concentration (a posteriori adjustment) is performed by ten tools, while nine are also able to propose a priori dosage regimens, based only on individual patient covariates such as age, sex and bodyweight. Among those applying Bayesian calculation, MM-USC*PACK© uses the non-parametric approach. The top two programs emerging from this benchmark were MwPharm© and TCIWorks. Most other programs evaluated had good potential while being less sophisticated or less user friendly. Programs vary in complexity and might not fit all healthcare settings. Each software tool must therefore be regarded with respect to the individual needs of hospitals or clinicians. Programs should be easy and fast for routine activities, including for non-experienced users. Computer-assisted TDM is gaining growing interest and should further improve, especially in terms of information system interfacing, user friendliness, data storage capability and report generation.
Resumo:
This paper presents two studies, both examining the efficacy of a computer programme (Captain's Log) in training attentional skills. The population of interest is the traumatically brain injured. Study #1 is a single-case design that offers recommendations for the second, .larger (N=5) inquiry. Study #2 is an eight-week hierarchical treatment programme with a multi-based testing component. Attention, memory, listening comprehension, locus-of-control, self-esteem, visuo-spatial, and general outcome measures are employed within the testing schedule. Results suggest that any improvement was a result of practice effects. With a few single-case exceptions, the participants showed little improvement in the dependent measures.
Resumo:
This study had three purposes related to the effective implem,entation and practice of computer-mediated online distance education (C-MODE) at the elementary level: (a) To identify a preliminary framework of criteria 'or guidelines for effective implementation and practice, (b) to identify areas ofC-MODE for which criteria or guidelines of effectiveness have not yet been developed, and (c) to develop an implementation and practice criteria questionnaire based on a review of the distance education literature, and to use the questionnaire in an exploratory survey of elementary C-MODE practitioners. Using the survey instrument, the beliefs and attitudes of 16 elementary C'- MODE practitioners about what constitutes effective implementation and practice principles were investigated. Respondents, who included both administrators and instructors, provided information about themselves and the program in which they worked. They rated 101 individual criteria statenlents on a 5 point Likert scale with a \. point range that included the values: 1 (Strongly Disagree), 2 (Disagree), 3 (Neutral or Undecided), 4 (Agree), 5 (Strongly Agree). Respondents also provided qualitative data by commenting on the individual statements, or suggesting other statements they considered important. Eighty-two different statements or guidelines related to the successful implementation and practice of computer-mediated online education at the elementary level were endorsed. Response to a small number of statements differed significantly by gender and years of experience. A new area for investigation, namely, the role ofparents, which has received little attention in the online distance education literature, emerged from the findings. The study also identified a number of other areas within an elementary context where additional research is necessary. These included: (a) differences in the factors that determine learning in a distance education setting and traditional settings, (b) elementary students' ability to function in an online setting, (c) the role and workload of instructors, (d) the importance of effective, timely communication with students and parents, and (e) the use of a variety of media.