947 resultados para REDUCED-ORDER
Resumo:
Durch steigende Energiekosten und erhöhte CO2 Emission ist die Forschung an thermoelektrischen (TE) Materialien in den Fokus gerückt. Die Eignung eines Materials für die Verwendung in einem TE Modul ist verknüpft mit der Gütezahl ZT und entspricht α2σTκ-1 (Seebeck Koeffizient α, Leitfähigkeit σ, Temperatur T und thermische Leitfähigkeit κ). Ohne den Leistungsfaktor α2σ zu verändern, soll ZT durch Senkung der thermischen Leitfähigkeit mittels Nanostrukturierung angehoben werden.rnBis heute sind die TE Eigenschaften von den makroskopischen halb-Heusler Materialen TiNiSn und Zr0.5Hf0.5NiSn ausgiebig erforscht worden. Mit Hilfe von dc Magnetron-Sputterdeposition wurden nun erstmals halbleitende TiNiSn und Zr0.5Hf0.5NiSn Schichten hergestellt. Auf MgO (100) Substraten sind stark texturierte polykristalline Schichten bei Substrattemperaturen von 450°C abgeschieden worden. Senkrecht zur Oberfläche haben sich Korngrößen von 55 nm feststellen lassen. Diese haben Halbwertsbreiten bei Rockingkurven von unter 1° aufgewiesen. Strukturanalysen sind mit Hilfe von Röntgenbeugungsexperimenten (XRD) durchgeführt worden. Durch Wachstumsraten von 1 nms 1 konnten in kürzester Zeit Filmdicken von mehr als einem µm hergestellt werden. TiNiSn zeigte den höchsten Leistungsfaktor von 0.4 mWK 2m 1 (550 K). Zusätzlich wurde bei Raumtemperatur mit Hilfe der differentiellen 3ω Methode eine thermische Leitfähigkeit von 2.8 Wm 1K 1 bestimmt. Es ist bekannt, dass die thermische Leitfähigkeit mit der Variation von Massen abnimmt. Weil zudem angenommen wird, dass sie durch Grenzflächenstreuung von Phononen ebenfalls reduziert wird, wurden Übergitter hergestellt. Dabei wurden TiNiSn und Zr0.5Hf0.5NiSn nacheinander abgeschieden. Die sehr hohe Kristallqualität der Übergitter mit ihren scharfen Grenzflächen konnte durch Satellitenpeaks und Transmissionsmikroskopie (STEM) nachgewiesen werden. Für ein Übergitter mit einer Periodizität von 21 nm (TiNiSn und Zr0.5Hf0.5NiSn jeweils 10.5 nm) ist bei einer Temperatur von 550 K ein Leistungsfaktor von 0.77 mWK 2m 1 nachgewiesen worden (α = 80 µVK 1; σ = 8.2 µΩm). Ein Übergitter mit der Periodizität von 8 nm hat senkrecht zu den Grenzflächen eine thermische Leitfähigkeit von 1 Wm 1K 1 aufgewiesen. Damit hat sich die Reduzierung der thermischen Leitfähigkeit durch die halb-Heusler Übergitter bestätigt. Durch die isoelektronischen Eigenschaften von Titan, Zirkonium und Hafnium wird angenommen, dass die elektrische Bandstruktur und damit der Leistungsfaktor senkrecht zu den Grenzflächen nur schwach beeinflusst wird.rn
Resumo:
In Alzheimer's disease (AD) patients, episodic memory impairments are apparent, yet semantic memory difficulties are also observed. While the episodic pathology has been thoroughly studied, the neurophysiological mechanisms of the semantic impairments remain obscure. Semantic dementia (SD) is characterized by isolated semantic memory deficits. The present study aimed to find an early marker of mild AD and SD by employing a semantic priming paradigm during electroencephalogram recordings. Event-related potentials (ERP) of early (P1, N1) and late (N400) word processing stages were obtained to measure semantic memory functions. Separately, baseline cerebral blood flow (CBF) was acquired with arterial spin labeling. Thus, the analysis focused on linear regressions of CBF with ERP topographical similarity indices in order to find the brain structures that showed altered baseline functionality associated with deviant ERPs. All participant groups showed semantic priming in their reaction times. Furthermore, decreased CBF in the temporal lobes was associated with abnormal N400 topography. No significant CBF clusters were found for the early ERPs. Taken together, the neurophysiological results suggested that the automatic spread of activation during semantic word processing was preserved in mild dementia, while controlled access to the words was impaired. These findings suggested that N400-topography alterations might be a potential marker for the detection of early dementia. Such a marker could be beneficial for differential diagnosis due to its low cost and non-invasive application as well as its relationship with semantic memory dysfunctions that are closely associated to the cortical deterioration in regions crucial for semantic word processing.
Resumo:
OBJECTIVES: Spousal caregivers of Alzheimer's disease patients are at increased risk for cardiovascular disease, possibly via sympathetic response to stressors and subsequent catecholamine surge. Personal mastery (i.e., belief that one can manage life's obstacles) may decrease psychological and physiological response to stressors. This study examines the relationship between mastery and sympathetic arousal in elderly caregivers, as measured by norepinephrine (NE) reactivity to an acute psychological stressor. DESIGN: Cross-sectional. SETTING: Data were collected by a research nurse in each caregiver's home. PARTICIPANTS: Sixty-nine elderly spousal Alzheimer caregivers (mean age: 72.8 years) who were not taking beta-blocking medication. INTERVENTION: After assessment for mastery and objective caregiving stressors, caregivers underwent an experimental speech task designed to induce sympathetic arousal. MEASUREMENTS: Mastery was assessed using Pearlin's Personal Mastery scale and Alzheimer patient functioning was assessed using the Clinical Dementia Rating Scale, Problem Behaviors Scale, and Activities of Daily Living Scale. Plasma NE assays were conducted using pre- and postspeech blood draws. RESULTS: Multiple regression analyses revealed that mastery was significantly and negatively associated with NE reactivity (B = -9.86, t (61) = -2.03, p = 0.046) independent of factors theoretically and empirically linked to NE reactivity. CONCLUSIONS: Caregivers with higher mastery had less NE reactivity to the stressor task. Mastery may exert a protective influence that mitigates the physiological effects of acute stress, and may be an important target for psychosocial interventions in order to reduce sympathetic arousal and cardiovascular stress among dementia caregivers.
Resumo:
An experimental short-term acidification with HCl at a first-order stream in central Maine, USA was used to study processes controlling the changes in stream chemistry and to assess the ability of stream substrate to buffer pH. The streambed exerted a strong buffering capacity against pH change by ion exchange during the 6-hour acidification. Streambed substrates had substantial cation and anion exchange capacity in the pH range of 4.1 to 6.5. The ion exchange for cations and SO42- were rapid and reversible. The speed of release of cations from stream substrates was Na1+ > Ca2+ > Mg2+ > Aln+ > Be2+, perhaps relating to charge density of these cations. Ca2+ desorption dominated neutralisation of excess H+ for the first 2 hr. As the reservoir of exchangeable Ca diminished, desorption land possibly dissolution) of Al3+ became the dominant neutralising mechanism. The exchangeable land possibly soluble) reservoir of Al was not depleted during the 6-hour acidification. Sulphate adsorption during the acidification reduced the concentration of SO42- in stream water by as much as 20 mu eq L-1 (from 70 mu eq L-1). Desorption of SO42- and adsorption of base cat ions after the artificial acidification resulted in a prolongation of the pH depression. The streambed had the capacity to buffer stream water chemistry significantly during an acidifying event affecting the entire upstream catchment.
Resumo:
We study the strength of the electroweak phase transition in models with two light Higgs doublets and a light SU(3)c triplet by means of lattice simulations in a dimensionally reduced effective theory. In the parameter region considered the transition on the lattice is significantly stronger than indicated by a 2-loop perturbative analysis. Within some ultraviolet uncertainties, the finding applies to MSSM with a Higgs mass mh ≈ 126 GeV and shows that the parameter region useful for electroweak baryogenesis is enlarged. In particular (even though only dedicated analyses can quantify the issue), the tension between LHC constraints after the 7 TeV and 8 TeV runs and frameworks where the electroweak phase transition is driven by light stops, seems to be relaxed.
Resumo:
One-dimensional dynamic computer simulation was employed to investigate the separation and migration order change of ketoconazole enantiomers at low pH in presence of increasing amounts of (2-hydroxypropyl)-β-cyclodextrin (OHP-β-CD). The 1:1 interaction of ketoconazole with the neutral cyclodextrin was simulated under real experimental conditions and by varying input parameters for complex mobilities and complexation constants. Simulation results obtained with experimentally determined apparent ionic mobilities, complex mobilities, and complexation constants were found to compare well with the calculated separation selectivity and experimental data. Simulation data revealed that the migration order of the ketoconazole enantiomers at low (OHP-β-CD) concentrations (i.e. below migration order inversion) is essentially determined by the difference in complexation constants and at high (OHP-β-CD) concentrations (i.e. above migration order inversion) by the difference in complex mobilities. Furthermore, simulations with complex mobilities set to zero provided data that mimic migration order and separation with the chiral selector being immobilized. For the studied CEC configuration, no migration order inversion is predicted and separations are shown to be quicker and electrophoretic transport reduced in comparison to migration in free solution. The presented data illustrate that dynamic computer simulation is a valuable tool to study electrokinetic migration and separations of enantiomers in presence of a complexing agent.
Resumo:
Abstraction-Carrying Code (ACC) has recently been proposed as a framework for mobile code safety in which the code supplier provides a program together with an abstraction whose validity entails compliance with a predefined safety policy. The abstraction plays thus the role of safety certifícate and its generation is carried out automatically by a fixed-point analyzer. The advantage of providing a (fixedpoint) abstraction to the code consumer is that its validity is checked in a single pass of an abstract interpretation-based checker. A main challenge is to reduce the size of certificates as much as possible while at the same time not increasing checking time. We introduce the notion of reduced certifícate which characterizes the subset of the abstraction which a checker needs in order to validate (and re-construct) the full certifícate in a single pass. Based on this notion, we instrument a generic analysis algorithm with the necessary extensions in order to identify the information relevant to the checker. We also provide a correct checking algorithm together with sufficient conditions for ensuring its completeness. The experimental results within the CiaoPP system show that our proposal is able to greatly reduce the size of certificates in practice.
Resumo:
Abstraction-Carrying Code (ACC) has recently been proposed as a framework for mobile code safety in which the code supplier provides a program together with an abstraction whose validity entails compliance with a predefined safety policy. The abstraction plays thus the role of safety certifícate and its generation is carried out automatically by a fixed-point analyzer. The advantage of providing a (fixedpoint) abstraction to the code consumer is that its validity is checked in a single pass of an abstract interpretation-based checker. A main challenge is to reduce the size of certificates as much as possible while at the same time not increasing checking time. In this paper, we first introduce the notion of reduced certifícate which characterizes the subset of the abstraction which a checker needs in order to validate (and re-construct) the full certifícate in a single pass. Based on this notion, we then instrument a generic analysis algorithm with the necessary extensions in order to identify the information relevant to the checker.
Resumo:
Abstraction-Carrying Code (ACC) has recently been proposed as a framework for mobile code safety in which the code supplier provides a program together with an abstraction (or abstract model of the program) whose validity entails compliance with a predefined safety policy. The abstraction plays thus the role of safety certifícate and its generation is carried out automatically by a fixed-point analyzer. The advantage of providing a (fixed-point) abstraction to the code consumer is that its validity is checked in a single pass (i.e., one iteration) of an abstract interpretation-based checker. A main challenge to make ACC useful in practice is to reduce the size of certificates as much as possible while at the same time not increasing checking time. The intuitive idea is to only include in the certifícate information that the checker is unable to reproduce without iterating. We introduce the notion of reduced certifícate which characterizes the subset of the abstraction which a checker needs in order to validate (and re-construct) the full certifícate in a single pass. Based on this notion, we instrument a generic analysis algorithm with the necessary extensions in order to identify information which can be reconstructed by the single-pass checker. Finally, we study what the effects of reduced certificates are on the correctness and completeness of the checking process. We provide a correct checking algorithm together with sufficient conditions for ensuring its completeness. Our ideas are illustrated through a running example, implemented in the context of constraint logic programs, which shows that our approach improves state-of-the-art techniques for reducing the size of certificates.
Resumo:
El requerimiento de proveer alta frecuencia de datos en los modernos sistema de comunicación inalámbricos resulta en complejas señales moduladas de radio-frequencia (RF) con un gran ancho de banda y alto ratio pico-promedio (PAPR). Para garantizar la linealidad del comportamiento, los amplificadores lineales de potencia comunes funcionan típicamente entre 4 y 10 dB de back-o_ desde la máxima potencia de salida, ocasionando una baja eficiencia del sistema. La eliminación y restauración de la evolvente (EER) y el seguimiento de la evolvente (ET) son dos prometedoras técnicas para resolver el problema de la eficiencia. Tanto en EER como en ET, es complicado diseñar un amplificador de potencia que sea eficiente para señales de RF de alto ancho de banda y alto PAPR. Una propuesta común para los amplificadores de potencia es incluir un convertidor de potencia de muy alta eficiencia operando a frecuencias más altas que el ancho de banda de la señal RF. En este caso, la potencia perdida del convertidor ocasionado por la alta frecuencia desaconseja su práctica cuando el ancho de banda es muy alto. La solución a este problema es el enfoque de esta disertación que presenta dos arquitecturas de amplificador evolvente: convertidor híbrido-serie con una técnica de evolvente lenta y un convertidor multinivel basado en un convertidor reductor multifase con control de tiempo mínimo. En la primera arquitectura, una topología híbrida está compuesta de una convertidor reductor conmutado y un regulador lineal en serie que trabajan juntos para ajustar la tensión de salida para seguir a la evolvente con precisión. Un algoritmo de generación de una evolvente lenta crea una forma de onda con una pendiente limitada que es menor que la pendiente máxima de la evolvente original. La salida del convertidor reductor sigue esa forma de onda en vez de la evolvente original usando una menor frecuencia de conmutación, porque la forma de onda no sólo tiene una pendiente reducida sino también un menor ancho de banda. De esta forma, el regulador lineal se usa para filtrar la forma de onda tiene una pérdida de potencia adicional. Dependiendo de cuánto se puede reducir la pendiente de la evolvente para producir la forma de onda, existe un trade-off entre la pérdida de potencia del convertidor reductor relacionada con la frecuencia de conmutación y el regulador lineal. El punto óptimo referido a la menor pérdida de potencia total del amplificador de evolvente es capaz de identificarse con la ayuda de modelo preciso de pérdidas que es una combinación de modelos comportamentales y analíticos de pérdidas. Además, se analiza el efecto en la respuesta del filtro de salida del convertidor reductor. Un filtro de dampeo paralelo extra es necesario para eliminar la oscilación resonante del filtro de salida porque el convertidor reductor opera en lazo abierto. La segunda arquitectura es un amplificador de evolvente de seguimiento de tensión multinivel. Al contrario que los convertidores que usan multi-fuentes, un convertidor reductor multifase se emplea para generar la tensión multinivel. En régimen permanente, el convertidor reductor opera en puntos del ciclo de trabajo con cancelación completa del rizado. El número de niveles de tensión es igual al número de fases de acuerdo a las características del entrelazamiento del convertidor reductor. En la transición, un control de tiempo mínimo (MTC) para convertidores multifase es novedosamente propuesto y desarrollado para cambiar la tensión de salida del convertidor reductor entre diferentes niveles. A diferencia de controles convencionales de tiempo mínimo para convertidores multifase con inductancia equivalente, el propuesto MTC considera el rizado de corriente por cada fase basado en un desfase fijo que resulta en diferentes esquemas de control entre las fases. La ventaja de este control es que todas las corrientes vuelven a su fase en régimen permanente después de la transición para que la siguiente transición pueda empezar muy pronto, lo que es muy favorable para la aplicación de seguimiento de tensión multinivel. Además, el control es independiente de la carga y no es afectado por corrientes de fase desbalanceadas. Al igual que en la primera arquitectura, hay una etapa lineal con la misma función, conectada en serie con el convertidor reductor multifase. Dado que tanto el régimen permanente como el estado de transición del convertidor no están fuertemente relacionados con la frecuencia de conmutación, la frecuencia de conmutación puede ser reducida para el alto ancho de banda de la evolvente, la cual es la principal consideración de esta arquitectura. La optimización de la segunda arquitectura para más alto anchos de banda de la evolvente es presentada incluyendo el diseño del filtro de salida, la frecuencia de conmutación y el número de fases. El área de diseño del filtro está restringido por la transición rápida y el mínimo pulso del hardware. La rápida transición necesita un filtro pequeño pero la limitación del pulso mínimo del hardware lleva el diseño en el sentido contrario. La frecuencia de conmutación del convertidor afecta principalmente a la limitación del mínimo pulso y a las pérdidas de potencia. Con una menor frecuencia de conmutación, el ancho de pulso en la transición es más pequeño. El número de fases relativo a la aplicación específica puede ser optimizado en términos de la eficiencia global. Otro aspecto de la optimización es mejorar la estrategia de control. La transición permite seguir algunas partes de la evolvente que son más rápidas de lo que el hardware puede soportar al precio de complejidad. El nuevo método de sincronización de la transición incrementa la frecuencia de la transición, permitiendo que la tensión multinivel esté más cerca de la evolvente. Ambas estrategias permiten que el convertidor pueda seguir una evolvente con un ancho de banda más alto que la limitación de la etapa de potencia. El modelo de pérdidas del amplificador de evolvente se ha detallado y validado mediante medidas. El mecanismo de pérdidas de potencia del convertidor reductor tiene que incluir las transiciones en tiempo real, lo cual es diferente del clásico modelos de pérdidas de un convertidor reductor síncrono. Este modelo estima la eficiencia del sistema y juega un papel muy importante en el proceso de optimización. Finalmente, la segunda arquitectura del amplificador de evolvente se integra con el amplificador de clase F. La medida del sistema EER prueba el ahorro de energía con el amplificador de evolvente propuesto sin perjudicar la linealidad del sistema. ABSTRACT The requirement of delivering high data rates in modern wireless communication systems results in complex modulated RF signals with wide bandwidth and high peak-to-average ratio (PAPR). In order to guarantee the linearity performance, the conventional linear power amplifiers typically work at 4 to 10 dB back-off from the maximum output power, leading to low system efficiency. The envelope elimination and restoration (EER) and envelope tracking (ET) are two promising techniques to overcome the efficiency problem. In both EER and ET, it is challenging to design efficient envelope amplifier for wide bandwidth and high PAPR RF signals. An usual approach for envelope amplifier includes a high-efficiency switching power converter operating at a frequency higher than the RF signal's bandwidth. In this case, the power loss of converter caused by high switching operation becomes unbearable for system efficiency when signal bandwidth is very wide. The solution of this problem is the focus of this dissertation that presents two architectures of envelope amplifier: a hybrid series converter with slow-envelope technique and a multilevel converter based on a multiphase buck converter with the minimum time control. In the first architecture, a hybrid topology is composed of a switched buck converter and a linear regulator in series that work together to adjust the output voltage to track the envelope with accuracy. A slow envelope generation algorithm yields a waveform with limited slew rate that is lower than the maximum slew rate of the original envelope. The buck converter's output follows this waveform instead of the original envelope using lower switching frequency, because the waveform has not only reduced slew rate but also reduced bandwidth. In this way, the linear regulator used to filter the waveform has additional power loss. Depending on how much reduction of the slew rate of envelope in order to obtain that waveform, there is a trade-off between the power loss of buck converter related to the switching frequency and the power loss of linear regulator. The optimal point referring to the lowest total power loss of this envelope amplifier is identified with the help of a precise power loss model that is a combination of behavioral and analytic loss model. In addition, the output filter's effect on the response is analyzed. An extra parallel damping filter is needed to eliminate the resonant oscillation of output filter L and C, because the buck converter operates in open loop. The second architecture is a multilevel voltage tracking envelope amplifier. Unlike the converters using multi-sources, a multiphase buck converter is employed to generate the multilevel voltage. In the steady state, the buck converter operates at complete ripple cancellation points of duty cycle. The number of the voltage levels is equal to the number of phases according the characteristics of interleaved buck converter. In the transition, a minimum time control (MTC) for multiphase converter is originally proposed and developed for changing the output voltage of buck converter between different levels. As opposed to conventional minimum time control for multiphase converter with equivalent inductance, the proposed MTC considers the current ripple of each phase based on the fixed phase shift resulting in different control schemes among the phases. The advantage of this control is that all the phase current return to the steady state after the transition so that the next transition can be triggered very soon, which is very favorable for the application of multilevel voltage tracking. Besides, the control is independent on the load condition and not affected by the unbalance of phase current. Like the first architecture, there is also a linear stage with the same function, connected in series with the multiphase buck converter. Since both steady state and transition state of the converter are not strongly related to the switching frequency, it can be reduced for wide bandwidth envelope which is the main consideration of this architecture. The optimization of the second architecture for wider bandwidth envelope is presented including the output filter design, switching frequency and the number of phases. The filter design area is restrained by fast transition and the minimum pulse of hardware. The fast transition needs small filter but the minimum pulse of hardware limitation pushes the filter in opposite way. The converter switching frequency mainly affects the minimum pulse limitation and the power loss. With lower switching frequency, the pulse width in the transition is smaller. The number of phases related to specific application can be optimized in terms of overall efficiency. Another aspect of optimization is improving control strategy. Transition shift allows tracking some parts of envelope that are faster than the hardware can support at the price of complexity. The new transition synchronization method increases the frequency of transition, allowing the multilevel voltage to be closer to the envelope. Both control strategies push the converter to track wider bandwidth envelope than the limitation of power stage. The power loss model of envelope amplifier is detailed and validated by measurements. The power loss mechanism of buck converter has to include the transitions in real time operation, which is different from classical power loss model of synchronous buck converter. This model estimates the system efficiency and play a very important role in optimization process. Finally, the second envelope amplifier architecture is integrated with a Class F amplifier. EER system measurement proves the power saving with the proposed envelope amplifier without disrupting the linearity performance.
Resumo:
Nonlinear analysis tools for studying and characterizing the dynamics of physiological signals have gained popularity, mainly because tracking sudden alterations of the inherent complexity of biological processes might be an indicator of altered physiological states. Typically, in order to perform an analysis with such tools, the physiological variables that describe the biological process under study are used to reconstruct the underlying dynamics of the biological processes. For that goal, a procedure called time-delay or uniform embedding is usually employed. Nonetheless, there is evidence of its inability for dealing with non-stationary signals, as those recorded from many physiological processes. To handle with such a drawback, this paper evaluates the utility of non-conventional time series reconstruction procedures based on non uniform embedding, applying them to automatic pattern recognition tasks. The paper compares a state of the art non uniform approach with a novel scheme which fuses embedding and feature selection at once, searching for better reconstructions of the dynamics of the system. Moreover, results are also compared with two classic uniform embedding techniques. Thus, the goal is comparing uniform and non uniform reconstruction techniques, including the one proposed in this work, for pattern recognition in biomedical signal processing tasks. Once the state space is reconstructed, the scheme followed characterizes with three classic nonlinear dynamic features (Largest Lyapunov Exponent, Correlation Dimension and Recurrence Period Density Entropy), while classification is carried out by means of a simple k-nn classifier. In order to test its generalization capabilities, the approach was tested with three different physiological databases (Speech Pathologies, Epilepsy and Heart Murmurs). In terms of the accuracy obtained to automatically detect the presence of pathologies, and for the three types of biosignals analyzed, the non uniform techniques used in this work lightly outperformed the results obtained using the uniform methods, suggesting their usefulness to characterize non-stationary biomedical signals in pattern recognition applications. On the other hand, in view of the results obtained and its low computational load, the proposed technique suggests its applicability for the applications under study.
Resumo:
Loss of genetic diversity and increased population differentiation from source populations are common problems associated with translocation programmes established from captive-bred stock or a small number of founders. The bridled nailtail wallaby is one of the most endangered macropods in Australia, having been reduced to a single remnant population in the last 100 years. A translocated population of bridled nailtail wallabies was established using animals sourced directly from the remnant population (wild-released) as well as the progeny of animals collected for a captive breeding programme (captive-bred). The aims of this study were to compare genetic diversity among released animals and their wild-born progeny to genetic diversity observed in the remnant population, and to monitor changes in genetic diversity over time as more animals were released into the population. Heterozygosity did not differ between the translocated and remnant population; however, allelic diversity was significantly reduced across all released animals and their wild-born progeny. Animals bred in captivity and their wild-born progeny were also significantly differentiated from the source population after just four generations. Wild-released animals, however, were representative of the source population and several alleles were unique to this group. Both heterozygosity and allelic diversity among translocated animals decreased over time with the additional release of captive-bred animals, as no new genetic stock was added to the population. Captive breeding programmes can provide large numbers of animals for release, but this study highlights the importance of sourcing animals directly from remnant populations in order to maintain genetic diversity and minimise genetic drift.
Resumo:
The present thesis tested the hypothesis of Stanovich, Siegel, & Gottardo (1997) that surface dyslexia is the result of a milder phonological deficit than that seen in phonological dyslexia coupled with reduced reading experience. We found that a group of adults with surface dyslexia showed a phonological deficit that was commensurate with that shown by a group of adults with phonological dyslexia (matched for chronological age and verbal and non-verbal IQ) and normal reading experience. We also showed that surface dyslexia cannot be accounted for by a semantic impairment or a deficit in the verbal learning and recall of lexical-semantic information (such as meaningful words), as both dyslexic subgroups performed the same. This study has replicated the results of our published study that surface dyslexia is not the consequence of a mild retardation or reduced learning opportunities but a separate impairment linked to a deficit in written lexical learning, an ability needed to create novel lexical representations from a series of unrelated visual units, which is independent from the phonological deficit (Romani, Di Betta, Tsouknida & Olson, 2008). This thesis also provided evidence that a selective nonword reading deficit in developmental dyslexia persists beyond poor phonology. This was shown by finding a nonword reading deficit even in the presence of normal regularity effects in the dyslexics (when compared to both reading and spelling-age matched controls). A nonword reading deficit was also found in the surface dyslexics. Crucially, this deficit was as strong as in the phonological dyslexics despite better functioning of the sublexical route for the former. These results suggest that a nonword reading deficit cannot be solely explained by a phonological impairment. We, thus, suggested that nonword reading should also involve another ability relating to the processing of novel visual orthographic strings, which we called 'orthographic coding'. We then investigated the ability to process series of independent units within multi-element visual arrays and its relationship with reading and spelling problems. We identified a deficit in encoding the order of visual sequences (involving both linguistic and nonlinguistic information) which was significantly associated with word and nonword processing. More importantly, we revealed significant contributions to orthographic skills in both dyslexic and control individuals, even after age, performance IQ and phonological skills were controlled. These results suggest that spelling and reading do not only tap phonological skills but also order encoding skills.
Resumo:
Purpose - We performed a study of laser panretinal photocoagulation in 20 patients with proliferative retinopathy. We compared short exposure, high-energy laser settings with conventional settings, using a 532?nm, frequency doubled, Neodymium–Yag laser and assessed the patients in terms of pain experienced and effectiveness of treatment. Methods - Twenty patients having panretinal photocoagulation for the first time underwent random allocation to treatment of the superior and inferior hemi-retina. Treatment A used ‘conventional’ parameters: exposure time 0.1?s, power sufficient to produce a visible grey-white burns, spot size 300?µm. The other hemi- retina was treated with treatment B using exposure 0.02?s, 300?µm and sufficient power to have similar endpoint. All patients were asked to evaluate severity of pain on a visual analogue scale. (0=no pain, 10=most severe pain). All patients were masked as to the type of treatment and the order of carrying out the treatment on each patient was randomised. Patients underwent fundus photography and were followed up for 6–45 months. Results - Seventeen patients had proliferative diabetic retinopathy, two had ischaemic central retinal vein occlusion and one had ocular ischaemic syndrome. The mean response to treatment A was 5.11, compared to 1.40 treatment B, on the visual analogue scale, which was statistically significant (P=0.001). All patients preferred treatment B. Further treatments, if required, were performed with treatment B parameters and long-term follow-up has shown no evidence of undertreatment. Conclusions - Shortening exposure time of retinal laser is significantly less painful but equally effective as conventional parameters.
Resumo:
We report the performance of a group of adult dyslexics and matched controls in an array-matching task where two strings of either consonants or symbols are presented side by side and have to be judged to be the same or different. The arrays may differ either in the order or identity of two adjacent characters. This task does not require naming – which has been argued to be the cause of dyslexics’ difficulty in processing visual arrays – but, instead, has a strong serial component as demonstrated by the fact that, in both groups, Reaction times (RTs) increase monotonically with position of a mismatch. The dyslexics are clearly impaired in all conditions and performance in the identity conditions predicts performance across orthographic tasks even after age, performance IQ and phonology are partialled out. Moreover, the shapes of serial position curves are revealing of the underlying impairment. In the dyslexics, RTs increase with position at the same rate as in the controls (lines are parallel) ruling out reduced processing speed or difficulties in shifting attention. Instead, error rates show a catastrophic increase for positions which are either searched later or more subject to interference. These results are consistent with a reduction in the attentional capacity needed in a serial task to bind together identity and positional information. This capacity is best seen as a reduction in the number of spotlights into which attention can be split to process information at different locations rather than as a more generic reduction of resources which would also affect processing the details of single objects.