989 resultados para algorithm development


Relevância:

60.00% 60.00%

Publicador:

Resumo:

La medicina y la ingeniería del siglo XXI han dado como fruto numerosos avances para la sociedad aunque en la mayoría de los casos los tratamientos suelen ser costosos e invasivos. La educación que recibe la sociedad sobre la salud es escasa, ya que sólo vamos al médico cuando realmente estamos enfermos. Este trabajo presenta nuestra apuesta por las terapias complementarias, para el desarrollo de una metodología terapéutica no invasiva y con un costo muy bajo. La finalidad de esta Tesis, que se enmarca en un equipo multidisciplinar, fruto de la estrecha colaboración en el que participan psicopedagogos, ingenieros y médicos, es perfilar una metodología que luego pueda ser aplicable a patologías neurológicas. Aquí, dejamos sentadas las bases. Faltarán nuevos investigadores que continúen este camino para tener una base de datos lo suficientemente extensa de registros de sujetos que hayan sido sometidos a terapia binaural, para poder sacar unas conclusiones sólidas. La aportación de esta Tesis deja cubierta la aplicación, selección, procesado de señal y desarrollo de algoritmos, test cognitivos indicados para el caso específico que nos ocupa, cálculo de incertidumbre del sistema utilizado para la aplicación del estímulo y desarrollo de un test psicoacústico específico. EL empleo del sonido en medicina como es la musicoterapia o sonoterapia ha experimentado una gran difusión en los últimos años, más de 100.000 nuevas citas bibliográficas han aparecido con respecto al año anterior. Sin embargo, son escasísimas las que hacen referencia a las características físico acústicas del sonido empleado, tan sólo hemos encontrado una par de ellas que correlacionan las características físicas del sonido con el tipo de respuesta terapéutica. No encontramos citas bibliográficas específicas que planteen un modelo experimental científico capaz de reproducir las mismas respuestas ante los mismos parámetros y estímulos. En esta Tesis proponemos el uso de estimulación sonora binaural, que consiste en la utilización de dos tonos puros idénticos pero ligeramente diferentes en frecuencia que se presentan de manera separada cada uno en un oído, como consecuencia, la persona que recibe la estimulación percibe un tercer tono, llamado tono binaural, formado por la diferencia de frecuencia de ambos variando su amplitud. Existen estudios que sugieren que dichas frecuencias binaurales pueden modificar los patrones eléctricos de la actividad cerebral y los niveles de arousal, conociéndose en la literatura bajo el nombre de “entrainment”. Tras la revisión bibliográfica del estado del arte, podemos concluir que es necesario el desarrollo de estudios doble ciego bien diseñados, con el objetivo de establecer una base sólida sobre los efectos de este tipo de estimulación, ya que la mayoría de los beneficios documentados se refieren a muestras muy pequeñas y con poco rigor científico, siendo los resultados positivos obtenidos debidos al efecto placebo. La tecnología binaural es barata siendo cualquier avance en esta dirección de interés público. El objetivo concreto de la investigación es estudiar el potencial de las ondas binaurales en un área en particular: tareas que requieren atención y concentración. Se busca obtener cualquier cambio en las ondas cerebrales que se puedan correlar con la mejoras. A la vista de los resultados de estas investigaciones se intentará aplicar esta metodología en neuropatologías que presenten alguna deficiencia en el área de atención como es el Trastorno de espectro Autista. En esta Tesis presentamos los resultados de dos estudios independientes, el primero para sentar las bases del método (tiempos, diseño de estimulaciones, procesado) en una muestra de 78 adultos sanos, el segundo a partir de los resultados obtenidos en el primero, afinando la metodología y para un grupo de 20 niños entre 8 y 12 años, los resultados del segundo estudio sirven para justificar su aplicación en niños con TEA que presenten déficit de atención. ABSTRACT Medicine and engineering in the 21st century have resulted in advances for society but in most cases the treatments are often costly and invasive. The health education society receive is scarce, since only go to the doctor when we are really sick. With this work I present my commitment to complementary therapies, my little grain of sand in the development of a noninvasive therapeutic approach and very low cost, well and can be used in a preventive manner resulting in a society with less sick. The purpose of this thesis is to outline a methodology that can then be applied to neurological diseases, here we lay the groundwork. New researchers are needed to continue this path for a sufficiently extensive records database of subjects who have undergone binaural therapy, and so to draw firm conclusions. The contribution of this thesis includes: the application, selection, signal processing and algorithm development, indicated cognitive tests for the specific case at hand, calculation of system uncertainty of the system and development of a specific psychoacoustic test. The use of sound in medicine, such as music therapy or sound therapy has experienced a great diffusion in recent years, more than 100,000 new citations have appeared over the previous year but very few are those referring to acoustic physical characteristics of sound employee, we have only found a couple of them that physical sound characteristics are correlated with the therapeutic response. We found no specific citations posing a scientific experimental model capable of reproducing the same answers to the same parameters and stimuli. In this thesis we propose the use of binaural sound stimulation which involves the use of two identical but slightly different in frequency pure tones presented separately each in one ear, as a result the subject perceives a third tone, called binaural tone, formed by the difference in frequency with amplitude variations Studies suggest that these binaural frequencies can modify the electrical patterns of brain activity and arousal levels, being known in the literature under the name of “entrainment”. After the literature review of the state of the art, we conclude, it is necessary to develop well-designed double-blind studies, in order to establish a solid foundation on the effects of such stimulation, since most of the documented benefits relate to very small samples and unscientific may be obtained positive results due to the placebo effect. The binaural technology is cheap being any progress in this direction in the public interest. The specific objective of the research is to study the potential of binaural waves in a particular area: tasks requiring attention and concentration also we want to get any change in brain waves that can correlate with improvements. In view of the results of this research we seek to apply this methodology in neuropathology presenting any deficiency in the area of attention such as Autism Spectrum Disorder. In this thesis we present the results of two independent studies, the first to lay the foundation of the method (times, stimulation design, processing) in a sample of 78 healthy adults, the second from the results obtained in the first, refine the methodology for a group of 20 children between 8 and 12 years, the results of the second study used to justify its use in children with ASD that present attention deficit.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

El transporte aéreo constituye un sector estratégico para el crecimiento económico de cualquier país. El sistema de gestión de tráfico aéreo ATM tiene como objetivo el movimiento seguro y eficiente de las aeronaves dentro del espacio aéreo y de los aeropuertos, siendo la seguridad, en la fase táctica, gestionada por el servicio de control de la circulación aérea. Mediante los procesos de control el tráfico aéreo es vigilado a través de sensores, regulado y guiado de forma organizada y segura. Es precisamente sobre la vigilancia donde se enfoca el contenido de la tesis, en el desarrollo de nuevos conceptos que proporcionen información de vigilancia de ‘bajo coste’ basados en las señales existentes proporcionadas por la infraestructura actual de radar secundario y por los sistemas de posicionamiento basados en satélite que utiliza la ADS-B. El conocimiento y acceso en tiempo real a las trayectorias de las aeronaves es un elemento de valor añadido no sólo para la provisión de los servicios de control de tránsito aéreo, sino para todos los actores del transporte aéreo o de la investigación, siendo uno de los elementos clave en el concepto operacional de los dos grandes proyectos tecnológicos, SESAR en Europa y NextGen en EE.UU.. En las últimas décadas el control de la circulación aérea en espacios aéreos de media y alta densidad de tráfico se ha basado en tecnologías complejas que requieren importantes infraestructuras como son el radar primario de vigilancia (PSR) y el radar secundario de vigilancia (SSR). La filosofía de los programas SESAR y NextGen siguiendo las directrices de la OACI es la de alejarse de las tecnologías basadas en tierra para evolucionar hacia nuevas tecnologías más dinámicas basadas en satélite como la ADS-B. Pero hasta que la implementación y operación de la ADS-B sea completa, existirá un período de transición que implica la coexistencia de aeronaves equipadas o no con ADS-B. El objetivo de la presente Tesis es determinar las metodologías y algoritmos más adecuados para poder hibridar las dos tecnologías descritas anteriormente, utilizando para ello un receptor de bajo coste con antena estática omnidireccional, que analice todas las señales presentes en el canal que comparten el SSR y ADS-B. Mediante esta hibridación se podrá obtener la posición de cualquier aeronave que transmita respuestas a interrogaciones SSR, en cualquiera de sus modos de trabajo, o directamente mensajes de posición ADS-B. Para desarrollar los algoritmos propuestos, además del hardware correspondiente, se han utilizado las aplicaciones LabVIEW para funciones de adquisición de datos reales, y el software MATLAB® para el desarrollo de algoritmos y análisis de datos. La validación de resultados se ha realizado mediante los propios mensajes de posición ADS-B y a través de las trazas radar proporcionadas por la entidad pública empresarial ENAIRE. La técnica desarrollada es autónoma, y no ha requerido de ninguna otra entrada que no sea la recepción omnidireccional de las señales. Sin embargo para la validación de resultados se ha utilizado información pública de las ubicaciones de la red de estaciones SSR desplegadas sobre territorio español y portugués y trazas radar. Los resultados obtenidos demuestran, que con técnicas basadas en superficies de situación definidas por los tiempos de llegada de las respuestas, es posible determinar con una precisión aceptable la posición de las estaciones SSR y la posición de cualquier aeronave que responda mediante el Modo A a éstas. ABSTRACT Air transport is a strategic sector for the economic growth of any country. The air traffic management system (ATM) aims at the safe and efficient movement of aircraft while operating within the airspace and airports, where safety, in the tactical phase, is managed by the air traffic control services. Through the air traffic control processes, aircraft are monitored by sensors, regulated and guided in an organized and safe manner. It is precisely on surveillance where this thesis is focused, developing new concepts that provide a 'low cost' surveillance information based on existing signals provided by currently secondary radar infrastructure and satellite-based positioning systems used by ADS-B. Having a deeper knowledge and a real-time access to the trajectories of the aircraft, is an element of added value not only for the provision of air traffic control services, but also for all air transport or research actors. This is one of the key elements in the operational concept proposed by the two large scale existing technological projects, SESAR in Europe and NextGen in the US. In recent decades, air traffic control in medium and high traffic density areas has been based on complex technologies requiring major infrastructures, such as the primary surveillance radar (PSR) and secondary surveillance radar (SSR). The philosophy of SESAR and NextGen programs, both following the guidelines of ICAO, is to move away from land-based technologies and evolving into some new and more dynamic satellite-based technologies such as ADS-B. Nevertheless, until the ADS-B implementation and operation is fully achieved, there will be a transitional period where aircraft with and without ADS-B equipment will have to coexist. The main objective of this thesis is to determine those methodologies and algorithms which are considered more appropriate to hybridize those two technologies, by using a low cost omnidirectional receiver, which analyzes all signals on the SSR and ADS-B shared channel. Through this hybridization, it is possible to obtain the position of any aircraft answering the SSR interrogations, in any of its modes of operation, or through the emission of ADS-B messages. To develop the proposed algorithms, LabVIEW application has been used for real-time data acquisition, as well as MATLAB software for algorithm development and data analysis, together with the corresponding hardware. The validation of results was performed using the ADS-B position messages and radar tracks provided by the Public Corporate Entity ENAIRE The developed technique is autonomous, and it does not require any other input other than the omnidirectional signal reception. However, for the validation of results, not only radar records have been used, but also public information regarding the position of SSR stations spread throughout the Spanish and Portuguese territory. The results show that using techniques based in the definition of positioning surfaces defined by the responses’ times of arrival, it is possible to determine with an acceptable level of accuracy both the position of the SSR stations as well as the position of any aircraft which transmits Mode A responses.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Time series of brightness temperatures (T(B)) from the Advanced Microwave Scanning Radiometer-Earth Observing System (AMSR-E) are examined to determine ice phenology variables on the two largest lakes of northern Canada: Great Bear Lake (GBL) and Great Slave Lake (GSL). T(B) measurements from the 18.7, 23.8, 36.5, and 89.0 GHz channels (H- and V- polarization) are compared to assess their potential for detecting freeze-onset/melt-onset and ice-on/ice-off dates on both lakes. The 18.7 GHz (H-pol) channel is found to be the most suitable for estimating these ice dates as well as the duration of the ice cover and ice-free seasons. A new algorithm is proposed using this channel and applied to map all ice phenology variables on GBL and GSL over seven ice seasons (2002-2009). Analysis of the spatio-temporal patterns of each variable at the pixel level reveals that: (1) both freeze-onset and ice-on dates occur on average about one week earlier on GBL than on GSL (Day of Year (DY) 318 and 333 for GBL; DY 328 and 343 for GSL); (2) the freeze-up process or freeze duration (freeze-onset to ice-on) takes a slightly longer amount of time on GBL than on GSL (about 1 week on average); (3) melt-onset and ice-off dates occur on average one week and approximately four weeks later, respectively, on GBL (DY 143 and 183 for GBL; DY 135 and 157 for GSL); (4) the break-up process or melt duration (melt-onset to ice-off) lasts on average about three weeks longer on GBL; and (5) ice cover duration estimated from each individual pixel is on average about three weeks longer on GBL compared to its more southern counterpart, GSL. A comparison of dates for several ice phenology variables derived from other satellite remote sensing products (e.g. NOAA Interactive Multisensor Snow and Ice Mapping System (IMS), QuikSCAT, and Canadian Ice Service Database) show that, despite its relatively coarse spatial resolution, AMSR-E 18.7 GHz provides a viable means for monitoring of ice phenology on large northern lakes.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Thesis (Ph.D.)--University of Washington, 2016-08

Relevância:

60.00% 60.00%

Publicador:

Resumo:

The only method used to date to measure dissolved nitrate concentration (NITRATE) with sensors mounted on profiling floats is based on the absorption of light at ultraviolet wavelengths by nitrate ion (Johnson and Coletti, 2002; Johnson et al., 2010; 2013; D’Ortenzio et al., 2012). Nitrate has a modest UV absorption band with a peak near 210 nm, which overlaps with the stronger absorption band of bromide, which has a peak near 200 nm. In addition, there is a much weaker absorption due to dissolved organic matter and light scattering by particles (Ogura and Hanya, 1966). The UV spectrum thus consists of three components, bromide, nitrate and a background due to organics and particles. The background also includes thermal effects on the instrument and slow drift. All of these latter effects (organics, particles, thermal effects and drift) tend to be smooth spectra that combine to form an absorption spectrum that is linear in wavelength over relatively short wavelength spans. If the light absorption spectrum is measured in the wavelength range around 217 to 240 nm (the exact range is a bit of a decision by the operator), then the nitrate concentration can be determined. Two different instruments based on the same optical principles are in use for this purpose. The In Situ Ultraviolet Spectrophotometer (ISUS) built at MBARI or at Satlantic has been mounted inside the pressure hull of a Teledyne/Webb Research APEX and NKE Provor profiling floats and the optics penetrate through the upper end cap into the water. The Satlantic Submersible Ultraviolet Nitrate Analyzer (SUNA) is placed on the outside of APEX, Provor, and Navis profiling floats in its own pressure housing and is connected to the float through an underwater cable that provides power and communications. Power, communications between the float controller and the sensor, and data processing requirements are essentially the same for both ISUS and SUNA. There are several possible algorithms that can be used for the deconvolution of nitrate concentration from the observed UV absorption spectrum (Johnson and Coletti, 2002; Arai et al., 2008; Sakamoto et al., 2009; Zielinski et al., 2011). In addition, the default algorithm that is available in Satlantic sensors is a proprietary approach, but this is not generally used on profiling floats. There are some tradeoffs in every approach. To date almost all nitrate sensors on profiling floats have used the Temperature Compensated Salinity Subtracted (TCSS) algorithm developed by Sakamoto et al. (2009), and this document focuses on that method. It is likely that there will be further algorithm development and it is necessary that the data systems clearly identify the algorithm that is used. It is also desirable that the data system allow for recalculation of prior data sets using new algorithms. To accomplish this, the float must report not just the computed nitrate, but the observed light intensity. Then, the rule to obtain only one NITRATE parameter is, if the spectrum is present then, the NITRATE should be recalculated from the spectrum while the computation of nitrate concentration can also generate useful diagnostics of data quality.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Executing a cloud or aerosol physical properties retrieval algorithm from controlled synthetic data is an important step in retrieval algorithm development. Synthetic data can help answer questions about the sensitivity and performance of the algorithm or aid in determining how an existing retrieval algorithm may perform with a planned sensor. Synthetic data can also help in solving issues that may have surfaced in the retrieval results. Synthetic data become very important when other validation methods, such as field campaigns,are of limited scope. These tend to be of relatively short duration and often are costly. Ground stations have limited spatial coverage whilesynthetic data can cover large spatial and temporal scales and a wide variety of conditions at a low cost. In this work I develop an advanced cloud and aerosol retrieval simulator for the MODIS instrument, also known as Multi-sensor Cloud and Aerosol Retrieval Simulator (MCARS). In a close collaboration with the modeling community I have seamlessly combined the GEOS-5 global climate model with the DISORT radiative transfer code, widely used by the remote sensing community, with the observations from the MODIS instrument to create the simulator. With the MCARS simulator it was then possible to solve the long standing issue with the MODIS aerosol optical depth retrievals that had a low bias for smoke aerosols. MODIS aerosol retrieval did not account for effects of humidity on smoke aerosols. The MCARS simulator also revealed an issue that has not been recognized previously, namely,the value of fine mode fraction could create a linear dependence between retrieved aerosol optical depth and land surface reflectance. MCARS provided the ability to examine aerosol retrievals against “ground truth” for hundreds of thousands of simultaneous samples for an area covered by only three AERONET ground stations. Findings from MCARS are already being used to improve the performance of operational MODIS aerosol properties retrieval algorithms. The modeling community will use the MCARS data to create new parameterizations for aerosol properties as a function of properties of the atmospheric column and gain the ability to correct any assimilated retrieval data that may display similar dependencies in comparisons with ground measurements.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Prokaryotic organisms are one of the most successful forms of life, they are present in all known ecosystems. The deluge diversity of bacteria reflects their ability to colonise every environment. Also, human beings host trillions of microorganisms in their body districts, including skin, mucosae, and gut. This symbiosis is active for all other terrestrial and marine animals, as well as plants. With the term holobiont we refer, with a single word, to the systems including both the host and its symbiotic microbial species. The coevolution of bacteria within their ecological niches reflects the adaptation of both host and guest species, and it is shaped by complex interactions that are pivotal for determining the host state. Nowadays, thanks to the current sequencing technologies, Next Generation Sequencing, we have unprecedented tools for investigating the bacterial life by studying the prokaryotic genome sequences. NGS revolution has been sustained by the advancements in computational performance, in terms of speed, storage capacity, algorithm development and hardware costs decreasing following the Moore’s Law. Bioinformaticians and computational biologists design and implement ad hoc tools able to analyse high-throughput data and extract valuable biological information. Metagenomics requires the integration of life and computational sciences and it is uncovering the deluge diversity of the bacterial world. The present thesis work focuses mainly on the analysis of prokaryotic genomes under different aspects. Being supervised by two groups at the University of Bologna, the Biocomputing group and the group of Microbial Ecology of Health, I investigated three different topics: i) antimicrobial resistance, particularly with respect to missense point mutations involved in the resistant phenotype, ii) bacterial mechanisms involved in xenobiotic degradation via the computational analysis of metagenomic samples, and iii) the variation of the human gut microbiota through ageing, in elderly and longevous individuals.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Descriptors based on Molecular Interaction Fields (MIF) are highly suitable for drug discovery, but their size (thousands of variables) often limits their application in practice. Here we describe a simple and fast computational method that extracts from a MIF a handful of highly informative points (hot spots) which summarize the most relevant information. The method was specifically developed for drug discovery, is fast, and does not require human supervision, being suitable for its application on very large series of compounds. The quality of the results has been tested by running the method on the ligand structure of a large number of ligand-receptor complexes and then comparing the position of the selected hot spots with actual atoms of the receptor. As an additional test, the hot spots obtained with the novel method were used to obtain GRIND-like molecular descriptors which were compared with the original GRIND. In both cases the results show that the novel method is highly suitable for describing ligand-receptor interactions and compares favorably with other state-of-the-art methods.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

The primary goal of this project is to demonstrate the accuracy and utility of a freezing drizzle algorithm that can be implemented on roadway environmental sensing systems (ESSs). The types of problems related to the occurrence of freezing precipitation range from simple traffic delays to major accidents that involve fatalities. Freezing drizzle can also lead to economic impacts in communities with lost work hours, vehicular damage, and downed power lines. There are means for transportation agencies to perform preventive and reactive treatments to roadways, but freezing drizzle can be difficult to forecast accurately or even detect as weather radar and surface observation networks poorly observe these conditions. The detection of freezing precipitation is problematic and requires special instrumentation and analysis. The Federal Aviation Administration (FAA) development of aircraft anti-icing and deicing technologies has led to the development of a freezing drizzle algorithm that utilizes air temperature data and a specialized sensor capable of detecting ice accretion. However, at present, roadway ESSs are not capable of reporting freezing drizzle. This study investigates the use of the methods developed for the FAA and the National Weather Service (NWS) within a roadway environment to detect the occurrence of freezing drizzle using a combination of icing detection equipment and available ESS sensors. The work performed in this study incorporated the algorithm developed initially and further modified for work with the FAA for aircraft icing. The freezing drizzle algorithm developed for the FAA was applied using data from standard roadway ESSs. The work performed in this study lays the foundation for addressing the central question of interest to winter maintenance professionals as to whether it is possible to use roadside freezing precipitation detection (e.g., icing detection) sensors to determine the occurrence of pavement icing during freezing precipitation events and the rates at which this occurs.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

OBJECTIVE: To review the available knowledge on epidemiology and diagnoses of acute infections in children aged 2 to 59 months in primary care setting and develop an electronic algorithm for the Integrated Management of Childhood Illness to reach optimal clinical outcome and rational use of medicines. METHODS: A structured literature review in Medline, Embase and the Cochrane Database of Systematic Review (CDRS) looked for available estimations of diseases prevalence in outpatients aged 2-59 months, and for available evidence on i) accuracy of clinical predictors, and ii) performance of point-of-care tests for targeted diseases. A new algorithm for the management of childhood illness (ALMANACH) was designed based on evidence retrieved and results of a study on etiologies of fever in Tanzanian children outpatients. FINDINGS: The major changes in ALMANACH compared to IMCI (2008 version) are the following: i) assessment of 10 danger signs, ii) classification of non-severe children into febrile and non-febrile illness, the latter receiving no antibiotics, iii) classification of pneumonia based on a respiratory rate threshold of 50 assessed twice for febrile children 12-59 months; iv) malaria rapid diagnostic test performed for all febrile children. In the absence of identified source of fever at the end of the assessment, v) urine dipstick performed for febrile children <2 years to consider urinary tract infection, vi) classification of 'possible typhoid' for febrile children >2 years with abdominal tenderness; and lastly vii) classification of 'likely viral infection' in case of negative results. CONCLUSION: This smartphone-run algorithm based on new evidence and two point-of-care tests should improve the quality of care of <5 year children and lead to more rational use of antimicrobials.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

Coherent anti-Stokes Raman scattering is the powerful method of laser spectroscopy in which significant successes are achieved. However, the non-linear nature of CARS complicates the analysis of the received spectra. The objective of this Thesis is to develop a new phase retrieval algorithm for CARS. It utilizes the maximum entropy method and the new wavelet approach for spectroscopic background correction of a phase function. The method was developed to be easily automated and used on a large number of spectra of different substances.. The algorithm was successfully tested on experimental data.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

A novel optimising controller is designed that leads a slow process from a sub-optimal operational condition to the steady-state optimum in a continuous way based on dynamic information. Using standard results from optimisation theory and discrete optimal control, the solution of a steady-state optimisation problem is achieved by solving a receding-horizon optimal control problem which uses derivative and state information from the plant via a shadow model and a state-space identifier. The paper analyzes the steady-state optimality of the procedure, develops algorithms with and without control rate constraints and applies the procedure to a high fidelity simulation study of a distillation column optimisation.

Relevância:

40.00% 40.00%

Publicador:

Resumo:

A dosing algorithm including genetic (VKORC1 and CYP2C9 genotypes) and nongenetic factors (age, weight, therapeutic indication, and cotreatment with amiodarone or simvastatin) explained 51% of the variance in stable weekly warfarin doses in 390 patients attending an anticoagulant clinic in a Brazilian public hospital. The VKORC1 3673G>A genotype was the most important predictor of warfarin dose, with a partial R(2) value of 23.9%. Replacing the VKORC1 3673G>A genotype with VKORC1 diplotype did not increase the algorithm`s predictive power. We suggest that three other single-nucleotide polymorphisms (SNPs) (5808T>G, 6853G>C, and 9041G>A) that are in strong linkage disequilibrium (LD) with 3673G>A would be equally good predictors of the warfarin dose requirement. The algorithm`s predictive power was similar across the self-identified ""race/color"" subsets. ""Race/color"" was not associated with stable warfarin dose in the multiple regression model, although the required warfarin dose was significantly lower (P = 0.006) in white (29 +/- 13 mg/week, n = 196) than in black patients (35 +/- 15 mg/week, n = 76).

Relevância:

40.00% 40.00%

Publicador:

Resumo:

In this paper, we present an algorithm for full-wave electromagnetic analysis of nanoplasmonic structures. We use the three-dimensional Method of Moments to solve the electric field integral equation. The computational algorithm is developed in the language C. As examples of application of the code, the problems of scattering from a nanosphere and a rectangular nanorod are analyzed. The calculated characteristics are the near field distribution and the spectral response of these nanoparticles. The convergence of the method for different discretization sizes is also discussed.