993 resultados para Systematic Errors


Relevância:

60.00% 60.00%

Publicador:

Resumo:

There are conventional methods to calculate the centroid of spatial units and distance among them with using Geographical Information Systems (GIS). The paper points out potential measurement errors of this calculation. By taking Indian district data as an example, systematic errors concealed in such variables are shown. Two comparisons are examined; firstly, we compare the centroid obtained from the spatial units, polygons, and the centre of each city where its district headquarters locates. Secondly, between the centres represented in the above, we calculate the direct distances and road distances obtained from each pair of two districts. From the comparison between the direct distances of centroid of spatial units and the road distances of centre of district headquarters, we show the distribution of errors and list some caveats for the use of conventional variables obtained from GIS.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

La fisuración iniciada en la superficie de los pavimentos asfálticos constituye uno de los más frecuentes e importantes modos de deterioro que tienen lugar en los firmes bituminosos, como han demostrado los estudios teóricos y experimentales llevados a cabo en la última década. Sin embargo, este mecanismo de fallo no ha sido considerado por los métodos tradicionales de diseño de estos firmes. El concepto de firmes de larga duración se fundamenta en un adecuado seguimiento del proceso de avance en profundidad de estos deterioros y la intervención en el momento más apropiado para conseguir mantenerlos confinados como fisuras de profundidad parcial en la capa superficial más fácilmente accesible y reparable, de manera que pueda prolongarse la durabilidad y funcionalidad del firme y reducir los costes generalizados de su ciclo de vida. Por lo tanto, para la selección de la estrategia óptima de conservación de los firmes resulta esencial disponer de metodologías que posibiliten la identificación precisa in situ de la fisuración descendente, su seguimiento y control, y que además permitan una determinación fiable y con alto rendimiento de su profundidad y extensión. En esta Tesis Doctoral se presentan los resultados obtenidos mediante la investigación sistemática de laboratorio e in situ llevada a cabo para la obtención de datos sobre fisuración descendente en firmes asfálticos y para el estudio de procedimientos de evaluación de la profundidad de este tipo de fisuras empleando técnicas de ultrasonidos. Dichos resultados han permitido comprobar que la metodología no destructiva propuesta, de rápida ejecución, bajo coste y sencilla implementación (principalmente empleada hasta el momento en estructuras metálicas y de hormigón, debido a las dificultades que introduce la naturaleza viscoelástica de los materiales bituminosos) puede ser aplicada con suficiente fiabilidad y repetibilidad sobre firmes asfálticos. Las medidas resultan asimismo independientes del espesor total del firme. Además, permite resolver algunos de los inconvenientes frecuentes que presentan otros métodos de diagnóstico de las fisuras de pavimentos, tales como la extracción de testigos (sistema destructivo, de alto coste y prolongados tiempos de interrupción del tráfico) o algunas otras técnicas no destructivas como las basadas en medidas de deflexiones o el georradar, las cuales no resultan suficientemente precisas para la investigación de fisuras superficiales. Para ello se han realizado varias campañas de ensayos sobre probetas de laboratorio en las que se han estudiado diferentes condiciones empíricas como, por ejemplo, distintos tipos de mezclas bituminosas en caliente (AC, SMA y PA), espesores de firme y adherencias entre capas, temperaturas, texturas superficiales, materiales de relleno y agua en el interior de las grietas, posición de los sensores y un amplio rango de posibles profundidades de fisura. Los métodos empleados se basan en la realización de varias medidas de velocidad o de tiempo de transmisión del pulso ultrasónico sobre una única cara o superficie accesible del material, de manera que resulte posible obtener un coeficiente de transmisión de la señal (mediciones relativas o autocompensadas). Las mediciones se han realizado a bajas frecuencias de excitación mediante dos equipos de ultrasonidos diferentes dotados, en un caso, de transductores de contacto puntual seco (DPC) y siendo en el otro instrumento de contacto plano a través de un material especialmente seleccionado para el acoplamiento (CPC). Ello ha permitido superar algunos de los tradicionales inconvenientes que presenta el uso de los transductores convencionales y no precisar preparación previa de las superficies. La técnica de autocalibración empleada elimina los errores sistemáticos y la necesidad de una calibración local previa, demostrando el potencial de esta tecnología. Los resultados experimentales han sido comparados con modelos teóricos simplificados que simulan la propagación de las ondas ultrasónicas en estos materiales bituminosos fisurados, los cuales han sido deducidos previamente mediante un planteamiento analítico y han permitido la correcta interpretación de dichos datos empíricos. Posteriormente, estos modelos se han calibrado mediante los resultados de laboratorio, proporcionándose sus expresiones matemáticas generalizadas y gráficas para su uso rutinario en las aplicaciones prácticas. Mediante los ensayos con ultrasonidos efectuados en campañas llevadas a cabo in situ, acompañados de la extracción de testigos del firme, se han podido evaluar los modelos propuestos. El máximo error relativo promedio en la estimación de la profundidad de las fisuras al aplicar dichos modelos no ha superado el 13%, con un nivel de confianza del 95%, en el conjunto de todos los ensayos realizados. La comprobación in situ de los modelos ha permitido establecer los criterios y las necesarias recomendaciones para su utilización sobre firmes en servicio. La experiencia obtenida posibilita la integración de esta metodología entre las técnicas de auscultación para la gestión de su conservación. Abstract Surface-initiated cracking of asphalt pavements constitutes one of the most frequent and important types of distress that occur in flexible bituminous pavements, as clearly has been demonstrated in the technical and experimental studies done over the past decade. However, this failure mechanism has not been taken into consideration for traditional methods of flexible pavement design. The concept of long-lasting pavements is based on adequate monitoring of the depth and extent of these deteriorations and on intervention at the most appropriate moment so as to contain them in the surface layer in the form of easily-accessible and repairable partial-depth topdown cracks, thereby prolonging the durability and serviceability of the pavement and reducing the overall cost of its life cycle. Therefore, to select the optimal maintenance strategy for perpetual pavements, it becomes essential to have access to methodologies that enable precise on-site identification, monitoring and control of top-down propagated cracks and that also permit a reliable, high-performance determination of the extent and depth of cracking. This PhD Thesis presents the results of systematic laboratory and in situ research carried out to obtain information about top-down cracking in asphalt pavements and to study methods of depth evaluation of this type of cracking using ultrasonic techniques. These results have demonstrated that the proposed non-destructive methodology –cost-effective, fast and easy-to-implement– (mainly used to date for concrete and metal structures, due to the difficulties caused by the viscoelastic nature of bituminous materials) can be applied with sufficient reliability and repeatability to asphalt pavements. Measurements are also independent of the asphalt thickness. Furthermore, it resolves some of the common inconveniences presented by other methods used to evaluate pavement cracking, such as core extraction (a destructive and expensive procedure that requires prolonged traffic interruptions) and other non-destructive techniques, such as those based on deflection measurements or ground-penetrating radar, which are not sufficiently precise to measure surface cracks. To obtain these results, extensive tests were performed on laboratory specimens. Different empirical conditions were studied, such as various types of hot bituminous mixtures (AC, SMA and PA), differing thicknesses of asphalt and adhesions between layers, varied temperatures, surface textures, filling materials and water within the crack, different sensor positions, as well as an ample range of possible crack depths. The methods employed in the study are based on a series of measurements of ultrasonic pulse velocities or transmission times over a single accessible side or surface of the material that make it possible to obtain a signal transmission coefficient (relative or auto-calibrated readings). Measurements were taken at low frequencies by two short-pulse ultrasonic devices: one equipped with dry point contact transducers (DPC) and the other with flat contact transducers that require a specially-selected coupling material (CPC). In this way, some of the traditional inconveniences presented by the use of conventional transducers were overcome and a prior preparation of the surfaces was not required. The auto-compensating technique eliminated systematic errors and the need for previous local calibration, demonstrating the potential for this technology. The experimental results have been compared with simplified theoretical models that simulate ultrasonic wave propagation in cracked bituminous materials, which had been previously deduced using an analytical approach and have permitted the correct interpretation of the aforementioned empirical results. These models were subsequently calibrated using the laboratory results, providing generalized mathematical expressions and graphics for routine use in practical applications. Through a series of on-site ultrasound test campaigns, accompanied by asphalt core extraction, it was possible to evaluate the proposed models, with differences between predicted crack depths and those measured in situ lower than 13% (with a confidence level of 95%). Thereby, the criteria and the necessary recommendations for their implementation on in-service asphalt pavements have been established. The experience obtained through this study makes it possible to integrate this methodology into the evaluation techniques for pavement management systems.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

La Fotogrametría, como ciencia y técnica de obtención de información tridimensional del espacio objeto a partir de imágenes bidimensionales, requiere de medidas de precisión y en ese contexto, la calibración geométrica de cámaras ocupa un lugar importante. El conocimiento de la geometría interna de la cámara es fundamental para lograr mayor precisión en las medidas realizadas. En Fotogrametría Aérea se utilizan cámaras métricas (fabricadas exclusivamente para aplicaciones cartográficas), que incluyen objetivos fotográficos con sistemas de lentes complejos y de alta calidad. Pero en Fotogrametría de Objeto Cercano se está trabajando cada vez con más asiduidad con cámaras no métricas, con ópticas de peor calidad que exigen una calibración geométrica antes o después de cada trabajo. El proceso de calibración encierra tres conceptos fundamentales: modelo de cámara, modelo de distorsión y método de calibración. El modelo de cámara es un modelo matemático que aproxima la transformación proyectiva original a la realidad física de las lentes. Ese modelo matemático incluye una serie de parámetros entre los que se encuentran los correspondientes al modelo de distorsión, que se encarga de corregir los errores sistemáticos de la imagen. Finalmente, el método de calibración propone el método de estimación de los parámetros del modelo matemático y la técnica de optimización a emplear. En esta Tesis se propone la utilización de un patrón de calibración bidimensional que se desplaza en la dirección del eje óptico de la cámara, ofreciendo así tridimensionalidad a la escena fotografiada. El patrón incluye un número elevado de marcas, lo que permite realizar ensayos con distintas configuraciones geométricas. Tomando el modelo de proyección perspectiva (o pinhole) como modelo de cámara, se realizan ensayos con tres modelos de distorsión diferentes, el clásico de distorsión radial y tangencial propuesto por D.C. Brown, una aproximación por polinomios de Legendre y una interpolación bicúbica. De la combinación de diferentes configuraciones geométricas y del modelo de distorsión más adecuado, se llega al establecimiento de una metodología de calibración óptima. Para ayudar a la elección se realiza un estudio de las precisiones obtenidas en los distintos ensayos y un control estereoscópico de un panel test construido al efecto. ABSTRACT Photogrammetry, as science and technique for obtaining three-dimensional information of the space object from two-dimensional images, requires measurements of precision and in that context, the geometric camera calibration occupies an important place. The knowledge of the internal geometry of the camera is fundamental to achieve greater precision in measurements made. Metric cameras (manufactured exclusively for cartographic applications), including photographic lenses with complex lenses and high quality systems are used in Aerial Photogrammetry. But in Close Range Photogrammetry is working increasingly more frequently with non-metric cameras, worst quality optical components which require a geometric calibration before or after each job. The calibration process contains three fundamental concepts: camera model, distortion model and method of calibration. The camera model is a mathematical model that approximates the original projective transformation to the physical reality of the lenses. The mathematical model includes a series of parameters which include the correspondents to the model of distortion, which is in charge of correcting the systematic errors of the image. Finally, the calibration method proposes the method of estimation of the parameters of the mathematical modeling and optimization technique to employ. This Thesis is proposing the use of a pattern of two dimensional calibration that moves in the direction of the optical axis of the camera, thus offering three-dimensionality to the photographed scene. The pattern includes a large number of marks, which allows testing with different geometric configurations. Taking the projection model perspective (or pinhole) as a model of camera, tests are performed with three different models of distortion, the classical of distortion radial and tangential proposed by D.C. Brown, an approximation by Legendre polynomials and bicubic interpolation. From the combination of different geometric configurations and the most suitable distortion model, brings the establishment of a methodology for optimal calibration. To help the election, a study of the information obtained in the various tests and a purpose built test panel stereoscopic control is performed.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

Establishing accurate extragalactic distances has provided an immense challenge to astronomers since the 1920s. The situation has improved dramatically as better detectors have become available, and as several new, promising techniques have been developed. For the first time in the history of this difficult field, relative distances to galaxies are being compared on a case-by-case basis, and their quantitative agreement is being established. New instrumentation, the development of new techniques for measuring distances, and recent measurements with the Hubble Space telescope all have resulted in new distances to galaxies with precision at the ±5–20% level. The current statistical uncertainty in some methods for measuring H0 is now only a few percent; with systematic errors, the total uncertainty is approaching ±10%. Hence, the historical factor-of-two uncertainty in the value of the H0 is now behind us.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

El objetivo de este trabajo consiste en examinar la rivalidad intra y entre grupos estratégicos de una industria. La literatura sobre la materia sugiere que las evidencias empíricas detectadas pueden verse afectadas por problemas operativos relacionados con la medición de la rivalidad. Como novedad, la metodología aplicada propone la utilización de dos enfoques de medición de la rivalidad, uno indirecto mediante la estimación de la variación conjetural, y una evaluación directa a través de las noticias sobre acciones y reacciones estratégicas detectadas en las publicaciones. La aplicación empírica realizada en el mercado español de depósitos bancarios no permite asegurar nada acerca de la mayor o menor rivalidad intra que entre grupos; pero los grupos estratégicos permiten predecir la forma en que cada entidad compite con las demás.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

This raster layer represents surface elevation and bathymetry data for the Boston Region, Massachusetts. It was created by merging portions of MassGIS Digital Elevation Model 1:5,000 (2005) data with NOAA Estuarine Bathymetric Digital Elevation Models (30 m.) (1998). DEM data was derived from the digital terrain models that were produced as part of the MassGIS 1:5,000 Black and White Digital Orthophoto imagery project. Cellsize is 5 meters by 5 meters. Each cell has a floating point value, in meters, which represents its elevation above or below sea level.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

JPRS: R-105-N/31; R-107-N/31; R-107-N/72; R-113-N/1; R-104-N/25; R-104-N/2; R-105-N/25; R-105-N/24.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

We discuss the construction of a photometric redshift catalogue of luminous red galaxies (LRGs) from the Sloan Digital Sky Survey (SDSS), emphasizing the principal steps necessary for constructing such a catalogue: (i) photometrically selecting the sample, (ii) measuring photometric redshifts and their error distributions, and (iii) estimating the true redshift distribution. We compare two photometric redshift algorithms for these data and find that they give comparable results. Calibrating against the SDSS and SDSS-2dF (Two Degree Field) spectroscopic surveys, we find that the photometric redshift accuracy is sigma similar to 0.03 for redshifts less than 0.55 and worsens at higher redshift (similar to 0.06 for z < 0.7). These errors are caused by photometric scatter, as well as systematic errors in the templates, filter curves and photometric zero-points. We also parametrize the photometric redshift error distribution with a sum of Gaussians and use this model to deconvolve the errors from the measured photometric redshift distribution to estimate the true redshift distribution. We pay special attention to the stability of this deconvolution, regularizing the method with a prior on the smoothness of the true redshift distribution. The methods that we develop are applicable to general photometric redshift surveys.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

O objetivo deste estudo prospectivo foi avaliar os efeitos do aparelho Forsus® nos incisivos centrais superiores e inferiores. A amostra constituiu-se de 22 tomografias computadorizadas de 11 pacientes (sexo masculino e feminino) idade média de 15,8 anos com má oclusão de Classe II que foram tratados com o aparelho Forsus® na clínica do programa de pós-graduação em Odontologia, área de concentração Ortodontia, da Universidade Metodista de São Paulo. As tomografias foram obtidas em dois momentos T1 (final de nivelamento e antes da instalação do Forsus® e T2 (remoção do Forsus®). Para avaliar a distância do ápice até a tábua óssea, as imagens a serem examinadas foram obtidas com o auxílio do viewer do próprio i-CAT® , o iCATVision® e examinadas com o CorelDRAW X5® já para as medidas cefalométricas IMPA e 1.PP as imagens cefalométricas ortogonais foram obtidas em proporção 1:1 com auxílio do software Dolphin 3D® (Dolphin Imaging and Management Solutions, Chatsworth, EUA) e em seguida examinadas com o software Radiocef Studio 2 (Radio Memory, Belo Horizonte, Brasil). Para a obtenção do erro intra-examinador foi feito o teste t de Student pareado para o erro sistemático e a fórmula de DAHLBERG para estimar a ordem de grandeza dos erros casuais e na análise estatística dos resultados utilizou-se: o teste t para a determinação das diferenças entres as fases de observação e o teste de correlação de Pearson para avaliar a correlação entres as alterações. Observou-se: um aumento significativo (p<0,05) tanto no IMPA quanto no 1.PP, aproximação do ápice dos incisivos inferiores da tábua óssea lingual, aproximação do ápice dos incisivos superiores da tábua óssea vestibular, uma correlação negativa muito forte entre o IMPA e a distância do ápice do incisivo até a tábua óssea lingual e uma correlação negativa moderada entre 1.PP e a distância do ápice do incisivo até a tábua óssea vestibular. Sendo assim o aparelho Forsus® no tratamento da Classe II teve como efeito: vestibularização significativa dos incisivos centrais inferiores, uma verticalização significativa dos incisivos centrais superiores, aproximação do ápice dos incisivos inferiores da cortical óssea lingual e aproximação do ápice dos incisivos superiores da cortical óssea vestibular.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

The case for monitoring large-scale sea level variability is established in the context of the estimation of the extent of anthropogenic climate change. Satellite altimeters are identified as having the potential to monitor this change with high resolution and accuracy. Possible sources of systematic errors and instabilities in these instruments which would be hurdles to the most accurate monitoring of such ocean signals are examined. Techniques for employing tide gauges to combat such inaccuracies are proposed and developed. The tide gauge at Newhaven in Sussex is used in conjunction with the nearby satellite laser ranger and high-resolution ocean models to estimate the absolute bias of the TOPEX, Poseidon, ERS 1 and ERS 2 altimeters. The theory which underlies the augmentation of altimeter measurements with tide gauge data is developed. In order to apply this, the tide gauges of the World Ocean Circulation Experiment are assessed and their suitability for altimeter calibration is determined. A reliable subset of these gauges is derived. A method of intra-altimeter calibration is developed using these tide gauges to remove the effect of variability over long time scales. In this way the long-term instability in the TOPEX range measurement is inferred and the drift arising from the on-board ultra stable oscillator is thus detected. An extension to this work develops a method for inter-altimeter calibration, allowing the systematic differences between unconnected altimeters to be measured. This is applied to the TOPEX and ERS 1 altimeters.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

This dissertation presents a study of the D( e, e′p)n reaction carried out at the Thomas Jefferson National Accelerator Facility (Jefferson Lab) for a set of fixed values of four-momentum transfer Q 2 = 2.1 and 0.8 (GeV/c)2 and for missing momenta pm ranging from pm = 0.03 to pm = 0.65 GeV/c. The analysis resulted in the determination of absolute D(e,e′ p)n cross sections as a function of the recoiling neutron momentum and it's scattering angle with respect to the momentum transfer [vector] q. The angular distribution was compared to various modern theoretical predictions that also included final state interactions. The data confirmed the theoretical prediction of a strong anisotropy of final state interaction contributions at Q2 of 2.1 (GeV/c)2 while at the lower Q2 value, the anisotropy was much less pronounced. At Q2 of 0.8 (GeV/c)2, theories show a large disagreement with the experimental results. The experimental momentum distribution of the bound proton inside the deuteron has been determined for the first time at a set of fixed neutron recoil angles. The momentum distribution is directly related to the ground state wave function of the deuteron in momentum space. The high momentum part of this wave function plays a crucial role in understanding the short-range part of the nucleon-nucleon force. At Q2 = 2.1 (GeV/c)2, the momentum distribution determined at small neutron recoil angles is much less affected by FSI compared to a recoil angle of 75°. In contrast, at Q2 = 0.8 (GeV/c)2 there seems to be no region with reduced FSI for larger missing momenta. Besides the statistical errors, systematic errors of about 5–6 % were included in the final results in order to account for normalization uncertainties and uncertainties in the determi- nation of kinematic veriables. The measurements were carried out using an electron beam energy of 2.8 and 4.7 GeV with beam currents between 10 to 100 &mgr; A. The scattered electrons and the ejected protons originated from a 15cm long liquid deuterium target, and were detected in conicidence with the two high resolution spectrometers of Hall A at Jefferson Lab.^

Relevância:

60.00% 60.00%

Publicador:

Resumo:

The objective of this study was to determine the seasonal and interannual variability and calculate the trends of wind speed in NEB and then validate the mesoscale numerical model for after engage with the microscale numerical model in order to get the wind resource at some locations in the NEB. For this we use two data sets of wind speed (weather stations and anemometric towers) and two dynamic models; one of mesoscale and another of microscale. We use statistical tools to evaluate and validate the data obtained. The simulations of the dynamic mesoscale model were made using data assimilation methods (Newtonian Relaxation and Kalman filter). The main results show: (i) Five homogeneous groups of wind speed in the NEB with higher values in winter and spring and with lower in summer and fall; (ii) The interannual variability of the wind speed in some groups stood out with higher values; (iii) The large-scale circulation modified by the El Niño and La Niña intensified wind speed for the groups with higher values; (iv) The trend analysis showed more significant negative values for G3, G4 and G5 in all seasons and in the annual average; (v) The performance of dynamic mesoscale model showed smaller errors in the locations Paracuru and São João and major errors were observed in Triunfo; (vi) Application of the Kalman filter significantly reduce the systematic errors shown in the simulations of the dynamic mesoscale model; (vii) The wind resource indicate that Paracuru and Triunfo are favorable areas for the generation of energy, and the coupling technique after validation showed better results for Paracuru. We conclude that the objective was achieved, making it possible to identify trends in homogeneous groups of wind behavior, and to evaluate the quality of both simulations with the dynamic model of mesoscale and microscale to answer questions as necessary before planning research projects in Wind-Energy area in the NEB

Relevância:

60.00% 60.00%

Publicador:

Resumo:

In this work, it was developed and validated methodologies that were based on the use of Infrared Spectroscopy Mid (MIR) combined with multivariate calibration Square Partial Least (PLS) to quantify adulterants such as soybean oil and residual soybean oil in methyl and ethyl palm biodiesels in the concentration range from 0.25 to 30.00 (%), as well as to determine methyl and ethyl palm biodiesel content in their binary mixtures with diesel in the concentration range from 0.25 to 30.00 (%). The prediction results showed that PLS models constructed are satisfactory. Errors Mean Square Forecast (RMSEP) of adulteration and content determination showed values of 0.2260 (%), with mean error (EM) with values below 1.93 (%). The models also showed a strong correlation between actual and predicted values, staying above 0.99974. No systematic errors were observed, in accordance to ASTM E1655- 05. Thus the built PLS models, may be a promising alternative in the quality control of this fuel for possible adulterations or to content determination.

Relevância:

60.00% 60.00%

Publicador:

Resumo:

A tenet of modern radiotherapy (RT) is to identify the treatment target accurately, following which the high-dose treatment volume may be expanded into the surrounding tissues in order to create the clinical and planning target volumes. Respiratory motion can induce errors in target volume delineation and dose delivery in radiation therapy for thoracic and abdominal cancers. Historically, radiotherapy treatment planning in the thoracic and abdominal regions has used 2D or 3D images acquired under uncoached free-breathing conditions, irrespective of whether the target tumor is moving or not. Once the gross target volume has been delineated, standard margins are commonly added in order to account for motion. However, the generic margins do not usually take the target motion trajectory into consideration. That may lead to under- or over-estimate motion with subsequent risk of missing the target during treatment or irradiating excessive normal tissue. That introduces systematic errors into treatment planning and delivery. In clinical practice, four-dimensional (4D) imaging has been popular in For RT motion management. It provides temporal information about tumor and organ at risk motion, and it permits patient-specific treatment planning. The most common contemporary imaging technique for identifying tumor motion is 4D computed tomography (4D-CT). However, CT has poor soft tissue contrast and it induce ionizing radiation hazard. In the last decade, 4D magnetic resonance imaging (4D-MRI) has become an emerging tool to image respiratory motion, especially in the abdomen, because of the superior soft-tissue contrast. Recently, several 4D-MRI techniques have been proposed, including prospective and retrospective approaches. Nevertheless, 4D-MRI techniques are faced with several challenges: 1) suboptimal and inconsistent tumor contrast with large inter-patient variation; 2) relatively low temporal-spatial resolution; 3) it lacks a reliable respiratory surrogate. In this research work, novel 4D-MRI techniques applying MRI weightings that was not used in existing 4D-MRI techniques, including T2/T1-weighted, T2-weighted and Diffusion-weighted MRI were investigated. A result-driven phase retrospective sorting method was proposed, and it was applied to image space as well as k-space of MR imaging. Novel image-based respiratory surrogates were developed, improved and evaluated.