981 resultados para Software 3D e 2D


Relevância:

30.00% 30.00%

Publicador:

Resumo:

As BIM adoption continues, the goal of a totally collaborative model with multiple contributors is attainable. Many initiatives such as the 2016 UK government level 2 BIM deadline are putting pressure on the construction industry to speed up the changeover. Clients and collaborators have higher expectations of using digital 3D models to communicate design ideas and solve practical problems. Contractors and clients are benefitting from cost saving scheduling and clash detection offered by BIM. Effective collaboration on the project will also give speed and efficiency gains. Despite this, many businesses of varying sizes are still having problems. The cost of the software and the training provides an obvious barrier for micro-enterprises and could explain a delay in adoption. Many studies have looked at these problems faced by SME and micro-enterprises. Larger companies have different problems. The efforts made by government to encourage them are quite comprehensive, but is anything being done to help smaller sectors and keep the industry cohesive? This limited study examines several companies of varying size and varying project type: architectural design businesses, main contractor, structural engineer and building consultancy. The study examines the barriers to a truly collaborative BIM workflow facing different specialities on a larger project and a contrasting small/medium project. The findings will establish that different barriers for each sector are actually pushing further apart, thus potentially creating a BIM-only construction elite, leaving the small companies remaining on 2D based drawing.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Since the beginning of 3D computer vision problems, the use of techniques to reduce the data to make it treatable preserving the important aspects of the scene has been necessary. Currently, with the new low-cost RGB-D sensors, which provide a stream of color and 3D data of approximately 30 frames per second, this is getting more relevance. Many applications make use of these sensors and need a preprocessing to downsample the data in order to either reduce the processing time or improve the data (e.g., reducing noise or enhancing the important features). In this paper, we present a comparison of different downsampling techniques which are based on different principles. Concretely, five different downsampling methods are included: a bilinear-based method, a normal-based, a color-based, a combination of the normal and color-based samplings, and a growing neural gas (GNG)-based approach. For the comparison, two different models have been used acquired with the Blensor software. Moreover, to evaluate the effect of the downsampling in a real application, a 3D non-rigid registration is performed with the data sampled. From the experimentation we can conclude that depending on the purpose of the application some kernels of the sampling methods can improve drastically the results. Bilinear- and GNG-based methods provide homogeneous point clouds, but color-based and normal-based provide datasets with higher density of points in areas with specific features. In the non-rigid application, if a color-based sampled point cloud is used, it is possible to properly register two datasets for cases where intensity data are relevant in the model and outperform the results if only a homogeneous sampling is used.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

This dataset consists of 2D footprints of the buildings in the metropolitan Boston area, based on tiles in the orthoimage index (orthophoto quad ID: 229890, 229894, 229898, 229902, 233886, 233890, 233894, 233898, 233902, 237890, 237894, 237898, 237902, 241890, 241894, 241898, 241902, 245898, 245902). This data set was collected using 3Di's Digital Airborne Topographic Imaging System II (DATIS II). Roof height and footprint elevation attributes (derived from 1-meter resolution LIDAR (LIght Detection And Ranging) data) are included as part of each building feature. This data can be combined with other datasets to create 3D representations of buildings and the surrounding environment.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

BACKGROUND AND PURPOSE In clinical diagnosis, medical image segmentation plays a key role in the analysis of pathological regions. Despite advances in automatic and semi-automatic segmentation techniques, time-effective correction tools are commonly needed to improve segmentation results. Therefore, these tools must provide faster corrections with a lower number of interactions, and a user-independent solution to reduce the time frame between image acquisition and diagnosis. METHODS We present a new interactive method for correcting image segmentations. Our method provides 3D shape corrections through 2D interactions. This approach enables an intuitive and natural corrections of 3D segmentation results. The developed method has been implemented into a software tool and has been evaluated for the task of lumbar muscle and knee joint segmentations from MR images. RESULTS Experimental results show that full segmentation corrections could be performed within an average correction time of 5.5±3.3 minutes and an average of 56.5±33.1 user interactions, while maintaining the quality of the final segmentation result within an average Dice coefficient of 0.92±0.02 for both anatomies. In addition, for users with different levels of expertise, our method yields a correction time and number of interaction decrease from 38±19.2 minutes to 6.4±4.3 minutes, and 339±157.1 to 67.7±39.6 interactions, respectively.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Underwater video transects have become a common tool for quantitative analysis of the seafloor. However a major difficulty remains in the accurate determination of the area surveyed as underwater navigation can be unreliable and image scaling does not always compensate for distortions due to perspective and topography. Depending on the camera set-up and available instruments, different methods of surface measurement are applied, which make it difficult to compare data obtained by different vehicles. 3-D modelling of the seafloor based on 2-D video data and a reference scale can be used to compute subtransect dimensions. Focussing on the length of the subtransect, the data obtained from 3-D models created with the software PhotoModeler Scanner are compared with those determined from underwater acoustic positioning (ultra short baseline, USBL) and bottom tracking (Doppler velocity log, DVL). 3-D model building and scaling was successfully conducted on all three tested set-ups and the distortion of the reference scales due to substrate roughness was identified as the main source of imprecision. Acoustic positioning was generally inaccurate and bottom tracking unreliable on rough terrain. Subtransect lengths assessed with PhotoModeler were on average 20% longer than those derived from acoustic positioning due to the higher spatial resolution and the inclusion of slope. On a high relief wall bottom tracking and 3-D modelling yielded similar results. At present, 3-D modelling is the most powerful, albeit the most time-consuming, method for accurate determination of video subtransect dimensions.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

The power required to operate large gyratory mills often exceeds 10 MW. Hence, optimisation of the power consumption will have a significant impact on the overall economic performance and environmental impact of the mineral processing plant. In most of the published models of tumbling mills (e.g. [Morrell, S., 1996. Power draw of wet tumbling mills and its relationship to charge dynamics, Part 2: An empirical approach to modelling of mill power draw. Trans. Inst. Mining Metall. (Section C: Mineral Processing Ext. Metall.) 105, C54-C62. Austin, L.G., 1990. A mill power equation for SAG mills. Miner. Metall. Process. 57-62]), the effect of lifter design and its interaction with mill speed and filling are not incorporated. Recent experience suggests that there is an opportunity for improving grinding efficiency by choosing the appropriate combination of these variables. However, it is difficult to experimentally determine the interactions of these variables in a full scale mill. Although some work has recently been published using DEM simulations, it was basically. limited to 2D. The discrete element code, Particle Flow Code 3D (PFC3D), has been used in this work to model the effects of lifter height (525 cm) and mill speed (50-90% of critical) on the power draw and frequency distribution of specific energy (J/kg) of normal impacts in a 5 m diameter autogenous (AG) mill. It was found that the distribution of the impact energy is affected by the number of lifters, lifter height, mill speed and mill filling. Interactions of lifter design, mill speed and mill filling are demonstrated through three dimensional distinct element methods (3D DEM) modelling. The intensity of the induced stresses (shear and normal) on lifters, and hence the lifter wear, is also simulated. (C) 2004 Elsevier Ltd. All rights reserved.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this paper, we present the correction of the geometric distortion measured in the clinical magnetic resonance imaging (MRI) systems reported in the preceding paper (Part 1) using a 3D method based on the phantom-mapped geometric distortion data. This method allows the correction to be made on phantom images acquired without or with the vendor correction applied. With the vendor's 2D correction applied, the method corrects for both the residual geometric distortion still present in the plane in which the correction method was applied (the axial plane) and the uncorrected geometric distortion along the axis non-nal to the plane. The evaluation of the effectiveness of the correction using this new method was carried out through analyzing the residual geometric distortion in the corrected phantom images. The results show that the new method can restore the distorted images in 3D nearly to perfection. For all the MRI systems investigated, the mean absolute deviations in the positions of the control points (along x-, y- and z-axes) measured on the corrected phantom images were all less than 0.2 mm. The maximum absolute deviations were all below similar to0.8 mm. As expected, the correction of the phantom images acquired with the vendor's correction applied in the axial plane performed equally well. Both the geometric distortion still present in the axial plane after applying the vendor's correction and the uncorrected distortion along the z-axis have all been restored. (C) 2004 Elsevier Inc. All rights reserved.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Recently, a 3D phantom that can provide a comprehensive and accurate measurement of the geometric distortion in MRI has been developed. Using this phantom, a full assessment of the geometric distortion in a number of clinical MRI systems (GE and Siemens) has been carried out and detailed results are presented in this paper. As expected, the main source of geometric distortion in modern superconducting MRI systems arises from the gradient field nonlinearity. Significantly large distortions with maximum absolute geometric errors ranged between 10 and 25 mm within a volume of 240 x 240 x 240 mm(3) were observed when imaging with the new generation of gradient systems that employs shorter coils. By comparison, the geometric distortion was much less in the older-generation gradient systems. With the vendor's correction method, the geometric distortion measured was significantly reduced but only within the plane in which these 2D correction methods were applied. Distortion along the axis normal to the plane was, as expected, virtually unchanged. Two-dimensional correction methods are a convenient approach and in principle they are the only methods that can be applied to correct geometric distortion in a single slice or in multiple noncontiguous slices. However, these methods only provide an incomplete solution to the problem and their value can be significantly reduced if the distortion along the normal of the correction plane is not small. (C) 2004 Elsevier Inc. All rights reserved.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Summarizing topological relations is fundamental to many spatial applications including spatial query optimization. In this article, we present several novel techniques to effectively construct cell density based spatial histograms for range (window) summarizations restricted to the four most important level-two topological relations: contains, contained, overlap, and disjoint. We first present a novel framework to construct a multiscale Euler histogram in 2D space with the guarantee of the exact summarization results for aligned windows in constant time. To minimize the storage space in such a multiscale Euler histogram, an approximate algorithm with the approximate ratio 19/12 is presented, while the problem is shown NP-hard generally. To conform to a limited storage space where a multiscale histogram may be allowed to have only k Euler histograms, an effective algorithm is presented to construct multiscale histograms to achieve high accuracy in approximately summarizing aligned windows. Then, we present a new approximate algorithm to query an Euler histogram that cannot guarantee the exact answers; it runs in constant time. We also investigate the problem of nonaligned windows and the problem of effectively partitioning the data space to support nonaligned window queries. Finally, we extend our techniques to 3D space. Our extensive experiments against both synthetic and real world datasets demonstrate that the approximate multiscale histogram techniques may improve the accuracy of the existing techniques by several orders of magnitude while retaining the cost efficiency, and the exact multiscale histogram technique requires only a storage space linearly proportional to the number of cells for many popular real datasets.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Tissue Doppler (TD) assessment of dysynchrony (DYS) is established in evaluation for bi-ventricular pacing. Time to regional minimal volume by real-time 3D echo (3D) has been applied to DYS. 3D offers simultaneous assessment of all segments and may limit errors in localization of maximum delay due to off-axis images.We compared TD and 3D for assessment of DYS. 27 patients with ischaemic cardiomyopathy (aged 60±11 years, 85% male) underwent TD with generation of regional velocity curves. The interval between QRS onset and maximal systolic velocity (TTV) was measured in 6 basal and 6 mid-cavity segments. Onthe same day,3Dwas performed and data analysed offline with Q-Lab software (Philips, Andover, MA). Using 12 analogous regional time-volume curves time to minimal volume (T3D)was calculated. The standard deviation (S.D.) between segments in TTV and T3D was calculated as a measure ofDYS. In 7 patients itwas not possible to measureT3D due to poor images. In the remaining 20, LV diastolic volume, systolic volume and EF were 128±35 ml, 68±23 ml and 46±13%, respectively. Mean TTV was less than mean T3D (150±33ms versus 348±54 ms; p < 0.01). The intrapatient range was 20–210ms for TTV and 0–410ms for T3D. Of 9 patients (45%) with significantDYS (S.D. TTV > 32 ms), S.D. T3D was 69±37ms compared to 48±34ms in those without DYS (p = ns). In DYS patients there was concordance of the most delayed segment in 4 (44%) cases.Therefore, different techniques for assessing DYS are not directly comparable. Specific cut-offs for DYS are needed for each technique.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Deformable models are a highly accurate and flexible approach to segmenting structures in medical images. The primary drawback of deformable models is that they are sensitive to initialisation, with accurate and robust results often requiring initialisation close to the true object in the image. Automatically obtaining a good initialisation is problematic for many structures in the body. The cartilages of the knee are a thin elastic material that cover the ends of the bone, absorbing shock and allowing smooth movement. The degeneration of these cartilages characterize the progression of osteoarthritis. The state of the art in the segmentation of the cartilage are 2D semi-automated algorithms. These algorithms require significant time and supervison by a clinical expert, so the development of an automatic segmentation algorithm for the cartilages is an important clinical goal. In this paper we present an approach towards this goal that allows us to automatically providing a good initialisation for deformable models of the patella cartilage, by utilising the strong spatial relationship of the cartilage to the underlying bone.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

In this paper, we demonstrate the integration of a 3D hydrogel matrix within a hollow core photonic crystal fibre (HC-PCF). In addition, we also show the fluorescence of Cy5-labelled DNA molecules immobilized within the hydrogel formed in two different types of HC-PCF. The 3D hydrogel matrix is designed to bind with the amino groups of biomolecules using an appropriate cross-linker, providing higher sensitivity and selectivity than the standard 2D coverage, enabling a greater number of probe molecules to be available per unit area. The HC-PCFs, on the other hand, can be designed to maximize the capture of fluorescence to improve sensitivity and provide longer interaction lengths. This could enable the development of fibre-based point-of-care and remote systems, where the enhanced sensitivity would relax the constraints placed on sources and detectors. In this paper, we will discuss the formation of such polyethylene glycol diacrylate (PEGDA) hydrogels within a HC-PCF, including their optical properties such as light propagation and auto-fluorescence.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Quantitative structure-activity relationship (QSAR) analysis is a cornerstone of modern informatics. Predictive computational models of peptide-major histocompatibility complex (MHC)-binding affinity based on QSAR technology have now become important components of modern computational immunovaccinology. Historically, such approaches have been built around semiqualitative, classification methods, but these are now giving way to quantitative regression methods. We review three methods--a 2D-QSAR additive-partial least squares (PLS) and a 3D-QSAR comparative molecular similarity index analysis (CoMSIA) method--which can identify the sequence dependence of peptide-binding specificity for various class I MHC alleles from the reported binding affinities (IC50) of peptide sets. The third method is an iterative self-consistent (ISC) PLS-based additive method, which is a recently developed extension to the additive method for the affinity prediction of class II peptides. The QSAR methods presented here have established themselves as immunoinformatic techniques complementary to existing methodology, useful in the quantitative prediction of binding affinity: current methods for the in silico identification of T-cell epitopes (which form the basis of many vaccines, diagnostics, and reagents) rely on the accurate computational prediction of peptide-MHC affinity. We have reviewed various human and mouse class I and class II allele models. Studied alleles comprise HLA-A*0101, HLA-A*0201, HLA-A*0202, HLA-A*0203, HLA-A*0206, HLA-A*0301, HLA-A*1101, HLA-A*3101, HLA-A*6801, HLA-A*6802, HLA-B*3501, H2-K(k), H2-K(b), H2-D(b) HLA-DRB1*0101, HLA-DRB1*0401, HLA-DRB1*0701, I-A(b), I-A(d), I-A(k), I-A(S), I-E(d), and I-E(k). In this chapter we show a step-by-step guide into predicting the reliability and the resulting models to represent an advance on existing methods. The peptides used in this study are available from the AntiJen database (http://www.jenner.ac.uk/AntiJen). The PLS method is available commercially in the SYBYL molecular modeling software package. The resulting models, which can be used for accurate T-cell epitope prediction, will be made are freely available online at the URL http://www.jenner.ac.uk/MHCPred.

Relevância:

30.00% 30.00%

Publicador:

Resumo:

Biometrics is afield of study which pursues the association of a person's identity with his/her physiological or behavioral characteristics.^ As one aspect of biometrics, face recognition has attracted special attention because it is a natural and noninvasive means to identify individuals. Most of the previous studies in face recognition are based on two-dimensional (2D) intensity images. Face recognition based on 2D intensity images, however, is sensitive to environment illumination and subject orientation changes, affecting the recognition results. With the development of three-dimensional (3D) scanners, 3D face recognition is being explored as an alternative to the traditional 2D methods for face recognition.^ This dissertation proposes a method in which the expression and the identity of a face are determined in an integrated fashion from 3D scans. In this framework, there is a front end expression recognition module which sorts the incoming 3D face according to the expression detected in the 3D scans. Then, scans with neutral expressions are processed by a corresponding 3D neutral face recognition module. Alternatively, if a scan displays a non-neutral expression, e.g., a smiling expression, it will be routed to an appropriate specialized recognition module for smiling face recognition.^ The expression recognition method proposed in this dissertation is innovative in that it uses information from 3D scans to perform the classification task. A smiling face recognition module was developed, based on the statistical modeling of the variance between faces with neutral expression and faces with a smiling expression.^ The proposed expression and face recognition framework was tested with a database containing 120 3D scans from 30 subjects (Half are neutral faces and half are smiling faces). It is shown that the proposed framework achieves a recognition rate 10% higher than attempting the identification with only the neutral face recognition module.^