932 resultados para Vital Registration


Relevância:

20.00% 20.00%

Publicador:

Resumo:

ELEFAN 0, ELEFAN I and ELEFAN II were used to estimate vital statistics of Nemipterus japonicus from length-frequency data sampled along the coast of Bangladesh. The parameters L and K were estimated at 24.5 cm and 0.94 year super(-1). The values of M and F were found to be 1.81 and 1.58 year super(-1), respectively. The fish recruit to the fishery during May-June and September-October.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

Vital statistics are presented for 38 marine species of Vanuatu based on previous studies conducted in the area, with parameters describing growth (6 species, 13 sets of parameters), mortality (estimates of M for 6 species), length-weight relationship (32 species), and reproduction (length at first maturity for 26 species, months of reproduction for 18 species). The species covered belong mainly to the family Lutjanidae.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We present a novel method to perform an accurate registration of 3-D nonrigid bodies by using phase-shift properties of the dual-tree complex wavelet transform (DT-CWT). Since the phases of DT-\BBCWT coefficients change approximately linearly with the amount of feature displacement in the spatial domain, motion can be estimated using the phase information from these coefficients. The motion estimation is performed iteratively: first by using coarser level complex coefficients to determine large motion components and then by employing finer level coefficients to refine the motion field. We use a parametric affine model to describe the motion, where the affine parameters are found locally by substituting into an optical flow model and by solving the resulting overdetermined set of equations. From the estimated affine parameters, the motion field between the sensed and the reference data sets can be generated, and the sensed data set then can be shifted and interpolated spatially to align with the reference data set. © 2011 IEEE.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

We present a video-based system which interactively captures the geometry of a 3D object in the form of a point cloud, then recognizes and registers known objects in this point cloud in a matter of seconds (fig. 1). In order to achieve interactive speed, we exploit both efficient inference algorithms and parallel computation, often on a GPU. The system can be broken down into two distinct phases: geometry capture, and object inference. We now discuss these in further detail. © 2011 IEEE.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This paper presents a method for vote-based 3D shape recognition and registration, in particular using mean shift on 3D pose votes in the space of direct similarity transforms for the first time. We introduce a new distance between poses in this spacethe SRT distance. It is left-invariant, unlike Euclidean distance, and has a unique, closed-form mean, in contrast to Riemannian distance, so is fast to compute. We demonstrate improved performance over the state of the art in both recognition and registration on a real and challenging dataset, by comparing our distance with others in a mean shift framework, as well as with the commonly used Hough voting approach. © 2011 IEEE.

Relevância:

20.00% 20.00%

Publicador:

Resumo:

This chapter presents a method for vote-based 3D shape recognition and registration, in particular using mean shift on 3D pose votes in the space of direct similarity transformations for the first time. We introduce a new distance between poses in this spacethe SRT distance. It is left-invariant, unlike Euclidean distance, and has a unique, closed-form mean, in contrast to Riemannian distance, so is fast to compute. We demonstrate improved performance over the state of the art in both recognition and registration on a (real and) challenging dataset, by comparing our distance with others in a mean shift framework, as well as with the commonly used Hough voting approach. © 2013 Springer-Verlag Berlin Heidelberg.