3 resultados para Descriptive quantitative analysis

em CiencIPCA - Instituto Politécnico do Cávado e do Ave, Portugal


Relevância:

90.00% 90.00%

Publicador:

Resumo:

Within the development of motor vehicles, crash safety (e.g. occupant protection, pedestrian protection, low speed damageability), is one of the most important attributes. In order to be able to fulfill the increased requirements in the framework of shorter cycle times and rising pressure to reduce costs, car manufacturers keep intensifying the use of virtual development tools such as those in the domain of Computer Aided Engineering (CAE). For crash simulations, the explicit finite element method (FEM) is applied. The accuracy of the simulation process is highly dependent on the accuracy of the simulation model, including the midplane mesh. One of the roughest approximations typically made is the actual part thickness which, in reality, can vary locally. However, almost always a constant thickness value is defined throughout the entire part due to complexity reasons. On the other hand, for precise fracture analysis within FEM, the correct thickness consideration is one key enabler. Thus, availability of per element thickness information, which does not exist explicitly in the FEM model, can significantly contribute to an improved crash simulation quality, especially regarding fracture prediction. Even though the thickness is not explicitly available from the FEM model, it can be inferred from the original CAD geometric model through geometric calculations. This paper proposes and compares two thickness estimation algorithms based on ray tracing and nearest neighbour 3D range searches. A systematic quantitative analysis of the accuracy of both algorithms is presented, as well as a thorough identification of particular geometric arrangements under which their accuracy can be compared. These results enable the identification of each technique’s weaknesses and hint towards a new, integrated, approach to the problem that linearly combines the estimates produced by each algorithm.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Purpose – Castings defects are usually easy to characterize, but to eradicate them can be a difficult task. In many cases, defects are caused by the combined effect of different factors, whose identification is often difficult. Besides, the real non-quality costs are usually unknown, and even neglected. This paper aims to describe the development of a modular tool for quality improvement in foundries, and its main objective is to present the application potential and the foundry process areas that are covered and taken into account. Design/methodology/approach – The integrated model was conceived as an expert system, designated Qualifound, which performs both qualitative and quantitative analyses. For the qualitative analyses mode, the nomenclature and the description of defects are based on the classification suggested by the International Committee of the Foundry Technical Association. Thus, a database of defects was established, enabling one to associate the defects with the relevant process operations and the identification of their possible causes. The quantitative analysis mode deals with the number of produced and rejected castings and includes the calculation of the non-quality costs. Findings – The validation of Qualifound was carried out in a Portuguese foundry, whose quality system had been certified according to the ISO 9000 standards. Qualifound was used in every management area and it was concluded that the application had the required technological requisites to provide the necessary information for the foundry management to improve process quality. Originality/value – The paper presents a successful application of an informatics tool on quality improvement in foundries.

Relevância:

80.00% 80.00%

Publicador:

Resumo:

Quantitative analysis of cine cardiac magnetic resonance (CMR) images for the assessment of global left ventricular morphology and function remains a routine task in clinical cardiology practice. To date, this process requires user interaction and therefore prolongs the examination (i.e. cost) and introduces observer variability. In this study, we sought to validate the feasibility, accuracy, and time efficiency of a novel framework for automatic quantification of left ventricular global function in a clinical setting.