Professional Documents
Culture Documents
INTRODUCTION
Non-metallic inclusions play a major role in defining the properties and cleanliness of steel. They are an inevitable product of
the chemical reactions occurring during steel processing, and if not controlled they can drastically affect the properties of
steel, and in some cases be detrimental to in-use performance. Research on the evolution and development of inclusions
throughout steel processing is a constantly ongoing study [1]. Thus, understanding the types, chemistries, and size distributions
of non-metallic inclusions is vital for optimizing the properties and processing of steel. Numerous techniques have been
developed for detecting and analyzing inclusions, such as conventional optical microscopy, laser induced spectrometry, ultra
sound testing, X-ray computed tomography, and scanning electron microscopy (SEM), just to name a few [2]. An integral part
of the analysis is the identification or classification of the inclusions, since certain types of inclusions can have adverse
effects on the steel’s properties. Various methods have been applied to identify the types of inclusions present within a steel
sample, using user defined rules. Story et al. [3] utilized an automated steel cleanliness analysis tool (ASCAT) to examine the
chemistry, content, and size of inclusions, which in turn helped to improve the castability and properties of steel.
Statistical analysis has been widely used in all fields of research as it provides further insight and explores correlations
between data by combining the collection, analysis, interpretation, presentation, and organization of the data [4]. Generally,
the aim of statistical analysis is to identify patterns or trends in data sets, that are otherwise difficult to recognize. For
relatively small sample sizes, traditional methods of sampling data and interpreting results have been used, and for multi-
dimensional and extremely large data sets today’s powerful computers and advanced algorithms enable statistical analysis to
be a valuable research tool. The output of inclusion analysis is in this category of large, multivariate data sets. The variables
in each observation can correspond to the chemistries, dimensions, morphology, or the location of each inclusion, therefore
examining the correlations and trends between these variables can prove to be very informative.
Nowadays, investigation of inclusions is widely carried out using scanning electron microscopes (SEM) equipped with
energy dispersive x-ray (EDX) spectroscopes, using automated inclusion characterization (AIC). By scanning steel samples
in electron microscopes, automated analysis identifies inclusions within the steel matrix and determines their size,
morphology, and chemistry [5]. This study investigates ways to optimize data analysis and increase the amount of information
obtained from automated inclusion studies, by using statistical analyses to identify correlations in the multivariable data sets
generated by AIC. So far statistical techniques carried out in this study have been mainly focused on cluster analyses.
Not to be confused with the clustering of inclusions, cluster analysis, or sometimes termed as clustering, is basically an
automated method of arranging data into groups (referred to as clusters). Accordingly, the analysis referred to in this study
pertains to the cluster of data, where a cluster is comprised of inclusions with similar chemistries and not a physical cluster of
inclusions. Clustering groups observations by similarity, such that observations in one cluster are related to each other, in one
way or another, more than observations in other clusters. Clustering itself is not one specific algorithm that can be applied to
any set of data, rather it’s a general term used that serves a specific task. There are numerous clustering models, each with its
EXPERIMENTAL PROCEDURE
The current aim of this study, is to utilize the output results obtained from the automated analysis to determine and propose
an optimum clustering model that will identify the types of inclusions present in a steel sample.
The input data set to be used for clustering, is the number of inclusions in a sample along with the chemical composition of
each inclusion. Where the number of inclusions identified are considered as the number of observations in the data set, and
their chemistries represent the number of variables (dimensionality). The desired output is the number of clusters along with
their chemistries, corresponding to the type of inclusion each cluster represents, such as spinel (MgAl2O3), Al2O3, CaS etc.
For the purpose of this study, data sets were limited to 5 dimensions, the Mg, Al, Ca, S and Mn contents of each inclusion,
obtained from the automated analysis.
Sample Preparation and SEM Analysis
All samples investigated under this study are from an industrial production facility. They were two ladle samples and one
tundish sample, all from the same heat. The lollipop samples were cut to less than 1 inch in length and mounted to
accommodate the SEM, samples were then ground and polished in incremental steps, to achieve a smooth polished cross-
section ready for the automated analysis.
The automated analysis was conducted in an FEI/Aspex Explorer SEM utilizing a back-scattered electron detector, equipped
with an EDX analyzer. The operating parameters were all held constant for all samples to ensure consistency, the accelerating
voltage was 10 kV, the spot size set to 40%, and the working distance to 16 - 17 mm. Based on previous literature [7],
automated inclusion analyses conducted at 10 kV are more precise due to steel matrix distortion effects produced at higher
voltages, such as 20 kV. Accordingly, a spot size of 40% is a good compromise between image resolution and analysis speed
for an accelerating voltage of 10 kV. The selected working distance is the specified optimal distance for EDX analysis, as
per the SEM’s manufacturer, FEI [8].
The EDX analysis recorded the counts of each chemical element within the inclusions, therefore counts were converted to
estimated chemical compositions using the Merlet algorithm [9] in an Excel sheet. Thereafter, the results were filtered out to
ensure that all the readings to be analyzed are from actual non-metallic inclusions, and not from contaminants on the surface
of the samples or from erroneous readings. The filtering criteria excluded any reading were the [Fe] plus the [Mn] is larger
than 85%, in addition to filtering out any reading with more than 100 Si counts, and any reading with an area greater than
19.6 μm2. An area of 19.6 μm2 was specified, since 19.6 μm2 corresponds to an equivalent diameter of 5 μm, the expected
inclusion size [10], therefore any feature with a diameter larger than 5 μm was omitted.
Clustering Algorithms
Initially, k-means, a popular algorithm for clustering, was applied to the data. The k-means algorithm is a centroid-based
clustering technique, where each cluster is assigned a mean vector (centroid) [6]. It is an iterative technique where the number
, (1)
= number of observations
= number of clusters
, = nth observation of the kth cluster
= mean of the kth cluster
As increases, the sum of squares within each group of clusters decreases, and generally the optimal number of clusters is
when there’s a sharp change in the slope, at the “elbow” of the k vs WSS plot.
Another disadvantage is the tendency for k-means to produce equally sized clusters, which is not always realistic especially
when considering non-metallic inclusions in steel. As a result, the next approach was to apply the expectation-maximization
(EM) algorithm. The features that distinguish the EM algorithm from k-means are that the EM algorithm allows for varying
size and shape of clusters, and the EM algorithm is considered as a soft clustering technique. Soft clustering is when each
observation is assigned a probabilistic mix proportion to belong to each cluster, which is beneficial when observations lie
midway between clusters and useful for assessing the quality of the classification. Contrary to k-means, a hard clustering
technique, where each observation is assigned to one cluster.
When considering the EM algorithm, the data set is assumed to be a Gaussian Mixture Model (GMM), where each cluster has
its own normal distribution and the data set’s mix distribution represents the sum of the cluster distributions. This is
illustrated in Figure 2, where the blue curves are the individual clusters’ Gaussian distributions and the red curve is the GMM
of the whole data set.
The BIC is the value for the maximized log likelihood, considering the models, observations, parameters, data dimensions,
and number of clusters [13]. Generally, the optimal model and number of clusters are given by the largest BIC value computed.
However, in some cases singularities can occur and the log likelihood converges to infinity, as a result the BIC value cannot
be computed. Such instances arise when a cluster’s mean converges on a data point, or when the data set has very few
scattered observations [12, 14].
Sample Information
The samples investigated in this study are from an aluminum killed heat, with no calcium treatment; the samples were
processed to be clean with a relatively low inclusion density due to slightly higher Al content, with the steel’s final sulfur
content at 10 ppm. Table 1 below explains the samples identification and acquisition time. The average inclusion chemistries
are given in Figure 3.
100%
80% Mn
60% Ca
40% S
Al
20%
Mg
0%
Sample A Sample B Sample C
Figure 4 Sample-A: Elbow plot, to identify optimal number of clusters for k-means
The elbow plot, Figure 4, generated for sample A, does not show a distinct slope change that can clearly specify the best fit
number of clusters, rather there’s a compromise between 2 or 3 clusters. This demonstrates the ambiguity of the k-means
algorithm in identifying the optimal number of clusters. Therefore, results were generated for 2 and 3 clusters for
comparison. A summary of the cluster analysis is provided in Table 2. The color of each cluster in the table corresponds to
the colors on the ternary plots, Figure 5 and Figure 6.
Three ternary plots were generated for each sample, a “Number Fraction”, an “Area Fraction”, and a “Cluster Means” plot.
The number fraction plot simply represents each inclusion as a point on the ternary scales. On the area fraction plot,
inclusions are represented as circles where the size of each circle is relative to the area of the inclusion in μm2. The cluster
means diagram plots the cluster centroids as large filled circles, where the size of each circle corresponds to the inclusion area
Figure 5 Sample-A: Ternary Plots for 2 Clusters. Number Fraction - inclusions chemistry. Area Fraction – inclusion
chemistry relative to size. Cluster Means – cluster centroids with respect to their inclusion area
Figure 6 Sample A: Ternary Plots for 3 Clusters. Number Fraction - inclusions chemistry. Area Fraction – inclusion
chemistry relative to size. Cluster Means – cluster centroids with respect to their inclusion area
Based on the ternary plots, the sample is comprised mainly of alumina and spinel inclusions, with slight calcium pick up in
some inclusions. When 3 clusters are selected, rather than 2, the black cluster (cluster no. 1) is broken down into 2 clusters,
forming a more defined cluster in the spinel region and a small cluster with higher Ca contents, shown in green.
EM Algorithm Clustering
Clustering using the EM algorithm was carried out on all 3 samples. The optimal number of clusters along with its best fit
model were identified by the BIC algorithm mentioned earlier. The BIC is computed for 14 different models, on a range of 1
to 9 clusters for each model. The 14 models represent the covariance structure of the data set, they include different variations
in distribution, shape, volume, and orientation of the clusters [12]. The optimal model with the best fit number of clusters is the
one that generates the highest BIC value. As shown in Figure 7, the optimal number of clusters is 4, 4, and 5 clusters for
samples A, B, and C, respectively. For all the samples, the best fit model identified by the BIC technique was the VEV
model, representing clusters with an ellipsoidal distribution and same shape but varying in volume and orientation.
CONCLUSION
Clustering enabled the reduction of broad inclusion data sets to a handful of observations that can adequately define a
samples chemistry. The final aim of this study is to determine an ideal clustering technique, that can accurately identify types
of inclusions within a steel sample. Thus, proving to be an effective tool in steel processing, by providing a quick method to
evaluate the quality of the heat being processed.
ACKNOWLEDGEMENTS
We would like to thank the industrial members of the Center for Iron and Steelmaking Research for their continuous support
in this study, as well as their provision of samples.
REFERENCES
[1] S. Yang, J. Li, L. Zhang, K. Peaslee and Z. Wang, "Evolution of MgO.Al2O3 Based Inclusions in Alloy Steel During
Refinig Process," Metallurgical and Mining Industry, vol. 2, no. 2, pp. 87-92, 2010.
[2] B. G. Bartosiaki, J. A. M. Pereira, W. V. Bielefeldt and A. C. F. Vilela, "Assessment of inclusion analysis via manual
and automated SEM and total oxygen content of steel," Journal of Materials Research and Technology, vol. 4, no. 3,
pp. 235-240, 2015.
[3] S. R. Stroy, G. E. Goldsmith, R. J. Fruehan, G. S. Casuccio, M. S. Potter and D. M. Williams, "Study of Casting Issues
using Rapid Inclusion Identification and Analysis," in AISTech, Cleveland, OH, 2006.
[4] A. B. Schmiedt, U. Kamps and E. Cramer, "Statistical Modeling of Non-Metallic Inclusions in Steels and Extreme
Value Analysis," Hochschulbibliothek der Rheinisch-Westfälischen Technischen Hochschule Aachen, 2013.
[5] P. C. Pistorius, A. Patadia and J. Lee, "Correction of Matrix Effects on Microanalysis of Calcium Aluminate
Inclusions".
[6] D. Binu, "Cluster analysis using optimization algorithms with newly designed objective functions," Expert Systems with
Applications, vol. 42, pp. 5848-5859, 2015.
[7] D. Tang and P. C. Pistorius, "Optimizing Speed and Quality of Automated Inclusion Analysis," in Iron and Steel
Technology, Cleveland, OH, 2015.
[9] C. Merlet, "An Accurate Computer Correction Program for Quantitative Electron Probe Microanalysis," Mikrochmica
Acta, no. 114-115, pp. 363-376, 1994.
[10] P. C. Pistorius and N. Verma, "Matrix Effects in the Energy Dispersive X-Ray Analysis of CaO-Al2O3-MgO Inclusions
in Steel," Microscopy and Microanalysis, vol. 17, no. 6, pp. 963-971, 2011.
[11] M. A. Peeples, "R script for K-means Cluster Analysis," 2011. [Online]. Available:
http://www.mattpeeples.net/kmeans.html. [Accessed 25 5 2016].
[12] C. Fraley and A. Raftery, "Bayesian Regularization for Normal Mixture Estimation and Model-Based Clustering,"
Journal of Classification, no. 24, pp. 155-181, 2007.
[13] C. Fraley, A. E. Raftery, T. B. Murphy and L. Scrucca, "MCLUST Version 4 for R: Normal Mixture Modeling for
Model-Based Clustering, Classification, and Density Estimation," University of Washington, Seattle, June 2012.
[15] "The R Project for Statistical Computing," The R Foundation, [Online]. Available: http://www.r-project.org.
[16] W.-C. Chen, R. Maitra and V. Melnykov, "EM Algorithm for Model-Based Clustering of Finite Mixture Gaussian
Distribution," CRAN, 2015.
[17] N. Hamilton, "An extension to 'ggplot2', for the Creation of Ternary Diagrams," CRAN, 2016.