This paper was converted on www.awesomepapers.org from LaTeX by an anonymous user.
Want to know more? Visit the Converter page.

Interpretable Recognition of Fused Magnesium Furnace Working Conditions with Deep Convolutional Stochastic Configuration Networks

Weitao Li, Xinru Zhang, Dianhui Wang, Qianqian Tong, and Tianyou Chai This work was supported by National Key R&D Program of China under Grant (2018AAA0100304), National Natural Science Foundation of China (62173120, 62103092), Anhui Provincial Natural Science Foundation (2108085UD11) and 111 Project (BP0719039). (Corresponding author: Dianhui Wang.)Weitao Li, Xinru Zhang, and Qianqian Tong are with the Department of Electric Engineering and Automation of Hefei University of Technology, Hefei 230009, China (e-mail: [email protected]; [email protected]; [email protected]).Dianhui Wang is with the Institute of Artificial Intelligence, China University of Mining and Technology, Xuzhou 221116, China (e-mail: [email protected]).Tianyou Chai is with the State Key Laboratory of Synthetical Automation for Process Industries, Northeastern University, Shenyang 110004, Liaoning, China (e-mail: [email protected]).
Abstract

To address the issues of a weak generalization capability and interpretability in working condition recognition model of a fused magnesium furnace, this paper proposes an interpretable working condition recognition method based on deep convolutional stochastic configuration networks (DCSCNs). Firstly, a supervised learning mechanism is employed to generate physically meaningful Gaussian differential convolution kernels. An incremental method is utilized to construct a DCSCNs model, ensuring the convergence of recognition errors in a hierarchical manner and avoiding the iterative optimization process of convolutional kernel parameters using the widely used backpropagation algorithm. The independent coefficient of channel feature maps is defined to obtain the visualization results of feature class activation maps for the fused magnesium furnace. A joint reward function is constructed based on the recognition accuracy, the interpretable trustworthiness evaluation metrics, and the model parameter quantity. Reinforcement learning (RL) is applied to adaptively prune the convolutional kernels of the DCSCNs model, aiming to build a compact, highly performed and interpretable network. The experimental results demonstrate that the proposed method outperforms the other deep learning approaches in terms of recognition accuracy and interpretability.

Index Terms:
Deep convolutional stochastic configuration networks, Gaussian differential convolution kernel, Class activation mapping map, Interpretability, Fused magnesium furnace working condition recognition.

I Introduction

Magnesium oxide, as the main component of fused magnesium sand (also known as fused magnesia), is an alkaline refractory raw material widely used in aerospace, nuclear furnaces, electronics, and other fields. As the world’s largest producer and supplier of fused magnesium, China faces challenges such as low grade, large composition fluctuations, and complex mineral composition in its magnesite ore. Therefore, a unique three-phase alternating current fused magnesium furnace is required for smelting. The smelting process of an fused magnesium furnace involves simultaneous feeding and smelting. Raw materials are poured into the furnace by a machine, and high-temperature arcs in the furnace heat the raw materials to generate magnesium oxide crystals [1, 2, 3].

To ensure the quality of electric fused magnesia, it is necessary to monitor the production process and prevent possible abnormal working conditions during smelting, including underburn, overheating, and abnormal exhaust [4]. The underburn condition is caused by impurities and complex minerals in the raw materials, resulting in incomplete combustion in the fused magnesium furnace. During this condition, the furnace wall in certain areas becomes red and bright. The overheating condition is characterized by a brighter flame at the furnace mouth, which may produce adverse substances such as magnesium smoke and magnesium oxide. The abnormal exhaust working condition involves the ejection of high-temperature molten metal from the furnace mouth, resulting in drastic changes in current. When these abnormal working conditions occur, they need to be promptly detected and addressed to avoid energy waste, reduction of the utilization rate and grade of magnesia sand, furnace lining burn-through, and potential hazards to operators due to raw material leakage. Presently, the inability to directly observe the internal molten pool in the furnace due to the presence of burning flames at the furnace mouth renders the diagnosis of abnormal working conditions in magnesium furnaces contingent upon manual decision-making [5]. However, each inspector is responsible for multiple furnaces, and subjective factors such as their experience, sense of responsibility, and labor intensity, as well as objective factors such as high temperature, noise, dust, water mist, and inherent white spots on the furnace wall in a complex firing environment, make it difficult to make manual decisions. This can easily lead to missed or erroneous inspections, resulting in irreversible losses such as the burn-through of fused magnesium furnaces and difficulties in meeting the real-time inspection requirements for operation and maintenance.

In recent years, with the continuous development of artificial intelligence technology, the use of machine learning and deep learning for the abnormal working condition diagnosis of fused magnesium furnaces has received widespread attention. These methods collect and process production data from the furnaces, extract various feature parameters, and utilize algorithms for the automatic diagnosis of abnormal working conditions. For example, the work in [6] proposes a magnesium furnace abnormal working condition diagnosis method based on Bayesian networks, which introduces transfer learning to address the problem of limited abnormal working condition data. The work in [7] applies semi-supervised learning to automatically label unlabeled current data and train classifiers constructed within a semi-supervised learning framework. The work in [8] combines convolutional neural networks (CNNs) and long short-term memory (LSTM) networks to extract the spatial and temporal features of fused magnesium furnaces for underburn condition recognition. The work in [9] uses convolutional networks, bi-directional long short-term memory networks (Bi-LSTM), and stacked autoencoders to respectively extract image, sound, and current features of fused magnesium furnaces, and then combines these different features to train a condition classifier. However, the unique operational conditions in fused magnesium furnaces present a challenge for the generalization capabilities of recognition models. Models that have been otherwise well-trained exhibit a marked decline in performance when exposed to these datasets, revealing a significant weakness in their ability to generalize.

Although deep learning has been widely used in the field of image analysis, there are still some challenges that need to be addressed. The black box nature of deep neural networks is essentially due to the inconsistency between the features learned by internal neurons and the semantic concepts understood by humans. The interpretability of deep learning refers to the extent to which the decision-making process of a model can be clearly understood and explained. It is an important indicator of the level of understanding of the model and the degree of trust in its decisions. Currently, the interpretability of deep learning models can be presented in different ways [10, 11], including feature visualization, analysis visualization, local and global interpretability, and the design of interpretable network structures. Of these, interpretable network structures can help users understand the model. They not only allow observation of the model’s prediction results but also provide insights into the reasons behind the model’s decisions. In the case of model errors, users can make adjustments to the model themselves [12]. Traditional methods for network construction usually rely on iterative trial and error based on human prior knowledge and experience, which consumes a considerable amount of time. In recent years, researchers have proposed some automatic network structure search methods [13]. The work in [14] introduces a graph-based neural architecture encoding scheme (GATES), which employs an intuitive modeling approach that mirrors the data processing flow of neural networks. This enhances the representation capability of neural architectures, leading to a substantial improvement in predictor performance across various cell-based search spaces. The work in [15] proposes a block-wise network generation pipeline (BlockQNN), which employs the Q-Learning paradigm with epsilon-greedy exploration strategy to automatically construct the network structure. However, these methods have a large search space and low computational efficiency. Additionally, training neural networks using backpropagation and gradient descent suffers from issues such as weight initialization, local minima, and the sensitivity of learning performance to learning rates, which also limit the performance of deep learning models. Therefore, it is crucial to research a fast and efficient automatic construction method for a network structure that possesses interpretability.

To address issues such as the long training time, susceptibility to local minima, and the difficulty in determining the number of hidden layer nodes closely related to the learning capability in neural networks, random learning algorithms have emerged. This algorithm randomly assigns input weights and biases and calculates output weights using least squares estimation. Distinguished from the random vector functional link networks [16], stochastic configuration networks (SCNs) are featured by using a supervisory mechanism to assign the random parameters and adaptively select the scope of the random parameters, ensuring the universal approximation property of the constructed randomized learners at algorithmic level. The work in [17] introduces a two-dimensional stochastic configuration network (2DSCNs) that can directly process two-dimensional data, improving the generalization performance of SCNs in image data modeling. The work in [18] proposes a deep stochastic configuration network (DeepSCNs) where hidden nodes in each layer are connected to the output, allowing the network to learn richer feature representations. Due to the merits of the SCNs framework, it has received considerable attention in the field of industrial data analysis [19, 20, 21, 22, 23].

Although SCNs construct randomized learners with the universal approximation property, they have a limited ability to extract image features, making them inadequate for representation of interest image information. In the construction process of CNNs, the network width is usually preset, i.e., the number of convolutional kernels in each layer. Due to the possible correlation between the convolutional kernels, coupled with the backpropagation to iteratively update the parameters of the convolutional kernels, the computation load for the training of CNNs is massive [24]. The work in [25] proposes an interpretable method for recognizing working conditions in fused magnesium furnace based on deep convolutional stochastic configuration networks (DCSCNs). This approach utilizes a supervised learning mechanism to generate Gaussian differential convolution kernel with physical meaning, and uses an incremental method to construct a deep convolutional neural network to ensure that the recognition error converges step by step and avoid the process of backpropagation algorithm to iteratively find the optimal convolutional kernel parameters. Additionally, channel feature map independence coefficients and interpretable credibility indexes are defined to visualize the activation mapping maps for fused magnesium furnace feature classes. Due to possible redundant information in the associated convolutional kernels, the pruning optimization of the convolutional kernels in the network is necessary. The work in [26] proposes a pruning pattern that removes consecutive NN output kernels sharing the same input channel index, preserving model accuracy while achieving substantial speedups on general CPUs. The work in [27] proposes a channel pruning method for the compression of deep neural networks, which introduces additional loss functions then performs layer-wise channel selection to recognize channels and kernels that meaningfully contribute to the discriminative capability. Applying reinforcement learning (RL) to prune the convolutional kernels enables the end-to-end learning of the optimal compressed feature combinations, thereby reducing network size and computational costs [28].

In summary, this study addresses the shortcomings related to the generalization and interpretability of recognition models in the specific operational environments of fused magnesium furnaces. Leveraging the explainable architecture of SCNs, an interpretable recognition method of fused magnesium furnace working conditions with deep convolutional stochastic configuration networks ((DCSCNs) is proposed. The major contributions of this study are as follows:

1) To overcome the limitations of traditional incremental convolutional networks, a stochastic configuration approach is used for the first time to construct a deep CNNs. The architecture starts with a single-layer network with a single convolutional kernel and incrementally expands until it meets the stopping criteria for iterations. Physically interpretable Gaussian differential convolutional kernel parameters are automatically configured through a data-correlated parameter selection strategy. This approach eliminates the need for iterative optimization of convolutional kernel parameters based on backpropagation algorithm, and ensures that the recognition error converges step by step. A proof of global convergence capabilities for the deep CNNs is also provided, further substantiating the model interpretability.

2) The feature maps generated by DCSCNs are upsampled to the size of the input image using bilinear interpolation. After overlaying the upsampled feature maps with the original image, they are re-inputted to the DCSCNs under the current level of stochastic convolutional configuration layers to obtain scores indicating the membership of the multi-modal conditions to different categories. The independent coefficient of channel feature maps is defined to measure the independence score of different channel features. The category scores, channel feature independence scores, and corresponding channel feature maps are linearly combined to generate class activation maps. By overlaying these maps onto the original image, the feature visualization results of the fused magnesium furnace under the current level condition can be obtained. An evaluative index for interpretable trustworthiness is defined to measure the consistency between the interpretability results and the ground truth.

3) The generated Gaussian differential convolutional kernels are sorted based on the independent scores of channel features. A joint reward function is defined, which includes an interpretable trustworthiness evaluative index, recognition accuracy, and model parameter quantity. RL is employed to adaptively prune the Gaussian differential convolutional kernel set to adjust the network width, obtaining optimal interpretable condition recognition results.

The remainder of this paper is organized as follows: Section II introduces the interpretable fused magnesium furnace working condition recognition model, outlining its three core components: the training, feedback, and testing layers, and explaining their synergistic interactions to improve performance. Section III describes an interpretable working condition recognition method based on DCSCNs, incorporating data augmentation, and a novel convolutional kernel parameter selection strategy. Section IV provides a rigorous theoretical proof of the convergence of the proposed DCSCNs. Section V leverages an activation mapping method based on feature independence to assess the interpretability of working condition recognition results in the fused magnesium furnace. Section VI proposes an adaptive pruning mechanism for convolutional kernels based on RL, which optimizes network width by ranking kernels based on feature independence scores. Section VII details the data, evaluation metrics, and experimental results, covering the DCSCNs performance, the interpretability evaluation, the convolutional kernel adaptive pruning, ablation studies, and performance comparisons. The final section summarizes the contributions, strengths, and potential directions for future research.

II Interpretable fused magnesium furnace working condition recognition model

The proposed interpretable working condition recognition model for a fused magnesium furnace based on DCSCNs utilizes a three-layer structure for information coupling, including the training layer, feedback layer, and testing layer. The model structure is illustrated in Figure 1.

Refer to caption
Figure 1: Structure of interpretable fused magnesium furnace working condition recognition model based on deep convolutional stochastic configuration networks.

The training layer consists of data augmentation, preprocessing, and DCSCNs modules. Initially, fused magnesium furnace training images undergo data augmentation and preprocessing to enlarge the dataset. These processed samples are then fed into the DCSCNs. A supervised mechanism is opted for the adaptive selection of new convolutional kernel parameters to avoids the need for the iterative backpropagation for updating convolutional kernel parameters. This process starts with a single-kernel and single-layer configuration (represented as ”Incremental convolutional feature extraction layer 1”, the yellow box as the constructed convolution, and the blue box as the convolution to be constructed). The training set is passed through this layer to obtain the feature map. The feature map is then used for classifying the working conditions of the fused magnesium furnace through a fully connected layer. If the recognition error meets the preset criteria, training is halted. Otherwise, the number of single-layer convolutional kernels continues to increase one by one at a time (indicated as the second yellow box of the incremental convolutional feature extraction layer 1) until it reaches the maximum allowable kernels (C1C_{1}) for that layer. If the error still does not meet the criteria, a second incremental convolutional feature extraction layer is added (represented as ”Incremental convolutional feature extraction layer 2”). This process is repeated until the error converges or the maximum number LmaxL_{\mathrm{max}} of convolutional layers is reached.

The feedback layer encompasses an interpretability assessment module and a network width adaptive adjustment module based on RL. After the construction of Lmax1L_{\mathrm{max}}-1 layers in the DCSCNs, the extracted feature maps are fed into the feedback layer (indicated by the blue arrow). These are upsampled to the original input image size and are superimposed with the original image. The superimposed image is re-inputted into the incremental convolutional network to obtain class scores. Channel feature independence scores are calculated, and a linear combination of the class scores, channel feature independence scores, and feature maps is used to construct a class activation mapping (CAM) and a measure index of interpretability trustworthiness. Meanwhile, the Gaussian differential convolutional kernels generated are then fed into a RL pruning module (indicated by the green arrow). The kernels are ranked based on their channel feature independence scores, and a joint reward function that combines accuracy, a measure index of interpretability trustworthiness, and the volume of network parameters is formulated. Convolutional kernels for each layer of the network are adaptively pruned through RL, enabling the self-adaptive adjustment of network width (indicated by the red arrow).

In the testing layer, the constructed interpretable deep stochastic configuration convolutional network is used to obtain the optimal recognition results for the test samples of fused magnesium furnace working conditions.

III Interpretable working condition recognition method based on deep convolutional stochastic configuration networks

III-A Data augmentation and data preprocessing

In the production process of fused magnesium furnaces, there are relatively few instances of abnormal working conditions. This leads to an imbalance issue in the image samples, which can cause over-fitting of the learner. To address this problem, this paper employs a non-generative approach to augment the image data. Techniques such as horizontal flipping, contrast and brightness adjustment, and adding noise are applied to increase the diversity of the data and alleviate over-fitting, enabling the learner to better adapt to the variations in magnesium furnace images across different scenarios.

Horizontal flipping can be described as follows:

I(x,y)=I(wx1,y)I(x,y)=I^{\prime}(w-x-1,y) (1)

where, I(x,y)I^{\prime}(x,y) represents the pixel value of the original image at coordinates (x,y)(x,y), I(x,y)I(x,y) denotes the pixel value of the data enhanced image at (x,y)(x,y), and ww denotes the width of the original image.

Contrast and brightness adjustment can be described as follows:

I(x,y)=1.5×(I(x,y)0.5)+0.5I(x,y)=1.5×(I^{\prime}(x,y)-0.5)+0.5 (2)

To prevent pixel value overflow, each value in the original image is subtracted by 0.5, multiplied by the contrast enhancement coefficient of 1.5, and then added by 0.5. This simultaneously enhances the brightness and contrast of the image.

Adding Gaussian noise can be described as follows:

I(x,y)=I(x,y)+N(0,η2)I(x,y)=I^{\prime}(x,y)+N(0,\eta^{2}) (3)

where, N(0,η2)N(0,\eta^{2}) represents Gaussian noise with a mean of 0 and a standard deviation of η2\eta^{2}. By adjusting the value of η\eta, the intensity of the noise can be controlled.

The collected images may contain some information unrelated to the fused magnesium furnace. To reduce the impact of such information, the center of the image is cropped to 1080×10801080\times 1080 and resized to 256×256256\times 256. The input values are normalized to the range of [1,1][-1,1] for subsequent image processing convenience.

III-B Deep convolutional stochastic configuration networks

To address the issues of structure design, hyperparameter tuning, and interpretability in CNNs, this paper proposes an efficient method for constructing interpretable CNNs as shown in Figure 2, called the DCSCNs, inspired by the stochastic configuration network [29, 30, 25]. This strategy starts from a single convolutional kernel in a single layer and incrementally generates new stochastic convolutional kernels with strong interpretability to construct the CNNs. It overcomes the tediousness of manually designing network structures and tuning hyperparameters using traditional gradient descent methods. The supervised learning mechanism is employed in the selection of convolutional kernel parameters, ensuring the global approximation capability of the deep learning model.

Refer to caption
Figure 2: Deep convolutional stochastic configuration network structure diagram.

III-B1 Generation strategies of randomized configuration convolutional kernel

The stochastic configuration convolutional generation strategy adaptively determines the range of convolutional kernel parameters and randomly generates new kernel parameters within this range, including weights and biases. The deep neural network is constructed starting from zero convolutional kernels, and interpretable new convolutional kernels are generated using the supervised learning mechanism, ensuring the improvement of network performance. Finally, the output weights of the network are updated using the least squares method, and the generation of convolutional kernels stops when the network error is below the predefined threshold.

Let F=[f1,f2,,fm]:d1×d2×d3mF=[f_{1},f_{2},…,f_{m}]:\mathbb{R}^{d_{1}\times d_{2}\times d_{3}}\to\mathbb{R}^{m} be a set of real-valued functions, where the L2L_{2} norm is defined as:

F=(q=1mD|fq(x)2|𝑑x)1/2<\small\left\|F\right\|=\left({\textstyle\sum_{q=1}^{m}}\int_{D}\left|f_{q}(x)^{2}\right|dx\right)^{1/2}<\infty\small (4)

The inner product between the real-valued functions FF and G=[g1,g2,,gm]:d1×d2×d3mG=[g_{1},g_{2},…,g_{m}]:\mathbb{R}^{d_{1}\times d_{2}\times d_{3}}\to\mathbb{R}^{m} can be represented as:

F,G=q=1mfq,gq=q=1mDfq(x)gq(x)𝑑x\langle F,G\rangle=\sum_{q=1}^{m}\left\langle f_{q},g_{q}\right\rangle=\sum_{q=1}^{m}\int_{D}f_{q}(x)g_{q}(x)dx (5)

Given an input matrix II and a convolutional kernel KK of size ρ×n\rho\times n, the element at position (i,j)(i,j) in the output matrix SS after the cross-correlation operation is defined as:

S(i,j)=(IK)(i,j)=ρnI(i+ρ,j+n)K(ρ,n)S(i,j)=(I*K)(i,j)=\sum_{\rho}\sum_{n}I(i+\rho,j+n)K(\rho,n)\quad (6)

where, * denotes the cross-correlation operator and I(i+ρ,j+n)K(ρ,n)I(i+\rho,j+n)K(\rho,n) represents the element at position (i+ρ,j+n)(i+\rho,j+n) in the input matrix II multiplied by the element at position (ρ,n)(\rho,n) in the convolutional kernel KK.

Given a target function F:dmF:\mathbb{R}^{d}\to\mathbb{R}^{m}, assuming a DCSCNs consists of LmaxL_{\mathrm{max}} convolutional layers, each layer with C1,,Cl,,CLmax(l[1,Lmax])C_{1},\ldots,C_{l},\ldots,C_{L_{\max}}\left(l\in\left[1,L_{\max}\right]\right) convolutional kernels with a size of k×kk\times k, the DCSCNs can be represented as:

Fl(x)=l=1LmaxC=1ClOClACl(WCl,bCl,Il)F^{l}(x)=\sum_{l=1}^{L_{\max}}\sum_{C=1}^{C_{l}}O_{C}^{l}A_{C}^{l}\left(W_{C}^{l},b_{C}^{l},I^{l}\right) (7)

where, WClW_{C}^{l} and bClb_{C}^{l} are the parameters of the CthC^{th} convolutional kernel in the lthl^{th} layer, WCl=W_{C}^{l}= [WC,[1,1]lWC,[1,k]lWC,[k,1]lWC,[k,k]l]\left[\begin{array}[]{ccc}W_{C,[1,1]}^{l}&\cdots&W_{C,[1,k]}^{l}\\ \vdots&\ddots&\vdots\\ W_{C,[k,1]}^{l}&\cdots&W_{C,[k,k]}^{l}\end{array}\right], IlI^{l} is the input of the lthl^{th} layer, AClA_{C}^{l} represents the convolutional function, including convolutional operation, activation function, and downsampling operation, OClO_{C}^{l} is the output weight, and the error el=FFl=[e1l,e2l,,eml]e^{l}=F-F^{l}=\left[e_{1}^{l},e_{2}^{l},\ldots,e_{m}^{l}\right].

Assuming the function space span(Γ)\operatorname{span}(\Gamma) composed of Γ\Gamma is dense in the L2L_{2} space, AΓ\forall A\in\Gamma, 0<A<b0<\|A\|<b, where b+b\in\mathbb{R}^{+}. Given p>0p>0 and a non-negative contraction sequence {uC}\left\{u_{C}\right\}, {uC}\left\{u_{C}\right\} satisfying limC+uC=\lim_{C\rightarrow+\infty}u_{C}= 0 and C=1uC=\sum_{C=1}^{\infty}u_{C}=\infty, the parameters of the CthC^{th} convolutional kernel in the lthl^{th} layer are randomly generated if they satisfy:

eC,ql,ACl2puCb2eC1,ql2\quad\left\langle e_{C,q}^{l},A_{C}^{l}\right\rangle^{2}\geq pu_{C}b^{2}\left\|e_{C-1,q}^{l}\right\|^{2} (8)

where, q=1,2,,mq=1,2,\ldots,m. Then, the condition liml+FFl=0\lim_{l\rightarrow+\infty}\left\|F-F^{l}\right\|=0 holds; otherwise, the parameters of the CthC^{th} convolutional kernel in the lthl^{th} layer are regenerated. When the number of convolutional kernels in the lthl^{th} layer increases to ClC_{l} if the error ele^{l} is still greater than the predetermined value, a new convolutional kernel is added as the first kernel of the l+1thl+1^{th} layer, and it satisfies:

eCl,ql,A1l+12ξuClb2eCl1,ql2\left\langle e_{C_{l},q}^{l},A_{1}^{l+1}\right\rangle^{2}\geq\xi u_{C_{l}}b^{2}\left\|e_{C_{l}-1,q}^{l}\right\|^{2} (9)

Then, the condition liml+FFl=0\lim_{l\rightarrow+\infty}\left\|F-F^{l}\right\|=0; otherwise, the parameters of the first convolutional kernel in the (l+1)th(l+1)^{th} layer are regenerated until the termination condition for kernel augmentation is met.

Therefore, the construction problem of a DCSCNs can be described as follows: Given training image data X={x1,x2,,xN}X=\left\{x_{1},x_{2},\ldots,x_{N}\right\}, where xid1×d2×d3x_{i}\in\mathbb{R}^{d_{1}\times d_{2}\times d_{3}}, and the corresponding outputs Y={y1,y2,,yN}Y=\left\{y_{1},y_{2},\ldots,y_{N}\right\}, where yimy_{i}\in\mathbb{R}^{m} represents the image category label. Let ItlI_{t}^{l} denote the ttht^{th} channel of the lthl^{th} layer convolution of input data IlI^{l}, for t=1,2,,Cl1t=1,2,\ldots,C_{l-1}. The feature map MClM_{C}^{l}, representing the output of the CthC^{th} convolutional kernel in the lthl^{th} layer, can be expressed as:

MCl=g(t=1Cl1WClItl+bCl)M_{C}^{l}=g\left(\sum_{t=1}^{C_{l-1}}W_{C}^{l}*I_{t}^{l}+b_{C}^{l}\right) (10)

where, g()g(\cdot) is the activation function, and MClM_{C}^{l} has dimensions H×WH\times W. After max pooling, MClM_{C}^{l} obtains the downsampled feature map ACIA_{C}^{I}:

ACl=maxm=1k1maxn=1k1MC,Ψ,ΩlA_{C}^{l}=\max_{m=1}^{k-1}\max_{n=1}^{k-1}M_{C,\Psi,\Omega}^{l} (11)

where, Ψ[h,h+m],h=1,2,\quad\Psi\in[h,h+m],h=1,2,\ldots, HH represents the length range of the feature map, and Ω[w,w+n]\Omega\in[w,w+n] with w=1,2,,W\quad w=1,2,\ldots,W represents the width range of the feature map.

Based on Equation (11), the feature map set Al=[A1l,A2l,,ACl]A^{l}=\left[A_{1}^{l},A_{2}^{l},\ldots,A_{C}^{l}\right] of the lthl^{th} layer convolution can be obtained. Multiplying AClA_{C}^{l} by the output weights OClO_{C}^{l} of the convolutional network yields the output FlF^{l} for the CthC^{th} convolutional kernel in the lthl^{th} layer:

Fl=lC=1ClOClAClF^{l}=\sum_{l}\sum_{C=1}^{C_{l}}O_{C}^{l}A_{C}^{l} (12)

The output weights OClO_{C}^{l} are updated using the least squares method:

OCl=[OC,1l,,OC,ml]=argminOYlC=1ClOClACl\begin{gathered}O_{C}^{l}=\left[O_{C,1}^{l},\ldots,O_{C,m}^{l}\right]=\arg\min_{O}\left\|Y-\sum_{l}\sum_{C=1}^{C_{l}}O_{C}^{l}A_{C}^{l}\right\|\end{gathered} (13)

The error eCle_{C}^{l} for the CthC^{th} convolutional kernel in the lthl^{th} convolutional layer is defined as:

eCl=YFl=[eC,1l,eC,2l,,eC,ml]e_{C}^{l}=Y-F^{l}=\left[e_{C,1}^{l},e_{C,2}^{l},\ldots,e_{C,m}^{l}\right] (14)

If the L2L_{2} norm eCl2\left\|e_{C}^{l}\right\|_{2} of eCle_{C}^{l} is greater than the desired error limit e¯\bar{e}, new parameters for the convolutional kernel with weights WC+1lW_{C+1}^{l} and biases bC+1lb_{C+1}^{l} are generated until the stopping criterion is met.

III-B2 Convolutional kernel parameter selection strategy

The construction of convolutional kernels in deep learning models directly affects the correlation between the learned model and input data. In a deep convolutional network with good performance, the convolutional kernels should follow a Gaussian distribution [31]. The highlight regions of the furnace wall and furnace mouth exhibit distinct texture characteristics, with significant brightness variations at the boundaries. Gaussian differential convolutional kernels, which involve Gaussian smoothing and differencing operations on images, can enhance texture information to distinguish image regions where texture is weakened or lost due to fog occlusion. Therefore, Gaussian differential convolutional kernels are selected for the DCSCNs, defined as follows [25]:

ψ(x,y)=12π[ex2+y22ξ21rex2+y22r2ξ2]\psi(x,y)=\frac{1}{2\pi}\left[e^{-\frac{x^{2}+y^{2}}{2\xi^{2}}}-\frac{1}{r}e^{-\frac{x^{2}+y^{2}}{2r^{2}\xi^{2}}}\right] (15)

where, ξ\xi represents the standard deviation, controlling the width range of the convolutional kernel, and rr represents the scale factor. Smaller rr values can detect fine edges and details, while larger rr values can detect rough edges and contour structures. Different combinations of ξ\xi and rr yield different information about the magnesium furnace conditions.

Unlike traditional methods that iteratively update convolutional kernel parameters using gradient descent algorithms, this study adopts a strategy of randomly selecting parameters to generate different convolutional kernels, where ξ\xi and rr follow uniform distributions. The weights WClW_{C}^{l} of the convolutional kernels follow a Gaussian differential distribution as shown in Equation (15), and the biases bClb_{C}^{l} are selected from a uniform distribution as shown in the following equation:

bClU(0,1)b_{C}^{l}\sim U(0,1) (16)

By generating TmaxT_{\max} candidate Gaussian differential convolutional kernels based on Equation (15) and Equation (16), the convergence score σC,q\sigma_{C,q} of each candidate convolutional kernel can be evaluated using Equation(17):

σC,ql=((eC,ql)TAC,ql)2(AC,ql)TAC,ql(ξ+r)uC(eC,ql)TeC,ql\sigma_{C,q}^{l}=\frac{\left(\left(e_{C,q}^{l}\right)^{\mathrm{T}}A_{C,q}^{l}\right)^{2}}{\left(A_{C,q}^{l}\right)^{\mathrm{T}}A_{C,q}^{l}}-(\xi+r)u_{C}\left(e_{C,q}^{l}\right)^{\mathrm{T}}e_{C,q}^{l} (17)

where, AC,ql=[AC,1l,AC,2l,,AC,ml,]A_{C,q}^{l}=\left[A_{C,1}^{l},A_{C,2}^{l},\cdots,A_{C,m}^{l},\right]. Candidate Gaussian differential convolutional kernels satisfying Equation (18) are retained:

q=1mσC,ql>0\quad\sum_{q=1}^{m}\sigma_{C,q}^{l}>0\quad (18)

The candidate Gaussian differential convolutional kernel with the highest convergence score q=1mσC,ql\sum_{q=1}^{m}\sigma_{C,q}^{l} is selected as the CthC^{th} convolutional kernel in the lthl^{th} layer. If none of the TmaxT_{\max} candidate Gaussian differential convolutional kernels meet the requirements, new ξ\xi and rr values are selected based on Equation (15) to generate a new Gaussian differential convolutional kernel. This process continues until a convolutional kernel in the lthl^{th} layer satisfies the condition in Equation (18).

III-B3 Proof of convergence of deep convolutional stochastic configuration networks

Assuming that the number of convolutional kernels in the lthl^{th} layer of DCSCNs is ClC_{l}, it can be derived that the lthl^{th} layer convolution satisfies [25]:

eCll2(ξ+r)uCleCl1l2\left\|e_{C_{l}}^{l}\right\|^{2}\leq(\xi+r)u_{C_{l}}\left\|e_{C_{l-1}}^{l}\right\|^{2} (19)

The output error e1l+1e_{1}^{l+1} of the first Gaussian differential convolutional kernel generated in the l+1thl+1^{th} layer convolution is given by:

e1l+12=eCllO1l+1A1l+12\quad\left\|e_{1}^{l+1}\right\|^{2}=\left\|e_{C_{l}}^{l}-O_{1}^{l+1}A_{1}^{l+1}\right\|^{2} (20)

where, O1l+1=[O1,1l,O1,2l,,O1,ml]O_{1}^{l+1}=\left[O_{1,1}^{l},O_{1,2}^{l},\ldots,O_{1,m}^{l}\right]. Let O¯1l+1=\bar{O}_{1}^{l+1}= O¯1,1l,O¯1,2l,,O¯1,ml\left\lceil\bar{O}_{1,1}^{l},\bar{O}_{1,2}^{l},\ldots,\bar{O}_{1,m}^{l}\right\rceil represent the optimal output weights. Then O¯1,ql+1\bar{O}_{1,q}^{l+1} can be expressed as:

O¯1,ql+1=eCl,ql,A1l+1A1l+12\bar{O}_{1,q}^{l+1}=\frac{\left\langle e_{C_{l},q}^{l},A_{1}^{l+1}\right\rangle}{\left\|A_{1}^{l+1}\right\|^{2}} (21)

By transforming e1l+12\left\|e_{1}^{l+1}\right\|^{2} using Equations (18-21), we have:

e1l+12eCllO¯1l+1A1l+12=eCll2q=1mO¯1,ql+1eCll2\begin{array}[]{r}\left\|e_{1}^{l+1}\right\|^{2}\leq\left\|e_{C_{l}}^{l}-\bar{O}_{1}^{l+1}A_{1}^{l+1}\right\|^{2}\\ =\left\|e_{C_{l}}^{l}\right\|^{2}-\sum_{q=1}^{m}\bar{O}_{1,q}^{l+1}\leq\left\|e_{C_{l}}^{l}\right\|^{2}\\ \end{array} (22)

Thus

eCl+1l+12e1l+12eCll2(ξ+r)u1(lCl)1e112\begin{array}[]{r}\left\|e_{C_{l+1}}^{l+1}\right\|^{2}\leq\left\|e_{1}^{l+1}\right\|^{2}\leq\left\|e_{C_{l}}^{l}\right\|^{2}\\ \leq(\xi+r)u_{1}^{\left(\sum_{l}C_{l}\right)-1}\left\|e_{1}^{1}\right\|^{2}\end{array} (23)

From this, it can be proven that the output error e1l+1e_{1}^{l+1} of the DCSCNs monotonically decreases, indicating the convergence of the network.

III-C Interpretability evaluation

To ensure the reliability and interpretability of the recognition results for working conditions in the fused magnesium furnace, it is important for operators to understand the trustworthy decision-making process of deep learning models. The DCSCNs ensures global convergence to obtain reliable recognition results [25]. The Gaussian distribution convolutional kernels generated based on supervised learning have a strong correlation with the input data, ensuring the interpretability of the model mechanism. Class activation mapping visualization, as a visual interpretability verification, represents the distribution of importance in the data. It utilizes the gradient information of the last convolutional layer as the category score for channel feature maps and calculates the category importance by weighting the feature maps. Finally, the feature activation mapping is superimposed on the original image to obtain the interpretable visualization of feature data as shown in Figure 3.

Refer to caption
Figure 3: Schematic diagram of the class activation mapping based on feature map independence scores.

The feature maps set AlA^{l} in the lthl^{th} layer is upsampled to the original input size using bilinear interpolation and normalized to the range [0,1][0,1]. The ϱth\varrho^{th} channel feature map AϱlA_{\varrho}^{l} from AlA^{l}, when overlaid with the original image xix_{i}, can be denoted as A¯ϱl\bar{A}_{\varrho}^{l} :

A¯ϱl=Aϱlxi\bar{A}_{\varrho}^{l}=A_{\varrho}^{l}\otimes x_{i} (24)

where, \otimes represents element-wise multiplication. A¯ϱl\bar{A}_{\varrho}^{l} is then fed into a DCSCNs model Δ\Delta, and the softmax function produces category scores SqlS_{q}^{l} :

Sql=Softmax(Δ(A¯ϱl))S_{q}^{l}=\operatorname{Softmax}\left(\Delta\left(\bar{A}_{\varrho}^{l}\right)\right) (25)

The DCSCNs model Δ\Delta predicts the category result yqy^{q} for xix_{i}, where q=[1,,m]q=[1,\cdots,m], and the category score for the ϱth\varrho^{th} channel feature map is denoted as Sϱ,qlS_{\varrho,q}^{l}. The definition of the channel feature independence coefficient is as follows:

FCϱl=AlΞϱlAlAlFC_{\varrho}^{l}=\frac{\left\|A^{l}\right\|_{*}-\left\|\Xi_{\varrho}^{l}\odot A^{l}\right\|_{*}}{\left\|A^{l}\right\|_{*}} (26)

where, FCϱlFC_{\varrho}^{l} represents the independence score of each channel feature map, AlH×W×ClA^{l}\in\mathbb{R}^{H\times W\times C_{l}}, and AϱlH×WA_{\varrho}^{l}\in\mathbb{R}^{H\times W} represents the ϱ\varrho, ϱ[1,Cl]\varrho\in\left[1,C_{l}\right] feature channel. \|\cdot\|_{*} represents the nuclear norm, \odot represents the Hadamard product, and Ξϱl\Xi_{\varrho}^{l} represents the convolutional kernel mask matrix, where the ϱth\varrho^{th} row is zero and the other rows are 1. By multiplying all the feature maps in AlA^{l} by Sϱ,qlS_{\varrho,q}^{l} and FCϱlFC_{\varrho}^{l}, and then summing them, the CAM for sample xix_{i} under the lthl^{th} layer DCSCNs model can be obtained:

q=ReLU(ϱFCϱlSϱ,qlAϱl)\mathcal{L}^{q}=\operatorname{ReLU}\left(\sum_{\varrho}FC_{\varrho}^{l}S_{\varrho,q}^{l}A_{\varrho}^{l}\right) (27)

Based on q\mathcal{L}^{q}, the interpretable trustworthiness evaluation index can be defined to measure the deviation between the predicted target and the true target, determining whether the interpretability result is consistent with the true result. The interpretable trustworthiness evaluation index for the ithi^{th} sample is defined as the intersection over union (IoU) between the highlighted region did_{i} of the interpretability result and the true annotated region d¯i\bar{d}_{i}:

IoUi=did¯idid¯i{IoU}_{i}=\frac{d_{i}\cap\bar{d}_{i}}{d_{i}\cup\bar{d}_{i}} (28)

where, did¯id_{i}\cap\bar{d}_{i} represents the area of intersection between did_{i} and d¯i\bar{d}_{i}, and did¯id_{i}\cup\bar{d}_{i} represents the area of union between did_{i} and d¯i\bar{d}_{i}. The interpretable trustworthiness evaluation index IoUl{IoU}^{l} for the lthl^{th} layer network can be obtained from IoUi{IoU}_{i} :

IoUl=i=1NIoUiN{IoU}^{l}=\frac{\sum_{i=1}^{N}{IoU}_{i}}{N} (29)

where, NN represents the number of samples.

III-D Adaptive pruning mechanism for convolutional kernels based on RL

During the construction process of DCSCNs, the maximum number of convolutional kernels in each layer network is manually set. However, the correlations between these kernels may introduce redundant features [32]. These features not only increase the complexity of the network but also reduce the computational efficiency. Therefore, the independent scores of the channel feature maps are used to evaluate the independence of feature map corresponding to each convolutional kernel. When a feature map has low independence with other channel feature maps, it means that its information has been largely encoded in other feature maps. Based on the independent scores of each channel feature map, the convolutional kernels can be sorted according to their scores. To enable the deep learning model to learn sparse representations of features, the DDPG (deep deterministic policy gradient) RL algorithm, as shown in Figure 4, is adopted for the adaptive pruning of the number of convolutional kernels [33]. A joint reward function based on recognition accuracy, interpretability reliability evaluation index, and parameter count is defined to balance the accuracy and interpretability reliability of the deep learning model. Through RL, the network width is adaptively adjusted to select feature maps with low correlations, thus improving the model’s inference speed and reducing the risk of over-fitting. The settings of actions, states, rewards, and training strategies in the deep RL model are as follows:

Refer to caption
Figure 4: Structure of convolutional kernel adaptive pruning based on feature map independence.

Action Set: A continuous action set A={a}A=\{a\} represents the pruning ratio of each layer convolutional kernels, where a(0,1)a\in(0,1).

State Set: The state set of the lthl^{th} layer convolutional layer is denoted as S={l,Cl,Cl1,h,w,al1}S=\left\{l,C_{l},C_{l-1},h,w,a_{l-1}\right\}, where ll represents the current convolutional layer, ClC_{l} represents the number of convolutional kernels in the current layer, Cl1C_{l-1} represents the number of convolutional kernels in the previous layer, hh represents the length of the input feature map, ww represents the width of the input feature map, and al1a_{l-1} represents the action of the previous layer.

Reward Function: To balance feature independence, model accuracy, and interpretable trustworthiness, the joint reward function is defined as follows:

R=ACCval+IoUvalβ×PAR=ACC_{val}+IoU_{val}-\beta\times PA (30)

where, ACCvalACC_{val} is the accuracy of the pruned model on the validation set, IoUvalIoU_{val} is the interpretable trustworthiness evaluation index of the final layer feature map in the pruned model on the validation set, PAPA is the parameter count after pruning, and β\beta is the weight coefficient.

Training Strategy: The DDPG RL algorithm is used, where the actor network performs action selection, including the policy network μ\mu and the target policy network μ\mu^{\prime}. The critic network evaluates values, including the value network QQ and the target value network QQ^{\prime}. The objective function of the critic value network is:

Loss=1NtN(ytQ(st,atω))2Loss=\frac{1}{N}\sum_{t}^{N}\left(y_{t}-Q\left(s_{t},a_{t}\mid\omega\right)\right)^{2} (31)

where, NN is the number of samples, Q(st,atω)Q\left(s_{t},a_{t}\mid\omega\right) is the value of evaluating state sts_{t} and action ata_{t} in the critic value network with parameters ω\omega, yty_{t} is the target value:

yt=Rt+γQ(st+1,μ(st+1ω)ω)y_{t}=R_{t}+\gamma Q^{\prime}\left(s_{t+1},\mu^{\prime}\left(s_{t+1}\mid\omega\right)\mid\omega^{\prime}\right) (32)

where, μ(st+1ω)\mu^{\prime}\left(s_{t+1}\mid\omega\right) is the action prediction for state st+1s_{t+1} at time t+1t+1 using the actor target policy network, ω\omega^{\prime} is the parameter of QQ^{\prime}, and γ\gamma is the discount factor.

The actor policy network is updated using policy gradients:

θμJ1Ntθμμ(sθμ)|s=staQ(s,aω)|s=s,a=μ(st)\begin{gathered}\quad\nabla_{\theta^{\mu}}J\approx\\ \left.\left.\frac{1}{N}\sum_{t}\nabla_{\theta^{\mu}}\mu(s\mid\theta^{\mu})\right|_{s=s_{t}}\nabla_{a}Q(s,a\mid\omega)\right|_{s=s,a=\mu\left(s_{t}\right)}\end{gathered} (33)

where θμ\theta^{\mu} is the parameter of the policy network μ\mu.

Update the target policy network μ\mu^{\prime} and the target value network Q:Q^{\prime}:

θμτθμ+(1τ)θμ\theta^{\mu^{\prime}}\leftarrow\tau\theta^{\mu}+(1-\tau)\theta^{\mu^{\prime}} (34)
ωτω+(1τ)ω\omega^{\prime}\leftarrow\tau\omega+(1-\tau)\omega^{\prime} (35)

where, τ\tau is the learning rate, τ(0,1)\tau\in(0,1).

IV Experimental analysis

IV-A Data description and processing

To verify the effectiveness of the proposed method, production videos were selected from an fused magnesium furnace factory in Liaoning Province. The videos were frame-separated to obtain images with a resolution of 1080×19201080\times 1920. After image data augmentation, a total of 12,000 image samples were obtained. Figures 5 to 8 show the partial results of data augmentation for normal, underburn, overheated, and abnormal exhaust working conditions, respectively. The working conditions of the fused magnesium furnace images were annotated by experts. Randomly, 60% of the data (7,200 images in total) were selected as the training set, with each of the four condition samples accounting for a quarter of the training set. Additionally, 20% of the data (2,400 images in total) were chosen as the validation set, and the remaining 20% (2,400 samples) were used as the test set. The experimental setup included an Intel i9-10900K processor, 16GB of memory, and an RTX 3060 graphics card.

Refer to caption
Figure 5: Results after image data enhancement for normal working conditions.
Refer to caption
Figure 6: Results after image data enhancement for underburn working conditions.
Refer to caption
Figure 7: Results after image data enhancement for overheating working conditions.
Refer to caption
Figure 8: Results after image data enhancement for abnormal exhaust working conditions.

IV-B Performance evaluation metrics

The accuracy (Ra)(R_{a}) is defined as the ratio of the number of correctly classified samples to the total number of samples, and it is calculated as follows:

Ra=TP+TNNR_{a}=\frac{TP+TN}{N} (36)

where, TPTP represents the number of samples that are truly positive and are correctly predicted as positive, and TNTN represents the number of samples that are truly negative and are correctly predicted as negative.

The parameter count (PA)(PA) measures the memory size occupied by the model, which is the sum of the parameter count of the convolutional layers (CA)(CA) and the fully connected layers (FA)(FA), and it is calculated as follows:

PA\displaystyle PA =i=1Lmax(CAl+FAl)\displaystyle=\sum_{i=1}^{L_{\max}}\left(CA_{l}+FA_{l}\right) (37)
=i=1Lmax(k×k×Cl1+1)×Cl+(min+1)×m\displaystyle=\sum_{i=1}^{L_{\max}}\left(k\times k\times C_{l-1}+1\right)\times C_{l}+\left(m_{in}+1\right)\times m

where, Cl1C_{l-1} is the number of convolutional kernels in the l1thl-1^{th} layer, ClC_{l} is the number of convolutional kernels in the lthl^{th} layer, minm_{\text{in}} is the number of input neurons in the fully connected layer, and mm is the dimension of the network output.

IV-C Experiment results and analysis

IV-C1 Experiment results of the deep convolutional stochastic configuration networks

The parameter settings for the DCSCNs are as follows: the preferred error limit is e¯=0.01\bar{e}=0.01, the maximum number of candidate Gaussian differential convolutional kernels is Tmax=100T_{\max}=100, the standard deviation ξ\xi ranges from 0.5 to 5 , the scale factor rr ranges from 0.8 to 1.5 , and the sizes of the Gaussian differential convolutional kernels is k={3,5,7}k=\{3,5,7\}. The maximum number of convolutional layers in the network is Lmax=10L_{\max}=10, the maximum number of convolutional kernels per layer is CLmax =50C_{L_{\text{max }}}=50, a sigmoid activation function is used, the pooling layer adopts the maximum pooling method with a kernel size of kp=2k_{p}=2, and a non-negative contraction sequence is defined as uC=1Cu_{C}=\frac{1}{C}, where C{C} ranges from 1 to CmaxC_{\max}. A pooling layer is inserted after every two convolutional layers.

Figure 9 shows the recognition accuracy curves of the training and testing samples of the fused magnesium furnace under different sizes of convolutional kernels when the network has only one convolutional layer (l=1)(l=1). It can be observed that during the training and testing processes, as convolutional kernels with different sizes and meaningful configurations are generated one by one, the DCSCNs model exhibits rapid convergence. After 50 convolutional operations, the model’s accuracy reaches a plateau. When the number of convolutional kernels is too small, the network fails to extract sufficient information. On the other hand, as the number of convolutional kernels continues to increase, the improvement in network performance becomes limited while the computational complexity increases. Therefore, this study sets the maximum number of convolutional kernels per layer to 50. Since the performance of a single convolutional layer is limited, multiple convolutional layers are necessary to extract abstract features and achieve the desired classification performance. Additionally, while different sizes of convolutional kernels can achieve similar model performance, they also bring additional computational complexity. Two 3×\times3 convolutional kernels have the same receptive field as one 5×\times5 convolutional kernel, but the two 3×\times3 kernels can perform two nonlinear transformations, providing better nonlinear transformation capability with lower computational complexity. Therefore, a kernel size of 3 can better balance model performance and complexity.

[Uncaptioned image]
Refer to caption
Figure 9: Single-layer DCSCNs recognition accuracy curves with different convolutional kernel sizes.

By fixing the kernel size to k=3k=3, Figure 10 shows the recognition accuracy curves of the training and testing sets for an 8-layer DCSCNs model. It can be observed that the multi-layer network exhibits better convergence compared to a single-layer network, with an accuracy of 94.53% on the training set and 92.78% on the testing set. The multi-layer DCSCNs achieves an improvement of 14.78% and 19.89% in accuracy compared to the single-layer DCSCNs on the training and testing sets, respectively, while the training time increases by 73.21%.

Refer to caption
Figure 10: Recognition accuracy curve of multi-layer DCSCNs network.

IV-C2 Results of the interpretability evaluation experiment

Figure 11 depicts the class activation mapping for the four working conditions in the deep convolutinal stochastic configuration network, with varying numbers of convolutional layers. The highlighted regions of the figure denotes the activation region, which represents the area of focus for this convolutional layer. When l=1l=1, the highlighted regions are small and dispersed, indicating that the network is unable to effectively focus on the target region. This suggests that the network is unable to extract sufficient feature information at this number of layers. When l=4l=4, the highlighted region becomes larger and is in closer proximity to the target region, indicating an improvement in feature extraction. When ll is further increased to 8, the highlighted region encompasses the entire target region, indicating that as the number of convolutional layers increases and the amount of computation increases, the network is able to extract more critical and complete feature information. Therefore, in this paper, ll is set to 8.

The class activation mapping maps for the four working conditions under l=8l=8 are presented in Fig. 12(a)-12(d). It can be observed that the highlighted regions with the various working conditions exhibit complete and continuous characteristics, allowing for the clear recognition of the target region. This suggests that when ll is 8, the DCSCNs are capable of adapting to the image data of different working conditions, effectively extracting the key features of each working condition, and accurately locating the target region of interest. Specifically, under normal working conditions, the network focuses on the position of the flame at the furnace opening. Under under-burning working conditions, the network is able to focus on the position of the furnace wall burning red and the furnace opening. Under overheating working conditions, the network accurately locates the position of the flame at the furnace opening. Under abnormal exhaust working conditions, the network is able to recognise the region where the solution overflows from the furnace opening. In conclusion, the proposed model provides reliable technical support for practical species for fused magnesium furnace working condition recognition.

Refer to caption
Figure 11: Class activation mapping for different working conditions with different number of convolutional layers
Refer to caption
Figure 12: Class activation mapping for different working conditions.

IV-C3 Convolutional kernel adaptive pruning experimental results

The parameter settings for the deep RL model are as follows: both the actor and critic networks are two-layer fully connected networks with 300 neurons, the discount factor γ\gamma is set to 0.9, the experience replay pool size is 2000, the learning rate γ\gamma is 0.005, and the training consists of 400 epochs.

Figure 13 shows the average reward curve after the Gaussian differential convolutional kernel set is fed into the RL module in the 8-layer DCSCNs. From this figure, it can be observed that after 300 training rounds, the average reward value stabilizes around 1.55, indicating the convergence of the RL module and the selection of a convolutional kernel set that balances model accuracy, interpretability, and parameter size.

Refer to caption
Figure 13: The average reward curve of the RL training process.

To further demonstrate the advantages of the proposed algorithm compared to the feature selection algorithm without RL, a comparative analysis of the interpretability of the models with and without the RL module is presented in Figure 14. From the figure, it can be observed that the proposed algorithm achieves more accurate and clear interpretability compared to the algorithm without RL. Specifically, the algorithm without RL exhibits unnecessary highlights in other parts of the image, despite having high attention on the flame of the furnace wall. These redundant highlights negatively impact the final decision, reducing the reliability and accuracy of the algorithm. In contrast, the proposed algorithm can accurately highlight the anomalous regions with higher precision and concentration, while reducing noise. This indicates that the features selected through RL exhibit higher independence and effectively eliminate redundant features, resulting in a more accurate and reliable algorithm. This advantage not only enhances the algorithm performance but also makes the internal workings of the algorithm more transparent and easier to understand, meeting the requirement for interpretability in industrial production.

Refer to caption
Figure 14: Comparison of the algorithm proposed in this paper with the class activation mapping without RL algorithm.

IV-D Ablation Experiments

To validate the effectiveness of each module proposed in this study, a comparative analysis is conducted against three variations: without Gaussian convolutional kernels, without the adaptive kernel pruning mechanism, and without both Gaussian convolutional kernels and the adaptive kernel pruning mechanism. An 8-layer DCSCNs is employed for these experiments. The results are presented in Table 1.

TABLE I: Ablation experiment results
Model Training RaR_{a}/% Testing RaR_{a}/% PAPA/MB Training time/s/s Inference time/s/s
Proposed method 93.96 92.57 48.85 19338.943 0.014
Without Gaussian convolutional kernels 93.04 91.86 48.85 17157.832 0.014
Without adaptive kernel pruning mechanisms 94.53 92.78 80.25 31475.854 0.019
Without Gaussian convolutional kernels
and adaptive kernel pruning mechanisms 93.38 92.01 80.25 23656.331 0.016

As can be inferred from Table 1, the proposed algorithm shows improvements in both training and testing accuracy rates compared to the methods without Gaussian convolutional kernels and those also without the adaptive kernel pruning mechanism module. Specifically, the training and testing accuracy rates are enhanced by 0.92%, 0.58%, 0.71%, and 0.56%, respectively. Furthermore, compared to the method lacking the adaptive kernel pruning mechanism, our approach reduced the parameter size by 31.4MB, and the training and inference time are decreased by 12136.911s and 0.005s, respectively. Therefore, the utilization of meaningful Gaussian differential convolutional kernels enables the model to better adapt to the complex scenarios encountered in fused magnesium furnace working conditions, thereby ensuring continual improvements in accuracy. CAM is employed for visualizing feature importance, and the interpretable trustworthiness index is defined, thereby making the working condition recognition more accurate and trustworthy. Moreover, the adaptive pruning mechanism proposed in this algorithm manages to reduce the loss in accuracy within an acceptable range, effectively reducing the parameter size and training time, and subsequently enhancing inference speed.

IV-E Performance Comparison Experiment

To validate the effectiveness of the proposed algorithm, it is compared with the performance of the SCNs [29], 2DSCNs [34], DeepSCNs [18], CNNs [35], Bayesian Network, and CNNs+LSTM working condition recognition models. For SCNs and 2DSCNs, the number of hidden layers is set to 1 and the number of nodes in the hidden layer is set to 2000. For DeepSCNs, the number of hidden layers is 4, and each hidden layer has 500 nodes. The activation function used is sigmoid. The range of the hidden layer node parameters λ\lambda is {1,3,5,7,9,10,25,50,100}\left\{1,3,5,7,9,10,25,50,100\right\}, ε\varepsilon is set to 0.01, and the contraction sequence rr is {0.9,0.99,0.999,0.9999,0.99999,0.999999}\left\{0.9,0.99,0.999,0.9999,0.99999,0.999999\right\}. The CNNs model consists of 8 convolutional layers, 8 sigmoid activation layers, 4 pooling layers, and 1 fully connected layer, and it is trained for 100 epochs.

Figure 15 shows the training recognition accuracy curves of different models. From the figure, it can be observed that the training recognition accuracy curves of the four models converge as the number of convolutional kernels/hidden layer nodes increases. SCNs, 2DSCNs, and DeepSCNs converge slowly around 2000 hidden layer nodes with recognition accuracies of approximately 0.77, 0.79, and 0.84, respectively. In contrast, the proposed method exhibits more dramatic changes in recognition accuracy, reaching 0.94 after 400 convolutional operations.

Refer to caption
Figure 15: Recognition accuracy curves of training samples for different network models.

Table 2 presents a comparison of the testing recognition accuracy, parameter size, training time, and inference time between the proposed algorithm and the other randomized methods. The proposed algorithm achieves the highest accuracy while maintaining the smallest number of parameters compared to SCNs, 2DSCNs, and DeepSCNs. In terms of training time, the multi-layer models of DCSCNs and DeepSCNs require more training time compared to the single-layer models of SCNs and 2DSCNs, while the inference time is at a similar order of magnitude. These findings suggest that the hidden layer nodes in SCNs, 2DSCN, and DeepSCNs exhibit constrained capabilities in feature extraction when processing complex image data directly. The proposed algorithm effectively extracts features of different working conditions in the fused magnesium furnace, thereby ensuring the convergence of recognition errors. This markedly enhances the accuracy of recognition outcomes, diminishes complexity and mitigates the risk of model overfitting. With regard to parameter size, the advantage of parameter sharing in the convolutional operations of the proposed algorithm, in conjunction with multiple convolutions and downsampling operations in deep networks, results in a reduction of the size of the feature maps input to the fully connected layer by one to two orders of magnitude compared to SCNs, 2DSCNs, and DeepSCNs. Furthermore, the proposed method employs three-channel colour images as input and replaces the hidden layer nodes with convolutional operations, resulting in a longer training time in comparison to SCN, 2DSCN, and DeepSCNs, which utilize grayscale images as input.

TABLE II: Performance comparison of different randomized methods
Model Testing RaR_{a}/% PAPA/MB Training time/s/s Inference time/s/s
SCNs 76.14 500.038 10278.834 0.011
2DSCNs 77.99 1000.038 12352.771 0.013
DeepSCNs 83.64 127.899 15411.081 0.013
Proposed method 92.57 12.854 18218.021 0.015

Moreover, Table 3 provides a comparative analysis of between the proposed algorithm and other fused magnesium furnace working condition recognition methods, including CNNs, Bayesian networks, Swin Transformer, and CNNs+LSTM. The proposed algorithm achieves accuracy improvements of 2.22%, 2.65%, 3.81%, and 3.00% over these methods, respectively. However, it exhibits an increase in parameter size by 12.136 MB, 12.769 MB, and 8.727 MB compared to CNNs, Swin Transformer, and CNNs+LSTM, respectively. The proposed method has the shortest training time, while its inference time remains comparable to other methods. This demonstrates that the proposed methods, which employ a randomized configuration of Gaussian differential convolutional kernel parameters with physical meanings based on a supervised learning mechanism, are capable of effectively extracting the features of different working conditions of a fused magnesium furnace. Furthermore, the definition of an interpretable trustworthiness index and the use of adaptive width adjustment through RL facilitate the attainment of optimal working conditions, thereby enhancing the recognition results. Since each feature map collection is connected to the output layer, the accuracy is enhanced, although the number of parameters is higher compared to CNN, Swin Transformer, and CNN+LSTM. Unlike these methods, which are trained using the gradient descent backpropagation method, the proposed approach utilizes an incremental convolutional kernel generation process driven by supervised learning. This approach effectively circumvents the potential issues associated with weight initialization, local minimums, and sensitivity to learning rate, significantly reducing the training time.

TABLE III: Performance comparison of different fused magnesium furnace working condition recognition methods
Model Testing RaR_{a}/% PAPA/MB Training time/s/s Inference time/s/s
CNNs 90.35 0.718 20714.322 0.015
Bayesian network 89.92 - - -
Swin transformer 88.76 0.085 31123.652 0.016
CNNs+LSTM 89.57 4.127 20159.642 0.015
Proposed method 92.57 12.854 18218.021 0.015

V Conclusion

To address the issues of poor generalization and weak interpretability in existing methods for working condition recognition in fused magnesium furnaces, this paper proposes an interpretable working condition recognition model. The innovations of this model are as follows:

1) Based on a supervised learning mechanism, an incremental stochastic configuration strategy is adopted to generate Gaussian differential convolutional kernels with physical meanings, which are used to construct a deep convolutional network. This ensures that the recognition errors converge gradually and the model exhibits transparency and interpretability in its network structure.

2) The class activation mapping method is employed to analyze the interpretability of the model, recognizing the furnace features that require attention. Interpretable trustworthiness evaluation metrics are defined, a joint reward function is constructed, and RL is used for the adaptive pruning of the convolutional kernels to obtain the optimal network width.

3) The proposed method achieves a test recognition accuracy of 92.57% for working condition in fused magnesium furnaces, outperforming other recognition methods.

In the future, stochastic configuration machines (SCMs) [36] will be employed to further improve the modeling quality in terms of the models’s light weight and interpretability.

References

  • [1] Q. Liu, Y. Zhang, G. Wu, and Z. Fan, “Disturbance robust abnormality diagnosis of fused magnesium furnaces using deep neural networks,” IEEE Transactions on Artificial Intelligence, vol. 4, no. 4, pp. 669–678, 2023.
  • [2] T. Chai, J. Zhang, and T. Yang, “Demand forecasting of the fused magnesia smelting process with system identification and deep learning,” IEEE Transactions on Industrial Informatics, vol. 17, no. 12, pp. 8387–8396, 2021.
  • [3] J. Yang, T. Chai, C. Luo, and W. Yu, “Intelligent demand forecasting of smelting process using data-driven and mechanism model,” IEEE Transactions on Industrial Electronics, vol. 66, no. 12, pp. 9745–9755, 2019.
  • [4] Z. Wu, Y. Wu, T. Chai, and J. Sun, “Data-driven abnormal condition identification and self-healing control system for fused magnesium furnace,” IEEE Transactions on Industrial Electronics, vol. 62, no. 3, pp. 1703–1715, 2015.
  • [5] W. Wang, T. Chai, H. Wang, and Z. Wu, “Signal-compensation-based adaptive PID control for fused magnesia smelting processes,” IEEE Transactions on Industrial Electronics, vol. 70, no. 9, pp. 9441–9451, 2023.
  • [6] H. Li, F. Wang, and H. Li, “Abnormal condition identification and safe control scheme for the electro-fused magnesia smelting process,” ISA Transactions, vol. 76, pp. 178–187, 2018.
  • [7] S. Lu and Y. Wen, “Semi-supervised condition monitoring and visualization of fused magnesium furnace,” IEEE Transactions on Automation Science and Engineering, vol. 19, no. 4, pp. 3471–3482, 2022.
  • [8] G. Wu, Q. Liu, and T. Chai, “Abnormal condition diagnosis through deep learning of image sequences for fused magnesium furnaces,” Acta Automatica Sinica, vol. 45, no. 8, pp. 1475–1485, 2019.
  • [9] K. Bu, L. Yan, and F. Wang, “Operating performance assessment based on multi-source heterogeneous information with deep learning for smelting process of electro-fused magnesium furnace,” ISA Transactions, vol. 128, pp. 357–371, 2022.
  • [10] S. Kim and D. Chae, “What does a model really look at?: Extracting model-oriented concepts for explaining deep neural networks,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 46, no. 7, pp. 4612–4624, 2024.
  • [11] W. Samek, A. Binder, G. Montavon, S. Lapuschkin, and K. Müller, “Evaluating the visualization of what a deep neural network has learned,” IEEE Transactions on Neural Networks and Learning Systems, vol. 28, no. 11, pp. 2660–2673, 2017.
  • [12] W. Chen, X. Gong, J. Wu, Y. Wei, H. Shi, Z. Yan, Y. Yang, and Z. Wang, “Understanding and accelerating neural architecture search with training-free and theory-grounded metrics,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 46, no. 2, pp. 749–763, 2024.
  • [13] H. Niknazar and S. C. Mednick, “A multi-level interpretable sleep stage scoring system by infusing experts’knowledge into a deep network architecture,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 46, no. 7, pp. 5044–5061, 2024.
  • [14] X. Ning, Y. Zheng, Z. Zhou, T. Zhao, H. Yang, and Y. Wang, “A generic graph-based neural architecture encoding scheme with multifaceted information,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 45, no. 7, pp. 7955–7969, 2023.
  • [15] Z. Zhong, Z. Yang, B. Deng, J. Yan, W. Wu, J. Shao, and C. Liu, “BlockQNN: Efficient block-wise neural network architecture generation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 43, no. 7, pp. 2314–2328, 2021.
  • [16] Y. H. Pao and Y. Takefuji, “Functional-link net computing: Theory, system architecture, and functionalities,” Computer, vol. 25, no. 5, pp. 76–79, 1992.
  • [17] M. Li and D. Wang, “2-D stochastic configuration networks for image data analytics,” IEEE Transactions on Cybernetics, vol. 51, no. 1, pp. 359–372, 2021.
  • [18] D. Wang and M. Li, “Deep stochastic configuration networks with universal approximation property,” in 2018 International Joint Conference on Neural Networks (IJCNN), 2018, pp. 1–8.
  • [19] K. Li, J. Qiao, and D. Wang, “Online self-learning stochastic configuration networks for nonstationary data stream analysis,” IEEE Transactions on Industrial Informatics, vol. 20, no. 3, pp. 3222–3231, 2024.
  • [20] D. Wang, P. Tian, W. Dai, and G. Yu, “Predicting particle size of copper ore grinding with stochastic configuration networks,” IEEE Transactions on Industrial Informatics, vol. 20, no. 11, pp. 12 969–12 978, 2024.
  • [21] K. Li, J. Qiao, and D. Wang, “Fuzzy stochastic configuration networks for nonlinear system modeling,” IEEE Transactions on Fuzzy Systems, vol. 32, no. 3, pp. 948–957, 2024.
  • [22] J. Lu, J. Ding, X. Dai, and T. Chai, “Ensemble stochastic configuration networks for estimating prediction intervals: A simultaneous robust training algorithm and its application,” IEEE Transactions on Neural Networks and Learning Systems, vol. 31, no. 12, pp. 5426–5440, 2020.
  • [23] J. Lu, J. Ding, C. Liu, and T. Chai, “Hierarchical-bayesian-based sparse stochastic configuration networks for construction of prediction intervals,” IEEE Transactions on Neural Networks and Learning Systems, vol. 33, no. 8, pp. 3560–3571, 2022.
  • [24] Y. LeCun, Y. Bengio, and G. Hinton, “Deep learning,” Nature, vol. 521, no. 7553, pp. 436–444, 2015.
  • [25] W. Li, Q. Tong, D. Wang, and G. Wu, “Research on working condition recognition method of fmf based on deep convolutional stochastic configuration network,” Acta Automatica Sinica, vol. 50, no. 3, pp. 527–544, 2024.
  • [26] M. Lin, Y. Zhang, Y. Li, B. Chen, F. Chao, M. Wang, S. Li, Y. Tian, and R. Ji, “1xN pattern for pruning convolutional neural networks,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 45, no. 4, pp. 3999–4008, 2023.
  • [27] J. Liu, B. Zhuang, Z. Zhuang, Y. Guo, J. Huang, J. Zhu, and M. Tan, “Discrimination-aware network pruning for deep model compression,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 44, no. 8, pp. 4035–4051, 2022.
  • [28] K. Chung, C. Lee, Y. Tsang, C. Wu, and A. Asadipour, “Multi-objective evolutionary architectural pruning of deep convolutional neural networks with weights inheritance,” Information Sciences, vol. 685, no. 121265, pp. 1–18, 2024.
  • [29] D. Wang and M. Li, “Stochastic configuration networks: Fundamentals and algorithms,” IEEE Transactions on Cybernetics, vol. 47, no. 10, pp. 3466–3479, 2017.
  • [30] W. Li, Q. Zhang, D. Wang, W. Sun, and Q. Li, “Stochastic configuration networks for self-blast state recognition of glass insulators with adaptive depth and multi-scale representation,” Information Sciences, vol. 604, pp. 61–79, 2022.
  • [31] M. L. Bailon and D. J. Horntrop, “On the calculation of convolutions with gaussian kernels,” Applied Mathematics and Computation, vol. 176, no. 1, pp. 383–387, 2006.
  • [32] J. Qiu, C. Chen, S. Liu, H. Y. Zhang, and B. Zeng, “SlimConv: Reducing channel redundancy in convolutional neural networks by features recombining,” IEEE Transactions on Image Processing, vol. 30, pp. 6434–6445, 2021.
  • [33] V. Mnih, K. Kavukcuoglu, D. Silver, A. A. Rusu, J. Veness, M. G. Bellemare et al., “Human-level control through deep reinforcement learning,” Nature, vol. 518, no. 7540, pp. 529–533, 2015.
  • [34] J. Li and D. Wang, “2D convolutional stochastic configuration networks,” Knowledge-Based Systems, vol. 300, no. 112249, pp. 1–12, 2024.
  • [35] Z. Li, F. Liu, W. Yang, S. Peng, and J. Zhou, “A survey of convolutional neural networks: Analysis, applications, and prospects,” IEEE Transactions on Neural Networks and Learning Systems, vol. 33, no. 12, pp. 6999–7019, 2022.
  • [36] D. Wang and M. J. Felicetti, “Stochastic configuration machines for industrial artificial intelligence,” 2023, arXiv:2308.13570.
[Uncaptioned image] Weitao Li received the Ph.D. degree is industrial automation from Northeastern University, Shenyang, China, in 2012. Since 2012, he has been an Associate Professor with the School of Electrical Engineering and Automation, Hefei University of Technology. His research interests include image processing, pattern recognition, and deep stochastic configuration networks for big data analytics. Dr. Li is an Associate Editor for Industrial Artificial Intelligence.
[Uncaptioned image] Xinru Zhang received the B.S. degree in automation from Hunan University, Hunan, China, in 2018. She is currently pursuing a Ph.D. in Control Science and Engineering as a Master’s-Ph.D. combined student at the School of Electrical Engineering and Automation, Hefei University of Technology. Her research interest covers deep learning, image processing, and industrial artificial intelligence.
[Uncaptioned image] Dianhui Wang (Senior Member, IEEE) received the Ph.D. degree in industrial automation from Northeastern University, Shenyang, China, in 1995. From 1995 to 2001, he was a Postdoctoral Fellow with Nanyang Technological University, Singapore, and a Researcher with The Hong Kong Polytechnic University, Hong Kong. He was with La Trobe University, Australia, in 2001 and was a Reader and an Associate Professor with the Department of Computer Science and Information Technology until the end of 2020. Since 2017, he has been a Visiting Professor with the State Key Laboratory of Synthetical Automation of Process Industries, Northeastern University, China. Since July 2021, he has been with the AI Research Institute, China University of Mining and Technology, Xuzhou, China, working as the Dean of the AI Research Institute, a Distinguished Professor of CUMT, and the Director of the Research Center for Stochastic Configuration Machines. He is the primary funder of stochastic configuration networks and stochastic configuration machines, and the key advocate and practitioner of these randomized learner models in industrial data modeling, intelligent sensing system design, and other domain applications. Dr. Wang is the Editor-in-Chief for Industrial Artificial Intelligence, and an Associate Editor for IEEE TRANSACTIONS ON CYBERNETICS, IEEE TRANSACTIONS ON FUZZY SYSTEMS, Information Sciences, Artificial Intelligence Review, and WIREs Data Ming and Knowledge Discovery.
[Uncaptioned image] Qianqian Tong received the B.S. degree in automation from Hefei University, Hefei, China, in 2017. She is currently pursuing the M.S.degree in Control Science and Engineering at the School of Electrical Engineering and Automation, Hefei University of Technology. Her research interest covers intelligent cognition, adaptive dynamic programming and optimal control.
[Uncaptioned image] Tianyou Chai (Life Fellow, IEEE) received the Ph.D. degree in control theory and engineering from Northeastern University, Shenyang, China, in 1985. He has been a Professor with Northeastern University since 1988. He is the Founder and Director of the Center of Automation, which became a National Engineering and Technology Research Center and a State Key Laboratory. He has published 150 international journal papers. His current research interests include modeling, control, optimization and integrated automation of complex industrial processes. Prof. Chai is a Member of the Chinese Academy of Engineering, an IFAC and IEEE Fellow, the Director of the Department of Information Science of the National Natural Science Foundation of China. For his contributions, he has won four prestigious awards of National Science and Technology Progress and National Technological Innovation, the 2007 Industry Award for Excellence in Transitional Control Research from the IEEE Multiple-Conference on Systems and Control. His current research interests include modeling, control, optimization, and integrated automation of complex industrial processes.