JAMP  Vol.7 No.11 , November 2019
Forest-Fire Recognition by Sparse and Collaborative Subspace Clustering
Abstract: Traditional forest-fire recognition based on the characteristics of smoke, temperature and light fails to accurately detect and respond to early fires. By analyzing the characteristics of flame, the methods based on aerial image recognition have been widely used, such as RGB-based and HIS-based methods. However, these methods are susceptible to background factors causing interference and false detection. To alleviate these problems, we investigate two subspace clustering methods based on sparse and collaborative representation, respectively, to detect and locate forest fires. Firstly, subspace clustering segments flame from the whole image. Afterwards, sparse or collaborative representation is employed to represent most of the flame information in a dictionary with l1-regularization or l2-regularization term, which results in fewer reconstruction errors. Experimental results show that the proposed SSC and CSC substantially outperform the state-of-the art methods.

1. Introduction

Forest fires destroy forests and undergrowth plants, endanger human beings and wildlife, pollute soil and air, affect water cycle, and so on. In order to protect the ecological environment and human security, early fire detection is a necessary step for forest fire warning. Conventional detection methods use the characteristics of smoke, temperature and light to monitor fires. However, these methods can’t respond quickly to early fire when the forest environment is bad, and the fires may not be detected immediately if it is far away from the sensors [1]. The methods based upon aerial image recognition avoid the adverse effects (e.g. space, height and airflow), by analyzing the characteristics of flame (e.g. color, area and location), to detect and locate forest fires accurately. Arguably, flame recognition has been the key to the success of early fire detection [2].

RGB (red, green, and blue) and HIS (hue, intensity and saturation)-based methods mainly make use of color information in flame recognition recent years [3] [4]. RGB-based methods employ color features to recognize fire areas by segmenting objects in a specified color range in RGB images [5]. In these methods, RGB vector m is used to represent the average color estimation. If the color of some pixels is within a certain range of the vector m, these pixels will be recognized and classified. The following comparation method RGB employs the Euclidean distance [6] as the classification range. On the other hand, HIS-based methods employ hue, light intensity and saturation as features for flame recognition. In general, the flame colors are usually yellow, red and orange, and the flame hue is distributed in 0 - 60 degrees. Due to the light intensity of background, the flame saturation may be affected, causing interference and false detection [7].

Since the sparse coefficients can result in fewer reconstruction errors and can represent most of the important information in a dictionary, we investigate sparse subspace clustering (SSC) for flame recognition to improve the recognition accuracy. Given a set of pixels drawn from a union of linear or affine subspaces, the task is to find segmentation of forest-fire images. Because each pixel in a union of subspaces can always be written as a linear or affine combination of all other pixels, the representation of pixels in the same subspace can be obtained by searching for the sparsest combination. Thus, a similarity matrix will be established for flame segmentation by sparse subspace clustering. Since the combinatorial l 0 -norm minimization is NP-hard, SSC estimates similarity matrix by imposing l 1 -norm optimization [8]. We also studied collaborative representation clustering (CSC) [9] whose similarity matrix is employed by imposing l 2 -norm optimization instead of l 1 -norm optimization, which means that the “collaborative” nature is instead of “competitive” nature of forest-fire images.

The main contributions of this paper are summarized as follows. 1) The proposed approaches are unsupervised methods requiring no initialization. 2) Our work is the first one to employ the spare or collaborative representation of vectors in subspace clustering for forest-fire recognition. 3) The proposed approaches, whose objective function is optimized by l 1 -norm optimization or l 2 -norm optimization fully exploiting the internal relationship among pixels, produces stronger ability of discrimination compared with traditional methods (i.e., RGB and HIS) by conducting experiments on three forest-fire images.

The remainder of this paper is organized as follows. In Section 2, we present the proposed approaches. In Section 3, we provide a description of the three forest-fire data for performance evaluation, as well as a detailed comparison with the state-of-the-art methods. Section 4 concludes the presented work.

2. Approach

The basic idea of subspace clustering is to extract a set of points from linear or affine subspace and to find the segments of data. Since each data point in the union set of a subspace can always be written as a linear or affine combination of other points, the sparse or collaborative expression-based subspace clustering method uses the low-dimensional representation of the data in the high-dimensional space by searching the most sparse or collaborative combination. This enables us to establish a similarity matrix and segment flame by spectral clustering. Since the combinatorial l 0 -norm minimization is NP-hard, the l 1 -norm minimization and l 2 -norm minimization employed in sparse subspace clustering and collaborative subspace clustering, respectively. Thus, spectral clustering method is applied to the similarity graph formed by sparse coefficient or collaborative coefficient to realize the detection and recognition of forest-fire images. Figure 1 shows an example of data drawn from two subspaces, the corresponding adjacency matrix and similarity graph. For each data point, the l 1 -minimization or l 2 -minimization gives subspace-preserving representation, hence the adjacency matrix has a block-diagonal structure and the similarity graph has two connected components. Using spectral clustering methods, we can recover the components of the similarity graph, obtaining the segmentation of forest-fire data.

The proposed sparse subspace clustering (SSC) algorithm will be discussed in the following. Let X = { X 1 , , X N } be N subspaces of dimensions { d 1 , , d N } embedded in D dimensional space. Consider a given collection of M points ψ = { ψ 1 , , ψ M } drawn from the N subspaces. Wherein, Y can be represented by sparse representation if Y is a linear combination of at most K ( K < < N ) basis vectors in ψ . In practice, the signal is K-sparse when it has at most K large nonzero coefficients and the remaining coefficients are very small [10]. We can write Y as

Y = j = 1 M X j ψ j (1)

Y and X belong to the space domain and the subspace domain, respectively. The idea is based on the observation that each data point can be described as a linear combination of all the other data points in subspace domain.

(a) (b) (c)

Figure 1. An example: (a) Data drawn from two subspaces; (b) Corresponding matrix of coefficients; (c) Similarity graph.

For separating X into different class-specific representations, we need to calculate

x i = X w i (2)

W is the similarity matrix whose ith column corresponds to the sparse representation of x i . For each pixel x i in the dictionary X , the sparse representation vector is calculated by solving the l 1 -norm optimization problem

min w i w i 1 s . t . X i w i x i 2 ε (3)

where X i does not include x i itself, w i = [ w i 1 , w i 2 , , w i N ] is the sparse representation vector corresponding to x i of size N × 1 , and ε > 0 is a small tolerance. Thus, the similarity (or affinity) matrix W is described as

( W s ) i , i ' = { 0 , i = i ' w i , i ' , i > i ' w i , i ' 1 , i < i ' (4)

The generalized eigenvalue problem can be expressed as

min w i { x i X i w i 2 2 + λ w i 1 } (5)

where λ is a Lagrange multiplier balancing the minimization between regularization term and residual part [11]. The procedure of SSC is outlined in Algorithm 1.

In the next, collaborative subspace clustering (CSC) is also proposed to investigate the collaborative nature of forest-fire images. Collaborative representation is studied to estimate the similarity matrix by replacing the l 1 -norm minimization in solving the weight matrix with an l 2 -norm minimization. The segmentation of forest-fire images is found by applying spectral clustering to a similarity graph formed by the collaborative coefficients. The collaborative representation vector for x i is calculated by solving the l 2 -norm optimization problem

min w i w i 2 s . t . X i w i = x i (6)

where X i does not include x i itself, and w i = [ w i 1 , w i 2 , , w i N ] is the collaborative representation vector corresponding to x i of size ( N 1 ) × 1 . The objective function can be reformulated as

min w i { x i X i w i 2 2 + λ w i 2 2 } (7)

The regularization parameter λ does two things: it makes the least square solution stable; and it introduces a weaker “sparsity” than that by l 1 -norm minimization to the solution depicted as Equation (6). The collaborative representation vector can be calculated as

w i = ( X i T X i + λ I ) 1 X i T x i . (8)

Similarity matrix W = [ w 1 , w 2 , , w N ] represents the graph weight matrix of size N × N and d i a g ( W ) is a zero vector. The graph weight matrix can be calculated by within-class samples to generate class-specific collaborative coefficients. The procedure of the proposed CSC is outlined in Algorithm 2.

3. Experimental Results

Experiments are conducted using three forest-fire images, which are represent a forest-fire scenario with 55 ´ 59 pixels, 79 ´ 63 pixels, and 58 ´ 51 pixels. To quantify the efficacy of the proposed two approaches, we study the classification performance of them and two state-of-the-art methods (HIS and RGB) for classifying forest-fire data. Table 1 shows the classification results with the overall classification accuracy (OA). From this table, we can see that the proposed methods are superior to HIS and RGB. In particular, the proposed SSC has more significant advantage in classification accuracy. For example, the OA of SSC has improvements of 10.31% and 1.05% for image 1 compared with HIS and RGB, respectively; the OA of SSC has improvements of 3.8% and 3.32% for image 2 compared with HIS and RGB, respectively.

Figures 2-4 provide the classification maps of the aforementioned methods for the forest-fire images. From the classification maps, it can be seen that the proposed SSC and CSC have better classification performance of details, which can play a positive role in the assessment of fire severity and areas.

Table 1. Classification results for the four methods.

(a) (b) (c) (d) (e) (f)

Figure 2. Classification maps of four approaches for image 1. (a) Image 1; (b) Ground-truth; (c) HIS; (d) RGB; (e) SSC; (f) CSC.

(a) (b) (c) (d) (e) (f)

Figure 3. Classification maps of four approaches for image 3. (a) Image 3; (b) Ground-truth; (c) HIS; (d) RGB; (e) SSC; (f) CSC.

(a) (b) (c) (d) (e) (f)

Figure 4. Classification maps of four approaches for image 2. (a) Image 2; (b) Ground-truth; (c) HIS; (d) RGB; (e) SSC; (f) CSC.

4. Conclusion

In this paper, we propose two unsupervised classification methods based on sparse or collaborative representation of vectors in subspace clustering for forest-fire recognition. These strategies employing l 1 -norm or l 2 -norm optimization exploits the internal relationship among pixels and produces stronger ability of discrimination. Experimental results show that the proposed methods substantially outperform the state-of-the-art methods. We hence conclude that the proposed SSC and CSC approach are two effective approaches for forest-fire recognition.


This research was supported by the National Natural Science Foundation of China (41601344, 61803042).

Cite this paper: Ye, Z. , Jiang, Y. , Shi, S. , Yan, J. and Bai, L. (2019) Forest-Fire Recognition by Sparse and Collaborative Subspace Clustering. Journal of Applied Mathematics and Physics, 7, 2883-2890. doi: 10.4236/jamp.2019.711197.

[1]   Wang, Y.B. and Ma, X.M. (2014) Early Fire Detection for High Space Based on Video-Image Processing. 2014 International Symposium on Computer, Consumer and Control, Taichung, 10-12 June 2014.

[2]   Qiu, T., Yan, Y. and Lu, G. (2012) An Autoadaptive Edge-Detection Algorithm for Flame and Fire Image Processing. IEEE Transactions on Instrumentation and Measurement, 61, 1486-1493.

[3]   Wang, L. and Li, A.G. (2017) Early fire recognition based on Multi-Feature Fusion of Video Smoke. The 36th Chinese Control Conference, Dalian, 26-28 July 2017.

[4]   Chen, T.-H., Wu, P.-H. and Chiou, Y.-C. (2004) An Early Fire-Detection Method Based on Image Processing. 2004 International Conference on Image Processing, Singapore, 24-27 October 2004.

[5]   Gonzalez, R.C. and Woods, R.E. (2017) Digital Image Processing. Publishing House of Electronics In-dustry, Beijing.

[6]   Chamorro-Martines, J., Soto-Hidalgo, J.M., Martinez-Jimenez, P.M., et al. (2017) Fuzzy Color Spaces: A Conceptual Approach to Color Vision. IEEE Transaction on Fuzzy Systems, 25, 1264-1280.

[7]   Celik, T. and Demirel, H. (2009) Fire Safety in Video Sequence Using a Generic Color Model. Fire Safety, 44, 951-61.

[8]   Elhamifar, E. and Vidal, R. (2013) Sparse Subspace Clustering: Algorithm, Theory, and Applications. IEEE Transactions on Pattern Analysis and Machine Intelligence, 35, 2765-2781.

[9]   Zhang, Y.B., Zhang, Y.L., Zhang, J. and Dai, Q.H. (2016) CCR: Clustering and Collaborative Representation for Fast Single Image Super-Resolution. IEEE Transactions on Multimedia, 18, 405-417.

[10]   Elhamifar, E. and Vidal, R. (2009) Sparse Subspace Clustering. IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Miami, 20-25 June 2009.

[11]   Zhang, L., Yang, M. and Feng, X.C. (2011) Sparse Representation or Collaborative Representation: Which Helps Face Recognition? IEEE International Conference on Computer Vision, Barcelona, Spain, November 2011, 471-478.