MULTI-SOURCE HIERARCHICAL CONDITIONAL RANDOM FIELD MODEL FOR FEATURE FUSION OF REMOTE SENSING IMAGES AND LIDAR DATA
Z. Zhang
a
, M.Y. Yang
b
, M. Zhou
a a
Key Laboratory of Quantitative Remote Sensing Information Technology, Academy of Opto-Electronics, Chinese Academy of Sciences, Beijing, China
zhangzheng, zhoumeiaoe.ac.cn
b
Institute for Information Processing TNT, Leibniz University Hannover, Germany yangtnt.uni-hannover.de
Commission WG VI4, WG III3
KEY WORDS: Feature fusion, Conditional Random Field, Image Classification, Multi-source Data, Hierarchical model ABSTRACT:
Feature fusion of remote sensing images and LiDAR points cloud data, which have strong complementarity, can effectively play the advantages of multi-class features to provide more reliable information support for the remote sensing applications, such as object
classification and recognition. In this paper, we introduce a novel multi-source hierarchical conditional random field MSHCRF model to fuse features extracted from remote sensing images and LiDAR data for image classification. Firstly, typical features are
selected to obtain the interest regions from multi-source data, then MSHCRF model is constructed to exploit up the features, category compatibility of images and the category consistency of multi-source data based on the regions, and the outputs of the
model represents the optimal results of the image classification. Competitive results demonstrate the precision and robustness of the proposed method.
1. INTRODUCTION
Nowadays, there are many different sources of earth observation data which reflect the different characteristics of targets on the
ground, so how to fuse the multi-source data reasonably and effectively for the application, such as object classification and
recognition, is a hot topic in the field of remote sensing applications. In all the data mentioned above, remote sensing
images and LiDAR points cloud have strong complementarity, so fusion of the two sources of data for object classification is
attached more and more attention and many methods were proposed. In general they can be classified into image fusion
Parmehr et al., 2012; Ge et al., 2012 and feature fusion Deng et al., 2012; Huang et al., 2011. The methods for image fusion
always include different resolution data sampling and registration, so the processing is time-consuming, and will
inevitably lose a lot of useful information, which reduces the accuracy of the subsequent image classification. In the feature
fusion methods, the features are usually extracted independently from different sources data, and the fusion lacks consideration
of correspondence of location and contextual information, so the classification results could be improved. In addition,
because the features selected in some methods are not invariant to rotation, scale, or affine, they are always poor in stability. In
order to overcome the shortages of former methods, this paper presents a novel multi-source hierarchical conditional random
field MSHCRF model to fuse features extracted from remote sensing images and LiDAR data for image classification. Firstly,
typical features are selected to obtain the interest regions from multi-source data. Then MSHCRF model is constructed to
exploit up the features, category compatibility of images and the category consistency of multi-source data based on the regions,
and the outputs of the model represents the optimal results of the image classification.
2. DESCRIPTION OF FEATURES SELECTED
In remote sensing images and LiDAR data, while the abundance of information offers more detailed information of interest
objects, it also enhances the noises. Selection of appropriate features in a reasonable way is important in our method.
In order to provide a reliable basis for subsequent processing, the proposed model contains five kinds of typical features: local
saliency feature LSF, line feature LF and texture feature TF are extracted from remote sensing images, mean shift feature
MSF and alpha shape feature ASF are from LiDAR data, so its robust to background interference, change of scale and
perspective, etc. The detector of KB Kadir et al., 2001 is a representative
LSF, which is invariant to viewpoint change, and sensitive to image perturbations. We utilize the detector of KB to
calculate saliency of each pixel in the images. LSD is a linear-time line segment detector that gives accurate
results, a controlled number of false detections, and requires no parameter tuning. In accordance with the method introduced in
Grompone et al., 2010, we can calculate the response value at each pixel.
As the basic unit of TF, Texton is utilized to distinguish between foreground and background regions effectively and
increase the accuracy of the results. Similar to the method in Shotton et al., 2009, we can obtain the response to Texton of
each pixel in the image. For the sparseness and discreteness of LiDAR points cloud data,
we utilize an adaptive mean shift algorithm which is a sample International Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences,
Volume XL-1W1, ISPRS Hannover Workshop 2013, 21 – 24 May 2013, Hannover, Germany
389
point estimation method based on data-driven. In our model, the specific process of achieving the MSF is introduced in
Georgescu et al., 2003. Based on the planar features obtained, the alpha shape
algorithm is used to extract the boundary contour of each target, and then the Delaunay triangulation is used to get the line
feature of LiDAR points cloud. The extraction of the ASF refers to Shen et al., 2011.
3. FEATURE FUSION USING MSHCRF