### Paper Menu >>

### Journal Menu >>

Journal of Signal and Information Processing, 2013, 4, 168-172 doi:10.4236/jsip.2013.43B030 Published Online August 2013 (http://www.scirp.org/journal/jsip) A Real Time Mosaic Method for Remote Sensing Video Images from UAV Yang Yang, Guangmin Sun, Dequn Zhao, Bo Peng Department of Electronic Engineering, Beijing University of Technology, Beijing, China. Email: young_job@163.com, gmsun@bjut.edu.cn Received May, 2013 ABSTRACT At present, in order to get a large field of view image, image mosaic technique has been widely applied in UAV remote sensing platform. The traditional mosaic system for UAV remote sensing image takes a lot of time and man power, it is difficult to complete the image stitching automatically. In the paper, an approach for geometric correction of remote sensing image without any ground control points is presented, and the SIFT algorithm is used to extract and match fea- ture points. Finally, the weighted average fusion method is used to smooth the image after splicing and an automatic mosaic system for UAV remote sensing video images is developed. In order to verify the system, some splicing ex- periments using UAV actual aerial photography images have been done and good results have been achieved. Keywords: Image Mosaic; SIFT Algorithm; Feature Matching 1. Introduction The UAV remote sensing system has been used widely in surveying and mapping, national defense, monitoring and other fields because it has the advantage in low cost, high flexibility and high efficiency, and the available remote sensing images obtained by it have the characteristics of rich information and high resolution. In a short period of time, the mass data of the target area can be obtained by the UAV remote sensing system, which have the charac- teristics of small picture format and large of data. Thus, in order to obtain large field of view image of the target area, the fast and accurate image mosaic is necessary. Generally, remote sensing image mosaic includes geometric correction, feature matching, image smoothing and other parts. Inclination and jitter will appear inevita- bly during UAV flight in process for the small UAV body result in the poor stability and bad wind resistance, consequently the geometric correction of UAV remote sensing image will affect the results of subsequent stitching directly. Matching algorithm is the key of im- proving the quality of image stitching, at present the fea- ture matching algorithm can be divided into two catego- ries: image feature based algorithm and image gray value based algorithm. The former make use of the photo- graphic field features which have the characteristics of relatively unchanging space position, such as edges, an- gular point and contour [1]. Although this geometric has the advantage of less calculation, it requires high quality images and the great ability of feature extraction, and it is sensitive to the interference of noise and geometric dis- tortion. The matching algorithm based on image gray value make use of the matching between the template area of image and the gray value of the search region [2,3], which has the relatively high matching accuracy, but has large calculation amount, and is unsuitable for real-time operation. So, it becomes the key problem need to be solved to improve the speed of matching under condition of ensuring the matching precision. In order to realize the real-time video mosaic, an automatic mosaic system suitable for UAV remote sens- ing image combining with different algorithm have been developed in the paper. Firstly, an original image is recti- fied in the system using geometric correction model without control points which is established by UAV flight status parameters. Then two images are matched by applying SIFT algorithm after correction. Finally, image is smoothed by the approach of weighted average. The system proposed in the paper have the advantage of high efficiency, anti geometric distortion and anti illumination change, which can realize the real-time dynamic mosaic of the remote sensing video images and ensure the stitching accuracy at the same time. 2. Image Rectification for Geometry Distortion The traditional geometric correction of remote sensing Copyright © 2013 SciRes. JSIP A Real Time Mosaic Method for Remote Sensing Video Images from UAV 169 image can be performed through the collinear equation and polynomial fitting; these methods need to set up a certain number of control points which are uniformly distributed on the ground. However, in surveillance, re- connaissance and other special fields, it is difficult to obtain the ground control points. Therefore, a UAV re- mote sensing geometric correction model without ground control point is necessary to established. The geometry distortion of UAV remote sensing im- age is caused mainly by the internal sensor error, ele- ments of exterior orientation changes and related geo- physical characteristics. The key of UAV remote sensing image geometric correction is to eliminate elements of exterior orientation changes if take no account of the effect of earth rotation, curvature and other physical characteristics. Therefore, in order to realize fast geomet- ric correction and get the orthographic projection, the geometric correction model which is established in this paper only considering the circumstance caused by the changes of flight attitude parameter[4]. Flight attitude parameters recorded during UAV flight process include pitching angle α, roll angle β and yaw angle γ. When α, β and γ are zeros, the remote sensing image from UAV can be considered to be an ideal image? Once α, β and γ changed, geometrical rectification image can be obtained through the establishment of coordinate system interconnections, and performed corresponding coordinate transformation on the original image. Through geometrical analysis, the correction model used in this paper can be described as following: ' ' ()()()() -- SS x x yRHRRRy f f (1) Figure 1 is two original images to be spliced. Figure 2 shows the rectified image of original image I1. 3. Image Matching Scale Invariant Feature Transform (SIFT) is a steady algorithm proposed and completed by D. G. Lower in 2004 [5]. The features extracted by SIFT algorithm are local features which are invariant to image rotation, scale and illumination changes. Therefore, SIFT algorithm is a quite suitable method which can application in remote sensing video images from UAV. In this paper, an im- proved SIFT feature matching algorithm was proposed. This algorithm mainly contains the following six steps, as shown in Figure 3: First of all, convert the two original images into the same coordinate system using the latitude and longitude information of image center point. In this way the over- lap region of two original images will be directly deter- mined, as shown in Figure 4. Figure 1. Original image I1 (left) and original image I2 (right). Figure 2. The rectified image of image I1. Figure 3. Flowchart of image matching. Figure 4. The schematic diagram of determinin g th e overlap region. In which, θ is the difference of the two images yaw angle, r is the distance between the two images center points s and s,. Secondly, establish the image scale space. The scale- space representation for image in multi-scales can be obtained by performing scale transform to the original image using Gaussian function, on which stable feature points can be extracted. It has been shown by Koender- ink [6] and Lindeberg [7] that under a variety of reason- able assumptions the only possible scale-space kernel is the Gaussian function. Consequently, the image scale space can be express as: (,, )(,, )(,)LxyGxy Ixy (2) Copyright © 2013 SciRes. JSIP A Real Time Mosaic Method for Remote Sensing Video Images from UAV 170 where (,, )Lxy is defined as scale-space, (, ) x y denotes the image pixel, and δ is scale factor. 22 2 -()/ 2 2 1 (,, )2 xy Gxy e , it is the Gaussian convolu- tion kernel, and is the input image. ),( yxI In order to improve the stability of feature point’s ex- traction, difference of Gaussian (DOG) scale space was defined as formula (3), which is the convolution of dif- ference of Gaussian kernel function and original image. ,,,,,, , ,, ,, D xyG xykG xyIxy Lxyk Lxy (3) Here k is a constant. Thirdly, determine the feature points. The position and scale of the candidate extreme point is determined by comparing each point with other 26 neighboring points (8 neighbors on the current scale space and 9*2 points on upper and lower adjacent scale space). Owing to DOG is more sensitive to the noise and edges, it is necessary to eliminate the points on the unstable edge of low contrast key points. Finally, the location and the DOG scale of extreme points are obtained accurately through three- dimensional quadratic function, which can enhance ac- curacy of extreme points and improve noise immunity. Fourthly, distribute the direction of the feature points. In practical computing, sampling is operated in the neighborhood window centre around the key point, which can ensure the direction of a gradient of neighborhood pixels through gradient histogram with ranges from 0 to 360 degrees, where each 10 degrees is a column, a total of 36 columns. So, the peak of histogram stands for the primary orientation of neighborhood gra- dient in this key point, namely the direction of key point. Moreover, another orientation can be identified as the auxiliary direction of the key point when there is another value of the gradient histogram equivalent to eighty per- cent of the value of main peak. Thus, the orientation of each key point is assigned using the gradient direction which make the operator has rotation invariance. Finally, the definition of a feature point descriptor is created according to the following steps, which can eliminate influences on illumination change and geomet- ric distortion. Firstly, for ensuring rotation invariance the coordinate axis is rotated to the direction of the key points. Then an 8*8 window centre around the key point is extracted, as shown in Figure 5. In which, each grid represents a pixel of feature point neighborhood in its scale space. The direction of arrow indicates the pixel gradient direction, and the length of arrow is used to indicate the gradient modulus. Then, eight directions of gradient orientation histogram are calculated in each 4*4 grid of image, and a seed point is produced by diagramming the accumulated value of each gradient direction. Finally, a 32-dimensional SIFT fea- ture vector is established, which has the advantage of anti-interference and better fault tolerance. Figure 6 is the results of extracting SIFT features in the overlap region of the original image. We can see the extracted features are concentrated in the overlap region of the original image. Figure 5. Feature vector generated by key pointneighborho od gradient information. Figure 6. The results of extracting SIFT feature vectors in the overlap region for I1 (left) and I2 (right). Figure 7. The results of image matching. After the two SIFT feature vector of images is gener- ated, the similarity measure of two key points is deter- mined by calculating the Euclidean distance between key point vectors in two images. One key point is chosen in the matching image 1, and then two key points can be found in the matching image 2 which have the nearest Euclidean distances to the key point in the matching im- age 1. Finally, it is considered successfully matched if the differences between the second nearest distance and the nearest distance greater than a certain threshold. Fig- ure 7 is the results of two images matching. Copyright © 2013 SciRes. JSIP A Real Time Mosaic Method for Remote Sensing Video Images from UAV 171 4. Image Smoothing Because of the effects of differences in illumination and angle during UAV flight in process, the image obtained by matching two original images will inevitably exist a seam [8]. Therefore, a mosaic image smoothing is neces- sary to be done. At present, the methods of image fusion may fall into three classes: the direct average fusion, weighted average fusion and Gauss distribution fusion. The weighted average fusion will be used in this paper because it has the advantages of simple calculation and good effect, and the smoothing process can be expressed as formula 4: 121 122 12 (, )(, )(, ) F nnAnn Bnn (4) where 12 (, ) A nn (, and 12 are the two original images, 12 (, )Bn n ) F nn is the result of fusion by the weighted average method, and the image resolution is n1*n2. The result of image smoothing is shown in the Figure 8. Figure 8. The result of image smoothing. Figure 9. The results of multiple frame image mosaic. 5. Result Analysis and Conclusions The experimental images in this paper come from UAV practical aerial photography. The algorithm has been implemented on VS2008 which includes OpenCV library. The images gotten from experiments are shown in Fig- ure 2, Figure 6, Figure 7 and Figure 8. As shown in Figure 6, there are 139 SIFT features extracted from the original image I1 and 168 features in I2. Time costs for feature extraction from image I1 and I2 are separate about 844ms and 875ms. Compared with features extrac- tion in the whole image, this method implemented on I1 can reduce 23 features and save time 312ms. When im- plemented on I2, it can reduce 212 features and save time 625 ms. After feature matching, we can get the result of image fusion, as shown in Figure 8, in which image I1 can be smoothly transited to image I2, and the seamless splicing is realized. Then, with method proposed in this paper, the results of multiple frames image mosaic is shown in Figure 9. Figure 10 is the result of whole mo- saic system, in which the left interface plays video, the right interface implements image stitching. Figure 10. The interface mosaic system. In this paper, an orthographic projection is obtained through the remote sensing image geometric correction without ground control points, the improved SIFT algo- rithm and weighted average fusion method is used to complete the seamless splitting. Experimental results show that, the systems implemented in this paper can im- prove the operation speed while ensuring the mosaic ef- fect, thus, it can apply to the real time mosaic system for UAV remote sensing video images. REFERENCES [1] G. Feng, J. Lu and Z. J. Lin, “An Image Matching Ap- proach Based on the Invariant Feature of Image Histo- gram,” Journal of Computer-Aided Design & Computer Graphics, Vol. 12, No. 2, 2000, pp. 146-148. [2] D. I. Barnea and H. F. Silverman, “A Class of Algorithms for Fast Digital Image Registration,” IEEE Transactions on Computer, Vol. 21, 1972, pp. 179-186. [3] X.-G. Yang, F. Cao and D. Miao, “Study on the Image Grayscale Matching Algorithm Based on Similarity Measures,” Systems Engineering and Electronics, Vol. 27, No. 5, 2005, pp. 918-921. [4] Y. X. Li, Z. Li, L. Tong, Y. T. Yan and D. Guo, “A Geo- metrical Rectification Algorithm of UAV Remote Sens- ing Images Based on Flight Attitude Parameters, 2011 IEEE Geoscience and Remote Sensing Symposium, 2011, [5] D. G. Lowe, “Distinctive Image Features from Scale-in- variant Keypoints,” International Journal of Computer Vision, Vol. 60, No. 2, 2004, pp.91-110. Copyright © 2013 SciRes. JSIP A Real Time Mosaic Method for Remote Sensing Video Images from UAV Copyright © 2013 SciRes. JSIP 172 [6] J. J. Koenderink, “The Structure of Images,” Biological Cybernetics, Vol. 50, No. 5, 1984, pp. 363-370. [7] T. Lindeberg, “Scale-Space Theory: A Basic Tool for Analyzing Structures at Different Scales,” Journal of Ap- plied Statistics, Vol. 21, No. 2, pp. 224-270. [8] S. L. Zhu, Z. B. Z. Qian, “The Seam-line Removal under Mosaicking of Remote Sensing Images,” Journal of Re- mote Sensing, Vol. 6, No. 3, 2002, pp. 183-187. |