Hyperspectral Image Segmentation Using a New Bayesian Approach With Active Learning
【最新精选】《数字图像处理》结课小论文题目汇总及要求

《数字图像处理》结课测试题目题目的路径:首先在Matlab的Command window中键入“demo”,进入demo 窗口。
然后在树形选择框中选择“Toolboxes\Image Processing”和“Blocksets\ Video and Image Processing”。
最后逐个查看并选择自己感兴趣的题目。
所有题目汇总如下:图像去模糊1. Deblurring Images Using the Blind Deconvolution Algorithm基于盲解卷算法的图像去模糊2. Deblurring Images Using the Lucy-Richardson Algorithm使用LR算法进行图像去模糊3. Deblurring Images Using a Regularized Filter使用正则滤波器进行图像去模糊4. Deblurring Images Using the Wiener Filter使用维纳滤波器进行图像去模糊图像增强5. Contrast Enhancement Techniques图像对比度增强技术6. Correcting Nonuniform Illumination如何对不均匀光照进行校正7. Enhancing Multispectral Color Composite Images多谱(卫星遥感) 图像增强技术图像配准8. Finding the Rotation and Scale of a Distorted Image计算失真图像的旋转参数和尺度参数9. Registering an Aerial Photo to an Orthophoto基于控制点的多幅航拍图像的配准10. Registering an Image Using Normalized Cross-Correlation使用归一化交叉相关法来配准图像图像分割11. Batch Processing Image Files Using Distributed Computing分布式计算对图像序列进行批处理12. Color-Based Segmentation Using the L*a*b* Color Space基于Lab色彩空间的彩色图像分割13. Color-Based Segmentation Using K-Means Clustering 基于K-均值聚类的彩色图像分割14. Detecting a Cell Using Image Segmentation使用图像分割技术来检测细胞15. Finding V egetation in a Multispectral Image多谱图像(卫星遥感)上的农作物区域分割16. Marker-Controlled Watershed Segmentation基于标记控制的分水岭分割算法17. Texture Segmentation Using Texture Filters基于纹理滤波器的纹理图像分割图像几何变换18. Creating a Gallery of Transformed Images常见的图像几何变换简介19. Exploring a Conformal Mapping图像的保角变换(共形映射)20. Extracting Slices from a 3-Dimensional MRI Data Set 如何从3维MRI数据集中提取切片图21. Padding and Shearing an Image Simultaneously图像的剪切变换和填充操作图像的测量22. Finding the Length of a Pendulum in Motion从单摆图像序列中计算摆长23. Granulometry of Snowflakes使用形态学方法对雪花的颗粒度进行测量24. Identifying Round Objects在图像中计算物体的“似圆度”25. Measuring Angle of Intersection在图像中计算钢梁的交叉角度26. Measuring the Radius of a Roll of Tape如何用图像方法测量胶带的半径图像的Radon变换27. Reconstructing an Image from Projection Data基于拉东(Radon)变换的CT图像重建视频检测和跟踪28. Abandoned Object Detection遗弃物体检测技术29. Motion Detection基于SAD的运动检测系统30. Lane Departure Warning System车道偏离预警系统31. Lane Detection and Tracking基于Hough变换的车道检测和跟踪32. Traffic Warning Sign Recognition交通警示牌自动识别技术33. People Tracking基于背景差分的行人检测技术34. Color Segmentation基于色彩分割的人体检测35. Tracking Cars Using Background Estimation 基于背景估计的汽车检测36. Tracking Cars Using Optical Flow基于光流法的汽车检测37. Surveillance Recording基于主帧检测的监控记录技术38. Pattern Matching基于模板匹配的PCB检测系统压缩技术39. V ideo Compression基于DCT变换的视频压缩技术40. Image Compression基于DCT变换的图像压缩技术视频分析技术41. Histogram Display图像直方图的实时显示42. Concentricity Inspection光纤的同心性检测系统43. Edge Detection边缘检测技术简介44. V ideo Focus Assessment视频自动聚焦参量计算视频增强45. V ideo Stabilization基于模板的电子稳像技术46. Periodic Noise Reduction针对周期噪声的图像降噪算法47. Histogram Equalization基于直方图均衡的图像增强48. Rotation Correction基于Hough变换的旋转图像校正基于形态学的视频分割技术49. Cell Counting细胞自动计数系统50. Feature Extraction如何自动计算视频中扇形的数目51. Object Counting如何自动计算订书钉的数目52. Object Extraction and Replacement视频目标的实时提取和替换视频回放处理53. Continuous Image Rotation图像连续旋转效果的实现54. Projecting Videos onto a Rotating Cube 如何将视频投影到旋转的立方体上55. V isual Effects图像浮雕效果的实现56. Picture in Picture画中画效果的实现57. Panorama Creation全景照片技术58. Bouncing Balls如何在图像上叠加动画《数字图像处理》结课测试报告规范1.内容要求(1)本报告(论文)的名字,系统功能、实现了什么结果。
A Fast and Accurate Plane Detection Algorithm for Large Noisy Point Clouds Using Filtered Normals

A Fast and Accurate Plane Detection Algorithm for Large Noisy Point CloudsUsing Filtered Normals and Voxel GrowingJean-Emmanuel DeschaudFranc¸ois GouletteMines ParisTech,CAOR-Centre de Robotique,Math´e matiques et Syst`e mes60Boulevard Saint-Michel75272Paris Cedex06jean-emmanuel.deschaud@mines-paristech.fr francois.goulette@mines-paristech.frAbstractWith the improvement of3D scanners,we produce point clouds with more and more points often exceeding millions of points.Then we need a fast and accurate plane detection algorithm to reduce data size.In this article,we present a fast and accurate algorithm to detect planes in unorganized point clouds usingfiltered normals and voxel growing.Our work is based on afirst step in estimating better normals at the data points,even in the presence of noise.In a second step,we compute a score of local plane in each point.Then, we select the best local seed plane and in a third step start a fast and robust region growing by voxels we call voxel growing.We have evaluated and tested our algorithm on different kinds of point cloud and compared its performance to other algorithms.1.IntroductionWith the growing availability of3D scanners,we are now able to produce large datasets with millions of points.It is necessary to reduce data size,to decrease the noise and at same time to increase the quality of the model.It is in-teresting to model planar regions of these point clouds by planes.In fact,plane detection is generally afirst step of segmentation but it can be used for many applications.It is useful in computer graphics to model the environnement with basic geometry.It is used for example in modeling to detect building facades before classification.Robots do Si-multaneous Localization and Mapping(SLAM)by detect-ing planes of the environment.In our laboratory,we wanted to detect small and large building planes in point clouds of urban environments with millions of points for modeling. As mentioned in[6],the accuracy of the plane detection is important for after-steps of the modeling pipeline.We also want to be fast to be able to process point clouds with mil-lions of points.We present a novel algorithm based on re-gion growing with improvements in normal estimation and growing process.For our method,we are generic to work on different kinds of data like point clouds fromfixed scan-ner or from Mobile Mapping Systems(MMS).We also aim at detecting building facades in urban point clouds or little planes like doors,even in very large data sets.Our input is an unorganized noisy point cloud and with only three”in-tuitive”parameters,we generate a set of connected compo-nents of planar regions.We evaluate our method as well as explain and analyse the significance of each parameter. 2.Previous WorksAlthough there are many methods of segmentation in range images like in[10]or in[3],three have been thor-oughly studied for3D point clouds:region-growing, hough-transform from[14]and Random Sample Consen-sus(RANSAC)from[9].The application of recognising structures in urban laser point clouds is frequent in literature.Bauer in[4]and Boulaassal in[5]detect facades in dense3D point cloud by a RANSAC algorithm.V osselman in[23]reviews sur-face growing and3D hough transform techniques to de-tect geometric shapes.Tarsh-Kurdi in[22]detect roof planes in3D building point cloud by comparing results on hough-transform and RANSAC algorithm.They found that RANSAC is more efficient than thefirst one.Chao Chen in[6]and Yu in[25]present algorithms of segmentation in range images for the same application of detecting planar regions in an urban scene.The method in[6]is based on a region growing algorithm in range images and merges re-sults in one labelled3D point cloud.[25]uses a method different from the three we have cited:they extract a hi-erarchical subdivision of the input image built like a graph where leaf nodes represent planar regions.There are also other methods like bayesian techniques. In[16]and[8],they obtain smoothed surface from noisy point clouds with objects modeled by probability distribu-tions and it seems possible to extend this idea to point cloud segmentation.But techniques based on bayesian statistics need to optimize global statistical model and then it is diffi-cult to process points cloud larger than one million points.We present below an analysis of the two main methods used in literature:RANSAC and region-growing.Hough-transform algorithm is too time consuming for our applica-tion.To compare the complexity of the algorithm,we take a point cloud of size N with only one plane P of size n.We suppose that we want to detect this plane P and we define n min the minimum size of the plane we want to detect.The size of a plane is the area of the plane.If the data density is uniform in the point cloud then the size of a plane can be specified by its number of points.2.1.RANSACRANSAC is an algorithm initially developped by Fis-chler and Bolles in[9]that allows thefitting of models with-out trying all possibilities.RANSAC is based on the prob-ability to detect a model using the minimal set required to estimate the model.To detect a plane with RANSAC,we choose3random points(enough to estimate a plane).We compute the plane parameters with these3points.Then a score function is used to determine how the model is good for the remaining ually,the score is the number of points belonging to the plane.With noise,a point belongs to a plane if the distance from the point to the plane is less than a parameter γ.In the end,we keep the plane with the best score.Theprobability of getting the plane in thefirst trial is p=(nN )3.Therefore the probability to get it in T trials is p=1−(1−(nN )3)ing equation1and supposing n minN1,we know the number T min of minimal trials to have a probability p t to get planes of size at least n min:T min=log(1−p t)log(1−(n minN))≈log(11−p t)(Nn min)3.(1)For each trial,we test all data points to compute the score of a plane.The RANSAC algorithm complexity lies inO(N(Nn min )3)when n minN1and T min→0whenn min→N.Then RANSAC is very efficient in detecting large planes in noisy point clouds i.e.when the ratio n minN is 1but very slow to detect small planes in large pointclouds i.e.when n minN 1.After selecting the best model,another step is to extract the largest connected component of each plane.Connnected components mean that the min-imum distance between each point of the plane and others points is smaller(for distance)than afixed parameter.Schnabel et al.[20]bring two optimizations to RANSAC:the points selection is done locally and the score function has been improved.An octree isfirst created from point cloud.Points used to estimate plane parameters are chosen locally at a random depth of the octree.The score function is also different from RANSAC:instead of testing all points for one model,they test only a random subset and find the score by interpolation.The algorithm complexity lies in O(Nr4Ndn min)where r is the number of random subsets for the score function and d is the maximum octree depth. Their algorithm improves the planes detection speed but its complexity lies in O(N2)and it becomes slow on large data sets.And again we have to extract the largest connected component of each plane.2.2.Region GrowingRegion Growing algorithms work well in range images like in[18].The principle of region growing is to start with a seed region and to grow it by neighborhood when the neighbors satisfy some conditions.In range images,we have the neighbors of each point with pixel coordinates.In case of unorganized3D data,there is no information about the neighborhood in the data structure.The most common method to compute neighbors in3D is to compute a Kd-tree to search k nearest neighbors.The creation of a Kd-tree lies in O(NlogN)and the search of k nearest neighbors of one point lies in O(logN).The advantage of these region growing methods is that they are fast when there are many planes to extract,robust to noise and extract the largest con-nected component immediately.But they only use the dis-tance from point to plane to extract planes and like we will see later,it is not accurate enough to detect correct planar regions.Rabbani et al.[19]developped a method of smooth area detection that can be used for plane detection.Theyfirst estimate the normal of each point like in[13].The point with the minimum residual starts the region growing.They test k nearest neighbors of the last point added:if the an-gle between the normal of the point and the current normal of the plane is smaller than a parameterαthen they add this point to the smooth region.With Kd-tree for k nearest neighbors,the algorithm complexity is in O(N+nlogN). The complexity seems to be low but in worst case,when nN1,example for facade detection in point clouds,the complexity becomes O(NlogN).3.Voxel Growing3.1.OverviewIn this article,we present a new algorithm adapted to large data sets of unorganized3D points and optimized to be accurate and fast.Our plane detection method works in three steps.In thefirst part,we compute a better esti-mation of the normal in each point by afiltered weighted planefitting.In a second step,we compute the score of lo-cal planarity in each point.We select the best seed point that represents a good seed plane and in the third part,we grow this seed plane by adding all points close to the plane.Thegrowing step is based on a voxel growing algorithm.The filtered normals,the score function and the voxel growing are innovative contributions of our method.As an input,we need dense point clouds related to the level of detail we want to detect.As an output,we produce connected components of planes in the point cloud.This notion of connected components is linked to the data den-sity.With our method,the connected components of planes detected are linked to the parameter d of the voxel grid.Our method has 3”intuitive”parameters :d ,area min and γ.”intuitive”because there are linked to physical mea-surements.d is the voxel size used in voxel growing and also represents the connectivity of points in detected planes.γis the maximum distance between the point of a plane and the plane model,represents the plane thickness and is linked to the point cloud noise.area min represents the minimum area of planes we want to keep.3.2.Details3.2.1Local Density of Point CloudsIn a first step,we compute the local density of point clouds like in [17].For that,we find the radius r i of the sphere containing the k nearest neighbors of point i .Then we cal-culate ρi =kπr 2i.In our experiments,we find that k =50is a good number of neighbors.It is important to know the lo-cal density because many laser point clouds are made with a fixed resolution angle scanner and are therefore not evenly distributed.We use the local density in section 3.2.3for the score calculation.3.2.2Filtered Normal EstimationNormal estimation is an important part of our algorithm.The paper [7]presents and compares three normal estima-tion methods.They conclude that the weighted plane fit-ting or WPF is the fastest and the most accurate for large point clouds.WPF is an idea of Pauly and al.in [17]that the fitting plane of a point p must take into consider-ation the nearby points more than other distant ones.The normal least square is explained in [21]and is the mini-mum of ki =1(n p ·p i +d )2.The WPF is the minimum of ki =1ωi (n p ·p i +d )2where ωi =θ( p i −p )and θ(r )=e −2r 2r2i .For solving n p ,we compute the eigenvec-tor corresponding to the smallest eigenvalue of the weightedcovariance matrix C w = ki =1ωi t (p i −b w )(p i −b w )where b w is the weighted barycenter.For the three methods ex-plained in [7],we get a good approximation of normals in smooth area but we have errors in sharp corners.In fig-ure 1,we have tested the weighted normal estimation on two planes with uniform noise and forming an angle of 90˚.We can see that the normal is not correct on the corners of the planes and in the red circle.To improve the normal calculation,that improves the plane detection especially on borders of planes,we propose a filtering process in two phases.In a first step,we com-pute the weighted normals (WPF)of each point like we de-scribed it above by minimizing ki =1ωi (n p ·p i +d )2.In a second step,we compute the filtered normal by us-ing an adaptive local neighborhood.We compute the new weighted normal with the same sum minimization but keep-ing only points of the neighborhood whose normals from the first step satisfy |n p ·n i |>cos (α).With this filtering step,we have the same results in smooth areas and better results in sharp corners.We called our normal estimation filtered weighted plane fitting(FWPF).Figure 1.Weighted normal estimation of two planes with uniform noise and with 90˚angle between them.We have tested our normal estimation by computing nor-mals on synthetic data with two planes and different angles between them and with different values of the parameter α.We can see in figure 2the mean error on normal estimation for WPF and FWPF with α=20˚,30˚,40˚and 90˚.Us-ing α=90˚is the same as not doing the filtering step.We see on Figure 2that α=20˚gives smaller error in normal estimation when angles between planes is smaller than 60˚and α=30˚gives best results when angle between planes is greater than 60˚.We have considered the value α=30˚as the best results because it gives the smaller mean error in normal estimation when angle between planes vary from 20˚to 90˚.Figure 3shows the normals of the planes with 90˚angle and better results in the red circle (normals are 90˚with the plane).3.2.3The score of local planarityIn many region growing algorithms,the criteria used for the score of the local fitting plane is the residual,like in [18]or [19],i.e.the sum of the square of distance from points to the plane.We have a different score function to estimate local planarity.For that,we first compute the neighbors N i of a point p with points i whose normals n i are close toFigure parison of mean error in normal estimation of two planes with α=20˚,30˚,40˚and 90˚(=Nofiltering).Figure 3.Filtered Weighted normal estimation of two planes with uniform noise and with 90˚angle between them (α=30˚).the normal n p .More precisely,we compute N i ={p in k neighbors of i/|n i ·n p |>cos (α)}.It is a way to keep only the points which are probably on the local plane before the least square fitting.Then,we compute the local plane fitting of point p with N i neighbors by least squares like in [21].The set N i is a subset of N i of points belonging to the plane,i.e.the points for which the distance to the local plane is smaller than the parameter γ(to consider the noise).The score s of the local plane is the area of the local plane,i.e.the number of points ”in”the plane divided by the localdensity ρi (seen in section 3.2.1):the score s =card (N i)ρi.We take into consideration the area of the local plane as the score function and not the number of points or the residual in order to be more robust to the sampling distribution.3.2.4Voxel decompositionWe use a data structure that is the core of our region growing method.It is a voxel grid that speeds up the plane detection process.V oxels are small cubes of length d that partition the point cloud space.Every point of data belongs to a voxel and a voxel contains a list of points.We use the Octree Class Template in [2]to compute an Octree of the point cloud.The leaf nodes of the graph built are voxels of size d .Once the voxel grid has been computed,we start the plane detection algorithm.3.2.5Voxel GrowingWith the estimator of local planarity,we take the point p with the best score,i.e.the point with the maximum area of local plane.We have the model parameters of this best seed plane and we start with an empty set E of points belonging to the plane.The initial point p is in a voxel v 0.All the points in the initial voxel v 0for which the distance from the seed plane is less than γare added to the set E .Then,we compute new plane parameters by least square refitting with set E .Instead of growing with k nearest neighbors,we grow with voxels.Hence we test points in 26voxel neigh-bors.This is a way to search the neighborhood in con-stant time instead of O (logN )for each neighbor like with Kd-tree.In a neighbor voxel,we add to E the points for which the distance to the current plane is smaller than γand the angle between the normal computed in each point and the normal of the plane is smaller than a parameter α:|cos (n p ,n P )|>cos (α)where n p is the normal of the point p and n P is the normal of the plane P .We have tested different values of αand we empirically found that 30˚is a good value for all point clouds.If we added at least one point in E for this voxel,we compute new plane parameters from E by least square fitting and we test its 26voxel neigh-bors.It is important to perform plane least square fitting in each voxel adding because the seed plane model is not good enough with noise to be used in all voxel growing,but only in surrounding voxels.This growing process is faster than classical region growing because we do not compute least square for each point added but only for each voxel added.The least square fitting step must be computed very fast.We use the same method as explained in [18]with incre-mental update of the barycenter b and covariance matrix C like equation 2.We know with [21]that the barycen-ter b belongs to the least square plane and that the normal of the least square plane n P is the eigenvector of the smallest eigenvalue of C .b0=03x1C0=03x3.b n+1=1n+1(nb n+p n+1).C n+1=C n+nn+1t(pn+1−b n)(p n+1−b n).(2)where C n is the covariance matrix of a set of n points,b n is the barycenter vector of a set of n points and p n+1is the (n+1)point vector added to the set.This voxel growing method leads to a connected com-ponent set E because the points have been added by con-nected voxels.In our case,the minimum distance between one point and E is less than parameter d of our voxel grid. That is why the parameter d also represents the connectivity of points in detected planes.3.2.6Plane DetectionTo get all planes with an area of at least area min in the point cloud,we repeat these steps(best local seed plane choice and voxel growing)with all points by descending order of their score.Once we have a set E,whose area is bigger than area min,we keep it and classify all points in E.4.Results and Discussion4.1.Benchmark analysisTo test the improvements of our method,we have em-ployed the comparative framework of[12]based on range images.For that,we have converted all images into3D point clouds.All Point Clouds created have260k points. After our segmentation,we project labelled points on a seg-mented image and compare with the ground truth image. We have chosen our three parameters d,area min andγby optimizing the result of the10perceptron training image segmentation(the perceptron is portable scanner that pro-duces a range image of its environment).Bests results have been obtained with area min=200,γ=5and d=8 (units are not provided in the benchmark).We show the re-sults of the30perceptron images segmentation in table1. GT Regions are the mean number of ground truth planes over the30ground truth range images.Correct detection, over-segmentation,under-segmentation,missed and noise are the mean number of correct,over,under,missed and noised planes detected by methods.The tolerance80%is the minimum percentage of points we must have detected comparing to the ground truth to have a correct detection. More details are in[12].UE is a method from[12],UFPR is a method from[10]. It is important to notice that UE and UFPR are range image methods and our method is not well suited for range images but3D Point Cloud.Nevertheless,it is a good benchmark for comparison and we see in table1that the accuracy of our method is very close to the state of the art in range image segmentation.To evaluate the different improvements of our algorithm, we have tested different variants of our method.We have tested our method without normals(only with distance from points to plane),without voxel growing(with a classical region growing by k neighbors),without our FWPF nor-mal estimation(with WPF normal estimation),without our score function(with residual score function).The compari-son is visible on table2.We can see the difference of time computing between region growing and voxel growing.We have tested our algorithm with and without normals and we found that the accuracy cannot be achieved whithout normal computation.There is also a big difference in the correct de-tection between WPF and our FWPF normal estimation as we can see in thefigure4.Our FWPF normal brings a real improvement in border estimation of planes.Black points in thefigure are non classifiedpoints.Figure5.Correct Detection of our segmentation algorithm when the voxel size d changes.We would like to discuss the influence of parameters on our algorithm.We have three parameters:area min,which represents the minimum area of the plane we want to keep,γ,which represents the thickness of the plane(it is gener-aly closely tied to the noise in the point cloud and espe-cially the standard deviationσof the noise)and d,which is the minimum distance from a point to the rest of the plane. These three parameters depend on the point cloud features and the desired segmentation.For example,if we have a lot of noise,we must choose a highγvalue.If we want to detect only large planes,we set a large area min value.We also focus our analysis on the robustess of the voxel size d in our algorithm,i.e.the ratio of points vs voxels.We can see infigure5the variation of the correct detection when we change the value of d.The method seems to be robust when d is between4and10but the quality decreases when d is over10.It is due to the fact that for a large voxel size d,some planes from different objects are merged into one plane.GT Regions Correct Over-Under-Missed Noise Duration(in s)detection segmentation segmentationUE14.610.00.20.3 3.8 2.1-UFPR14.611.00.30.1 3.0 2.5-Our method14.610.90.20.1 3.30.7308Table1.Average results of different segmenters at80%compare tolerance.GT Regions Correct Over-Under-Missed Noise Duration(in s) Our method detection segmentation segmentationwithout normals14.6 5.670.10.19.4 6.570 without voxel growing14.610.70.20.1 3.40.8605 without FWPF14.69.30.20.1 5.0 1.9195 without our score function14.610.30.20.1 3.9 1.2308 with all improvements14.610.90.20.1 3.30.7308 Table2.Average results of variants of our segmenter at80%compare tolerance.4.1.1Large scale dataWe have tested our method on different kinds of data.We have segmented urban data infigure6from our Mobile Mapping System(MMS)described in[11].The mobile sys-tem generates10k pts/s with a density of50pts/m2and very noisy data(σ=0.3m).For this point cloud,we want to de-tect building facades.We have chosen area min=10m2, d=1m to have large connected components andγ=0.3m to cope with the noise.We have tested our method on point cloud from the Trim-ble VX scanner infigure7.It is a point cloud of size40k points with only20pts/m2with less noise because it is a fixed scanner(σ=0.2m).In that case,we also wanted to detect building facades and keep the same parameters ex-ceptγ=0.2m because we had less noise.We see infig-ure7that we have detected two facades.By setting a larger voxel size d value like d=10m,we detect only one plane. We choose d like area min andγaccording to the desired segmentation and to the level of detail we want to extract from the point cloud.We also tested our algorithm on the point cloud from the LEICA Cyrax scanner infigure8.This point cloud has been taken from AIM@SHAPE repository[1].It is a very dense point cloud from multiplefixed position of scanner with about400pts/m2and very little noise(σ=0.02m). In this case,we wanted to detect all the little planes to model the church in planar regions.That is why we have chosen d=0.2m,area min=1m2andγ=0.02m.Infigures6,7and8,we have,on the left,input point cloud and on the right,we only keep points detected in a plane(planes are in random colors).The red points in thesefigures are seed plane points.We can see in thesefig-ures that planes are very well detected even with high noise. Table3show the information on point clouds,results with number of planes detected and duration of the algorithm.The time includes the computation of the FWPF normalsof the point cloud.We can see in table3that our algo-rithm performs linearly in time with respect to the numberof points.The choice of parameters will have little influence on time computing.The computation time is about one mil-lisecond per point whatever the size of the point cloud(we used a PC with QuadCore Q9300and2Go of RAM).The algorithm has been implented using only one thread andin-core processing.Our goal is to compare the improve-ment of plane detection between classical region growing and our region growing with better normals for more ac-curate planes and voxel growing for faster detection.Our method seems to be compatible with out-of-core implemen-tation like described in[24]or in[15].MMS Street VX Street Church Size(points)398k42k7.6MMean Density50pts/m220pts/m2400pts/m2 Number of Planes202142Total Duration452s33s6900sTime/point 1ms 1ms 1msTable3.Results on different data.5.ConclusionIn this article,we have proposed a new method of plane detection that is fast and accurate even in presence of noise. We demonstrate its efficiency with different kinds of data and its speed in large data sets with millions of points.Our voxel growing method has a complexity of O(N)and it is able to detect large and small planes in very large data sets and can extract them directly in connected components.Figure 4.Ground truth,Our Segmentation without and with filterednormals.Figure 6.Planes detection in street point cloud generated by MMS (d =1m,area min =10m 2,γ=0.3m ).References[1]Aim@shape repository /.6[2]Octree class template /code/octree.html.4[3] A.Bab-Hadiashar and N.Gheissari.Range image segmen-tation using surface selection criterion.2006.IEEE Trans-actions on Image Processing.1[4]J.Bauer,K.Karner,K.Schindler,A.Klaus,and C.Zach.Segmentation of building models from dense 3d point-clouds.2003.Workshop of the Austrian Association for Pattern Recognition.1[5]H.Boulaassal,ndes,P.Grussenmeyer,and F.Tarsha-Kurdi.Automatic segmentation of building facades using terrestrial laser data.2007.ISPRS Workshop on Laser Scan-ning.1[6] C.C.Chen and I.Stamos.Range image segmentationfor modeling and object detection in urban scenes.2007.3DIM2007.1[7]T.K.Dey,G.Li,and J.Sun.Normal estimation for pointclouds:A comparison study for a voronoi based method.2005.Eurographics on Symposium on Point-Based Graph-ics.3[8]J.R.Diebel,S.Thrun,and M.Brunig.A bayesian methodfor probable surface reconstruction and decimation.2006.ACM Transactions on Graphics (TOG).1[9]M.A.Fischler and R.C.Bolles.Random sample consen-sus:A paradigm for model fitting with applications to image analysis and automated munications of the ACM.1,2[10]P.F.U.Gotardo,O.R.P.Bellon,and L.Silva.Range imagesegmentation by surface extraction using an improved robust estimator.2003.Proceedings of Computer Vision and Pat-tern Recognition.1,5[11] F.Goulette,F.Nashashibi,I.Abuhadrous,S.Ammoun,andurgeau.An integrated on-board laser range sensing sys-tem for on-the-way city and road modelling.2007.Interna-tional Archives of the Photogrammetry,Remote Sensing and Spacial Information Sciences.6[12] A.Hoover,G.Jean-Baptiste,and al.An experimental com-parison of range image segmentation algorithms.1996.IEEE Transactions on Pattern Analysis and Machine Intelligence.5[13]H.Hoppe,T.DeRose,T.Duchamp,J.McDonald,andW.Stuetzle.Surface reconstruction from unorganized points.1992.International Conference on Computer Graphics and Interactive Techniques.2[14]P.Hough.Method and means for recognizing complex pat-terns.1962.In US Patent.1[15]M.Isenburg,P.Lindstrom,S.Gumhold,and J.Snoeyink.Large mesh simplification using processing sequences.2003.。
基于u-net的“高分五号”卫星高光谱图像土地类型分类

第40卷第6期航天返回与遥感2019年12月SPACECRAFT RECOVERY & REMOTE SENSING99基于U-net的“高分五号”卫星高光谱图像土地类型分类孙晓敏1,2,3郑利娟4吴军5陈前6徐崇斌1马杨1陈震1(1 北京空间机电研究所,北京 100094)(2 北京航天创智科技有限公司,北京 100076)(3 北京市航空智能遥感装备工程技术研究中心,北京 100094)(4 自然资源部国土卫星遥感应用中心,北京 100048)(5 国网湖北省电力有限公司直流运检公司,宜昌 443000)(6 中国资源卫星应用中心,北京 100094)摘要“高分五号”卫星是世界首颗实现对大气和陆地综合观测的全谱段高光谱卫星,对于土地利用类型分类具有重要的应用价值,如何利用深度学习技术开展高光谱图像分类是当前研究的热点问题。
深度学习中的语义分割方法在地面场景的图像中已经获得较好的应用,但是对于高光谱遥感图像的精度和适用性较差,无法准确获得精确的分类结果。
文章采用U-net模型开展高光谱土地利用类型分类研究,首先基于“高分五号”卫星高光谱数据,构建样本数据集,然后训练分类模型,进行土地利用类型分类,探讨语义分割方法在高分五号高光谱数据上的应用能力。
结果表明,采用深度学习中的语义分割方法能够有效提高精度水平,U-net模型的整体分类精度为0.9357,Kappa系数达到0.92,均高于SVM方法和CNN方法。
采用深度学习中的语义分割方法,可以为“高分五号”高光谱数据的土地利用分类提供技术支撑,有效提升“高分五号”卫星的应用能力。
关键词U-网络模型深度学习高光谱图像土地利用分类高分五号卫星应用中图分类号: TP75文献标志码: A 文章编号: 1009-8518(2019)06-0099-08DOI: 10.3969/j.issn.1009-8518.2019.06.012Land Classification of GF-5 Satellite HyperspectralImages Using U-net ModelSUN Xiaomin1,2,3 ZHENG Lijuan4 WU Jun5 CHEN Qian6 XU Chongbin1 MA Yang1CHEN Zhen1(1 Beijing Institute of Space Mechanics & Electricity, Beijing, 100094, China)(2 Beijing Aerospace Innovative Intelligence Science and Technology Co., Ltd, Beijing, 100076, China)(3 Beijing Engineering Technology Research Center of Aerial Intelligence Remote Sensing Equipments, Beijing, 100094, China)(4 Land Satellite Remote Sensing Application Center, Ministry of Natural Resources of P R China, Beijing, 100048, China)(5 State Grid Hubei DC Operation & Mainteance Company, Yichang, 443000, China)(6 China Centre for Resources Satellite Data and Application, Beijing, 100094, China)Abstract GF-5 satellite is the world's first full-spectrum hyper-spectral imagery satellite to observe the收稿日期:2019-07-12引用格式:孙晓敏, 郑利娟, 吴军, 等. 基于U-net的“高分五号”卫星高光谱图像土地类型分类[J]. 航天返回与遥感, 2019, 40(6): 99-106.SUN Xiaomin, ZHENG Lijuan, WU Jun, et al. Land Classification of GF-5 Satellite Hyperspectral Images Using U-net Model[J]. Spacecraft Recovery & Remote Sensing, 2019, 40(6): 99-106. (in Chinese)100航天返回与遥感2019年第40卷Earth’s atmosphere and surface. It is important for the classification of land use types. How to use deep learning technology to carry out hyperspectral image classification is a hot issue in current research. Semantic segmentation method in depth learning has been well applied in the image of ground scene, but the accuracy and applicability of hyperspectral remote sensing images are relatively poor, and the accurate classification results are difficult to be obtained. In this paper, the U-net model is used to study the classification of hyperspectral land use types. Firstly, based on the hyperspectral data of GF-5 satellite, the sample data set is constructed, then the classification model is trained, the land use type classification is carried out, and the application ability of semantic segmentation method on hyperspectral data of GF-5 satellite is discussed. The results show that the semantic segmentation method in deep learning can effectively improve the accuracy level. The overall classification accuracy of the U-net model is 0.9357, and the Kappa coefficient is 0.92, which is higher than the SVM method and CNN method. Using the semantic segmentation method in deep learning, it can provide technical support for land use classification of GF-5 satellite hyperspectral data, and effectively improve the application ability of GF-5 satellite.Keywords U-net model; deep learning; land use classification; hyper-spectral images; GF-5; satellite applications0引言“高分五号”(GF-5)卫星是我国高分辨率地球观测系统的最重要的遥感卫星之一,是我国实现高光谱分辨率对地观测能力的重要标志[1]。
基于高光谱成像的苹果病害无损检测方法

基于高光谱成像的苹果病害无损检测方法刘思伽;田有文;冯迪;张芳;崔博【摘要】Disease is easy to occur in apple fruit. Traditional detection of apple disease is not adapted to the requirement of apple grading on-line detection. In order to achieve the fast, effective online detection for the disease apple, hyperspectral imaging was adopted to study the nondestructive detection of the anthracnose, bitter pox disease and black fruit rot and leaf spot disease in Hanfu apple. According to the relative reflectance spectrum difference between disease area and normal area, the improved manifold distance method was proposed. The total improved manifold distance L value was comprehensive calculated by the relative reflectance spectra of the disease and normal area, disease with stem/calyx area, normal and stem/calyx area. So three feature wavelengths were selected respectively from the whole band wavelength, 700, 765, 904nm. In order to get the mask image, the image of the characteristic wave band at 700 nm was threshold segmented. The interested area was extracted after secondary threshold segmentation of the mask image. The relative reflectance spectra of the three characteristic wave bands were combined, respectively, as the BP neural network input vector, to detect whether apple fruit was diseased. Finally, the relative reflectance spectra under 700 nm to 904 nm band were selected as the best combination by comparing the detection results. A recognition rate of the normal apples and diseased apples respectively were 96.25%. Results showed that the twocharacteristics of band obtained by hyperspectral imaging technology can effectively detect disease for apple and provide the reference for the development of multispectral imaging of appleˊs quality detection and classification system.%苹果果实易发生病害,传统的苹果病害的检测不适应苹果分级在线检测的要求。
HSI Classification by Exploiting the Spectral-Spatial Correlations in the Sparse Coefficients

Hyperspectral Image Classification by Exploiting the Spectral-Spatial Correlations in the Sparse CoefficientsDan Li u, Sh u tao Li, and Ley u an Fan gColle g e of Electrical and Information En g ineerin g,H u nan University, Chan g sha, 410012, China{liudan1,shutao_li,leyuan_fang}@ Abstract. This paper proposes a novel hyperspectral ima g e (HSI) classificationmethod based on sparse model, which incorporates the spectral and spatial in-formation of the sparse coefficient. Firstly, a sparse dictionary is b u ilt by u sin gthe trainin g sampl es and the sparse coefficient is obtained thro ug h the sparserepresentation method. Secondly, a probability map for each class is establishedby s u mmin g the sparse coefficients of each class. Thirdly, the mean filterin g isapplied on each probability map to exploit the spatial information. Finally, wecompare the probabil ity map to find the maxim u m probabil ity for each pixeland then determine the class label of each pixel. Experimental res u l ts demon-strate the effectiveness of the proposed method.Keywords: Hyperspectral ima g e classification, sparse representation, spectral-spatial information, mean filter.1IntroductionHyperspectral ima g e (HSI) is formed by tens to h u ndreds of contin u o u s and s u bdivided spectral bands while reflectin g interested tar g et areas sim u ltaneo u sl y. In HSI, different materials have different spectral information, which can be u sed for classification.Many m u ltispectral ima g e classification methods, s u ch as s u pport vector machines (SVMs) [1], [2], ne u ral network [3], and adaptive artificial imm u ne network [4], have been applied to HSI classification. Generally, these methods have obtained g ood per-formance.Researchers show that HSI contains rich spatial information and the pixel s in a small nei g hborhood have similar spectral characteristics. If the pixels are in a small nei g hbor, they sho u ld belon g to the same material. Therefore, Some methods [5], [6], [7] have combined spectral information and spatial information, and the classification acc u racy has been improved. In partic u l ar, the se g mentation based method [8] first se g ment the HSI into many local re g ion with similar spectral characteristics and then cl assify each re g ion. After u sin g the spatial information, the cl assifiers can obtain improved performance.Recentl y, sparse representation has become a powerf u l tool to solve some prob-lems, s u ch as face reco g nition [9], tar g et detection [10], [11], remote sensin g ima g e S. Li et al. (Eds.): CC P R 2014, P art I, CCIS 483, pp. 151–158, 2014.© Sprin g er-Verla g Berlin Heidelber g 2014152 D. Li u , S. Li, and L. Fan gf u sion [12] and medical imag e reconstr u ction [13], [14]. Recently, the sparse repre-sentation method has also been extended to HSI classification [7], [15], [16]. Basical-l y, the previo u s sparse representation based HSI cl assification methods u til ize the reconstr u ction error for the classification. In this paper, we propose a novel method that can combines the spatial information and spectral information in the sparse coef-ficients for the cl assification. Firstl y, we u se the trainin g sampl es to constr u ct the trainin g dictionary and then u til ize the sim u ltaneo u s ortho g onal matchin g p u rs u it (SOM P ) to obtain the sparse coefficient of each spectral pixel. Differ from other sparse representation based methods which u ses the resid u al to determine the pixel ’s class, the proposed method first employs the coefficients to constr u ct several proba-bility maps. S u bseq u ently, we exploit the spatial information by filterin g every map and g ain a probabil ity map for each cl ass. Final l y, we can determine the pixel ’s class by comparin g the probability maps.The rest of this paper is constr u cted as follows. Section 2 introd u ces the proposed cl assification method. Section 3 shows the experimental res u lts and concl u sions are g iven in the section 4.2 The Proposed Classification MethodFi g . 1 shows the schematic of the proposed classification method. It is constr u cted by fo u r steps: Firstl y, the sparse representation method is adopted to obtain the sparse coefficients. Then, the coefficients bel on g in g to each cl ass are s u mmed to obtain probability map for each pixel. S u bseq u entl y, a mean filterin g is cond u cted on each probabil ity map to expl oit the spatial information. Final l y, cl assification is accom-plished by comparin g the maps. The details of each step are ill ustrated in the follows.x x x 4x x }}}Fig. 1. The scheme of the proposed classification methodHyperspectra l Ima g e Classification by Exploitin g the Spectral-Spatial Correlations 153 Step 1: In HSI, every spectral pixel can be re g arded as a vector i x and the trainin g pixels constr u ct a matrix =12n D [d ,d ,...,d ]which is called dictionary. Every pixel canbe represented by the dictionary.1212...n i i i n i i ααα=+++=x d d d D α (1)In the eq u ation (1), 12,,...,n d d d is cal l ed atom and 12[,,...,]n i i i i ααα=αis cal ed sparse coefficient vector. The sparse coefficient vector can be obtained by solvin g the optimization problem.200ˆar g min s u bject to i i i i K =−≤αx A αα (2)where 0K is the maxim u m val u e of the sparsity level. This optimization problem is aN P -hard and cannot be sol ved directl y. However, it can be sol ved by g reedy al g o-rithms approximately, s u ch as s u bspace p u rs u it (S P ) [17], ortho g onal matchin g p u r-s u it (OM P ) [18] and Sim u ltaneo u s OM P (SOM P ) [7]. In this paper, the SOM P isadopted to obtain the sparse coefficient vector ˆi αfor each spectral pixel i x . Step 2: In the sparse coefficient vector ˆi α, there are onl y a few nonzero sparse coefficients. The lar g er the nonzero coefficients val u es in one specific class, the more probability the test pixel belon g s to this class. We denote the nonzero coefficients in one cl ass as the ,i m α, where {1,2,...,}m M ∈, and M is the total n u mber of cl asses.Then, we s u m the nonzero coefficients ,i m αfor each class of each spectral pixel,(),,s u m ,{1,2,...,},and {1,2,...,}s um i m i m m M i N =∈∈αα (3) where N is the total n u mber of spectral pixels in the HSI. In each class, the s u mmed coefficients ,s um i m αfor al l the spectral pixel s in the HSI can constr u ct one probabil itymap m z .Step 3: As disc u ssed above, one coefficient in a class probability map m z can be re-g arded as the l ikel ihood for the correspondin g pixel bel on g in g to this cl ass. If the probability map m z is directly u sed for determinin g the class of each pixel, the spatialinformation in the probability map is not exploited. To exploit the spatial information, a mean filterin g operation is cond u cted on each m z ,()meanfilterin g ,{1,2,...,}m e an f m m z z m M =∈ (4)where the window for mean operation is selected to 3×3.Step 4: the cl ass l abel of each pixel i x is obtained by comparin g the coefficients in the fil tered probabil ity maps,,1,...,ˆmax (),{1,2,...,}m e an f i m i i m Mm z i N ==∈x (5) where max is the operation to comp u te the max coefficient amon g different maps.154 D.Li u, S. Li, and L. Fan g3Experimental ResultsThis section tests the effectiveness of the proposed classification method on two real HSIs (Indian pines and Salinas scene). The classification res u lts of the proposed me-thod are compared with those obtained by SVM [19], SVM-CK [20], OM P [7] and SOM P [7]. SVM [19] is desi g ned for the classification of the spectral pixel witho u t u tilizin g the spatial information. SVM-CK [20] is a method that incorporates spatial information via a composite kernel. OM P and SOM P are two sparse representation based methods.In o u r first experiment, we u sed Airborne Visible/Infrared Ima g in g Spectrometer (AVIRIS) ima g e Indian pines as testin g HSI. This ima g e is a widely u sed data set and was taken over Indiana’s Indian P ine test site in J u ne 1992. The Indian P ines has a size of 145×145×220, with 220 spectral bands. Beca u se 20 bands is water absorp-tion, these bands are removed. There are 16 g ro u nd-tr u th classes and the size is from 20 to 2455 pixels (the total pixels are 10249).We chose 10% of the samples for each class as trainin g sample and the remainder as testin g sampl es. For each method, we did five experiments and avera g ed the re-s u l ts. The n u mber of the trainin g sample and the testin g sample is presented in Table 1.In this table, we can see the overall acc u racy (OA), avera g e acc u racy (AA) and the kappa coefficient by u sin g different methods (the SOM P-P is denoted as o u r method). Table 1. Trainin g sets, testin g sets and cl assification acc u racy (%) obtained from different methods for the Indian P ines ima g el ass Train Test SVM SVM-CK OM P SOM P SOM P-P Cl fa 6 40 77.73 91.25 55.1292.26 95.04Al faCorn-N 144 1284 77.35 92.79 61.60 93.46 97.77Corn-M 84 746 78.56 93.98 58.62 90.22 97.4042.2187.32 95.1168.75Corn 24 21387.28Grass-M 50 433 88.87 94.90 87.29 95.20 94.04Grass-T 75 655 89.12 99.51 95.30 96.12 96.57 Grass-P 3 25 95.37 85.20 85.20 87.10 87.14Hay-W 49 429 95.09 99.91 96.44 99.10 99.8767.65Oats 2 1883.33 36.67 55.78 0Soybean-N 97 875 78.64 90.33 71.10 93.45 93.47Soybean-M 247 2208 81.19 96.25 74.11 95.10 99.20Soybean-C 62 531 79.74 89.04 51.05 87.49 97.61Wheat 22 183 92.26 99.07 96.85 88.20 97.76 Woods 130 1135 92.72 98.63 91.85 99.00 100B u ildin g s 38 348 69.79 92.64 41.67 83.05 97.7291.51 99.35stone 10 83 97.96 90.24 91.9093.6673.3894.82OA - -82.9197.4983.17AA - -92.77 71.06 89.83 91.010.6960.9310.8050.941k - -0.971Hyperspectra l Ima g e Classification by Exploitin g the Spectral-Spatial Correlations 155 The Table 1 shows the trainin g sets, testin g sets and classification maps obtained by SVM, SVM-CK, OM P, SOM P and SOM P-P and the res u lt is the avera g e of five experiments. From the Tabl e 1, we can see that o u r al g orithm has the best perfor-mance in terms of overall acc u racy and kappa coefficient. As for its avera g e acc u racy, it is only a little worse than the classifier SVM-CK.(a)(b)(c)(d)(e)(f)(g)Fig. 2. Indian P ines: (a) Train samples, (b)Test samples, and the classification res u lts obtainedby (c) SVM, (d) SVM-CK, (e) OM P, (f) SOM P, (g) SOM P-PTable 2. Trainin g sets, testin g sets and cl assification acc u racy (%)obtained from different methods for the Salinas scene ima g eCl ass TrainTestSVMOM P SOM P SOM P-P Weed_1 20198999.8898.68100 100 Weed_2 37368998.5298.7899.7299.95Fall ow 20195692.4894.5598.70 98.41Fallow plow 14 1380 97.46 99.35 96.93 99.69Fallow smooth 27 2651 97.19 93.26 97.45 99.24 St u bbl e 40391999.9899.7299.97100 Cel ery 36354398.1499.4099.55100 Grapes 1131115876.1172.8384.5094.12Soil 62614198.6397.4199.37100 Corn 33324589.2988.1495.2498.04Lett u ce 4wk 11 1057 92.82 96.18 99.26 100Lett u ce 5wk 19 1908 96.16 99.77 96.73 99.73 Lett u ce 6wk 9 907 94.99 98.05 92.53 99.15Lett u ce 7wk 11 1059 94.85 90.87 97.40 99.43 Vineyard u ntrained 73 7195 71.90 57.77 85.24 83.15Vineyard trellis 18 1789 98.87 95.06 98.91 98.92 OA --89.1686.4893.4796.21AA --93.5992.5396.1398.11k --0.8900.8490.92740.958156 D.Li u, S. Li, and L. Fan gIn the Fi g. 3, (a) and (b) are an example of the trainin g and testin g samples. (c) is the classification map obtained from SVM, similarly, (d), (e), (f) are the classification maps of SVM-CK, OM P, SOM P and SOM P-P respectively.In o u r second experiment, we u se the HSI Salinas scene which was collected by 224-band over Sal inas Val ey and Cal ifornia. The size of the Sal inas ima g e is 512×217×224.Al so, beca u se 20 bands is water absorption which is the same as Indian P ines, the n u mber of bands is red u ced to 204. There are 16 g ro u nd-tr u th classes containin g ve g etables, bare soils, and vineyard fields and the size is from 916 to11271 pixels (the total pixels are 54129).We chose 1% of the samples for each class as trainin g sample and the rest as test-in g sample. The n u mber of the trainin g sample and the testin g sample is presented in Table 2. In this table, we can see the overall acc u racy (OA), avera g e acc u racy (AA) and the kappa coefficient by u sin g different methods (the SOM P-P is o u r method). It is easy to see that the performance of the proposed methods is fine. The Fi g. 2 shows the classification maps.Fig. 3. Salinas scene: (a) Train samples, (b) Test samples, and the classification res u lts obtained by(c) SVM, (d) OM P, (e) SOM P, (f) SOM P-P4ConclusionsIn this paper, we have proposed a novel HSI cl assification method base on sparse representation. Differ from other traditional sparse classification technolo g ies which expl oit the sparse coefficient and resid u al to cl assify directl y, this method u ses the sparse coefficient to constr u ct probability maps and then exploits the spatial informa-tion in the maps for classification. Experimental res u lts show that the proposed me-thod has better performance than several well-known classifiers.Hyperspectra l Ima g e Classification by Exploitin g the Spectral-Spatial Correlations 157 Acknowledgement. This work was s u pported in part by the National Nat u ral Science Fo u ndation of China u nder Grant No. 61172161, the National Nat u ral Science Fo u n-dation for Distin gu ished Yo u n g scholars of China u nder Grant No. 61325007.References1.G u altieri, J.A., Cromp, R.F.: S u pport Vector Machines for Hyperspectral Remote Sensin gClassification. In: P roc. S P IE, vol. 3584, pp. 221–232 (1998)2.Mel g ani, F., Br u zzone, L.: Cl assification of Hyperspectral Remote Sensin g Ima g e withS u pport Vector Machines. IEEE. Trans. Geosci. Remote Sens. 42(8), 1778–1790 (2004) 3.Ratle, F., Camps, G.V., Weston, J.: Semis u pervised Ne u ral Networks for Efficient Hyper-spectral Ima g e Classification. IEEE Trans. Geosci. Remote Sens. 48(5), 2271–2282 (2010) 4.Zhon g, Y., Zhan g, L.: An Adaptive Artificial Imm u ne Network for S u pervised Classifica-tion of M u ti-/Hyperspectra Remote Sensin g Ima g ery. IEEE Trans. Geosci. Remote Sens. 50(3), 894–909 (2012)5.Rand, R.S., Keenan, D.M.: Spatial l y smooth partitionin g of hyperspectral ima g ery u sin gspectral/spatial meas u res of disparity. IEEE Trans. Geosci. Remote Sens. 41(6), 1479–1490 (2003)6.Kan g, X., Li, S., Fan g, L.: Extended Random Walker-Based Classification of Hyperspec-tral Ima g es. IEEE Trans. Geosci. Remote Sens., 1–10 (May 2014)7.Chen, Y., Nasrabadi, N.M., Tran, T.D.: Hyperspectral Ima g e Classification Usin g Dictio-nary-Based sparse Representation. IEEE Trans. Geosci. Remote Sens. 49(10), 3973–3985 (2011)8.Driesen, J., Thoonen, G., Sche u nders, P.: Spatial Hyperspectral Ima g e Cl assification byP rior Se g mentation. In: IEEE Geosci. Remote Sens. Symp., vol. 3, pp. 709–712 (2009)9.John, W., Yan g, A.Y., Arvind, G., Sastry, S.S., Ma, Y.: Rob u st Face Reco g nition viaSparse Representation. IEEE Trans. P attern Anal. 31(2), 210–227 (2009)10.Chen, Y., Nasrabadi, N.M., Tran, T.D.: Sparse Representation for Tar g et Detection inHyperspectral Ima g ery. IEEE Jo u rnal of Selected Topics in Si g nal P rocessin g 5(3), 629–640 (2011)11.Fan g, L., Li, S., H u, J.: M u ltitemporal ima g e chan g e detection with compressed sparse re-presentation. IEEE Ima g e P rocessin g, 2673–2676 (2011)12.Li, S., Yin, H., Fan g, L.: Remote Sensin g Ima g e F u sion via Sparse Representations OverLearned Dictionaries. IEEE Trans. Geosci. Remote Sens. 51(9), 4779–4789 (2013)13.Fan g, L., Li, S., Kan g, X., Benediktsson, J.A.: Spectral-Spatial Hyperspectral Ima g e Clas-sification via M u l tiscal e Adaptive Sparse Representation. IEEE Trans. Geosci. Remote Sens., 1–12 (2014)14.Fan g, L., Li, S., Ryan, M., Qin g, N., Anthony, K.: Fast Acq u isition and Reconstr u ction ofOptical Coherence Tomo g raphy Ima g e via Sparse Representation. IEEE Trans. Med. Im-a g. 32(11), 2034–2049 (2013)15.Fan g, L., Li, S., Kan g, X., Benediktsson, J.: Spectral-Spatial Hyperspectral Ima g e Classifi-cation via M u ltiscale Adaptive Sparse Representation. IEEE Trans. Geosci. Remote Sens., 1–12 (2014)16.Fan g, L., Li, S., Kan g, X.: Spectral-Spatial Hyperspectral Ima g e Classification via M u ltis-cal e Adaptive Sparse Representation. IEEE Trans. Geosci. Remote Sens. 52(12), 7738–7749 (2014)17.Dai, W., Mil enkovic, O.: S u bspace Pu rs u it for Compressive Sensin g Si g nal Reconstr u c-tion. IEEE Trans. Inf. Theory 55(5), 2230–2249 (2009)158 D.Li u, S. Li, and L. Fan g18.Tropp, J., Gilbert, A.: Si g nal recovery from random meas u rements via ortho g onal match-in g p u rs u it. IEEE Trans. Inf. Theory 53(12), 4655–4666 (2007)19.G u altieri, J.A., Cromp, R.F.: S u pport Vector machines for Hyperspectral Remote Sensin gClassification. In: P roc. S P IE, vol. 3584, pp. 221–232 (1998)20.Fa u vel, M., Chan u ssot, J., Benediktsson, J.A.: Adaptive P ixel Nei g hborhood Definition forthe Classification of Hyperspectral Ima g es with S u pport Vector Machines and Composite Kernel. In: P roc. IEEE Int. Conf. Ima g e P rocess., pp. 1884–1887 (2008)。
遥感erdas界面翻译

一、主页(一)Information1、ContentsContentsRetrieverGeocoder2、MetadataView/Edit Image Metadata View/Edit Point Cloud Metadata View/Edit Vector Metadata View/Edit Annotation Metadata View/Edit NITF MetadataView/Edit IMAGINE HFAEdit Image Metadata3、SelectSelectSelect by BoxSelect by LineSelect by EllipseSelect by PolygonFollow HyperlinksSelector PropertiesPick Properties4、InquireInquire BoxInquire ShapeInquire Color5、MeasureMeasure(二)Edit1、Cut2、Copy3、Paste4、Delete5、Undo6、Paste From Selected Object (三)Extent1、Fit to Frame2、ResetReset (一)信息内容:内容检索地理编码元数据:查看/编辑影像元数据查看/编辑点云元数据查看/编辑矢量元数据查看/编辑注记元数据查看NITF元数据查看IMAGINE HFA文件内容元数据编辑选择:选择通过拉框选择通过画线选择通过画椭圆选择通过画多边形选择跟踪超链接选择器属性采集器属性查询:查询框查询光标形状查询光标颜色测量:测量(二)编辑剪切复制粘贴删除撤销由选中内容粘贴(三)内容全景显示重置:重置(七)Roam1、HorizontalHorizontalVerticalUser-Defined2、Speed Down3、Speed4、Speed UpSpeed UpSpeed Reset5、Go to Start6、Step Backwards7、Reverse8、Stop9、Start/Pause10、Step Forwards11、Go to End12、Snail TrailSnail TrailMerge Snail Trails13、Roam Properties(七)漫游水平:水平垂直自定义减速速度加速:加速重置速度从头开始快退倒退停止开始/暂停快进到最后追踪:追踪合并追踪漫游属性二、Manage Data管理数据(一)Catalog1、Hexagon ContentHexagon ContentImage Catalog(二)Conversion1、Coordinate Calculator2、Import Data3、Export Data4、GeoMedia ToolsShapefiles to WarehouseWarehouse to ShapefilesGeoMedia Utilities5、Pixels To ASC||6、ASC|| to Pixels7、Graphical Importer(一)目录海克斯康内容:海克斯康内容影像目录(二)转换坐标计算器数据导入数据导出空间数据仓库工具:Shp转仓库仓库转Shp空间数据仓库工具栅格转文本文本转栅格GeoRaster导入器8、GeoRaster Manager9、Imagizer Data Prep(三)VectorizeRaster to Shape to Shape to Annotation (四)Rasterize1、Vector to RasterVector to RasterAnnotation to Raster(五)Image1、Edit Image Metadata2、Pyramids &StatisticsCompute Pyramids and Statistics Create ThumbnailsProcess Footprints and RSETS3、Compare Images4、Create New Image5、Create ECW Transparency(六)NITF/NSIF1、NITFView NITF MetadataExtract Shape LASDPPDB WorkstationMake RPF TOCCIB JPEG 2000 ExporterRPF Frame SearchMake ECRG/ECIB TOCECRG/ECIB Frame Search(七)Office Tools1、Send to PowerPoint2、Send to Word3、Send to GeoPDF4、Send to JPEG GeoRaster管理器IMAGIZER数据预处理(三)矢量化栅格转shp:栅格转shp栅格转注记(四)栅格化矢量转栅格:矢量转栅格注记转栅格(五)影像编辑元数据金字塔和统计:计算金字塔和统计创建缩略图处理范围线和金字塔影像比较创建新影像创建ECW透明度(六)NITF/NSIFNITF:查看NITF元数据提取Shp文件提取LAS文件合并数字式目标定位数据库工作站生成影像目录CIB JPEG 2000输出RPF帧搜索制作ECRG/ECIB目录表ECRG/ECIB帧搜索(七)Office工具发送到PPT发送到Word发送到GeoPDF发送到JPEG三、Raster栅格(一)Resolution (一)分辨率·Two Layer Union Operators Zonal AttributesMatrix UnionSummary Report of Matrix Overlay by Min or Max Index by Weighted Sum (七)Scientific1、FunctionsTwo Image Functions Single Image Functions2、Fourier Analysis Fourier TransformFourier Transform Editor Inverse Fourier Transform Fourier Magnitude ·双层联合计算区域分析矩阵分析归纳分析叠加分析加权分析(七)科学的函数分析:两个影像函数单个影像函数傅里叶分析:傅里叶变换傅里叶变换编辑傅里叶逆变换傅里叶幅值计算四、Vector矢量(一)Manage1、Copy Vector Layer2、Rename Vector Layer3、Delete Vector Layer4、Buffer Analysis5、Attribute to Annotation(二)ER Mapper Vector to Shape 1、Reproject Shape Shape Elevation (三)Raster To VectorRaster to Shapefile (一)管理复制矢量重命名矢量删除矢量缓冲区分析属性转注记ER映射矢量转Shp Shape重投影Shp裁切高程重投影(二)栅格转矢量栅格转Shp五、Terrain地形(一)Manage (一)管理六、Toolbox工具箱(一)Common1、IMAGINE Photogrammetry2、Image Equalizer3、Spatial Model Editor Spatial Model EditorLaunch Spatial Model4、Model MakerModel MakerModel Librarian5、MosaicMosaicProMosaicPro from 2D View Mosaic ExpressUnchip NITF6、AutoSync Workstation AutoSync Workstation Georeferencing wizardEdge Match WizardOpen AutoSync Project7、Stereo AnalystStereo AnalystAuto-Texturize from Block Texel MapperExport 3D shape KML Extended Features to Ground 8、MapsMap Series ToolMap Database ToolEdit Composition Paths9、VirtualGISVirtual World EditorCreate MovieRecord Flight Path with GPS Create TIN Mesh (一)通用图像摄影测量影像匀光器空间模型编辑器:空间模型编辑器发射空间模型空间建模:空间建模模型库管理影像镶嵌:启动专业镶嵌镶嵌视窗显示影像镶嵌快车合并自动配准:自动配准地理参考向导边缘匹配向导打开自动配准工程立体分析:立体分析自动纹理纹理编辑器输出Shp为KML构建地面实体地图工具:图幅地图地图数据库工具修改制图文件路径虚拟GIS:虚拟世界编辑器录像录制通过GPS点定义飞行路径建立不规则三角网七、Help帮助(一)Reference Library1、Help2、About IMAGINE3、Reference booksHexGeoWiki4、WorkflowsCommon WorkflowsSpatial Modeler WorkflowsClassification WorkflowsPhotogrammetry WorkflowsPoint Cloud WorkflowsZonal Change WorkflowsRectification WorkflowsMap Making WorkflowsMosaic WorkflowsVector WorkflowsModel Maker WorkflowsNITF Workflows5、User GuidesAAIC User GuideAutonomous Spectral Image Processing User GuideAutoSync User GuideDeltaCue User GuideHyperspectral User GuideIMAGINE Objective User GuideIMAGIZER Data Prep User GuideIMAGIZER Viewer User Guide Photogrammetry Suite Contents Operational Radar User GuideRader InterferometryStereo Analyst User GuideSubpixel Classifier User GuideVirtual GIS User GuideInstallation and Configuration Guide6、Spatial ModelingSpatial Model EditorModel Maker(Legacy) (一)相关阅览帮助关于IMAGINE参考书:希格维基工作流:常见工作流空间建模器工作流分类工作流摄影测量工作流点云工作流分区变化工作流整流工作流地图制图工作流镶嵌工作流矢量工作流模型制作工作流NITF工作流使用指南:AAIC使用指南自主光谱图像处理用户指南自动配准使用指南变化检测使用指南高光谱使用指南IMAGINE面向对象使用指南IMAGIZER数据准备使用指南IMAGIZER视窗使用指南摄影测量套件目录操作雷达用户指南雷达干涉立体分析使用指南子像元分类使用指南虚拟GIS使用指南安装与配置指南空间建模:空间模型编辑器模型制作者(Legacy)Spatial Modeler Language(Legacy)Graphical Models Reference Guide7、Language ReferenceERDAS Macro Language8、Release NotesERDAS IMAGINE Issues ResolvedAutonomous Spectral Image ProcessingRelease Notes9、ERDAS IMAGINE Release Notes(二)Search Commands1、Search2、Search Box(三)Page1、Previous2、Next空间建模语言(Legacy)图估模型的参考指南建模和定制:ERDAS宏语言发布说明:ERDAS IMAGINE问题解决自主光谱图像处理发布说明ERDAS IMAGINE发行说明(二)搜索命令搜索搜索框(三)页码向前向后八、Multispectral(一)Enhancement1、Adjust RadiometryGeneral ContrastBrightness/ContrastPhotography EnhancementsPiecewise ContrastBreakpointsLoad BreakpointsSave BreakpointsData Scaling2、Discrete DRADiscrete DRADRA Properties(二)Brightness Contrast1、Contrast Down/Up2、Brightness Down/Up(三)Sharpness1、Sharpness Down/Up2、FilteringConvolution FilteringStatistical FilteringReset Convolution(四)Bands1、Sensor Types2、Common Band Combinations (五)View1、Set Resampling Method2、Pixel Transparency(六)Utilities1、Subset & ChipCreate Subset ImageNITF ChipMaskDice ImageImage Slicer2、Spectral Pro Pro Pro Pro Features3、Pyramids & Statistics Compute Pyramids && Statistics Compute Statistics on Window Generate RSETs(七)Transform & Orthocorrect 1、Transform & OrthoOrtho Using Existing ModelOrtho With Model Selection Transform Using Existing Model Create Affine CalibrationPerform Affine Resample Resample Pixel Size2、Control Points3、Single Point4、Check Accuracy(八)Edit1、Fill2、Offset3、Interpolate九、Drawing(一)Edit1、Cut2、Copy3、Paste4、Delete5、Undo6、Paste from Selected Object (二)Insert Geometry1、Point2、Insert Tic3、Arc4、Create Freehand Polyline5、Rectangle6、Polygon7、Ellipse8、Create Concentric Rings9、Text10、Place GeoPoint11、Place GeoPoint Properties12、GrowGrowGrowing Properties13、EasyTrace14、Lock15、Layer Creation Options (三)Modify1、Enable Editing2、SelectSelectSelect by BoxSelect by LineSelect by EllipseSelect by PolygonFollow HyperlinksSelector PropertiesPick Properties3、LineLineReshapeReplace a portion of a lineSplineDensifyGeneralizeJoinSplit4、AreaAreaReshapeSplit polygon with PolylineReplace a portion of a polygon Append to existing polygonInvert Region5、Vector Options(四)Insert Map Element1、Map GridMap GridUTM GridGeographic GridMGRS Grid Zone ElementDeconflict Grid TicmarksGrid Tic Modifier toolGrid Preferences2、Scale Bar3、Legend4、North ArrowNorth ArrowDefault North Arrow Style5、Dynamic ElementsDynamic ElementsDynamic Text EditorConvert to Text(五)Font/Size1、Font Face2、Font/Symbol Unit Type3、Font/Symbol Size4、Font/Symbol Units5、Bold6、Italic7、Underline8、Colors(六)Locking1、Lock Annotation OrientationLock Annotation OrientationReset Annotation Orientation to Screen Reset Annotation Orientation to Map Lock Annotation Orientation Set Default (七)Styles1、Object Style GalleryAdd to GalleryCustomize Styles2、Customize Styles(八)Shape1、Area Fill2、Line Color3、Line StyleLine Thickness1 pt2 pt4 pt6 ptLine PatternSolid LineDotted LineDashed LineDashed Dotted Line OutlineNo OutlineArrowsNo ArrowStart ArrowEnd ArrowBoth Ends(九)Arrange1、ArrangeOrder ObjectsBring to FrontBring ForwardSend To BackSend BackwardGroup ObjectsGroupUngroupPosition ObjectsRotate North十、Format(一)Insert Geometry1、point2、Insert Tic3、Arc4、Create Freehand Polyline5、Rectangle6、Polygon7、Ellipse8、Create Concentric Rings9、Text10、Place GeoPoint11、Place GeoPoint Properties12、GrowGrowGrowing Properties13、EasyTrace14、Lock15、Layer Creation Options (二)Text1、Text GalleryAdd to Gallery(三)Font1、Font Face2、Font Unit Type3、Font Size4、Font Units5、Bold6、Italic7、Underline8、Colors(四)Symbol1、Symbol Size2、Symbol Units3、Symbol Unit Type(五)Locking1、Lock Annotation OrientationLock Annotation OrientationReset Annotation Orientation to Screen Reset Annotation Orientation to Map Lock Annotation Orientation Set Default (六)Styles1、Object Style GalleryAdd to GalleryCustomize Styles2、Customize Styles(七)Shape1、Area Fill2、Line Color3、Line StyleLine Thickness1 pt2 pt4 pt6 ptLine PatternSolid LineDotted LineDashed LineDashed Dotted Line OutlineNo OutlineArrowsNo ArrowStart ArrowEnd ArrowBoth Ends(八)Arrange1、Bring to Front Bring to FrontBring Forward2、Send to Back Send to BackSend Backward3、GroupGroupUngroup4、AlignAlign Horizontal Left Align Horizontal Center Align Horizontal Right Align Vertically Top Align Vertically Center Align Vertically Bottom Distribute Horizontally Distribute Vertically Alignment..5、FlipFlip VerticallyFlip Horizontally6、Rotate North(十一)Table(一)View1、Show AttributesShow AttributesFrom View Attribute2、Switch Table View(二)Drive1、Drive Viewer to first selected item2、Drive to previous selected feature3、Drive to next selected feature4、Drive Viewer to last selected item5、Zoom to Item(三)Column1、Unselect Columns2、Select All Columns3、Invert Column Selection4、Add Class Name5、Add Area(四)Row1、Unselect Rows2、Select All Rows3、Invert Row Selection4、Criteria(五)Query1、Merge2、Colors3、Column Properties(六)Edit1、Edit Column Next2、Edit Row Next。
纹理物体缺陷的视觉检测算法研究--优秀毕业论文
摘 要
在竞争激烈的工业自动化生产过程中,机器视觉对产品质量的把关起着举足 轻重的作用,机器视觉在缺陷检测技术方面的应用也逐渐普遍起来。与常规的检 测技术相比,自动化的视觉检测系统更加经济、快捷、高效与 安全。纹理物体在 工业生产中广泛存在,像用于半导体装配和封装底板和发光二极管,现代 化电子 系统中的印制电路板,以及纺织行业中的布匹和织物等都可认为是含有纹理特征 的物体。本论文主要致力于纹理物体的缺陷检测技术研究,为纹理物体的自动化 检测提供高效而可靠的检测算法。 纹理是描述图像内容的重要特征,纹理分析也已经被成功的应用与纹理分割 和纹理分类当中。本研究提出了一种基于纹理分析技术和参考比较方式的缺陷检 测算法。这种算法能容忍物体变形引起的图像配准误差,对纹理的影响也具有鲁 棒性。本算法旨在为检测出的缺陷区域提供丰富而重要的物理意义,如缺陷区域 的大小、形状、亮度对比度及空间分布等。同时,在参考图像可行的情况下,本 算法可用于同质纹理物体和非同质纹理物体的检测,对非纹理物体 的检测也可取 得不错的效果。 在整个检测过程中,我们采用了可调控金字塔的纹理分析和重构技术。与传 统的小波纹理分析技术不同,我们在小波域中加入处理物体变形和纹理影响的容 忍度控制算法,来实现容忍物体变形和对纹理影响鲁棒的目的。最后可调控金字 塔的重构保证了缺陷区域物理意义恢复的准确性。实验阶段,我们检测了一系列 具有实际应用价值的图像。实验结果表明 本文提出的纹理物体缺陷检测算法具有 高效性和易于实现性。 关键字: 缺陷检测;纹理;物体变形;可调控金字塔;重构
Keywords: defect detection, texture, object distortion, steerable pyramid, reconstruction
II
基于分组双阶段双向卷积长短期方法的高光谱图像超分辨率网络
智城实践NO.04 20241智能城市 INTELLIGENT CITY基于分组双阶段双向卷积长短期方法的高光谱图像超分辨率网络林建君1侯钧译2杨翠云2(1.烟台职业学院信息工程系,山东 烟台 264670;2.青岛科技大学信息科学技术学院,山东 青岛 266000)摘要:文章提出基于分组的双阶段Bi-ConvLSTM网络(GDBN),可以充分利用图像的空间和光谱信息,通过使用以波段为单位的分组策略,有效缓解了计算负担,并对光谱信息进行保护。
在编码器的不同阶段,对浅层信息提取模块和深度特征提取模块进行不同层次信息的提取,浅层信息提取模块能够对不同尺度的浅层特征信息进行充分捕捉,深度特征提取模块能够捕捉图像的高频特征信息。
文章还引入通道注意力机制,增强网络对特征的组织能力,并在自然数据集cave上进行大量实验,效果普遍优于目前主流的深度学习方法。
关键词:双向卷积长短期记忆网络;高光谱图像超分辨率;通道注意力;神经网络;深度学习中图分类号:TP391 文献标识码:A 文章编号:2096-1936(2024)04-0001-03DOI:10.19301/ki.zncs.2024.04.001Hyperspectral image super-resolution network based on groupedtwo-stage biconvolution long-term and short-term methodLIN Jian-jun HOU Jun-yi YANG Cui-yunAbstract:In this paper, a two-stage Bi-ConvLSTM network based on grouping (GDBN) is proposed, which can make full use of the spatial and spectral information of images, and effectively relieve the computational burden and protect the spectral information by using the grouping strategy based on band units. At different stages of the encoder, the shallow information extraction module and the depth feature extraction module can extract different levels of information. The shallow information extraction module can fully capture the shallow feature information of different scales, and the depth feature extraction module can capture the high-frequency feature information of the image. The paper also introduces channel attention mechanism to enhance the network's ability to organize features, and conducts a large number of experiments on natural data set cave, and the effect is generally better than the current mainstream deep learning methods.Key words:bidirectional convolution long-term and short-term memory network; hyperspectral image super-resolution; channel attention; neural network; deep learning近年来,基于深度学习[1-2]的单图像超分辨率方法取得了广泛发展。
基于X射线相衬显微CT的肝肿瘤定量分析研究
TECHNOLOGY AND INFORMATION科学与信息化2023年5月下 175基于X射线相衬显微CT的肝肿瘤定量分析研究*林瑶 路文平 张耀中 郑焕圣 王坤(通讯作者)新疆第二医学院 新疆 克拉玛依 834000摘 要 新生血管对肿瘤提供无限生长的应用,但目前传统成像技术只能实现200μm的血管。
本研究收集乏血型和富血型肝肿瘤标本20例,进行X射线相衬断层图像中可清晰显示肝肿瘤组织微血管的分布特征,与对应病理切片表现基本相吻合;通过灰度直方图和灰度-梯度共生矩阵提取图像特征,肿瘤图像特征统计结果表明两组存在显著差异(P<0.05),富血型肿瘤相比于乏血型肿瘤组织图像灰度分布不均匀、变化不规则、图像复杂度升高、清晰度降低,为进一步对不同类型肝肿瘤的诊断和特征评价提供数据支持和科学依据。
关键词 肝肿瘤;X射线相衬显微CT;灰度直方图;灰度-梯度共生矩阵Quantitative Analysis of Liver Tumors Based on X-ray Phase Contrast Micro-CT Lin Yao, Lu Wen-ping, Zhang Yao-zhong, Zheng Huan-sheng, Wang Kun (corresponding author)Xinjiang Second Medical College, Karamay 834000, Xinjiang Uygur Autonomous Region, ChinaAbstract Neovascularization provides unlimited growth for tumors, but currently only 200μm vessels can be achieved by conventional imaging techniques. In this study, 20 cases of liver tumor specimen with poor blood type and rich blood type are collected. X-ray phase-contrast tomography images could clearly show the distribution characteristics of microvessels in liver tumor tissue, which are basically consistent with the corresponding pathological section manifestations. The image features are extracted by grayscale histogram and grayscale-gradient co-occurrence matrix, and the statistical results of tumor image features show that there are significant differences between the two groups (P<0.05). Compared with poor-blood tumor tissue image, the rich-blood tumor tissue image has uneven gray distribution, irregular changes, increased image complexity, and decreased clarity, which provide data support and scientific basis for further diagnosis and characteristic evaluation of different types of liver tumors.Key words liver tumor; X-ray phase-contrast micro-CT; grayscale histogram; grayscale-gradient co-occurrence matrix引言肝肿瘤是一种常见的恶性肿瘤,2016年肝肿瘤已位居我国常见的恶性肿瘤的第二位,因此对肝肿瘤的早期诊断和治疗成为临床工作中的重要环节[1]。
基于graph cuts 和主动轮廓的纹理感知图像分割
图的对应关系。左图图像中的像素看作是右图中的结点,邻接像素之间的相似性看作是边上的权值
。类似图,除了普通结点外,还包含两个称为“终点”的点 s、t。边集 E 中包含两种边,一种是 连接相邻结点之间的边(n-links),一种是连接普通结点和“终点”之间的边(t-links)。
假设整幅图像的标签label为L= {l1,l2,,,, lp },其中li为0(背景)或者1(目标)。那假设图像的 分割为L时,图像的能量可以表示为:
Active contour(主动轮廓)、 Color-texture(彩色纹理)、 Structure tensor(结构张量)
2. 本文的研究意义:由于graphcut 的缺点,本论文在四个方面推广延伸了graphcut 算法:
(1):把纹理考虑到分割过程中,我们设计了一个有效地、可靠的纹理探测器,用探测结果生成 一幅增强图像。然后对增强图像用graphcut算法,这样可以利用graphcu进行纹理感知以及有
决方法,对于图像我们适应它,并融合作为后期处理从而增强分割的平滑性、准确性。
(4):当分割比如伪彩色这样的复杂图像时,我们建议在图像分割过程中包含软约束,这样可以 允许用户去勾勒从而效地指导算法去找寻最满意的结果。
(1)给定输入图像和用户的前景/背景输入,算法分析图像并生成一幅增强图像,其中包含了原始 图像和纹理检测结果。
3. 纹理感知graphcut 分割
这一部分首先简单地介绍了graphcut分割作为我们接下来工作的基础,然后提出了两项研究,
构造了纹理增强图像来代替输入图像用于分割以及把结构张量融入到graphcut模型当中。
3.1 graphcut 分割 graph Cut交互式方法的主要思想是:将对图像的目标和背景进行标记作为硬约束,在满足这
- 1、下载文档前请自行甄别文档内容的完整性,平台不提供额外的编辑、内容补充、找答案等附加服务。
- 2、"仅部分预览"的文档,不可在线预览部分如存在完整性等问题,可反馈申请退款(可完整预览的文档不适用该条件!)。
- 3、如文档侵犯您的权益,请联系客服反馈,我们会尽快为您处理(人工客服工作时间:9:00-18:30)。
Abstract—This paper introduces a new supervised Bayesian approach to hyperspectral image segmentation with active learning, which consists of two main steps. First, we use a multinomial logistic regression (MLR) model to learn the class posterior probability distributions. This is done by using a recently introduced logistic regression via splitting and augmented Lagrangian algorithm. Second, we use the information acquired in the previous step to segment the hyperspectral image using a multilevel logistic prior that encodes the spatial information. In order to reduce the cost of acquiring large training sets, active learning is performed based on the MLR posterior probabilities. Another contribution of this paper is the introduction of a new active sampling approach, called modified breaking ties, which is able to provide an unbiased sampling. Furthermore, we have implemented our proposed method in an efficient way. For instance, in order to obtain the time-consuming maximum a posteriori segmentation, we use the α-expansion min-cut-based integer optimization algorithm. The state-of-the-art performance of the proposed approach is illustrated using both simulated and real hyperspectral data sets in a number of experimental comparisons with recently introduced hyperspectral image analysis methods. Index Terms—Active learning, graph cuts, hyperspectral image segmentation, ill-posed problems, integer optimization, mutual information (MI), sparse multinomial logistic regression (MLR).
I. I NTRODUCTION ITH THE recent developments in remote sensing instruments, hyperspectral images are now widely used in different application domains [1]. The special characteristics of hyperspectral data sets bring difficult processing problems. Obstacles, such as the Hughes phenomenon [2], come out as the data dimensionality increases. These difficulties have fostered the development of new classification methods, which are able to deal with ill-posed classification problems. For instance, several machine learning techniques are applied to extract relevant information from hyperspectral data sets [3]–[5]. However,
IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, VOL. 49, NO. 10, OCTOBER 2011
3947Βιβλιοθήκη Hyperspectral Image Segmentation Using a New Bayesian Approach With Active Learning
although many contributions have been made to this area, the difficulty in learning high-dimensional densities from a limited number of training samples (an ill-posed problem) is still an active area of research. Discriminative approaches, which learn the class distributions in high-dimensional spaces by inferring the boundaries between classes in feature space [6]–[8], effectively tackle the aforementioned difficulties. Specifically, support vector machines (SVMs) [9] are among the state-of-the-art discriminative techniques that can be applied to solve ill-posed classification problems. Due to their ability to deal with large input spaces efficiently and to produce sparse solutions, SVMs have been used successfully for supervised and semisupervised classifications of hyperspectral data using limited training samples [1], [3], [10]–[15]. On the other hand, multinomial logistic regression (MLR) [16] is an alternative approach to deal with ill-posed problems, which has the advantage of learning the class probability distributions themselves. This is crucial in the image segmentation step. As a discriminative classifier, MLR directly models the posterior densities instead of the joint probability distributions. The distinguishing features of discriminative classifiers have been reported in the literature before [7], [8], [17]. For instance, effective sparse MLR (SMLR) methods are available in the literature [18]. These ideas have been applied to hyperspectral image classification [5], [19], [20], yielding good performance. Another well-known difficulty in supervised hyperspectral image classification is the limited availability of training data, which are difficult to obtain in practice as a matter of cost and time. In order to effectively work with limited training samples, several methodologies have been proposed, including feature extraction methods such as principal component analysis (PCA), linear discriminant analysis (LDA), discriminant analysis feature extraction, multiple classifiers, and decision fusion [21], among many others [1]. Active learning, which is another active research topic, has been widely studied in the literature [22]–[28]. These studies are based on different principles, such as the evaluation of the disagreement between a committee of classifiers [25], the use of hierarchical classification frameworks [24], [27], unbiased query by bagging [28], or the exploitation of a local proximity-based data regularization framework [26]. In this paper, we use active learning to construct small training sets with high training utility, with the ultimate goal of systematically achieving noticeable improvements in classification results with regard to those found by randomly selected training sets of the same size. Since active learning is intrinsically biased sampling, an issue to be investigated in our experiments