Image Magnification Using Adaptive Interpolation by Pixel Level Data-Dependent Geometrical

合集下载

计算机科学与技术专业使用阈值技术的图像分割等毕业论文外文文献翻译及原文

计算机科学与技术专业使用阈值技术的图像分割等毕业论文外文文献翻译及原文

毕业设计(论文)外文文献翻译文献、资料中文题目: 1.使用阈值技术的图像分割2.最大类间方差算法的图像分割综述文献、资料英文题目:文献、资料来源:文献、资料发表(出版)日期:院(部):专业:计算机科学与技术班级:姓名:学号:指导教师:翻译日期: 2017.02.14毕业设计(论文)题目基于遗传算法的自动图像分割软件开发翻译(1)题目Image Segmentation by Using ThresholdTechniques翻译(2)题目A Review on Otsu Image Segmentation Algorithm使用阈值技术的图像分割 1摘要本文试图通过5阈值法作为平均法,P-tile算法,直方图相关技术(HDT),边缘最大化技术(EMT)和可视化技术进行了分割图像技术的研究,彼此比较从而选择合的阈值分割图像的最佳技术。

这些技术适用于三个卫星图像选择作为阈值分割图像的基本猜测。

关键词:图像分割,阈值,自动阈值1 引言分割算法是基于不连续性和相似性这两个基本属性之一的强度值。

第一类是基于在强度的突然变化,如在图像的边缘进行分区的图像。

第二类是根据预定义标准基于分割的图像转换成类似的区域。

直方图阈值的方法属于这一类。

本文研究第二类(阈值技术)在这种情况下,通过这项课题可以给予这些研究简要介绍。

阈分割技术可分为三个不同的类:首先局部技术基于像素和它们临近地区的局部性质。

其次采用全局技术分割图像可以获得图像的全局信息(通过使用图像直方图,例如;全局纹理属性)。

并且拆分,合并,生长技术,为了获得良好的分割效果同时使用的同质化和几何近似的概念。

最后的图像分割,在图像分析的领域中,常用于将像素划分成区域,以确定一个图像的组成[1][2]。

他们提出了一种二维(2-D)的直方图基于多分辨率分析(MRA)的自适应阈值的方法,降低了计算的二维直方图的复杂而提高了多分辨率阈值法的搜索精度。

这样的方法源于通过灰度级和灵活性的空间相关性的多分辨率阈值分割方法中的阈值的寻找以及效率由二维直方图阈值分割方法所取得的非凡分割效果。

【最新精选】《数字图像处理》结课小论文题目汇总及要求

【最新精选】《数字图像处理》结课小论文题目汇总及要求

《数字图像处理》结课测试题目题目的路径:首先在Matlab的Command window中键入“demo”,进入demo 窗口。

然后在树形选择框中选择“Toolboxes\Image Processing”和“Blocksets\ Video and Image Processing”。

最后逐个查看并选择自己感兴趣的题目。

所有题目汇总如下:图像去模糊1. Deblurring Images Using the Blind Deconvolution Algorithm基于盲解卷算法的图像去模糊2. Deblurring Images Using the Lucy-Richardson Algorithm使用LR算法进行图像去模糊3. Deblurring Images Using a Regularized Filter使用正则滤波器进行图像去模糊4. Deblurring Images Using the Wiener Filter使用维纳滤波器进行图像去模糊图像增强5. Contrast Enhancement Techniques图像对比度增强技术6. Correcting Nonuniform Illumination如何对不均匀光照进行校正7. Enhancing Multispectral Color Composite Images多谱(卫星遥感) 图像增强技术图像配准8. Finding the Rotation and Scale of a Distorted Image计算失真图像的旋转参数和尺度参数9. Registering an Aerial Photo to an Orthophoto基于控制点的多幅航拍图像的配准10. Registering an Image Using Normalized Cross-Correlation使用归一化交叉相关法来配准图像图像分割11. Batch Processing Image Files Using Distributed Computing分布式计算对图像序列进行批处理12. Color-Based Segmentation Using the L*a*b* Color Space基于Lab色彩空间的彩色图像分割13. Color-Based Segmentation Using K-Means Clustering 基于K-均值聚类的彩色图像分割14. Detecting a Cell Using Image Segmentation使用图像分割技术来检测细胞15. Finding V egetation in a Multispectral Image多谱图像(卫星遥感)上的农作物区域分割16. Marker-Controlled Watershed Segmentation基于标记控制的分水岭分割算法17. Texture Segmentation Using Texture Filters基于纹理滤波器的纹理图像分割图像几何变换18. Creating a Gallery of Transformed Images常见的图像几何变换简介19. Exploring a Conformal Mapping图像的保角变换(共形映射)20. Extracting Slices from a 3-Dimensional MRI Data Set 如何从3维MRI数据集中提取切片图21. Padding and Shearing an Image Simultaneously图像的剪切变换和填充操作图像的测量22. Finding the Length of a Pendulum in Motion从单摆图像序列中计算摆长23. Granulometry of Snowflakes使用形态学方法对雪花的颗粒度进行测量24. Identifying Round Objects在图像中计算物体的“似圆度”25. Measuring Angle of Intersection在图像中计算钢梁的交叉角度26. Measuring the Radius of a Roll of Tape如何用图像方法测量胶带的半径图像的Radon变换27. Reconstructing an Image from Projection Data基于拉东(Radon)变换的CT图像重建视频检测和跟踪28. Abandoned Object Detection遗弃物体检测技术29. Motion Detection基于SAD的运动检测系统30. Lane Departure Warning System车道偏离预警系统31. Lane Detection and Tracking基于Hough变换的车道检测和跟踪32. Traffic Warning Sign Recognition交通警示牌自动识别技术33. People Tracking基于背景差分的行人检测技术34. Color Segmentation基于色彩分割的人体检测35. Tracking Cars Using Background Estimation 基于背景估计的汽车检测36. Tracking Cars Using Optical Flow基于光流法的汽车检测37. Surveillance Recording基于主帧检测的监控记录技术38. Pattern Matching基于模板匹配的PCB检测系统压缩技术39. V ideo Compression基于DCT变换的视频压缩技术40. Image Compression基于DCT变换的图像压缩技术视频分析技术41. Histogram Display图像直方图的实时显示42. Concentricity Inspection光纤的同心性检测系统43. Edge Detection边缘检测技术简介44. V ideo Focus Assessment视频自动聚焦参量计算视频增强45. V ideo Stabilization基于模板的电子稳像技术46. Periodic Noise Reduction针对周期噪声的图像降噪算法47. Histogram Equalization基于直方图均衡的图像增强48. Rotation Correction基于Hough变换的旋转图像校正基于形态学的视频分割技术49. Cell Counting细胞自动计数系统50. Feature Extraction如何自动计算视频中扇形的数目51. Object Counting如何自动计算订书钉的数目52. Object Extraction and Replacement视频目标的实时提取和替换视频回放处理53. Continuous Image Rotation图像连续旋转效果的实现54. Projecting Videos onto a Rotating Cube 如何将视频投影到旋转的立方体上55. V isual Effects图像浮雕效果的实现56. Picture in Picture画中画效果的实现57. Panorama Creation全景照片技术58. Bouncing Balls如何在图像上叠加动画《数字图像处理》结课测试报告规范1.内容要求(1)本报告(论文)的名字,系统功能、实现了什么结果。

Image sequence fusion using a shift invariant wavelet transform

Image sequence fusion using a shift invariant wavelet transform
I Introduction
With the development of new imaging sensors arises the need of a meaningful combination of all employed imaging sensors, a problem which is addressed by image fusion. In this paper, we focus on the so-called pixel-level image fusion, e.g. a composite image has to be built of several spatially registered input images or image sequences. A possible application is the fusion of infrared and visible light image sequences in an airborne sensor platform to aid pilot navigation in poor weather conditions or darkness.
20
root mean square error (RMSE)
15
10
5
0
0
4
81216 Nhomakorabea20
24
28
32
shift in [pixels]
Fig. 1: Shift dependency of the multiresolution
fusion methods (see text for details)
The dotted line indicates the shift error when the Haar

threshold and

threshold and

Effective wavelet-based compression method with adaptive quantizationthreshold and zerotree codingArtur Przelaskowski, Marian Kazubek, Tomasz JamrógiewiczInstitute of Radioelectronics, Warsaw University of Technology, Nowowiejska 15/19, 00-665 Warszawa,PolandABSTRACTEfficient image compression technique especially for medical applications is presented. Dyadic wavelet decomposition by use of Antonini and Villasenor bank filters is followed by adaptive space-frequency quantization and zerotree-based entropy coding of wavelet coefficients. Threshold selection and uniform quantization is made on a base of spatial variance estimate built on the lowest frequency subband data set. Threshold value for each coefficient is evaluated as linear function of 9-order binary context. After quantization zerotree construction, pruning and arithmetic coding is applied for efficient lossless data coding. Presented compression method is less complex than the most effective EZW-based techniques but allows to achieve comparable compression efficiency. Specifically our method has similar to SPIHT efficiency in MR image compression, slightly better for CT image and significantly better in US image compression. Thus the compression efficiency of presented method is competitive with the best published algorithms in the literature across diverse classes of medical images. Keywords: wavelet transform, image compression, medical image archiving, adaptive quantization1. INTRODUCTIONLossy image compression techniques allow significantly diminish the length of original image representation at the cost of certain original data changes. At range of lower bit rates these changes are mostly observed as distortion but sometimes improved image quality is visible. Compression of the concrete image with its all important features preserving and the noise and all redundancy of original representation removing is do required. The choice of proper compression method depends on many factors, especially on statistical image characteristics (global and local) and application. Medical applications seem to be challenged because of restricted demands on image quality (in the meaning of diagnostic accuracy) preserving. Perfect reconstruction of very small structures which are often very important for diagnosis even at low bit rates is possible by increasing adaptability of the algorithm. Fitting data processing method to changeable data behaviour within an image and taking into account a priori data knowledge allow to achieve sufficient compression efficiency. Recent achievements clearly show that nowadays wavelet-based techniques can realise these ideas in the best way.Wavelet transform features are useful for better representation of the actual nonstationary signals and allow to use a priori and a posteriori data knowledge for diagnostically important image elements preserving. Wavelets are very efficient for image compression as entire transformation basis function set. This transformation gives similar level of data decorrelation in comparison to very popular discrete cosine transform and has additional very important features. It often provides a more natural basis set than the sinusoids of the Fourier analysis, enables widen set of solution to construct effective adaptive scalar or vector quantization in time-frequency domain and correlated entropy coding techniques, does not create blocking artefacts and is well suited for hardware implementation. Wavelet-based compression is naturally multiresolution and scalable in different applications so that a single decomposition provides reconstruction at a variety of sizes and resolutions (limited by compressed representation) and progressive coding and transmission in multiuser environments.Wavelet decomposition can be implemented in terms of filters and realised as subband coding approach. The fundamental issue in construction of efficient subband coding techniques is to select, design or modify the analysis and synthesis filters.1Wavelets are good tool to create wide class of new filters which occur very effective in compression schemes. The choice of suitable wavelet family, with such criteria as regularity, linearity, symmetry, orthogonality or impulse and step response of corresponding filter bank, can significantly improve compression efficiency. For compactly supported wavelets corresponding filter length is proportional to the degree of smoothness and regularity of the wavelet. Butwhen the wavelets are orthogonal (the greatest data decorrelation) they also have non-linear phase in the associated FIR filters. The symmetry, compact support and linear phase of filters may be achieved by biorthogonal wavelet bases application. Then quadrature mirror and perfect reconstruction subband filters are used to compute the wavelet transform. Biorthogonal wavelet-based filters occurred very efficient in compression algorithms. A construction of wavelet transformation by fitting local defined basis transformation function (or finite length filters) into image data characteristics is possible but very difficult. Because of nonstationary of image data, miscellaneous image futures which could be important for good reconstruction, significant various image quality (signal to noise level, spatial resolution etc.) from different imaging systems it is very difficult to elaborate the construction method of the optimal-for-compression filters. Many issues relating to the choice of the most efficient filter bank for image compression remain still unresolved.2The demands of preserving the diagnostic accuracy in reconstructed medical images are exacting. Important high frequency coefficients which appear at the place of small structure edges in CT and MR images should be saved. Accurate global organ shapes reconstruction in US images and strong noise reduction in MN images is also required. It is rather difficult to imagine that one filter bank can do it in the best way. Rather choosing the best wavelet families for each modality is expected.Our aim is to increase the image compression efficiency, especially for medical applications, by applying suitable wavelet transformation, adaptive quantization scheme and corresponding processed decomposition tree entropy coding. We want to achieve higher acceptable compression ratios for medical images by better preserving the diagnostic accuracy of images. Many bit allocation techniques applied in quantization scheme are based on data distribution assumptions, quantiser distortion function etc. All statistical assumptions built on global data characteristics do not cover exactly local data behaviour and important detail of original image, e.g., different texture small area may be lost. Thus we decided to build quantization scheme on the base of local data characteristics such a direct data context in two dimensions mentioned earlier. We do data variance estimation on the base of real data set as spatial estimate for corresponding coefficient positions in successive subbands. The details of quantization process and correlated coding technique as a part of effective simple wavelet-based compression method which allows to achieve high reconstructed image quality at low bit rates are presented.2. THE COMPRESSION TECHNIQUEScheme of our algorithm is very simple: dyadic, 3 levels decomposition of original image (256×256 images were used) done by selected filters. For symmetrical filters symmetry boundary extension at the image borders was used and for asymmetrical filters - a periodic (or circular) boundary extension.Figure 1. Dyadic wavelet image decomposition scheme. - horizontal relations, - parent - children relations. LL - the lowest frequency subband.Our approach to filters is utilitarian one, making use of the literature to select the proper filters rather than to design them. We conducted an experiment using different kinds of wavelet transformation in presented algorithm. Long list of wavelet families and corresponding filters were tested: Daubechies, Adelson, Brislawn, Odegard, Villasenor, Spline, Antonini, Coiflet, Symmlet, Beylkin, Vaid etc.3 Generally Antonini 4 filters occurred to be the most efficient. Villasenor, Odegard and Brislawn filters allow to achieve similar compression efficiency. Finally: Antonini 7/9 tap filters are used for MR and US image compression and Villasenor 18/10 tap filters for CT image compression.2.1 Adaptive space-frequency quantizationPresented space-frequency quantization technique is realised as entire data pre-selection, threshold selection and scalar uniform quantization with step size conditioned by chosen compression ratio. For adaptive estimation of threshold and quantization step values two extra data structure are build. Entire data pre-selection allows to evaluate zero-quantized data set and predict the spatial context of each coefficient. Next simple quantization of the lowest frequency subband (LL) allows to estimate quantized coefficient variance prediction as a space function across sequential subbands. Next the value of quantization step is slightly modified by a model build on variance estimate. Additionally, a set of coefficients is reduced by threshold selection. The threshold value is increased in the areas with the dominant zero-valued coefficients and the level of growth depends on coefficient spatial position according variance estimation function.Firstly zero-quantized data prediction is performed. The step size w is assumed to be constant for all coefficients at each decomposition level. For such quantization model the threshold value is equal to w /2. Each coefficient whose value is less than threshold is predicted to be zero-valued after quantization (insignificant). In opposite case coefficient is predicted to be not equal to zero (significant). It allows to create predictive zero-quantized coefficients P map for threshold evaluation in the next step. The process of P map creation is as follows:if c w then p else p i i i <==/201, (1)where i m n m n =⋅−12,,...,;, horizontal and vertical image size , c i - wavelet coefficient value. The coefficient variance estimation is made on the base of LL data for coefficients from next subbands in corresponding spatial positions. The quantization with mentioned step size w is performed in LL and the most often occurring coefficient value is estimated. This value is named MHC (mode of histogram coefficient). The areas of MHC appearance are strongly correlated with zero-valued data areas in the successive subbands. The absolute difference of the LL quantized data and MHC is used as variance estimate for next subband coefficients in corresponding spatial positions. We tested many different schemes but this model allows to achieve the best results in the final meaning of compression efficiency. The variance estimation is rather coarse but this simple adaptive model built on real data does not need additional information for reconstruction process and increases the compression efficiency. Let lc i , i =1,2,...,lm , be a set ofLL quantized coefficient values, lm - size of this set . Furthermore let mode of histogram coefficient MHC value be estimated as follows:f MHC f lc MHC Al lc Al i i ()max ()=∈∈ and , (2)where Al - alphabet of data source which describes the values of the coefficient set and f lc n lmi lc i ()=, n lc i - number of lc i -valued coefficients. The normalised values of variance estimate ve si for next subband coefficients in corresponding to i spatial positions (parent - children relations from the top to the bottom of zerotree - see fig. 1) are simply expressed by the following equation: ve lc MHC ve si i =−max . (3)These set of ve si data is treated as top parent estimation and is applied to all corresponding child nodes in wavelet hierarchical decomposition tree.9-th order context model is applied for coarser data reduction in ‘unimportant' areas (usually with low diagnostic importance). The unimportance means that in these areas the majority of the data are equal to zero and significant values are separated. If single significant values appear in these areas it most often suggests that these high frequency coefficients are caused by noise. Thus the coarser data reduction by higher threshold allows to increase signal to noise ratio by removing the noise. At the edges of diagnostically important structures significant values are grouped together and the threshold value is lower at this fields. P map is used for each coefficient context estimation. Noncausal prediction of the coefficient importance is made as linear function of the binary surrounding data excluding considered coefficient significance. The other polynomial, exponential or hyperbolic function were tested but linear function occurred the most efficient. The data context shown on fig. 2 is formed for each coefficient. This context is modified in the previous data points of processing stream by the results of the selection with the actual threshold values at these points instead of w /2 (causal modification). Values of the coefficient importance - cim are evaluated for each c i coefficient from the following equation:cim coeff p i i j j =⋅−=∑1199(),, where i m n =⋅12,,...,. (4)Next the threshold value is evaluated for each c i coefficient: th w cim w ve i i si =⋅+⋅⋅−/(())211, (5)where i m n =⋅12,,...,, si - corresponding to LL parent spatial location in lower decomposition levels.The modified quantization step model uses the LL-based variance estimate to slightly increase the step size for less variance coefficients. Threshold data selection and uniform quantization is made as follows: each coefficient value is firstly compared to its threshold value and then quantized using w step for LL and modified step value mw si for next subbands . Threshold selection and quantization for each c i coefficient can be clearly described by the following equations:LLif c then c c welse if c th then c else c c mw i i i i i i i i si∈=<==//0, (6)where mw w coeff ve si si =⋅+⋅−(())112. (7)The coeff 1 and coeff 2 values are fitted to actual data characteristic by using a priori image knowledge and performingentire tests on groups of similar characteristic images.a) b)Figure 2. a) 9-order coefficient context for evaluating the coefficient importance value in procedure of adaptive threshold P map context of single edge coefficient.2.2 Zerotrees construction and codingSophisticated entropy coding methods which can significantly improve compression efficiency should retain progressive way of data reconstruction. Progressive reconstruction is simple and natural after wavelet-based decomposition. Thus the wavelet coefficient values are coded subband-sequentially and spectral selection is made typically for wavelet methods. The same scale subbands are coded as follows: firstly the lowest frequency subband, then right side coefficient block, down-left and down-right block at the end. After that next larger scale data blocks are coded in the same order. To reduce a redundancy of such data representation zerotree structure is built. Zerotree describes well the correlation between data values in horizontal and vertical directions, especially between large areas with zero-valued data. These correlated fragments of zerotree are removed and final data streams for entropy coding are significantly diminish. Also zerotree structure allows to create different characteristics data streams to increase the coding efficiency. We used simple arithmetic coders for these data streams coding instead of applied in many techniques bit map (from MSB to LSB) coding with necessity of applying the efficient context model construction. Because of refusing the successive approximation we lost full progression. But the simplicity of the algorithm and sometimes even higher coding efficiency was achieved. Two slightly different arithmetic coders for producing ending data stream were used.2.2.1 Construction and pruning of zerotreeThe dyadic hierarchical image data decomposition is presented on fig. 1. Decomposition tree structure reflects this hierarchical data processing and strictly corresponds to created in transformation process data streams. The four lowest frequency subbands which belong to the coarsest scale level are located at the top of the tree. These data have not got parent values but they are the parents for the coefficients in lower tree level of greater scale in corresponding spatial positions. These correspondence is shown on the fig. 1 as parent-children relations. Each parent coefficient has got four direct children and each child is under one direct parent. Additionally, horizontal relations at top tree level are introduced to describe the data correlation in better way.The decomposition tree becomes zerotree when node values of quantized coefficients are signed by symbols of binary alphabet. Each tree node is checked to be significant (not equal to zero) or insignificant (equal to zero) - binary tree is built. For LL nodes way of significance estimation is slightly different. The MHC value is used again because of the LL areas of MHC appearance strong correlation with zero-valued data areas in the next subbands. Node is signed to be significant if its value is not equal to MHC value or insignificant if its value is equal to MHC. The value of MHC must be sent to a decoder for correct tree reconstruction.Next step of algorithm is a pruning of this tree. Only the branches to insignificant nodes can be pruned and the procedure is slightly other at different levels of the zerotree. Procedure of zerotree pruning starts at the bottom of wavelet zerotree. Sequential values of four children data and their parent from higher level are tested. If the parent and the children are insignificant - the tree branch with child nodes is removed and the parent is signed as pruned branch node (PBN). Because of this the tree alphabet is widened to three symbols. At the middle levels the pruning of the tree is performed if the parent value is insignificant and all children are recognised as PBN. From conducted research we found out that adding extra symbols to the tree alphabet is not efficient for decreasing the code bit rate. The zerotree pruning at top level is different. The checking node values is made in horizontal tree directions by exploiting the spatial correlation of the quantized coefficients in the subbands of the coarsest scale - see fig. 1. Sequentially the four coefficients from the same spatial positions and different subbands are compared with one another. The tree is pruned if the LL node is insignificant and three corresponding coefficients are PBN. Thus three branches with nodes are removed and LL node is signed as PBN. It means that all its children across zerotree are insignificant. The spatial horizontal correlation between the data at other tree levels is not strong enough to increase the coding efficiency by its utilisation.2.2.2 Making three data streams and codingPruned zerotree structure is handy to create data streams for ending efficient entropy coding. Instead of PBN zero or MHC values (nodes of LL) additional code value is inserted into data set of coded values. Also bit maps of PBN spatial distribution at different tree levels can be applied. We used optionally only PBN bit map of LL data to slightly increase the coding efficiency. The zerotree coding is performed sequentially from the top to the bottom to support progressive reconstruction. Because of various quantized data characteristics and wider alphabet of data source model after zerotree pruning three separated different data streams and optionally fourth bit map stream are produced for efficient data coding. It is well known from information theory that if we deal with a data set with significant variability of data statistics anddifferent statistics (alphabet and estimate of conditional probabilities) data may be grouped together it is better to separate these data and encode each group independently to increase the coding efficiency. Especially is true when context-based arithmetic coder is used. The data separation is made on the base of zerotree and than the following data are coded independently:- the LL data set which has usually smaller number of insignificant (MHC-valued) coefficients, less PBN and less spatial data correlation than next subband data (word- or charwise arithmetic coder is less efficient then bitwise coder);optionally this data stream is divided on PBN distribution bit map and word or char data set without PBNs,- the rest of top level (three next subbands) and middle level subband data set with a considerable number of zero-valued (insignificant) coefficients and PBN code values; level of data correlation is greater, thus word- or charwise arithmetic coder is efficient enough,- the lowest level data set with usually great number of insignificant coefficients and without PBN code value; data correlation is very high.Urban Koistinen arithmetic coder (DDJ Compression Contest public domain code accessible by internet) with simple bitwise algorithm is used for first data stream coding. For the second and third data stream coding 1-st order arithmetic coder built on the base of code presented in Nelson book 5 is applied. Urban coder occurred up to 10% more efficient than Nelson coder for first data stream coding. Combining a rest of top level data and the similar statistics middle level data allows to increase the coding efficiency approximately up to 3%.The procedure of the zerotree construction, pruning and coding is presented on fig. 3.Construction ofbinary zerotreeBitwise arithmetic codingFinal compressed data representationFigure 3. Quantized wavelet coefficients coding scheme with using zerotree structure. PBN - pruned branch node.3. TESTS, RESULTS AND DISCUSSIONIn our tests many different medical modality images were used. For chosen results presentation we applied three 256×256×8-bit images from various medical imaging systems: CT (computed tomography), MR (magnetic resonance) and US(ultrasound) images. These images are shown on fig. 4. Mean square error - MSE and peak signal to noise ratio - PSNR were assumed to be reconstructed image quality evaluation criteria. Subjective quality appreciation was conducted in very simple way - only by psychovisual impression of the non-professional observer.Application of adaptive quantization scheme based on modified threshold value and quantization step size is more efficient than simple uniform scalar quantization up to 10% in a sense of better compression of all algorithm. Generally applying zerotree structure and its processing improved coding efficiency up to 10% in comparison to direct arithmetic coding of quantized data set.The comparison of the compression efficiency of three methods: DCT-based algorithm,6,7 SPIHT 8 and presented compression technique, called MBWT (modified basic wavelet-based technique) were performed for efficiency evaluation of MBWT. The results of MSE and PSNR-based evaluation are presented in table 1. Two wavelet-based compression techniques are clearly more efficient than DCT-based compression in terms of MSE/PSNR and also in our subjective evaluation for all cases. MBWT overcomes SPIHT method for US images and slightly for CT test image at lower bit rate range.The concept of adaptive threshold and modified quantization step size is effective for strong reduction of noise but it occurs sometimes too coarse at lower bit rate range and very small details of the image structures are put out of shape. US images contain significant noise level and diagnostically important small structures do not appear (image resolution is poor). Thus these images can be efficiently compressed by MBWT with image quality preserved. It is clearly shown on fig.5. An improvement of compression efficiency in relatio to SPIHT is almost constant at wide range of bit rates (0.3 - 0.6 dB of PSNR).a) b)c)Figure 4. Examples of images used in the tests of compression efficiency evaluation. The results presented in table 1 and on fig. 5 were achieved for those images. The images are as follows: a ) echocardiography image, b) CT head image, c) MR head image.Table 1. Comparison of the three techniques compression efficiency: DCT-based, SPIHT and MBWT. The bit rates are chosen in diagnostically interesting range (near the borders of acceptance).Modality - bit rateDCT-based SPIHT MBWTMSE PSNR[dB] MSE PSNR[dB] MSE PSNR[db] MRI - 0.70 bpp8.93 38.62 4.65 41.45 4.75 41.36 MRI - 0.50 bpp13.8 36.72 8.00 39.10 7.96 39.12 CT - 0.50 bpp6.41 40.06 3.17 43.12 3.1843.11 CT - 0.30 bpp18.5 35.46 8.30 38.94 8.0639.07 US - 0.40 bpp54.5 30.08 31.3 33.18 28.3 33.61 US - 0.25 bpp 91.5 28.61 51.5 31.01 46.8 31.43The level of noise in CT and MR images is lower and small structures are often important in image analysis. That is the reason why the benefits of MBWT in this case are smaller. Generally compression efficiency of MBWT is comparable to SPIHT for these images. Presented method lost its effectiveness for higher bit rates (see PSNR of 0.7 bpp MR representation) but for lower bit rates both MR and CT images are compressed significantly better. Maybe the reason is that the coefficients are reduced relatively stronger because of its importance reduction in MBWT threshold selection at lower bits rate range.0,20,30,40,50,60,70,8Rate in bits/pixel PSNR in dBFigure 5. Comparison of SPIHT and presented in this paper technique (MBWT) compression efficiency at range of low bit rates. US test image was compressed.4. CONCLUSIONSAdaptive space-frequency quantization scheme and zerotree-based entropy coding are not time-consuming and allow to achieve significant compression efficiency. Generally our algorithm is simpler than EZW-based algorithms 9 and other algorithms with extended subband classification or space -frequency quantization models 10 but compression efficiency of presented method is competitive with the best published algorithms in the literature across diverse classes of medical images. The MBWT-based compression gives slightly better results than SPIHT for high quality images: CT and MR and significantly better efficiency for US images. Presented compression technique occurred very useful and promising for medical applications. Appropriate reconstructed image quality evaluation is desirable to delimit the acceptable lossy compression ratios for each medical modality. We intend to improve the efficiency of this method by: the design a construction method of adaptive filter banks and correlated more sufficient quantization scheme. It seems to be possible byapplying proper a priori model of image features which determine diagnostic accuracy. Also more efficient context-based arithmetic coders should be applied and more sophisticated zerotree structures should be tested.REFERENCES1.Hui, C. W. Kok, T. Q. Nguyen, …Image Compression Using Shift-Invariant Dydiadic Wavelet Transform”, subbmited toIEEE Trans. Image Proc., April 3nd, 1996.2.J. D. Villasenor, B. Belzer and J. Liao, …Wavelet Filter Evaluation for Image Compression”, IEEE Trans. Image Proc.,August 1995.3. A. Przelaskowski, M.Kazubek, T. Jamrógiewicz, …Optimalization of the Wavelet-Based Algorithm for Increasing theMedical Image Compression Efficiency”, submitted and accepted to TFTS'97 2nd IEEE UK Symposium on Applications of Time-Frequency and Time-Scale Methods, Coventry, UK 27-29 August 1997.4.M. Antonini, M. Barlaud, P. Mathieu and I. Daubechies, …Image coding using wavelet transform”, IEEE Trans. ImageProc., vol. IP-1, pp.205-220, April 1992.5.M. Nelson, The Data Compression Book, chapter 6, M&T Books, 1991.6.M. Kazubek, A. Przelaskowski and T. Jamrógiewicz, …Using A Priori Information for Improving the Compression ofMedical Images”, Analysis of Biomedical Signals and Images, vol. 13,pp. 32-34, 1996.7. A. Przelaskowski, M. Kazubek and T. Jamrógiewicz, …Application of Medical Image Data Characteristics forConstructing DCT-based Compression Algorithm”, Medical & Biological Engineering & Computing,vol. 34, Supplement I, part I, pp.243-244, 1996.8. A. Said and W. A. Pearlman, …A New Fast and Efficient Image Codec Based on Set Partitioning in Hierarchical Trees”,submitted to IEEE Trans. Circ. & Syst. Video Tech., 1996.9.J. M. Shapiro, …Embedded Image Coding Using Zerotrees of Wavelet Coefficients”, IEEE Trans. Signal Proces., vol.41, no.12, pp. 3445-3462, December 1993.10.Z. Xiong, K. Ramchandran and M. T. Orchard, …Space-Frequency Quantization for Wavelet Image Coding”, IEEETrans. Image Proc., to appear in 1997.。

Creating full view panoramic image mosaics and environment maps

Creating full view panoramic image mosaics and environment maps
The case of general camera rotation has been studied previously [12, 9, 18], using an 8-parameter planar perspective motion model. By contrast, our algorithm uses a 3-parameter rotational motion model, which is more robust since it has fewer unknowns. Since this algorithm requires knowing the camera’s focal length, we develop a method for computing an initial focal length estimate from a set of 8-parameter perspective registrations. We also investigate how to close the “gap” (or “overlap”) due to accumulated registration errors after a complete panoramic sequence has been assembled. To demonstrate the advantages of our algorithm, we apply it to a sequence of images taken with a handheld digital camera.
Copyright ©1997 by the Association for Computing Machinery, Inc. Permission to make digital or hard copies of part or all of this work for personal or classroom use is granted without fee provided that copies are not made or distributed for profit or commercial advantage and that copies bear this notice and the full citation on the first page. To copy otherwise, to republish, to post on servers, or to distribute to lists, requires prior specific permission and/or a fee.

改进的生成对抗网络图像去噪算法

改进的生成对抗网络图像去噪算法

数字图像在采集和网络传输的过程中,往往会受到一些随机信号的干扰而产生图像噪声,导致图像质量降低,从而影响人对图像的理解,所以有效地降低图像的噪声,提高图像的质量仍是图像处理领域的热点之一[1-2]。

图像去噪已经有很长的一段历史了,传统的图像去噪方法有空域滤波[3]和变换域滤波[4],空域滤波直接对图像的像素用滤波模板进行卷积,包括领域中值滤波[5]、均值滤波[6]等。

变换域滤波利用噪声图像和无噪图像在频域的分布差异,将图像转换到频域进行处理后再将结果变换回空间域,从而获得去噪后的图像,常见的变换域有小波变换域[7]、傅里叶变换域[8]等。

这些方法在一定程度上可以抑制图像的噪声,但修复结果往往会使图像纹理信息缺失,导致图像模糊。

BM3D [9](Block-Matching and 3D filtering )利用自然图像中存在的自相似性,通过对相似块转换并进行加权处理得到目标块,取得了不错的去噪效果。

CBM3D [10]是BM3D 改进的彩色图像去噪方法,该方法利用了亮度-色度颜色空间的每个通道中高度稀疏的局部3D 变换域中的滤波,这种去噪效果依赖相似块的选取,在图像去噪过程中常常存在一些复杂的优化问题。

近年来,深度学习在目标识别及检测等图像处理领域大放异彩,使得很多学者将深度学习模型应用于图像去噪。

深度卷积神经网络拥有很好的学习能力,通过对噪声样本的学习,能够实现图像去噪的自动化与智能化。

Burger 等[11]提出多层感知器MLP (Multi-Layer Per-改进的生成对抗网络图像去噪算法陈人和,赖振意,钱育蓉新疆大学软件学院,乌鲁木齐830046摘要:由于图像噪声的存在会干扰人对图像的理解,为了有效地去除噪声并获得比较好的视觉观感,提出一种基于生成对抗网络算法,该算法通过增加生成网络的宽度来获取更多的图像特征,并加入一个全局残差对输入的噪声图像进行特征的提取与学习,避免特征的丢失。

网络采用对抗损失和重建损失的加权和,在去除噪声的同时能够有效地保留图像的细节信息。

基于局部自适应色差阈值的彩色图像边缘检测

基于局部自适应色差阈值的彩色图像边缘检测

比度视觉阈值与对比灵敏度成倒数关系,对比灵敏
度与色差可视阈值的关系可用式(7)描述:
csf
=1 ct
=
1 ΔELab /
E(Y
)
=
E(Y ) ΔELab
(7)
Lab 空间的对比敏感函数根据调制方向的不同
可分为亮度对比敏感、红绿色对比敏感与黄蓝色对
比敏感,亮度 CSF 一般采用 Kiorpes 和 Movshon
2012-03-02 收到,2012-05-14 改回 国家自然科学基金(61105015),江苏省自然科学基金(BK2010366) 和江苏省科技支撑计划(BE2011747)资助课题 *通信作者:杨娴 yangxian0830@
息对人眼色差感知的影响,使得很多可忽略的边缘 被过检测,对噪声的鲁棒性较差。
based on the adaptive perceptual color difference is proposed. A weighting factor including luminance masking
effect and contrast sensitivity function is proposed, which combine the influence of local changes in luminance and
+ 60.59 ⋅ exp(−0.0037 ⋅ ϕ2.1677 )
(9)
CSFyb (ϕ) = 35.0328 ⋅ exp(−0.0004 ⋅ ϕ4.2582 )
+ 40.6910 ⋅ exp(−0.1040 ⋅ ϕ1.6487 ) (10)
其中 ϕ 为 L, a, b 通道的空间角频率 w, u, v 的集合,

基于光照模型的细胞内镜图像不均匀光照校正算法

基于光照模型的细胞内镜图像不均匀光照校正算法

文章编号 2097-1842(2024)01-0160-07基于光照模型的细胞内镜图像不均匀光照校正算法邹鸿博1,章 彪1,王子川1,陈 可2,王立强2,袁 波1 *(1. 浙江大学 光电科学与工程学院, 浙江 杭州 310027;2. 之江实验室类人感知研究中心, 浙江 杭州 311100)摘要:细胞内镜需实现最大倍率约500倍的连续放大成像,受光纤照明及杂散光的影响,其图像存在不均匀光照,且光照分布会随放大倍率的变化而变化。

这会影响医生对病灶的观察及判断。

为此,本文提出一种基于细胞内镜光照模型的图像不均匀光照校正算法。

根据图像信息由光照分量和反射分量组成这一基础,该算法通过卷积神经网络学习图像的光照分量,并基于二维Gamma 函数实现不均匀光照校正。

实验表明,经本文方法进行不均匀光照校正后,图像的光照分量平均梯度和离散熵分别为0.22和7.89,优于自适应直方图均衡化、同态滤波和单尺度Retinex 等传统方法以及基于深度学习的WSI-FCN 算法。

关 键 词:细胞内镜;不均匀光照;光照模型;卷积神经网络中图分类号:TN29;TP391.4 文献标志码:A doi :10.37188/CO.2023-0059Non-uniform illumination correction algorithm for cytoendoscopyimages based on illumination modelZOU Hong-bo 1,ZHANG Biao 1,WANG Zi-chuan 1,CHEN Ke 2,WANG Li-qiang 2,YUAN Bo 1 *(1. College of Optical Science and Engineering , Zhejiang University , Hangzhou 310027, China ;2. Research Center for Humanoid Sensing , Zhejiang Lab., Hangzhou 311100, China )* Corresponding author ,E-mail : **************.cnAbstract : Cytoendoscopy requires continuous amplification with a maximum magnification rate of about 500 times. Due to optical fiber illumination and stray light, the image has non-uniform illumination that changes with the magnification rate, which affects the observation and judgement of lesions by doctors.Therefore, we propose an image non-uniform illumination correction algorithm based on the illumination model of cytoendoscopy. According to the principle that image information is composed of illumination and reflection components, the algorithm obtains the illumination component of the image through a convolution-al neural network, and realizes non-uniform illumination correction based on the two-dimensional Gamma function. Experiments show that the average gradient of the illumination channel and the discrete entropy of the image are 0.22 and 7.89, respectively, after the non-uniform illumination correction by the proposed method, which is superior to the traditional methods such as adaptive histogram equalization, homophobic收稿日期:2023-04-04;修订日期:2023-05-15基金项目:国家重点研发计划项目(No. 2021YFC2400103);之江实验室科研项目(No. 2019MC0AD02,No. 2022MG0AL01)Supported by the National Key Research and Development Program of China (No. 2021YFC2400103); Key Research Project of Zhejiang Lab (No. 2019MC0AD02, No. 2022MG0AL01)第 17 卷 第 1 期中国光学(中英文)Vol. 17 No. 12024年1月Chinese OpticsJan. 2024filtering, single-scale Retinex and the WSI-FCN algorithm based on deep learning.Key words: cytoendoscopy;non-uniform illumination;illumination model;convolutional neural network1 引 言细胞内镜是一种具有超高放大倍率的内窥镜[1-4],可实现常规倍率到细胞级放大倍率的连续放大观察。

  1. 1、下载文档前请自行甄别文档内容的完整性,平台不提供额外的编辑、内容补充、找答案等附加服务。
  2. 2、"仅部分预览"的文档,不可在线预览部分如存在完整性等问题,可反馈申请退款(可完整预览的文档不适用该条件!)。
  3. 3、如文档侵犯您的权益,请联系客服反馈,我们会尽快为您处理(人工客服工作时间:9:00-18:30)。

Image Magnification Using Adaptive Interpolation by Pixel Level Data-Dependent Geometrical ShapesMuhammad Sajjad, Naveed Khattak, and Noman JafriAbstract—World has entered in 21st century. The technology of computer graphics and digital cameras is prevalent. High resolution display and printer are available. Therefore high resolution images are needed in order to produce high quality display images and high quality prints. However, since high resolution images are not usually provided, there is a need to magnify the original images. One common difficulty in the previous magnification techniques is that of preserving details, i.e. edges and at the same time smoothing the data for not introducing the spurious artefacts. A definitive solution to this is still an open issue. In this paper an image magnification using adaptive interpolation by pixel level data-dependent geometrical shapes is proposed that tries to take into account information about the edges (sharp luminance variations) and smoothness of the image. It calculate threshold, classify interpolation region in the form of geometrical shapes and then assign suitable values inside interpolation region to the undefined pixels while preserving the sharp luminance variations and smoothness at the same time.The results of proposed technique has been compared qualitatively and quantitatively with five other techniques. In which the qualitative results show that the proposed method beats completely the Nearest Neighbouring (NN), bilinear(BL) and bicubic(BC) interpolation. The quantitative results are competitive and consistent with NN, BL, BC and others.Keywords—Adaptive, digital image processing, image magnification, interpolation, geometrical shapes, qualitative & quantitative analysis.I. I NTRODUCTIONODAY, there is a huge amount of digital images availableto computer users. This is caused by the rapid growth both in computer hardware and software technologies. Low price digital cameras are now common, and as a result users are able to buy them and take as many digital images as desired. The significant development in the field of computer graphics has also boosted the production of digital images. As computer users become more familiar with digital images, the need to display and print them also increases. In an era where high-resolution display and printing devices are common, it is vital that high-resolution images are available in order to produce high quality displayed images and high quality prints. This is particularly important for desktop publishing, large artistic printing, etc. The problem is that high-resolution images are not usually provided. In these cases, there is a need to Authors are with Department of computer Science, College of signals, National University of Sciences and Technology, Rawalpindi, Pakistan (e-mail: qazi.msajjad@, {khattakn, mnjafri}@.pk). magnify the original images. Therefore, the development of a good image magnification algorithm is very important.Until now, a large number of interpolation techniques for magnifying images have been proposed. A typical problem with most interpolation techniques is that although smoothing the data and keeping the low frequencies in the new zoomed picture, they are not able to enhance the high frequencies or preserve the edges equally well. Visually those problems will result in either blurring or blocking artifacts. A possible solution would need a sort of non-linear interpolation, taking into account the directional variation for maintaining the sharpness of the new enlarged image and smoothness as well. The simplest method to magnify images is the pixel replication. However, the resulting magnified images have aliasing effect in the form of jagged edges. Nearest neighbor interpolation is the simplest method and basically makes the pixels bigger. The color of a pixel in the new image is the color of the nearest pixel of the original image[4]. Most image viewing and editing software use this type of interpolation to enlarge a digital image for the purpose of closer examination because it does not change the color information of the image and does not introduce any anti-aliasing. For the same reason, it is not suitable to enlarge photographic images because it increases the visibility of jaggies. More elaborate approaches use the bilinear or the bicubic interpolation. Bilinear Interpolation determines the value of a new pixel based on a weighted average of the 4 pixels in the nearest 2 x 2 neighborhood of the pixel in the original image [4]. The averaging has an anti-aliasing effect and therefore produces relatively smooth edges with hardly any jaggies. Bicubic interpolation is more sophisticated and produces smoother edges than bilinear interpolation. Here, a new pixel is a bicubic function using 16 pixels in the nearest 4 x 4 neighborhood of the pixel in the original image [1, 4]. This is the method most commonly used by image editing software, printer drivers and many digital cameras for resampling images. Commercial software Adobe Photoshop [1] provides these two functions for interpolating images. Other methods, using the B-spline interpolators [8, 11] or the cubic convolution methods [10] have also been proposed. However, these methods tend to blur the edges and cause them to be jagged.Research on interpolating images taking into account the edges, has gained much attention. Allebach and Wong [2] proposed methods that search for edges in the input image and use them to assure that the interpolation does not crossTthem. The problem is one of how to define and find the important edges in the input image. Other edge-adaptive methods have been proposed by Jensen and Anastassiou [9], Li and Orchard [12], and Muresan and Parks [15, 16, 17, 18] . Currently, the methods presented in [12, 18] are the most widely known edge-adaptive methods. They can well enough avoid jagged edges, but a limitation is that they sometimes introduce highly visible artifacts into the magnified images, especially in areas with small size repetitive patterns. X. Yu, B. Morse, T.W Sederberg[19] presented a method that computes a triangulation of the input image where the pixels in the input image are the vertices of the triangles in the triangulation. The input image at any arbitrary scale is reconstructed by rendering its triangulation. However, since the edges in the input image are approximated using piecewise linear segments, curved edges cannot be properly reconstructed especially when the scaling factor is a large number. Morse and Schwartzwald [14] presented a level-set reconstruction method to solve the problem of jagged edges. Their approach starts by magnifying the input image using the bicubic interpolation method, then iteratively smoothing the contours in the image. This approach, however, does not overcome the blurring problem found in the bicubic interpolation method. Hertzmann et al. [7] and Freeman et al.[3] proposed methods that learn the correspondences between low and high resolution images from a set of training data. The advantage of these approaches is that fine details can be added when producing high-resolution images when the input image is in the same class of image as the training data. The disadvantages of these approaches are that they will fail if the input image is not in the same class as the training data and that the computational cost is high. Variational based approaches for image magnification have been presented by Malgouyres and Guichard [13]. The magnified images obtained using these methods are better than those obtained using the bicubic interpolation method. However, since these methods solve optimization problems where all the pixels in the magnified image are unknowns, these methods, too, have high computational costs. As a result, they are not suitable for practical use. Another good approach is presented by Henry Johan and Tomoyuki Nishita [5] which produces a high quality magnified image. They have proposed a progressive refinement approach for image magnification. They also claimed for sharp magnified image without generating distinct artifacts but some times it also produce jaggeis to magnify the image. Muneeb, Naveed khattak (KM) [6] and . Battiato, G. Gallo, and F. Stanco[20] also proposed an adaptive approach for image magnification which produces a high quality magnified image while preserving information contents of original image but it has also some defects. These [6, 20] has also been considered in qualitative and quantitative analysis of this paper.The main contribution of this paper is to produce an enlarge image with high visual quality of the original image. The basic idea of the proposed technique is to separate the interpolation region in the form of geometrical shape and then the interpolator assign a proper intensity value to the undefined pixel inside this interpolation region. It does not require a preliminary gradient computation because the relevant information is collected during magnification process. The proposed technique is so intelligent that it can differentiate between low and high frequency content of the source image and thus it preserve the visual sharpness of the original image. The proposed technique is very easy to implement and efficient as well than other proposed techniques.The quantitative and qualitative analysis of proposed technique shows that it has covered some of the weakness of other proposed technique up to some extent with respect to quality and efficiency.II. T HRESHOLDINGBecause of its intuitive properties and simplicity of implementation, image thresholding play a vital roll in application of digital image processing.In image magnification thresholding also plays a key roll in preservation of edges. So it’s most important to select a suitable threshold during interpolation to preserve the fine detail of the image. To preserve the visual quality of the image, the threshold on the basis of safe color [21] is calculated. There are 16 true gray shades from 0 to 255 which can be differentiated visually.Fig. 1 Sixteen Safe Gray Colors in the 256-Color RGB SystemAs in Fig. 1 there are 16 safe colors out of 256. If ‘256’ is divided by ‘16’ we will get 16 as a Quotient. It means that after adding 16 to any gray shad then it will change its visual depiction. To calculate the threshold ‘T‘ for the preservation of the edge using above concept. If N is equal to 16 and where X1=0,X2=2,.......... X N=15and Median denoted by M d is defined as:M d = (X N/2 + X N/2+1)/2 (1)T= M d (2)It has been proved experimentally that it gives excellent result in preservation of edges during magnification and this threshold is also considered during magnification process.III. C LASSIFICATION OF I NTERPOLATION R EGIONS ANDG EOMETRICAL S HAPESEdge preservation plays very important roll in magnification because it specifies the interpolation region in which interpolator adopt itself according to the region. To consider all possible interpolation region during magnification algorithm and assign a suitable value to the undefined pixel is very important. The concept relies on using the low resolution(LR) image to find zero crossing that indicate the presence of an edge passing through LR unit cell. These zero-crossing are then linked by straight line segments to obtain an estimate of the edge which divide the region in different interpolation region. The zero crossing are determined by applying Second order derivative on the LR image then for every LR unit cell after applying 2nd derivative the absolute value either on each side of the zero is compared to the threshold calculated in section 2. If the value is greater thanthreshold it mean that the point of high contrast is present as shown in Fig. 2.Fig. 2 Determination of zero-crossing and its parameterIn Fig. 2 sequence of LR samples from some function f are given. Now calculate the point of high contrast by applying ∂2f /∂2x on sequence of LR unit cell of some function f. After calculation the magnitude on either side of zero-crossing is compared as in equation 3.∂2f /∂2x = f (x LR2) + f (x LR0) - 2f (x LR1)│∂2f /∂2x │ >T(3) If the above statement is true then it mean that there is point of high contrast i.e. edge is present. So to make the above description as a base, a LR (original image) unit cell is analyzed consist on four pixels and determines the point of high contrast to categorize the edges, which split the LR unit cell into different interpolation and its orientation. In this way the interpolation regions are classified which make different geometrical shapes.A. Constant RegionThe LR unit cell which has not any point of high contrast among the pixels is a linear area. It consists only on one interpolation region.Fig. 3 Area where no edge passing through an LR unit cellIn Fig. 3 circles show pixels and their color content shows intensity level. All the pixels are at the same level of intensity, so the resultant region is linear and the total number of interpolation region is one, it has also been visualize in the form three dimensions as in Fig. 4.Fig. 4 3 D visualization of Fig. 3As in Fig. 4 all the pixels are at the same level and it create a shape of square which consist only on one interpolation region.B. LR Region with Corner EdgesThe LR unit cell where an edge separates one pixel from the other three pixels of the LR unit cell. When the point of high contrast separate one pixel from the other three i.e. this pixel will have a value substantially different to the other three. This also call the outlier, imagine in Fig. 5.Fig. 5 LR unit cell where edge separate one corner from the otherthree i.e. two interpolation regionsIn Fig. 5 color difference of pixels show intensity difference which has been explained above. This intensity difference split the LR unit cell in two interpolation region. It has been visualized in the following Fig. 6.Fig. 6 Triangulation in a four-pixel squareIn Fig. 6, pixels with same intensity level form a shape of triangle. This triangle has been considered one interpolation region. There are four possible cases of triangulation, which form triangle in different direction. So there are four different cases in which one pixel isolate from other three and in this way it divide the LR unit cell into two interpolation regions. The remaining cases have been shown in the next Fig. 7.Fig. 7 Triangulations in a four-pixel squareIn Fig. 7 these triangle in different direction are considered different interpolation region. If one pixel isolate itself form other pixels due to intensity difference in all scenarios, whatever the intensity difference, all three pixels will be considered in same region of triangle which has been shown above.C. LR Region with Horizontal and Vertical EdegesIn this scenario edge separate a pair of pixels from other pair of pixels i.e. it split the LR cell in two horizontal or vertical regions having contrast intensity.Fig. 8 Edge which split the LR unite cell verticallyIn Fig. 9, one pair of pixels is on same side having different intensity level from other pair of pixels is on opposite side. So in such type of scenarios LR unit cell will be split horizontally or vertically.Fig. 9 Square split by edge into two interpolation regionThe same process can be repeated for the horizontal case. Here important is the splitting of two poles, whether it is vertically or horizontally. If the intensity contrast occurs horizontally the shape of the LR unit cell will be become as inFig. 10.Fig. 10 Edge which split the LR unit cell into two horizontalinterpolation regionsNow all the information is at hand about the interpolation regions. This information will be used in the implantation of algorithm. Due to this basic information the result of proposed algorithm has improved qualitatively and quantitatively as well. The concept of this section will be used in the coming section ‘The Basic Concept of Algorithm’.IV. T HE B ASIC C ONCEPT OF A LGORITHMIn this section the detail description of the basic concept of proposed algorithm is described. First the proposed algorithm is explained for gray scale images and then it is generalized for colour images. Algorithm works in four phases. This has been described in this section.In the first phase of the proposed algorithm the input image is expanded. Suppose the size of the input image is n x m where ‘n’ is number of rows and m is the number of columns. The image will be expanded to size of (2n-1) x (2m-1). The question arises that why one is subtracted from rows and columns. If it is not subtracted then there will be one additional row and column of undefined pixels which will have the intensity value of the adjacent row and column respectively. This is a sort of replication and the replication has been avoided completely in proposed algorithm.Fig. 11 Expansion phase showing source image (n x m) and expandedimage (2n-1) x (2m-1)In Fig. 11, solid circles show original pixels and hollow circles show undefined pixels. In the remaining three phases these undefined pixels will be filled with proper intensity values while preserving details and edges of original image and data smoothing.The second phase of the algorithm is most important one. In this phase the interpolator assign value to the undefined pixel by pixel level data dependent geometrical shapes.Fig. 12 HR unit cell with undefined pixels Top, Center, Bottom, Left, Right denoted by T, B, C, L, R respectivelyAs it has been mentioned that the assignment of proper intensity value to the undefined pixel is depend on the pixel level data dependent geometrical shapes. In this phase the algorithm scan the image and each time it consider the group of pixels as shown in Fig. 12 and checks that what type ofgeometrical shape form here. After confirmation of geometrical shape, it assigns value to the undefined pixel.If the region is constant and there is no point of high contrast among the defined pixels of the High resolution (HR) cell as shown in Fig. 12. It will be confirmed by calculating the standard deviation of defined pixels in HR cell.()()()()NX X X X X X X X i i i i 2322212−−−−+−+−+−=σ (4)2 * σ < T (5)Where X is mean of defined pixels in HR cell and σ is standard deviation. If equation 5 is true it mean the HR cell is consist on constant region. No point of high contrast, then it form a square, and all the undefined pixels T, B, L, R, C which are inside the region of this square, will be assign the average value of the defined pixels in the HR cell and it will adopt a form as in Fig. 13.Fig. 13 HR cell which adopt the shape of Square. All undefinedpixels inside this square, will have average intensity value of definedpixels If the pixel x i in the HR cell as shown in Fig. 12, isolate from other defined pixels due to intensity difference then there is a top-left corner edge, the detail also present in section 3.2 and it forms a shape of triangle. For confirmation, the standard deviation will be calculated of the defined pixels inside atriangle and then it will be compared with threshold as in equation 5 if it is true then it is a triangle with top-Left corneredge. After confirmation, all undefined pixels inside triangle which are C, R, B are assigned the average value of the defined pixels in the triangle which are x i-1, x i-2, x i-3 and other two undefined pixels of HR cell are left undefined and will beconsidered in next phase.Fig. 14 A triangle with Top-Left corner edgeIn Fig. 14 a circle ‘x i ’ has gray color which shows intensity difference from other three defined pixels of the triangle. The strength of employing triangles in this way is that edges are modeled in the image. In fact it tunes the interpolator to match the edges. In Fig. 14, when interpolating the HR pixel fallingin triangle x i-1, x i-2, x i-3, the interpolator will not use the value of x ipixel which is very different to this plateau and thus the sharpness of the edge is kept. The interpolator keeps smoothness as well, even across triangle boundaries. The same procedure has been repeated for other three cases of triangle which are bottom-left corner edge triangle, top-right corner edge triangle and bottom-right corner edge triangle, detailpresent in section 3.2. In case when pair of pixels on different poles have contrast intensity which split the HR square region into two vertical orhorizontal regions, also explained in section 3.3. The pixels onthe same side having intensity difference less then thresholdwill be considered in the same region.Fig. 15 HR cell which adopt the shape of Split Square which divideHR square into two vertical region of different intensityIn Fig. 15, when interpolating the HR pixel on different sides, the interpolator won’t use the value of the other side andthis vertical edge can be expanded to both side but one at thesame time as shown in the diagram. So pixels of HR T, C and B would take the average value of left two original pixels x i and x i-2 or right side pixels x i-1 and x i-3 because both have the same result. It has been proved experimentally duringdevelopment of the algorithm . The pixels L will take average vale of left and R will take average value of right. The same procedure can be repeated for HR square split by horizontaledge.This simply geometry suggest a way to guide the interpolator so that smoothness within the regions and sharpness between the flat region and cliff region can both be kept. In this 2ndphase of the algorithm, approximately 85% undefined Pixels of HR image are assigned proper intensity values.In third phase of the algorithm it scans magnified image lineby line and looking for those pixels which left undefined in theprevious phase. Solid pixels are original pixels in Fig. 16.Fig. 16 Report the layout referred in the description of phase three ofthe algorithmTwo scenarios are there in 3rd phase. 1st when c1 and c2 are not assigned in both forms of the Fig. 16. Then the intensity difference of the original pixels is calculated. If this difference less than threshold then it is assigned the average value of both original pixels to ‘a’ otherwise leave it undefined. In 2ndscenario when both c1 and c2 are defined then the direction of edge is calculated to specify the interpolation region and then assign proper intensity value to ‘a’. At the end of 3rd phase all pixels whose spatial dependence from the neighbourhood values is “sample” have been assigned. Using the information gathered insofar, in the next stage the remaining “holes” are eventually filled.In fourth and last phase of the proposed algorithm the holes are filled which are left undefined up to this phase. In this phase of algorithm, the median of the neighbouring defined pixels of the undefined pixel is calculated and then assign this calculated value to the undefined pixel. So in this way the more frequent value can be selected and it has been tested experimentally that it guarantees a better detail preservation in the magnified image.Eventually the algorithm scan image again by applying the 4thphase of the algorithm and look for undefined pixels whose value is still left undefined. But it has been observed that this last extra scanning of image unessential because most of the time, no pixels remain undefined up to this last extra scanning but it is only for care.The magnification of color image is performed by independently applying the proposed method to each color channel.V. A LGORITHM A NALYSISIn this section the complexity of proposed algorithm is analysed. Suppose the Image I has width n and height m. So it requires O (n x m) steps to magnify out of factor two of digital image. The algorithm does not require any preliminary information for geometrical shapes because it measures all the shapes during the execution of algorithm and therefore it also does not need any extra memory. Therefore the memory requirement is also simple. The algorithm requires only the storage space for the magnified image. Therefore proposed algorithm gives excellent result with respect to execution time and memory space as well.VI. R ESULT AND A NALYSISOnly through extensive testing can an algorithm be considered a success. For any image magnification algorithmto be considered successful, it needs to excel in both a qualitative and a quantitative analysis when compared with other magnification techniques. Proposed Interpolation technique was tested and compared with several common magnification techniques. These include Nearest neighbor(Pixel Replication), Bilinear Interpolation, Bicubic Interpolation, and Battiato, G. Gallo, and F. Stanco[20] and Muneeb,Naveed khattak [6]. Qualitative analysis provides aset of image comparisons to the reader for personal analysis. Aquantitative analysis is done utilizing several methods,including mean squared error, mean absolute error, cross-correlation coefficient. As is shown in the results, proposed Interpolation performs well in both qualitative and quantitativemeasurements, and can be considered successful in producing realistic higher resolution interpolations of real world images.A. Qualitative AnalysisIn qualitative analysis of proposed Interpolation comparison was made with series of image and was judge. Here only one comparison of the sub image of camera stand has been presented. Other comparison can be provided by demand. By this one comparison the reader can judge for themselves if proposed Interpolation produces results that are superior to common magnification techniques. The following pages include this image comparison, labeling with their corresponding interpolation algorithm. The magnification of aerial and color images are also given in Figs. 18 & 19 respectively.B. Quantitative AnalysisIt is difficult to rank a technique by just looking at its visual results. So there must be a mathematical method used to compare different underlying interpolation techniques. Proposed interpolation method can be compared with other interpolation algorithms like nearest neighbor (NN), bilinear (BL), bicubic (BC), Muneeb & Khattak(MK)[6] , and Battiato(B)[20] interpolation. The 1st three interpolation algorithms are most common techniques and best options available for comparison. Three different quantitative analysis measurements were used on four real world images. In order to obtain these measures, an image was first down sampled by a factor 2. This lower-resolution image was then magnified by a factor of 2 using a variety of magnification techniques, and then compared with the original image. MK’s code was provided by Khattak and pseudo code description of the Battiato’s algorithm was present in [20] and it was implanted to make it executable for analysis purposes. Three of the measurements reflect the accuracy of the magnification. These measurements are Mean Squared Error(MSE), Mean Absolute Error(MAE), and Cross-correlation Coefficient(CCC) The equation are given from 6 to 8.|..),(),( Î..),(),( Î|,1,122,1,122,1,1⎟⎟⎟⎟⎠⎞⎜⎜⎜⎜⎝⎛⎟⎟⎟⎠⎞⎜⎜⎜⎝⎛−⎟⎟⎟⎠⎞⎜⎜⎜⎝⎛−⎥⎥⎥⎦⎤⎢⎢⎢⎣⎡−=∑∑∑======N M y x N M y x N M y x b l k y x I na y x b a n y x I y x CCC. (6)()MNy)I(x,-y)Î(x,,1,12∑===NM y x MSE (7)MNy)I(x,-y)Î(x,,1,1∑===NM y x MAE (8)(a) Original Image(b) Magnified 4x with Pixel Replication (c) Magnified 4x with Bilinear(d) Magnified 4x with Bilinear (e) Magnified 4x with KM [6](f) Magnified 4x with Battiato [20] (g) Magnified 4x with Proposed AlgorithmFig. 17Magnifying sub-Image of Camera Stand。

相关文档
最新文档