HEBS Histogram Equalization for Backlight Scaling

合集下载

基于灰度归一化融合深度神经网络的水下图像增强方法

基于灰度归一化融合深度神经网络的水下图像增强方法

SR
l MSE
=
VGG1.9
网络
(4)
G
噪比。
基于 ReLU 的 VGG 损失函数如下所示:
W i j H i j
1
=
(ϕ ij (I HR) xy - ϕ ij (G θ (I HR)) xy )2
å
W ij H ij å
x=1y=1
G
式中 ϕ ij 表示某个卷积层的特征图,用于计算与生
一幅虚假的高分辨率图像,此时,将高分辨率的真
行一个调整,最终得到一个 4 倍分辨率的图片。
中判别真假,在对比过程中生成训练判别模型的损
倍,然后利用一个 3 通道的卷积对原来的通道数进
实图像与生成的虚假高分辨率图像传入判别模型
失函数;同时,真假图像传入 VGG 网络后,获取图
像的特征并进行对比,再得到生成模型的损失函
度较高的图片。白平衡方法是根据图像呈现的色
多种改进算法也获得了较好的结果。直方图均衡
化方法是将图像的灰度直方图中的灰度区间集中
所示。GAN 的优化过程是两个网络之间的博弈过
函数如式(2)所示:
min
max
{F(DG) = E x~p
G
D
部分均匀拉伸,扩大图像的灰度值范围并增加对比
度,从而突出细节效果。暗通道先验法由 He
good effect on the color correction of underwater images,while the effect of the anti-generation network on image pixel reconstruc⁃
tion has the advantages of more details and smooth transition than traditional enhancement methods. Experiments show that this

Thresholded and Optimized Histogram Equalization for contrast

Thresholded and Optimized Histogram Equalization for contrast

Thresholded and Optimized Histogram Equalization for contrastenhancement of imagesqP.Shanmugavadivu a ,⇑,K.Balasubramanian baDepartment of Computer Science and Applications,Gandhigram Rural Institute –Deemed University,Dindigul,Tamil Nadu,India b Department of Computer Applications,PSNA College of Engineering and Technology,Dindigul,Tamil Nadu,Indiaa r t i c l e i n f o Article history:Available online 20July 2013a b s t r a c tA novel technique,Thresholded and Optimized Histogram Equalization (TOHE)is pre-sented in this paper for the purpose of enhancing the contrast as well as to preserve theessential details of any input image.The central idea of this technique is to first segmentthe input image histogram into two using Otsu’s threshold,based on which a set of weigh-ing constraints are formulated.A decision is made whether to apply those constraints toany one of the sub-histograms or to both,with respect to the input image’s histogram pat-tern.Then,those two sub-histograms are equalized independently and their union pro-duces a contrast enhanced output image.While formulating the weighing constraints,Particle Swarm Optimization (PSO)is employed to find the optimal constraints in orderto optimize the degree of contrast enhancement.This technique is proved to have an edgeover the other contemporary methods in terms of Entropy and Contrast ImprovementIndex.Ó2013Elsevier Ltd.All rights reserved.1.IntroductionContrast enhancement techniques are used in image and video processing to achieve better visual interpretation.In gen-eral,histogram equalization based contrast enhancement is attained through the redistribution of intensity values of an in-put image.Histogram modification is the underlying strategy in most of the contrast enhancement techniques.Histogram equalization (HE)is one of the most widely used techniques to achieve contrast enhancement,due to its simplicity and effec-tiveness [1].The HE techniques use linear cumulative histogram of an input image and distribute its pixel values over its dynamic intensity range.HE based enhancement finds applications in medical image processing,speech recognition,texture synthesis,satellite image processing,etc.HE methods can be categorized into two as global and local.Global HE methods improve the quality of image by normal-izing the distribution of intensities over the dynamic range,using the histogram of the entire image.Histogram equalization is an ideal example of this approach that is widely used for contrast enhancement [1].It is achieved by manipulating the intensity distribution using its Cumulative Distribution Function (CDF)so that the resultant image may have a linear distri-bution of intensities.As HE modifies the mean of the original image,it tends to introduce washed-out effect in the output image.Local Histogram Equalization (LHE)methods use the histogram intensity statistics of the neighbourhood pixels of an im-age for equalization.These techniques usually divide the original image into several non-overlapped sub-blocks and perform histogram equalization on the individual sub-blocks.The resultant image is produced by merging the sub-blocks using the 0045-7906/$-see front matter Ó2013Elsevier Ltd.All rights reserved./10.1016/peleceng.2013.06.013qReviews processed and recommended for publication to Editor-in-Chief by Deputy Editor Dr.Ferat Sahin.⇑Corresponding author.Tel.:+919443736780.E-mail addresses:psvadivu67@ (P.Shanmugavadivu),ksbala75@ (K.Balasubramanian).bilinear interpolation method.The major drawback of these methods is the introduction of checkerboard effect which ap-pears near the boundaries of the sub-blocks.Histogram Specification(HS)is another enhancement method in which the ex-pected output of image histogram can be controlled by specifying the desired output histogram[1].However,specifying the output histogram pattern is not a simple task as it varies with the images.In this paper,a novel HE based method namely,Thresholded and Optimized Histogram Equalization(TOHE)is proposed that uses Otsu’s method to perform histogram thresholding.With respect to Otsu’s threshold,a set of weighing constraints are formulated and applied to the sub-histograms.The weighing constraints are optimized using Particle Swarm Optimiza-tion(PSO),which is a population-based optimization technique.TOHE is proved to exhibit better brightness preservation and contrast enhancement.In Section2,the traditional HE and a few contemporary HE based methods are described.Section3presents the principle of the proposed technique,TOHE,various image enhancement measures and the PSO algorithm.In Section4,the results are discussed and in Section5,the conclusion is given.2.Review of histogram equalization methodsThe traditional histogram equalization technique[1]is described below:Consider the input image,F(i,j)with a total number of‘n’pixels in the gray level range[X0,X NÀ1].The probability density function P(r k)for the level r k is given by:Pðr kÞ¼n kð1Þwhere n k represents the frequency of occurrence of the level r k in the input image,n is the total number of pixels in the image and k=0,1,...,NÀ1.A plot of n k against r k is known as histogram of the image F.Based on Eq.(1),the cumulative density function is calculated as:Cðr kÞ¼X ki¼0Pðr iÞð2ÞHE maps an image into the entire dynamic range,[X0,X NÀ1]using the cumulative density function which is given as: fðXÞ¼X0þðX NÀ1ÀX0ÞCðXÞð3ÞThus,HEflattens the histogram of an image and results in a significant change in the brightness.A new HE based brightness preservation method known as Brightness Preserving Bi-Histogram Equalization(BBHE)was proposed by Kim[2].BBHEfirst segments the histogram of input image into two,based on its mean;the one ranging from minimum gray level to mean and the other from mean to the maximum and then,it equalizes the two histograms indepen-dently.It has been clearly proved that BBHE can preserve the original brightness to a certain extent[3].Wan,Chen and Zhang proposed a method(1999)called equal area Dualistic Sub-Image Histogram Equalization(DSIHE)which is an extension of BBHE[4].DSIHE differs from BBHE only in the segmentation process.The input image is segmented into two,based on med-ian instead of mean.This method is suitable only for images which are not having uniform intensity distribution.But,the brightness preserving potential of DSIHE is not found to be significant.Minimum Mean Brightness Error Bi-Histogram Equal-ization(MMBEBHE)is also an extension of BBHE(2003)proposed by Chen and Ramli which performs the histogram sepa-ration based on the threshold level,which would yield minimum difference between input and output mean,called Absolute Mean Brightness Error(AMBE)[5].This technique is also not free from undesirable effects.Recursive Mean Separate Histogram Equalization(RMSHE)was proposed by Chen and Ramli(2003)in which the histo-gram of the given image is partitioned recursively[6].Each segment is equalized independently and the union of all the seg-ments gives the contrast-enhanced output.This technique has been clearly proved to be a better method among the recursive partitioning approaches.Sim,Tso and Tan proposed a similar method(2007)called Recursive Sub-Image Histo-gram Equalization(RSIHE)[7].This technique has the same characteristics as RMSHE in equalizing an input image,except that it separates the histogram based on gray level with cumulative probability density equal to0.5,whereas RMSHE uses mean-separation approach.This method is proved to have an edge over RMSHE.However,this recursive procedure increases the computational complexity and the resultant image is very similar to that of original input image as the recursion level increases.Moreover,finding a generic optimal level of recursion,applicable to all types of images is still a challenge for all of these methods.A fast and effective method for video and image contrast enhancement,known as Weighted Thresholded HE(WTHE)was proposed[8].This technique provides an adaptive mechanism to control the enhancement process.WTHE method provides twofold benefits such as,adaptivity to different images and ease of control,which are difficult to achieve in the GHE-based enhancement methods.In this method,the probability density function of an image is modified by weighing and threshold-ing prior to HE.A mean adjustment factor is then added to normalize the luminance changes.Two more weighing tech-niques,Weight Clustering HE(WCHE)[9]and Recursively Separated and Weighted HE(RSWHE)[10]were also developed (2008).Both these techniques use different weighing principles and have proved their own merits.Ibrahim and Kong pro-posed Sub-Regions Histogram Equalization(SRHE)[11]which partitions the input image based on its Gaussianfiltered, 758P.Shanmugavadivu,K.Balasubramanian/Computers and Electrical Engineering40(2014)757–768smoothed intensity values and outputs the sharpened image.Zuo,Chen and Sui recently developed Range Limited Bi-Histo-gram Equalization(RLBHE)for image contrast enhancement in which the input image’s histogram is divided into two inde-pendent sub-histograms by a threshold that minimizes the intra-class variance[12].Then,the range of the equalized image is calculated to yield minimum absolute mean brightness error between the original image and the equalized one.3.Thresholded and Optimized Histogram Equalization(TOHE)The proposed TOHE accomplishes the contrast enhancement of input images in three distinct phases as:Phase I:Segmentation of the input image’s histogram based on Otsu’s threshold.Phase II:Development of weighing constraints with respect to the threshold.Phase III:Optimizing the constraints using Particle Swarm Optimization(PSO).3.1.Segmentation of the input image’s histogram based on Otsu’s thresholdThresholding is an ideal method for image segmentation.A threshold is used to divide the input image’s histogram into two parts:the lower gray level of the object and the higher gray level of the background.Then the target region and the back-ground can be equalized separately so that the contrast of target and background can both be effectively improved.From the pattern recognition perspective,the optimal threshold should produce the best performance to separate the target class from the background class.This performance is characterized by the intra-class variance.Otsu’s method[13]is used to automat-ically perform thresholding based on the histogram shape of the image.Otsu’s method assumes that the image to be thres-holded contains two classes of pixels(e.g.,foreground and background)and calculates the optimum threshold separating those two classes.It exhaustively searches for the threshold that maximizes the inter-class variance which is defined as a weighted sum of variances of the two classes given below:r2ðtÞ¼W LðEðX LÞÀEðXÞÞ2þW UðEðX UÞÀEðXÞÞ2ð4Þwhere E(X L)and E(X U)are the average brightness of the two sub-images thresholded by‘t’.E(X)is the mean brightness of the whole image.W L and W U are the cumulative probabilities of class occurrences of the two classes of pixels of the image and are given as:W L¼X ti¼0pið5ÞandW U¼X NÀ1i¼tþ1pið6ÞFor bi-level thresholding,the optimal threshold tÃis chosen so as to maximize the between-class variance r2(t)as follows: tümax06t<NÀ1f r2ðtÞgð7Þ3.2.Development of weighing constraints with respect to the thresholdThe input image,F(i,j)is segmented into two,using the optimal threshold tÃobtained by Otsu’s method as:F L(i,j)and F U(-i,j).Then,the Probability Density Functions(PDF),P L(r k)and P U(r k)for the lower and upper sub-images respectively are com-puted.The mean PDF of lower and upper sub-images are found as m L and m U.3.2.1.Constraint for lower sub-imageThe PDFs of lower sub-image is suitably modified using the transformation function T(.)with the following constraints:P LCðr kÞ¼TðP Lðr kÞÞ¼a if P Lðr kÞ>aP Lðr kÞÀbaaÂa if b6P Lðr kÞ6a0if P Lðr kÞ<b8>><>>:9>>=>>;ð8Þwhere a=bÂmax(P L(r k)),0.1<b<1.0,b=0.0001and‘a’is the power factor,0.1<a<1.0.Then,the mean PDF of constrained lower sub-image,m LC is calculated.The mean error m eL is found as:m eL=m LCÀm L.The error,m eL is added to P LC(r k).Finally,the Cumulative Distribution Function(CDF),C L(F L(i,j))using P LC(r k)is computed and the HE procedure is applied using:F0 L ði;jÞ¼X0þðtÃÀX0ÞÂC LðF Lði;jÞÞð9ÞP.Shanmugavadivu,K.Balasubramanian/Computers and Electrical Engineering40(2014)757–768759The original PDFs of the lower sub-image are clamped to the upper threshold a and lower threshold b.The PDFs of the lower sub-image which are higher than a are normalized to a,based on the maximum probability level of this sub-image.The b value of lower sub-image isfixed as low as possible(0.0001)since the contribution of intensities of such PDFs is negligible in controlling the enhancement process.3.2.2.Constraint for upper sub-imageAnalogously,the following constraint is applied to the PDFs of upper sub-image.P UCðr kÞ¼TðP Uðr kÞÞ¼d if P Uðr kÞ>dP Uðr kÞÀ/cÂd if/6P Uðr kÞ6d/if P Uðr kÞ</8>><>>:9>>=>>;ð10Þwhere d=dÂmax(P U(r k)),0.1<d<1.0;/the mean(P U(r k))and‘c’is the power factor,0.1<c<1.0.The mean PDF of constrained upper sub-image,m UC and the mean error,m eU are calculated(m eU=m UCÀm U).The m eU is added to P UC(r k).The CDF,C U(F U(i,j))using P UC(r k)is calculated and the HE procedure is applied as: F0Uði;jÞ¼ðtÃþ1ÞþðX nÀðtÃþ1ÞÞÂC UðF Uði;jÞÞð11ÞThe PDFs of the upper sub-image are clamped with respect to the upper threshold d and lower threshold/.The/value of lower sub-image isfixed as the mean PDF value of the upper sub-image so that there is not much degradation in the mean of the output image with respect to the input image.The PDFs of the upper sub-image which are higher than d are normalized to d,based on the maximum probability level of this sub-image.This upper limit clamping procedure in both sub-images is used to avoid the dominating high probabilities,when allo-cating the output dynamic range.The values of the controlling parameters b and d arefixed in the range0.1and1.0.When those values go beyond this limit,image gets over-enhanced.The parameters a and c are the power factors to control the degree of enhancement.When a and c are less than1.0,the less probable levels in the corresponding sub-images are real-located with more probable levels.So,the important visual details of the entire input image are preserved.When the value of those power factors gradually approaches1.0,the effect of the proposed method tends to behave like the traditional HE. When these values are greater than1.0,more weight is shifted to the high-probability levels and TOHE would yield even stronger effect than the traditional HE.But,this also results in over-enhancement,yet it is still useful in specific applications where the levels with high probabilities(e.g.,the background)need to be enhanced with extra strength.The union of F0L ði;jÞand F0Uði;jÞproduces the contrast enhanced output image.F O¼F0L ði;jÞ[F0Uði;jÞð12Þ3.3.Optimizing the constraints using Particle Swarm Optimization(PSO)In this proposed TOHE algorithm,four major parameters namely a,b,c and d are identified.The optimal values of them are found using Particle Swarm Optimization(PSO)in which an objective function is defined which will maximize the con-trast of the input image.There are several measures such as Discrete Entropy(DE)[14]and Contrast Improvement Index(CII) [15],which are used to calculate the degree of image’s contrast enhancement.One such measure can be considered to be an objective function.In this paper,it is found that DE provides better trade-off than CII.Hence,DE is selected as the objective function while CII being a supportive measure to evaluate the degree of contrast enhancement.3.3.1.Discrete EntropyDiscrete Entropy E(X)is used to measure the richness of information in an image after enhancement.It is defined as: EðXÞ¼ÀX255k¼0pðX kÞlog2ðpðX kÞÞð13ÞWhen the entropy value of an enhanced image is closer to that of original input image,then the details of the input image are said to be preserved in the output image.3.3.2.Contrast Improvement IndexCII is a quantitative measure of the image contrast enhancement which is defined as:CII¼C pC oð14Þwhere C p and C o are the contrast values of the processed and original images respectively.The contrast C of an image is de-fined as:C¼ðfÀbÞðfþbÞð15Þ760P.Shanmugavadivu,K.Balasubramanian/Computers and Electrical Engineering40(2014)757–768where f and b are the mean gray levels of the foreground and background of the image.The higher values of CII signify the contrast improvement in the enhanced image.3.3.3.Particle Swarm OptimizationParticle Swarm Optimization(PSO)is a population based stochastic optimization technique developed by Dr.Eberhart and Dr.Kennedy in1995,inspired by social behaviour of birdflocking[16].It uses a number of agents(particles)that con-stitute a swarm moving around in the search space looking for the best solution.Each particle keeps track of its coordinates in the solution space which are associated with the best solution(fitness)that has achieved so far by that particle.This value is called personal best(pbest)and another best value that is tracked by the PSO is the best value obtained so far by any par-ticle in its neighbourhood,is called global best(gbest).The basic concept of PSO lies in accelerating each particle towards its pbest and the gbest locations,with a random weighted acceleration at each time as shown in Fig.1.S k:current searching point.S k+1:modified searching point.V k:current velocity.V k+1:modified velocity.V pbest:velocity based on pbest.V gbest:velocity based on gbest.Each particle tries to modify its position using the information such as current position,current velocity,the respective distance between the current position and pbest,the distance between the current position and the gbest.The velocity of each particle is suitably modified using following equation.V kþ1 i ¼WÂV kiþc1ÂrandðÞÂpbest iÀs kiÀÁþc2ÂrandðÞÂgbest iÀs kiÀÁð16Þwhere v k i is the velocity of agent i at iteration k(usually in the range,0.1–0.9);c1and c2are learning factors in the range,0–4;rand()is the uniformly distributed random number between0and1;s kiis the current position of agent i at k th iteration; pbest i is present best of agent i and gbest is global best of the group.The inertia weight,W is set to be in the range,0.1–0.9and is computed as:W¼W MaxÀ½ðW MaxÀW MinÞÂiterationIteration Maxð17ÞUsing the modified velocity,the particle’s position can be updated as:S kþ1 i ¼S kiþV kþ1ið18ÞThe optimal values of a,b,c and d are found using the following algorithm.PROCEDURE Optimize_Param ISInput:Image X(i,j)Output:Optimal values of a,b,c and dBEGINStep1:Initialize particles with random position and velocity vectorStep2:Loop until maximum generationStep2.1:Loop until the particles exhaustStep2.1.1:Evaluate the difference between Discrete Entropy values of original and TOHEed image(p)Step2.1.2:If p<pbest,then pbest=pStep2.1.3:GOTO Step2.1Step2.2:Set best of pbests as gbest and record the values of a,b,c and dStep2.3:Update particles velocity using Eq.(16)and position using Eq.(18)respectivelyStep3:GOTO Step2Step4:Stop–Giving gbest,the optimal solution with optimal a,b,c and d valuesENDSimilar kind of procedure is developed for two parameters’case also,so as to apply the constraints either to lower(a and b) or upper sub-image(c and d)only.The decision to use any one of these procedures(four parameters case or two parameters) depends on the nature of the input image and of the need of the hour.These procedures were experimented on three dif-ferent images namely,Elaine,Natural and Pirate.The Elaine image shown in Fig.2(a)is a light image which is applied with both PSO procedures.Fig.2(b)–(d)are the TOHEed images with the constraints applied to upper sub-image only,lower sub-image only and to both sub-images respectively.P.Shanmugavadivu,K.Balasubramanian/Computers and Electrical Engineering40(2014)757–768761Fig.1.Modification of a searching point by PSO.Elaine(original),(b)constraints applied to upper sub-image only,(c)constraints applied to lower sub-image only,(d)constraints appliedsub-images.Natural(original),(b)constraints applied to upper sub-image only,(c)constraints applied to lower sub-image only,(d)constraints applied sub-images.Fig.2(b)and(d)are almost equal to original where Fig.2(c)is clearly an enhanced image in terms of visual perception.A natural dark image is given in Fig.3(a).Fig.3(b)–(d)are the results of the proposed technique,TOHE where the constraints are applied to only upper sub-image,only lower sub-image and to both sub-images respectively.It can be noted that Fig.3(b) is better enhanced image than Fig.3(c)and(d).Similarly,in Fig.4,the Pirate and its TOHEed images are given.Among Fig.4(b)–(d),Fig.4(d)exhibits better contrast enhanced result.It is apparent from the above experimentation that it is enough to apply the optimal constraints to only lower sub-image, when the given image is light image.In case of dark image like Natural image given in Fig.3(a),applying optimal constraints only to upper sub-image will improve the contrast effectively.When the input image is having its histograms evenly distrib-uted like Pirate image(Fig.4(a)),it is better to apply the optimal constraints to both sub-images.Once the application of constraints to sub-image(s)is over,both sub-images are equalized independently and their union produces a contrast en-hanced output image.As the performance of this proposed technique is decided by the two dependent parameters,namely the number of gen-erations and the size of the particles,the execution time are thoroughly influenced by those two parameters.The computa-tional complexity of the proposed PSO algorithm is the order of n2,since two iterations are involved in the PSO algorithm; one is to process generations and the other is to process particles.4.Results and discussionThe performance of the proposed method,TOHE was tested on standard images such as Aircraft,Einstein,Airport,F16, Copter,Girl,Truck and Pirate.These images are subjected to PSO-based TOHE process with50generations.Figs.5and6 are the graphs showing the PSO-based optimum entropy search for Aircraft and Truck images.Both graphs clearly show the effective implementation of optimal search mechanism since the best entropy value of each generation revolves around the global best entropy without much deviation.The global best entropy value(optimal)for Aircraft and Truck images are obtained in11th and27th generations respectively.compare the performance of TOHE,the same images are enhanced with the contemporary enhancement techniques BBHE,HS,RMSHE,RLBHE,SRHE and WTHE.The performance of all these methods is measured qualitatively in human visual perception and quantitatively using Discrete Entropy and CII.qualitative performance of TOHE(constraints applied to both sub-images)and the contemporary methods are using Aircraft and Truck images which are given in Figs.7and8respectively.The enhanced images of the sameHS,RMSHE,RLBHE,SRHE and WTHE are shown in Figs.7(b)–(h)and8(b)–(h)respectively.In Fig.7(b)–(h),(a)Pirate(original),(b)constraints applied to upper sub-image only,(c)constraints applied to lower sub-image only,(d)constraints applied sub-images.Fig.6.PSO search for Truck image.Fig.5.PSO search for Aircraft image.764P.Shanmugavadivu,K.Balasubramanian/Computers and Electrical Engineering40(2014)757–768Fig.7.Enhancement of Aircraft image.encircled portions in the Aircraft image clearly show the brightness degradation and over-enhancement.The same abrupt brightness change and over-enhancement can be notified in the Truck images(Fig.8(b)–(h))also.Though there is not much brightness change in the results of RMSHE(Figs.7(e)and8(e)),these images are almost similar to original image even for the recursion level2and the marked portions of them are not found to be enhanced.Figs.7(i)and8(i)are the visual results of TOHE which are better than those of other HE techniques and are free from over-enhancement.The TOHE is found to pro-duce comparatively better results for the other test images too.The histogram patterns of original,HEed and TOHEed images of Aircraft and Truck are shown in Figs.9and10respec-tively.In Figs.9(b)and10(b),the abrupt changes in brightness as well as contrast can be well notified because of the uncon-trolled distribution of histograms,whereas Figs.9(c)and10(c)exhibit the controlled distribution,which results in the expected contrast enhancement and brightness preservation.The proposed technique,TOHE can be effectively applied on video frame image ually,the global HE methods result in the introduction of white saturation effect while enhancing the input video frame images.This is overcome by TOHE and can be suitably applied for video frame enhancement.Fig.11(a)shows the video frame of a mechanic.This image is enhanced by HE and TOHE and are presented in Fig.11(b) and(c)respectively.It is evident from Fig.11(b)that it is added with excessive white saturation.This is effectively avoided in Fig.11(c).Most of the HE based enhancement techniques,invariably attribute to mean shift in the output image due to the redis-tribution of intensity values during intensity normalization.However,the proposed technique TOHE is proved to preserve the mean of the input image.Fig.12shows the comparison of mean values of the original,HEed and TOHEed test images. It is apparent from thisfigure that HE drastically changes the original mean of the input image which always results inP.Shanmugavadivu,K.Balasubramanian/Computers and Electrical Engineering40(2014)757–768765Fig.8.Enhancement of Truck image.Fig.9.Histogram patterns of Aircraft image.the brightness degradation.But,the controlled procedure adopted by TOHE gives the mean values which are closer to that of original ones.This characteristic of TOHE preserves the brightness of the input image.Further,the qualities of the test images which are enhanced using the above mentioned techniques are measured in terms of DE and CII and are given in Tables1and2respectively.In Table1,TOHE is proved to produce better DE values which are very much closer to that of original.This confirms the retention of original image details in the enhanced image.The rel-atively higher CII values of TOHE than other HE methods presented in Table2is another quantitative proof for the merits of TOHE for being a better technique for contrast enhancement.。

一种非线性变换的双直方图红外图像增强方法

一种非线性变换的双直方图红外图像增强方法

一种非线性变换的双直方图红外图像增强方法李绘卓;范勇;唐峻;唐遵烈;熊平;周建勇【摘要】红外图像具有噪声大、对比度低等特点,红外图像增强是红外探测、识别和跟踪应用中的核心问题之一。

在红外图像增强技术中,直方图均衡方法简单、有效,但存在细节信息损失较大的缺陷。

提出一种对红外图像采用非线性变换分段直方图的增强方法,该方法对红外图像进行非线性变换,提高较暗区域的像素亮度,根据前背景区域特征将直方图分成两段,进行双直方图均衡化处理,对前景和背景分别进行图像的增强。

经过实验验证,该算法能有效提高图像亮度,扩大目标区的灰度范围,增强前景图像的细节部分。

%The infrared image has large noise and low contrast characteristic. The infrared image enhancement is one of core problems of infrared detection, recognition and tracking applications. Histogram equalization plays an important role in infrared image enhancement techniques, but large defect exists, for example the loss of detail information. This paper proposes an infrared image enhancement method based on subsection histogram with nonlinear transformation. The pixel luminance from the dark region is improved by nonlinear transformation. The histogram is divided into two sections based on foreground and background regional characteristics, and the two histograms are equalized to enhance the contrast of the foreground and background. The experiment proves that this method can effectively improve the brightness of the image;extend the gray range from the target region and enhance the detail in the foreground.【期刊名称】《计算机工程与应用》【年(卷),期】2014(000)009【总页数】5页(P155-159)【关键词】红外图像;图像增强;非线性变换;双直方图均衡【作者】李绘卓;范勇;唐峻;唐遵烈;熊平;周建勇【作者单位】西南科技大学计算机科学与技术学院,四川绵阳 621010;西南科技大学计算机科学与技术学院,四川绵阳 621010;西南科技大学计算机科学与技术学院,四川绵阳 621010;中国电子科技集团公司第四十四研究所,重庆 400060;中国电子科技集团公司第四十四研究所,重庆 400060;中国电子科技集团公司第四十四研究所,重庆 400060【正文语种】中文【中图分类】TP391.411 引言红外图像具有低分辨率,视觉效果模糊,信噪比低等特点,同时红外图像的图像动态变化范围大,自然环境与高热目标形成明显的暗亮区,超过人类的眼睛感知范围,导致部分图像细节难以被人眼觉察。

基于限制对比度颜色校正的水下图像增强

基于限制对比度颜色校正的水下图像增强

基于限制对比度颜色校正的水下图像增强余义德;周曼丽;王红萍【摘要】针对水下图像对比度低、颜色失真的问题,提出了限制对比度自适应的颜色校正模型算法,在灰度的基础上,先求出RGB各个通道的均值然后进行对比判断,根据均值的大小来决定是一端线性拉伸还是两端线性拉伸.该方法基于颜色平衡、RGB色彩模型和HSI颜色模型的对比度校正.基于增强水下图像质量的需求,该算法在CLAHE分块进行局部处理方式的基础上,对每小块区域进行限制对比度,然后运用颜色校正模型算法,并采用双线性插值来提高算法的效率,实验结果表明该算法更加优异.【期刊名称】《无线电工程》【年(卷),期】2017(047)009【总页数】5页(P16-20)【关键词】水下图像;图像增强;颜色校正模型;自适应算法【作者】余义德;周曼丽;王红萍【作者单位】中国人民解放军第91550部队,辽宁大连 116023;中国海洋大学信息科学与工程学院,山东青岛 266100;中国人民解放军第91550部队,辽宁大连116023【正文语种】中文【中图分类】TP391.41随着科技的进步,光学成像技术得到了广泛的应用,已逐步涉及到人类生活和社会发展的各个方面,但受到场景条件的影响,有些图像拍摄的视觉效果并不理想,需要采用图像处理技术来突出其中的重点目标特征、减弱或去除噪声,得到更适合人或机器进行分析处理的图像[1],进一步完成目标识别、跟踪和解算等。

图像增强作为图像处理的重要组成部分之一,对于改善图像的质量发挥了巨大作用。

早在20世纪60年代末和70年代初,图像增强技术已经在医学影像、地球遥感监测和天文学等领域崭露头角。

目前图像增强技术被广泛用于退化文档图像的增强[2];安控领域中的指纹图像增强[3];交通应用中的雾霭图像增强[4],车牌、路标等重要信息的识别;军事应用中提取我方感兴趣的敌方目标;以及海洋开发中越来越多的水下目标探测与图像处理。

水下环境复杂,对其进行观测、监控时较为直接有效的方法是采集图像视频,而由于水体本来的光学特性以及水中的悬浮颗粒、微生物和藻类等杂质的影响,导致采集到的视频、图像整体模糊,噪声较多,对比度降低,整体偏暗,彩色图片的颜色失真。

基于光照模型的细胞内镜图像不均匀光照校正算法

基于光照模型的细胞内镜图像不均匀光照校正算法

文章编号 2097-1842(2024)01-0160-07基于光照模型的细胞内镜图像不均匀光照校正算法邹鸿博1,章 彪1,王子川1,陈 可2,王立强2,袁 波1 *(1. 浙江大学 光电科学与工程学院, 浙江 杭州 310027;2. 之江实验室类人感知研究中心, 浙江 杭州 311100)摘要:细胞内镜需实现最大倍率约500倍的连续放大成像,受光纤照明及杂散光的影响,其图像存在不均匀光照,且光照分布会随放大倍率的变化而变化。

这会影响医生对病灶的观察及判断。

为此,本文提出一种基于细胞内镜光照模型的图像不均匀光照校正算法。

根据图像信息由光照分量和反射分量组成这一基础,该算法通过卷积神经网络学习图像的光照分量,并基于二维Gamma 函数实现不均匀光照校正。

实验表明,经本文方法进行不均匀光照校正后,图像的光照分量平均梯度和离散熵分别为0.22和7.89,优于自适应直方图均衡化、同态滤波和单尺度Retinex 等传统方法以及基于深度学习的WSI-FCN 算法。

关 键 词:细胞内镜;不均匀光照;光照模型;卷积神经网络中图分类号:TN29;TP391.4 文献标志码:A doi :10.37188/CO.2023-0059Non-uniform illumination correction algorithm for cytoendoscopyimages based on illumination modelZOU Hong-bo 1,ZHANG Biao 1,WANG Zi-chuan 1,CHEN Ke 2,WANG Li-qiang 2,YUAN Bo 1 *(1. College of Optical Science and Engineering , Zhejiang University , Hangzhou 310027, China ;2. Research Center for Humanoid Sensing , Zhejiang Lab., Hangzhou 311100, China )* Corresponding author ,E-mail : **************.cnAbstract : Cytoendoscopy requires continuous amplification with a maximum magnification rate of about 500 times. Due to optical fiber illumination and stray light, the image has non-uniform illumination that changes with the magnification rate, which affects the observation and judgement of lesions by doctors.Therefore, we propose an image non-uniform illumination correction algorithm based on the illumination model of cytoendoscopy. According to the principle that image information is composed of illumination and reflection components, the algorithm obtains the illumination component of the image through a convolution-al neural network, and realizes non-uniform illumination correction based on the two-dimensional Gamma function. Experiments show that the average gradient of the illumination channel and the discrete entropy of the image are 0.22 and 7.89, respectively, after the non-uniform illumination correction by the proposed method, which is superior to the traditional methods such as adaptive histogram equalization, homophobic收稿日期:2023-04-04;修订日期:2023-05-15基金项目:国家重点研发计划项目(No. 2021YFC2400103);之江实验室科研项目(No. 2019MC0AD02,No. 2022MG0AL01)Supported by the National Key Research and Development Program of China (No. 2021YFC2400103); Key Research Project of Zhejiang Lab (No. 2019MC0AD02, No. 2022MG0AL01)第 17 卷 第 1 期中国光学(中英文)Vol. 17 No. 12024年1月Chinese OpticsJan. 2024filtering, single-scale Retinex and the WSI-FCN algorithm based on deep learning.Key words: cytoendoscopy;non-uniform illumination;illumination model;convolutional neural network1 引 言细胞内镜是一种具有超高放大倍率的内窥镜[1-4],可实现常规倍率到细胞级放大倍率的连续放大观察。

图像增强算法综述

图像增强算法综述

图像增强算法综述①靳阳阳, 韩现伟, 周书宁, 张世超(河南大学 物理与电子学院, 开封 475001)通讯作者: 韩现伟摘 要: 图像增强算法主要是对成像设备采集的图像进行一系列的加工处理, 增强图像的整体效果或是局部细节,从而提高整体与部分的对比度, 抑制不必要的细节信息, 改善图像的质量, 使其符合人眼的视觉特性. 首先, 本文从图像增强算法的基本原理出发, 归纳了直方图均衡图像增强、小波变换图像增强、偏微分方程图像增强、分数阶微分的图像增强、基于Retinex 理论的图像增强和基于深度学习的图像增强算法, 并讨论了它们的改进算法. 然后,从视觉效果、对比度、信息熵等方面对几种算法进行了定性和定量的对比, 分析了它们的优势和劣势. 最后, 对图像增强算法的未来发展趋势作了简单的展望.关键词: 图像增强; 直方图均衡; 小波变换; 微分方程; Retinex 理论; 深度学习引用格式: 靳阳阳,韩现伟,周书宁,张世超.图像增强算法综述.计算机系统应用,2021,30(6):18–27. /1003-3254/7956.htmlReview on Image Enhancement AlgorithmsJIN Yang-Yang, HAN Xian-Wei, ZHOU Shu-Ning, ZHANG Shi-Chao(School of Physics and Electronics, Henan University, Kaifeng 475001, China)Abstract : Image enhancement algorithm mainly process the captured images to enhance the overall effect or local details,increasing the overall and partial contrast while suppressing unwanted details. As a result, the quality of the images is improved, conforming to the visual perception of the human eye. Firstly, according to the basic principles of image enhancement algorithms, this study analyzes those based on histogram equalization, wavelet transform, partial differential equations, fractional-order differential equations, the Retinex theory and deep learning, and their improved algorithms.Then, the qualitative and quantitative comparisons between image enhancement algorithms are carried out with regard to visual effect, contrast, and information entropy to indentify the advantages and disadvantages of them. Finally, the future development trend of image enhancement algorithms is briefly predicted.Key words : image enhancement; histogram equalization; wavelet transform; differential equation; Retinex theory; deep learning在全球信息化大幅发展的时代, 对于这个世界的认识越来越依靠于信息的爆炸性传递. 大部分人认识世界的主要途径还是眼睛的可视性, 人眼所看到的一切都可以化作图像的形式. 图像的获取、生成、压缩、存储、变换过程自然会受到各种状况的影响, 例如获取图像时会因为天气原因, 不同光照条件, 图像亮度也有着细微的变化, 同样由于仪器设备的质量, 参数的设置, 人员的操作都会使图像质量在一定程度上的损伤, 影响图像的质量. 图像增强算法的出现, 无疑是对受损的图像做一个“修补”的工作, 以此来满足各样的需求. 图像增强的目的是为了适应人眼的视觉特性,且易于让机器来进行识别. 近些年来, 图像增强的发展计算机系统应用 ISSN 1003-3254, CODEN CSAOBNE-mail: Computer Systems & Applications,2021,30(6):18−27 [doi: 10.15888/ki.csa.007956] ©中国科学院软件研究所版权所有.Tel: +86-10-62661041① 收稿时间: 2020-10-12; 修改时间: 2020-11-05; 采用时间: 2020-11-17; csa 在线出版时间: 2021-06-01涉及了很多领域, 其中包括了遥感卫星成像领域、医学影像领域、影视摄影等各领域[1].要想真正地实现图像增强的效果, 首先对于整个图像来讲, 要提高图像部分和整体的对比度, 细节也不能忽略; 其次应提高图像的信噪比, 抑制噪声的产生,对“降质”的图像处理; 然后是对于增强过的图像来讲,避免出现局部增强不适, 影响人眼的观看模式.下面我们将列出几类典型的且应用范围比较广的图像增强算法以及改进的算法. 直方图均衡(HE)技术原理是对原图像的灰度直方图从比较集中的某个灰度区间转换为全部灰度区域内的均匀分布[2]; 由此算法进行转化的局部直方图均衡化[3], 符合图像局部特性; Kim 等提出的保持亮度的双直方图均衡算法(BBHE)[4],最大亮度双直方图均衡(MMBEBHE)算法有效地保持图像亮度[5]; 迭代阈值的双直方图均衡算法(IBBHE)[6]用迭代的方法达到增强对比度和亮度保持的效果; 彩色图像直方图均衡算法[7], 运算复杂度很低, 合并图像的视觉效果很好. 基于偏微分方程(PDE)的增强方法是把图像作为水平集或高维空间中的曲面, 再根据曲线和曲面演化逐步来增强图像的对比度[8]; 基于全变分模型插值的图像增强方法[9], 保留原图像的细节, 提高了对比度; 基于HE的偏微分方程增强方法, 在梯度域增强对比度基础上[10]提出新梯度变换函数. 小波变换中增强本质是图像信号分解为不同频段图像分量[11]; 小波变换图像多聚集模糊增强方法[12], 增强后的图像较为清晰; 基于离散余弦变换(DCT)和离散小波变换(DWT)的图像增强方法, 提高图像的质量, 同时减少计算复杂度和内存使用量[13]; 基于小波分析和伪彩色处理的图像增强方法[14], 在降噪增强的同时进一步提高图像分辨率. 基于量子力学偏微分方程的缺陷图像增强的研究[15]. 基于PDE的红外图像增强, 很好改进了传统对比度增强方法的不足[16]; 基于PDE平滑技术是一种新兴的图像增强滤波技术, 实质性、开创性的研究在图像增强滤波中引入的尺度空间理论[17]. 基于LBPV (Local Binary Pattern Variance)的分数阶微分图像增强算法[18],在图像纹理和细节方面处理效果比现有分数阶算法效果更好; 自适应分数阶微分理论指纹图像增强算法改进了传统分数阶微分形式, 提高了计算精度[19]. 基于多尺度Retinex的HSV彩色快速图像增强算法, 在HSV 颜色模型中有与Multi-Scale Retinex (MSR)等同的结果, 处理时间短[20]; 基于多尺度Retinex的数字射线照相增强算法, 改善对比度, 抑制噪声[21]; MSR与颜色恢复(MSRCR)算法增强的图像在复杂的情况下进行识别物体[22]; 基于变分Retinex方法的图像增强, 良好结合了MSRCR和变分方法的优点, 保证图像自然度[23].近年来, 基于深度学习的图像处理算法迎来了一个新的时代[24]. Hu等利用超分辨卷积神经网络(SRCNN)方法提高了风云卫星亮温图像的峰值信噪比, 结果较传统方法更精细[25]; Li等利用深度学习来增强低光图像, 提出利用深度的卷积神经网络进行学习, 提高图像质量[26].1 图像增强算法的介绍1.1 直方图均衡算法直方图均衡化算法, 简言之就是对图像直方图的每个灰度级来进行统计[3]. 实现归一化的处理, 再对每一灰度值求累积分布的结果, 可求得它的灰度映射表,由灰度映射表, 可对原始图像中的对应像素来进行修正, 生成一个修正后的图像.1.1.1 传统标准直方图均衡算法f HE传统直方图均衡算法是通过图像灰度级的映射,在变换函数作用下, 呈现出相对均匀分布的输出图像灰度级, 增强了图像的对比度. 该算法是相对于图1中n=1, 均衡函数为的简化模型[27], 即:f HEX k= {X0,X1,···,X L−1}其中, 函数代表直方图均衡过程, 其大致过程为: 已知输入和输出图像为X和Y, 总灰度级为L, 则存在, 均衡后输出和输入图之间有如下变换关系:c(X k)其中, 展现的累积概率分布表示函数输入图像灰度级.图1 全局均衡算法的模型L=∞如果输入图像看作一个连续随机变量, 即,则输出图像自然是一个随机变量, 输出图像灰度级均衡后的概率分布将趋于均匀, 则输出图像的亮度均值为:2021 年 第 30 卷 第 6 期计算机系统应用得到均衡后图像的均值分布与原图像无关, 由此可知其不能有效保持原始图像的亮度, 由于原图像各灰度级概率密度的差异, 简并现象的产生明显变多.1.1.2 保持亮度的双直方图均衡算法BBHE 实质是利用两个独立的子图像的直方图等价性[4]. 两个子图像的直方图等价性是根据输入图像的均值对其进行分解得到, 其约束条件是得到均衡化后的子图像在输入均值附近彼此有界作为基于图像均值进行的分割, 均衡后图像均值偏离原始图像均值的现象不会出现, 达到了亮度保持的目的, 其算法流程如下:G mean 1)计算输入图像均值, 根据均值将原始直方图分为左右两个子直方图.P L (i )P R (i )2)分别计算左右两个子直方图的灰度分布概率直方图和, 即:N L N R 其中, 和分别表示左右两个子直方图的总像素数,L 表示图像总灰度级数.cd f L (i )cd f R (i )3)计算左右两子直方图的累积分布直方图和, 即:tab L (i )tab R (i )4)计算左右两个映射表和, 合并之后得到最终的映射表tab , 其中round 表示四舍五入取整, 即:对于一些低照度和高亮的图像, 均值会处于较低和较高的地方, 若此时基于均值进行分割并分别均衡的话, 很大程度上会导致一个有大量数据的子直方图在小范围内进行均衡的情况出现, 另一个只有少量数据的子直方图却在较宽的范围内均衡.1.2 小波变换图像增强算法19世纪80年代Morlet 提出小波变换的概念, 数学家Merey 在十几年后提出小波基构造思想, 随着Mallat 的加入, 两个人共同建立了小波变换算法. 通过小波逆变换将同态滤波处理的低频分量和经自应阈值噪、改进模糊增强的高频分量得到增强处理后的红外图像[28].1.2.1 标准小波变换图像增强小波理论具有低熵和多分辨率的性质, 处理小波系数对降噪有一定作用, 噪声主要在高通系数中呈现,对高低通子带均需要增强对比度和去噪处理. 标准小波变换图像增强(WT)将图像分解为1个低通子图像和3个具有方向性的高通子图像, 高通子图像包括水平细节图像、垂直细节图像和对角细节图像[29]. 小波变换最大的特点是能较好地用频率表示某些特征的局部特征, 而且小波变换的尺度可以不同[30].1.2.2 改进后的小波变换图像增强算法针对传统方法对图像多聚焦模糊特征进行增强会出现图像不清晰、细节丢失现象, 小波变换图像多聚焦模糊特征增强方法, 利用背景差分法将目标图像的前景区域提取出来, 背景区域亮度会随时间发生变化,进而完成背景区域特征更新; 根据全局像素点熵值和预设阈值校正加强模糊特征, 突出小波变换图像边界局部纹理细节信息, 完成增强变换. 基于小波变换域的医学图像增强方法[31], 是基于Shearlet 变换改进的Gamma 校正, 采用改进的伽玛校正对低频进行处理, 利用模糊对比函数增强图像细节, 增强图像的对比度.二进小波变换简单的对信号尺度参数实现了离散化, 不过仍具备和连续小波变换同样的平移不变特性.利用二进小波变换将指纹图像分解[32], 步骤如下:1)首先将获取的指纹图像进行尺度的分解, 这样得到的频率分量为一低三高;2)对低频分量进行直方图均衡;3)对3个高频分量先进行高斯拉普拉斯掩膜锐化, 得到锐化后的图像;4)直方图均衡后的低频分量和处理后的3个高频分量进行二进小波逆变换重构, 得到增强后的图像.1.3 偏微分方程图像增强算法u (x 1,x 2,···,x n )关于未知函数的偏微分方程是形如式(11)的等式:计算机系统应用2021 年 第 30 卷 第 6 期x =(x 1,x 2,···,x n )Du =u x 1,u x 2,···,u x n 其中, , , F 是关于x 和未知函数u 加上u 的有限多个偏导数的基础函数. 偏微分方程(Partial Differential Equation, PDE)是微分方程的一种, 如果一个微分方程出现多元函数的偏导数, 这种方程就是偏微分方程[33].1.3.1 标准偏微分方程图像增强V l o (p )V l (p )l o V l o (p )V l (p )l o l o l o 假设和分别为两幅图像和l 的对比度场, 若与在每一点上具有相同的梯度方向,但前者大小均大于后者, 则图像应该比l 具有更高的对比度, 可以将看作l 的增强图像. 实际上, 从图像l 到图像的过程就是标准PDE 图像增强实现的过程,可以由以下式子来描述它们的关系:V l o (p )式中, 为增强后图像的对比度场; k 为增强因子,一般情况下k >1, 过大的话会增大噪声. 对于式(12),图像l 是已知的, 其解为:φl o (p )式中, 是一个与坐标无关的常数. 可看到两幅图像之间的动态范围存在k 倍的差距. 对于可在计算机屏幕上显示的数字图像, 其动态范围为0 ~ 255. 我们要做到先要对的对比度场进行约束, 之后开始按照步骤运算, 最后才能得到比较准确的数据.1.3.2 改进的偏微分方程增强方法∇u max ∥∇u ∥min为避免增强图像梯度场同时造成噪声的危害加剧,寻找一种比较适合的增强方法. 定义原图像的数值梯度函数为, 梯度模的最大值为, 最小值为, 增强之后的图像梯度为S [10]:∥∇u ∥[min ∥∇u ∥,max ∥∇u ∥][0,max ∥∇u ∥]式中, 表示梯度场的方向信息. 经过改进的梯度函数梯度场从的区域内映射到内. 原本纹理突显出来的同时保留梯度值较大的边缘.基于量子力学偏微分方程的缺陷图像增强研究方法[15]. 航空材料缺陷的图像增强对缺陷的定性和定量性能起着至关重要的作用, 由于复合材料分布不均匀,将导致缺陷成像对比度不高, 会让识别和量化的难度加大. 算法主要分为两个步骤: 首先是根据量子力学理论, 计算图像边缘的量子概率; 在此基础上, 建立融合各向异性量子概率的偏微分方程来增强航空材料缺陷图像. 此算法可以在有效抑制噪声和减少成像不均匀性的同时, 更好保留缺陷的特征, 增强图像的对比度.1.4 分数阶微分方程增强算法近些年, 分数阶微积分在多领域都有了突破性进展[34]. 分数阶微分不仅可以提升图像中的高频分量, 还可以以一种非线性形式保留图像中低频分量所带有的性能. 常用的分数阶微分定义有G-L 、R-L 、Caputo 三种定义, 其中最常用的是采用非整型分数阶微积分的G-L 定义[35].1.4.1 图像增强的分数阶微分算子构造m ×n 让图像像素邻域中任一像素与对应系数进行乘法运算, 得到的结果再进行和运算, 得到像素点所在位置的回复, 当邻域的大小为, 要求的系数会很多. 这些系数被排列成一个矩阵, 称为滤波器、模板或者掩模[36].f (x ,y )在整数阶微分方程的增强算子中, 有一类是拉普拉斯算子, 对任一二元连续函数来讲, 其拉氏变换可表示为:f (x ,y )f (x ,y )f (x ,y )x ∈[x 1,x 2]y∈[y 1,y 2]n x =[x 2−x 1]n y =[y 2−y 1]由于在图像中, 两个相邻像素点之间灰度产生差异的距离最小, 因此图像在它的x 和y 方向上灰度值的变化只能以像素之间的最小距离为单位来进行数值度量和分析, 所以的最小等分间隔只能设为: h =1, 如果图像中x 和y 方向的持续区间分别为和, 则最大等分份数分别为和.将上式拉普拉斯变换写成离散的表示形式, 对x 方向和y 方向重新定义, 得到它的二阶微分表示:根据以上定义, 可以得到:拉氏算子还要对处理前后的图像完成进一步的叠加, 其方式如下:2021 年 第 30 卷 第 6 期计算机系统应用在雾天图像中应用算子增强图像, 边缘轮廓还有纹理部分的效果会很容易看到, 不过若是图像像素中某一范围灰度变化不明显, 细节可能受到损失. 因此,构建图像增强的分数阶微分算子, 将整数阶微分扩展到分数阶微分上并且应用于图像增强中[37].1.4.2 改进的分数阶微分算子增强图像相比传统的分数阶微分算法的不足, 提出新的改进算法, 在极端条件处理拍摄的交通图像时, 具有良好效果. 上文提到的指纹图像增强算法, 对传统形式加以改造, 在计算精度上有所提升, 进而构造了更加高精度的分数阶微分掩模. 通过对像素周围的纹理对比从而逐点选择微分阶, 明确的选择了具有二阶精度的分数阶微分形式来构造IRH 算子, 并对算子结构进行相应的改进, 之后利用图像的梯度信息和局部统计信息, 结合中心像素对相邻像素的影响, 建立自适应分数阶微分的自适应函数, 此法保留了指纹纹线和图像纹理细节, 对于降噪起到很好的作用.1.5 Retinex 图像增强算法S (x ,y )L (x ,y )R (x ,y )S (x ,y )L (x ,y )Retinex 是retina(视网膜)和cortexv(大脑皮层)组成的, Retinex 算法由美国物理学家提出[38]. Retinex 理论的基础是人类视觉系统的色彩恒常性, 人类视觉感知系统的色知觉存在“先入为主”的特性, 即光源条件发生改变, 视网膜接收到的彩色信息也会被人们的大脑驳回. Retinex 理论的依据就是是原始图像可以分解为照射图像和反射图像, 最重要的就是让摆脱的影响, 以便得到图像的反射属性.1.5.1 经典的Retinex 图像增强对数域进行操作可以把乘法运算变成简单的加法运算, 进而出现了多种Retinex 算法. 经典的有: 单尺度Retinex 算法(SSR)、多尺度Retinex 算法(MSR)和带色彩恢复的多尺度Retinex 算法(MSMCR)等[39].针对运算速度缓慢的问题, 在1986年, Jobson 等[40]将高斯低通滤波与Retinex 结合, 改进了Land 提出的中心环绕Retinex 算法(Center/Surround Retinex), 提出了单尺度Retinex(SSR)算法. 在SSR 算法中, Jobson 等创新的使用高斯函数与图像进行卷积的方式来近似实现了入射分量的表达. 它的数学表达式如式(20)表示:I i (x ,y )i ∈(R ,G ,B )G (x ,y ,c )∗L i (x ,y )其中, 表示原始图像的第i 个通道分量的像素值,颜色通道中的一个, 表示中心环绕函数, 是一种卷积操作表示, 入射分量的表达可以借用Jobson 等的成果, 则可以看做入射图像的第i 个通道分量. SSR 的实现过程如式(21)至式(23)所示:由于SSR 算法处理要对图像细节对比度和色彩的保留做到很好的发展, 而尺度c 又相对难做到极好的运用, MSR 算法的出现, 在很大程度上解决了这一问题, 起到了平衡图像色彩和细节的良好效果.1.5.2 改进的Retinex 图像增强Retinex 算法对于图像增强的效果需要经过精确且复杂的计算, 最后的结果精确度越高, 增强效果将会更好. 文献[20]中基于多尺度Retinex 的HSV 彩色快速图像增强算法. 在HSV 模型中用多尺度Retinex 进行图像增强, 由于颜色转换的非线性, 计算起来非常复杂. 使用亮度校正的MSR 算法基于HSV 颜色模型和修正的V 频道输出图像的RGB 分量的线性形式减少30–75%的平均处理时间, MSR 算法在Haar 小波变换低频区域应用亮度校正的处理速度有很明显优势, 平均加速度接近3倍. 文献[22,23]中介绍了MSRCR 算法. 由于传统均值移位算法有不少的不足, 改进后, 对要增强的图像可以在情况复杂下进行识别物体, 增强对比度的同时, 光晕现象的产生被消灭, 噪声得到抑制,保证图像自然度. 基于Retinex 提出一种自适应的图像增强方法, 其中包括如下4个步骤: (1)用引导滤波器估计其照度分量; (2)提取图像的反射分量; (3)对反射分量进行颜色恢复校正; (4)后处理. 由于雾霾和照度较低, 自然生成的图像质量比较差, 而此法不管是在定量还是定性上都突出了更好的优势. 此算法最终的结果图像具有清晰的对比度和生动自然的颜色[41].1.6 基于深度学习的图像增强算法在当今社会经济科技奋进之时, 深度学习的发展可谓是如日中天, 特别是在图像增强方面.1.6.1 卷积神经网络图像增强算法神经网络(neural networks)最基本的组成结构是计算机系统应用2021 年 第 30 卷 第 6 期神经元(neuron), 神经元概念源于生物神经网络[42]. 卷积神经网络(Convolutional Neural Network, CNN)在传统神经网络基础上, 引入了卷积(convolution)和池化(pooling), CNN 的建筑灵感来自于视觉感知[43]. CNN 是深度学习领域最重要的网络之一, CNN 在计算机视觉和自然语言处理等诸领域都有很大成就. 卷积神经网络的特性比较突出, 除了可以实现权值共享外, 可调的参数相对来说不多, 对二维图像这类的, 它的平移、倾斜、缩放包括其他形变都拥有着极高的不变性.CNN 相比于一般的神经网络, 具有很大优势[44]: (1)局部连接. 每个神经元只与少数神经元相连, 有效地减少了参数, 加快了收敛速度; (2)重量共享. 一组连接可做到同时分享相同的权值, 进一步降低了所需的参数;(3)降采样降维. 池化层利用图像部分相关的依据对图像进行降采样, 降低运算数据量, 留存有效的信息值.卷积神经网络大致包含4部分, 卷积层、池化层、全连接层以及反卷积层, 各自具有不同作用, 承担独自的工作. 深度越深, 网络性能越好; 随着深度增加, 网络性能逐渐饱和.1.6.2 基于深度学习图像增强的改进算法f o=F (g )F (g )Hu 等基于深度学习方法增强MMSI 亮温图像, 设计卷积神经网络重建风云四号卫星MMSI 的亮温图像和风云三号卫星微波成像仪亮温图像[25]. 在根据SRCNN进行实现映射函数, 式中, g 为监测的天线温度的图像, 可用于复原, 使其尽可能接近地面真实高分辨率亮温图像f . 映射函数F 的完成可以依据学习思想, 构建一种卷积神经网络, 为了让观测图像数据重新构建为理想的高分辨数据, 需要对卷积神经网络进行一系列特征变换, 此过程即达成卷积核的卷积操作.相比古老的插值方法而言, SRCNN 方法除了提高图像的峰值信噪比之外, 在提高图像细节较古老的方法也有很大的提高.2 图像增强算法的评价和对比2.1 各种算法增强效果的分析通过对论文文献研究比对, 以及对于其中的经典算法以及改进的算法, 对应用广泛的上述6大类图像增强算法进行较概括的研究分析.图2是几种不同算法得到的增强图像. 从增强图像的效果来看, HE 增强效果是对图像的动态范围进行拉大实现的, 增强效果随动态范围增加而变差. BBHE算法均衡后的图像在增强对比度的同时很好保持原图像的平均亮度. IBBHE 根据各子图像的直方图分别进行独立的均衡化处理, IBBHE 增强效果更好. WT 算法增强图像细节信息, 但是增加了噪声. 小波变换图像多聚集模糊增强方法, 对图像增强后, 图像较为清晰, 细节没有丢失, 效果较好. PDE 和TVPDE 算法放大了图像对比度场, 增强后图像都有较高对比度[45]. 自适应分数阶微分可以很好降噪. SSR 和MSR 算法去除了图像中照度分量影响, 还原景物本身的亮度信息, MSRCR 处理后的图像比原图像细节增加了, 亮度有所提高, 颜色有一定矫正, 对颜色的恢复存在失真现象. 基于深度学习的图像增强算法通过复杂的神经网络, 进行大量的训练, 得到的模型同时减少了训练时间, 取得了更好的精度.2.2 算法增强效果的对比对一幅图像的增强效果来讲, 需要对图像对比度和信息熵来进行评价和比较, 可以对图像有很好认识.图像对比度的计算公式:I i ,j 其中, 为中心像素点的灰度值, N 为图像局部块内像素点的个数. 为了计算一幅完整图像的对比度, 需要对图像中所有部分块对比度总体的平均值来表示.图像的信息熵公式如下:p (k )式中, 为灰度级k 的概率密度, M 为最大的灰度级.表1中为第一幅图通过不同算法得到的图像质量的客观结果评价, 评价指标为对比度和信息熵. 通过对文献中算法的研究以及本文中对增强算法的分析对比, 我们得到表2中对不同算法优缺点的总结.3 增强算法发展趋势及有意义的研究方向根据上文所介绍的不同图像增强算法及实验分析对比结果, 可预见未来的图像增强算法发展将有以下特点: 超分辨率、多维化、智能化和超高速.1)超分辨率, 对获得的低分辨率图像进行增强从而得到超高分辨率的图像, 重点是对采集分辨率以及显示分辨率做进一步的提升, 突破技术壁垒限制, 向时空感知超分辨率迈进.2021 年 第 30 卷 第 6 期计算机系统应用。

直方图均衡化(histogramequalization)

直方图均衡化(histogramequalization)

直⽅图均衡化(histogramequalization)⼀. 原理直⽅图均衡化是想要将聚集在某⼀区间内分布的灰度值,变为均匀的在所有区间内分布。

为了达到这⼀⽬的,我们需要找出⼀个函数T,将r(原图像灰度)映射到s(新图像灰度)上。

同时,由于不想将图像反转,我们需要保证函数单调不减(若需要逆运算,则要严格单调递增)s=T(r)设p r(r)为r的概率分布函数,p s(s)为s的概率分布函数,则两者关系如下p s(s)=p r(r)|dr ds|也就是说,T(r)这个函数现在要满⾜条件:单调递增值域范围为[0,L-1]p s(s)为⼀个常函数(表⽰每⼀个灰度值的概率都相等)现在给出⼀个函数(⾄于为什么是这个函数,我也不知道啊,哪位看官可以帮忙推导⼀下QAQ):s=T(r)=(L−1)∫r0p r(w)dw我们将逐步证明,以上三点成⽴由于此函数为积分所得,因此函数单调递增由于p r(r)为概率密度函数,因此,积分区间为[0,1],乘以常数,区间为[0,L-1]推导过程:ds=dT(r)=d((L−1)∫r0p r(w)dw)=(L−1)p r(r)drp s(s)=p r(r)|drds|=pr(r)|1(L−1)p r(r)|=1L−1此时p s(s)为⼀个均匀密度函数,符合对于离散值,计数每个不⼤于r的概率和,就是这个r对应的ss k=T(r k)=(L−1)k∑j=0p r(r j)⼆. 代码img = imread("14.jpeg");r = img(:,:,1);g = img(:,:,2);b = img(:,:,3);p = stat(r);map = cal(p);img_equa = my_hist_equation(r,map); % 画图%原图subplot(331)imshow(r);title('原图');subplot(332)imhist(r);title('原图直⽅图');%⾃⼰的直⽅图均衡化图subplot(334)imshow(img_equa);title('my直⽅图均衡化图');subplot(335)imhist(img_equa);title('my直⽅图均衡化直⽅图'); subplot(336)plot((0:255)/255,map/255);title('my函数');%标准的直⽅图均衡化图[std_img,T] = histeq(r);subplot(337)imshow(std_img);title('标准直⽅图均衡化图');subplot(338)imhist(std_img);title('标准直⽅图均衡化图直⽅图'); subplot(339)plot((0:255)/255,T);Processing math: 100%title('std函数');figureplot((0:255)/255,map/255);figureplot((0:255)/255,T);%计算图像的分布function p=stat(img)[r,c] = size(img);p = linspace(0,0,256);for i=1:1:rfor j=1:1:cp(img(i,j)+1)= p(img(i,j)+1)+1;endendp = p/(r*c);end%计算r与s的对应关系,矩阵坐标为r,值为s(动态规划) function s=cal(probality)s=probality;l = length(probality);for i=2:ls(i) = s(i-1)+s(i);ends = round(s*255);end%将图⽚根据映射函数,映射function img_equa=my_hist_equation(img,map)[r,c] = size(img);img_equa = zeros(r,c);for i=1:rfor j=1:cimg_equa(i,j) = map(img(i,j)+1);endendimg_equa = uint8(img_equa);end放⼤我的映射函数图⽚为:matlab⾃建的映射函数为:对⽐两个映射函数可以看出来,⾃⼰实现的为⽐较连续的,⽽matlab⾃建的为阶段性的。

DNA指纹图谱论文:DNA指纹图谱的自动识别与分析定位研究

DNA指纹图谱论文:DNA指纹图谱的自动识别与分析定位研究

DNA指纹图谱论文:DNA指纹图谱的自动识别与分析定位研究【中文摘要】DNA指纹图谱是通过实验使不同大小的DNA片断在凝胶底板上分离并显影而得到的图像。

DNA指纹图谱首先在法医、亲子鉴定及遇难人员身份确定等社会领域得到应用。

随后,当生物学家们将这种方法引入到生物学研究中时,其优越的表现使其成为目前分子生物学研究中一种普遍使用的手段,在种质资源保护、作物遗传育种等领域得到广泛的应用。

因此,寻求一种高效的DNA指纹图谱的鉴定方法可以更好的推动DNA指纹图谱在各个适用领域的应用。

本文在已有DNA指纹图谱分析系统基础上,对现有DNA指纹图谱分析系统进行了进一步改进,完善了DNA指纹图谱的数学模型及泳道中谱带的自动分析定位方法。

该数学模型表达了图谱中DNA限制性片段大小与其位移之间的函数关系。

基于数字图像处理的DNA指纹图谱分析定位的具体步骤包括:1)图像预处理。

主要是去除噪声和图像增强,具体手段包括直方图均衡化、图像锐化和中值滤波等。

本文采用的方法是采用直方图均衡化和去除小成分的结合方法,实验表明除小成分方法能很好的去除掉这种不关心的区域,而且很好地去除了噪声,并且虽然对图像边缘产生模糊效果,但极大的降低了泳道与泳道间的区域对泳道的影响,为后续精确的分割泳道奠定了基础。

2)泳道自动分割。

是将图谱中不同的泳道分割开来,形成一个个逻辑上相对独立的实体便于下一步分别进行分析。

根据DNA指纹图谱的特点,本文采用了利用两个泳道间灰度值之和最大的原理进行自动分割的模型,并发展出相应的方法。

实验证明,该方法分割效果较好,能够自动分割出DNA指纹图谱中所有泳道,即使泳道宽度略有不同,也可以准确的定位泳道的位置,并分割出来。

3)谱带自动定位。

这是DNA指纹图谱数字图像分析的最后一步,也是本文的核心部分,谱带的定位也是整个DNA指纹图谱分析过程最重要的部分,从一定程度上说,谱带定位的精确与否直接反映出DNA指纹图谱分析的结果正确与否。

  1. 1、下载文档前请自行甄别文档内容的完整性,平台不提供额外的编辑、内容补充、找答案等附加服务。
  2. 2、"仅部分预览"的文档,不可在线预览部分如存在完整性等问题,可反馈申请退款(可完整预览的文档不适用该条件!)。
  3. 3、如文档侵犯您的权益,请联系客服反馈,我们会尽快为您处理(人工客服工作时间:9:00-18:30)。

HEBS: Histogram Equalization for Backlight Scaling Ali Iranli, Hanif Fatemi, Massoud PedramDepartment of Electrical EngineeringUniversity of Southern California{iranli,fatemi,pedram}@ABSTRACT - In this paper, a method is proposed for finding a pixel transformation function that maximizes backlight dimming while maintaining a pre-specified image distortion level for a liquid crystal display. This is achieved by finding a pixel transformation function, which maps the original image histogram to a new histogram with lower dynamic range. Next the contrast of the transformed image is enhanced so as to compensate for brightness loss that would arise from backlight dimming. The proposed approach relies on an accurate definition of the image distortion which takes into account both the pixel value differences and a model of the human visual system and is amenable to highly efficient hardware realization. Experimental results show that the histogram equalization for backlight scaling method results in about 45% power saving with an effective distortion rate of 5% and 65% power saving for a 20% distortion rate. This is significantly higher power savings compared to previously reported backlight dimming approaches.1 IntroductionAs the portable electronic devices become more intertwined with everyday life of people, it becomes necessary to put more functionality into these devices, run them at higher circuit speeds, and have them consume a small amount of energy. These electronic devices are becoming smaller and lighter and often required to operate with fancy Liquid Crystal Displays (LCD’s) for increasing periods of time. Unfortunately, the battery capacities are increasing in much slower pace than the overall power dissipation of these kinds of devices. Therefore, it is essential to develop low power design techniques to reduce the overall power dissipation of these devices.Previous studies on battery powered electronic devices have pointed out that the energy consumption in the Cold Cathode Fluorescent Lamp (CCFL), which is the backlight of an LCD, dominates the overall energy consumption of the device [1]. In the SmartBadge system for instance, the display consumes 28.6%, 28.6%, and 50% of the total power in the active, idle, and standby modes, respectively [1]. Unfortunately, as for other resources, one cannot tackle this energy hungry component with some form of power shutdown. This is due to the fact that LCD subsystem must be continuously refreshed and cannot be turned off or put to sleep without a significant penalty in performance and Quality of Service (QoS).There are two main classes of techniques for lowering the power consumption of LCD subsystem. The first class of techniques is focused on the digital/analog interface between the graphics controller and the LCD controller [2][3]. These techniques try to minimize the energy consumption by taking advantage of different encoding schemes to minimize the switching activity of the electrical bus. For instance, reference [2] uses the spatial locality of the video data to reduce the number of transition on the DVI bus reducing its energy consumption by 75%. More recently, reference [3] has extended the previous work by using the limited intra-word transition codes instead of single intra-word transition codes used in [2] and including the DC balancing conditions in the coding scheme to achieve more than 60% energy saving, on average compared to the basic transmission protocol. The second class of techniques is focused on the video controller and the backlight to lower the energy consumption of the display system. The key idea of these techniques follows from the observation that eye’s perception of the light, which is emitted from the LCD panel, is a function of two parameters, 1) the light intensity of the backlight and 2) the transmittance of the LCD panel. Therefore, by carefully adjusting these two parameters one can achieve the same perception in human eyes at different values of the backlight intensity and the LCD transmittance.H owever, since the changes in energy consumption of the backlight lamp is higher than that of the LCD panel by orders of magnitude, one can save energy by simply dimming the backlight and increasing the LCD transmittance to compensate for the loss of backlight. Reference [4] was the first to propose a simple approach to the backlight scaling problem, where each pixel value is changed to achieve a higher transmittance through the LCD panel, thereby creating the possibility to reduce the backlight intensity and hence lowering the energy consumption. The main drawbacks of this approach are twofold; a) there is no measure of how distorted the new image would become, and b) manipulating pixels one by one is a time consuming process that limits the applicability of this approach. Image distortion after backlight luminance dimming is evaluated by the percentage of saturated pixels that exceed the range of pixel values, e.g., [0..255]. In contrast, reference [5] improved this basic approach by proposing the contrast fidelity as a measure of image distortion and solving the optimal transmittance scaling policy problem. Moreover, an elegant approach for hardware implementation of the transmissivity scaling function by using the built-in reference voltages is proposed which eliminates the pixel-by-pixel manipulation of the image. The authors report a 3.7X power saving with less than 10% contrast distortion in the image.In this paper, a method is proposed for finding a pixel transformation function that maximizes backlight dimming while maintaining a pre-specified distortion level. This is achieved with the following algorithm, called H EBS (for Histogram Equalization for Backlight Scaling):1)Given an original image, F, and an upper bound on the tolerableimage distortion, we determine the minimum dynamic range, R,of pixel values in a transformed image. The goal is to achieve themaximum power savings as a result of a follow-on backlightdimming. Therefore, this step also produces the optimumbacklight scaling factor, E.2)We determine a transformation function, ), which takes theoriginal image histogram (with dynamic range N) to a uniformdistribution histogram with a range of R. This is in turnaccomplished by mapping the N pixel values in the originalimage to the R target values of the transformed image so that thedifference between the resulting histogram and the uniformdistribution histogram is minimized.3)For reasons having to do with the existing infrastructure forgrayscale level assignment and the efficiency of a hardware-based implementation of the proposed transformation, weapproximate) with a piecewise linear function, / such that thesquared error difference between ) and / is minimized.4)We construct the transformed image by applying / to theoriginal image, F. At the same time, we dim the backlight byfactor E.The advantages of HEBS compared to previous backlight dimming techniques are:1) A more accurate definition of the image distortion which takesinto account both the pixel value differences and a model of thehuman visual system is used.2) A powerful image histogram “compression” technique whichallows us to reduce the dynamic range of an image subject to agiven image distortion level. This should be contrasted to priorworks which can only reduce the image dynamic range bysaturating the pixel values either at one end [4] or both ends [5]of the image histogram.3)Amenable to highly efficient hardware realization by makingminimal change to the existing Programmable LCD ReferenceDriver (PLRD).4)An additional power saving of 15% compared to the best of theexisting strategies. In a typical battery-powered mobile computersystem, this constitutes a total additional system power saving of2 BackgroundFigure 1a shows the typical architecture of the digital LCD subsystem in a microelectronic device. There are two main components in this subsystem: a) video controller and frame buffer memory and b) LCD controller and panel. The image data, which is received from the processing unit, is first saved into the frame buffer memory by the video controller and is subsequently transmitted to the LCD controller through an appropriate analog (e.g., VGA) or digital (e.g., DVI) interface. The LCD controller receives the video data and generates a proper grayscale – i.e., transmissivity of the panel – for each pixel based on its pixel value. All of the pixels on a transmissive LCD panel are illuminated from behind by the backlight. To the observer, a displayed pixel looks bright if its transmittance is high (i.e., it is in the 'on' state), meaning it passes the backlight. On the other hand, a displayed pixel looks dark if its transmittance is low (i.e., it is in the 'off' state), meaning that it blocks the backlight. For color LCD’s, different filters are used to generate shades of three main colors (i.e. red, blue, and green), and then color pixels are generated by mixing three sub-pixels together to produce different colors.Figure 1b depicts the LCD controller in more detail. The data received from the video bus is used to infer a row of pixels timing information and respective grayscale levels. Then, this timing information is used to select the appropriate row in the LCD matrix. Next, the pixel values are converted to the corresponding voltage levels to drive the thin-film-transistors (TFT’s) on different columns of the selected row. The backlight bulb is powered with the aid of a DC-AC converter, to provide the required illumination of the LCD matrix.Each pixel has an individual liquid crystal cell, a TFT, and a storage capacitor (cf. Figure 1c). The electrical field of the capacitor controls the transmittance of the liquid crystal cell. The capacitor is charged and discharged by the TFT. The gate electrode of the TFT controls the timing for charging/discharging of the capacitor when the pixel is scanned (or addressed) by the tracer for refreshing its content. The (drain-) source electrode of the TFT controls the amount of charge. The gate electrodes and source electrodes of all TFT’s are driven by a set of gate drivers and source drivers, respectively. A single gate driver (called a gate bus line) drives all gate electrodes of the pixels on the same row. The gate electrodes are enabled at the same time the row is traced. A single source driver (called a source bus line) drives all source electrodes of the pixels on the same column. The source driver supplies the desired voltage level (called grayscale voltage) according to the pixel value. In other words, ideally, the pixel value transmittance, t(X), is a linear function of the grayscale voltage v(X), which is in turn a linear function of the pixel value X. The transfer function of source driver which maps different pixel values, X, into different voltage levels, v(X) is called the grayscale-voltage function. If there are 256 grayscales, then the source driver must be able to supply 256 different grayscale voltage levels. For the source driver to provide a wide range of grayscales, a number of reference voltages are required. The source driver mixes different reference voltages to obtain the desired grayscale voltages. Typically, these different reference voltages are fixed and designed as a voltage divider. Mathematically speaking, in a transmissive TFT-LCD monitor, for a pixel with value X, the luminance I(X) of the pixel1 is:().()I X b t X(1a) where t(X) is the transmissivity of the TFT-LCD cell for pixel value X, and b [0,1] is the (normalized) backlight illumination factor with b=1 representing the maximum backlight illumination and b=0 representing no backlight. Note that t(X) is a linear mapping from [0,255] domain to [0,1] range. In backlight scaled TFT-LCD, b is scaled down and accordingly t(X) is increased to achieve the same image luminance.Reference [4] describes two backlight luminance dimming techniques. This technique dims the backlight and compensates for the luminance loss by adjusting the grayscale of the image to increase its brightness or contrast. More precisely,().((,))I X t XE E)(1b) where 0<E 1is the backlight scaling factor and )(X,E) is the pixel transformation function.Let x denote the normalized pixel value, i.e., assuming an 8-bit color depth, x=X/255. The authors of [4] scale the backlight luminance by a factor of E while increasing the pixel values from x to )(x,E) by two mechanisms. Clearly, )(x,E)=x denotes the identity pixel transformation function (cf. Figure 2a.) The “backlight luminance dimming with brightness compensation” technique uses the following pixel transformation function (cf. Figure 2b):(,)min(1,1)x xE E) (2a) whereas the “backlight luminance dimming with contrast enhancement” technique uses this transformation function (cf. Figure 2c):(,)min(1,)xx EE)(2b)In these schemes, the optimal backlight factor is determined by the backlight luminance dimming policy subject to the given distortion rate. To calculate the distortion rate, an image histogram estimator is required for calculating the statistics of the input image. Note that the image histogram simply denotes the marginal distribution function of the image pixel values.Reference [5] proposes a different approach in which the pixel values in both dark and bright regions of the image are used to enable a further dimming of the backlight. The key idea is to first truncate the image histogram on both ends to obtain a smaller dynamic range for the image pixel values and then to spread out the pixel values in this range (by applying an affine transformation) so as to enable a more aggressive backlight dimming while maintaining the contrast fidelity of the image. The pixel transformation function is given as (cf. Figure 2d):1 Illuminance can be used to characterize the luminous flux emitted froma surface. Most photographic light meters measure the illuminance. In terms of visual perception, we perceive luminance. It is an approximate measure of how “bright” a surface appears when we view it from a given direction. Luminance is measured in lumens per square meter per steradian. The maximum brightness of a CRT or LCD monitor isa. LCD subsystem architectureb. LCD component architecturec. TFT cell schematicFigure 1. TFT-LCD screen0,011,;1,1(,)x g l g l cx d g x g c d ul g g g g u u llg x u x EE ­°d d ° °d d® ° °°d d ¯) (3)where (g l ,0) and (g u ,1) are the points where )(x,E )=cx+d intersects)(x,E )=0 and )(x,E )=1, respectively.The implementation is quite simple requiring minimal change to the built-in Programmable LCD Reference Driver (PLRD). Notice that PLRD allows a class of linear transformations on the backlight-scaled image, resulting in both brightness scaling and contrast scaling. Clearly, brightness and contrast are the two most important properties of any image. Recall that for transmissive LCD monitors, the brightness control changes the backlight illumination and the contrast control changes the LCD transmittance function.The previous approaches cannot fully utilize the power saving potential of the dynamic backlight scaling scheme, due to the fact thattheir measure of distortion between the original and the backlight-scaled image is an overestimation. This is because these approaches simply either minimize the number of saturated pixel values [4] or maximize the number of pixel values that are preserved [5].It is well known that image distortion (more precisely, the difference between a pair of similar images) is a complex function of the visual perception, and hence, it cannot be correctly evaluated by comparing the images pixel by pixel (i.e., calculating the root mean squared error of the corresponding pixel values) or as a whole (i.e., using the integral of the absolute value of the histogram differences) [7][8]. A correct measure of distortion should appropriately combine the mathematical difference between pixel values (or histograms) and the characteristics of the human visual system. One approach presented in [6] is to perform some transformation on both the original and new (i.e., backlight-scaled) images according to a human visual system model [9], and then, compare the transformed results using quantitative measures.3 Histogram Equalization for Backlight Scaling (HEBS) Due to the coherence of the objects within an image, pixels describing a single object generally have similar luminous intensity values [9].2Thesingle-band grayscale spreading (cf. Figure2d) attempts to exploit this object coherence by enhancing the contrast within a band (sub-range) of the pixel values at the expense of the remaining ones, thereby, creating the opportunity for a more aggressive backlight dimming policy. H owever, a single band is used for the entire image. Intensity values lying above or below the band are mapped to the maximum and minimum possible intensity values, respectively. The result is that the2Luminous intensity is the photometric equivalent of radiance intensity.pixel values within the window are displayed with marginal loss of quality,while information in the other parts of the image is discarded.Let F and F '=)(F ,E ) denote the original and the transformed image data, respectively. Moreover, let D(F F ' and P(F ' E ) denote the distortion of the images F and F ' and the power consumption of the LCD-subsystem while displaying image F ' with backlight scaling factor, E .Dynamic Backlight Scaling (DBS) Problem: Given the original image F and the maximum tolerable image distortion D max,find t he backlight scaling factor E and the corresponding pixel transformation function F '=)(F ,E ) such that P(F ' E ) is minimized and D(F F ' d D max .The general form of DBS problem as stated above is difficult to solve due to the complexity of the distortion function, D , and also the non-linear function minimization step that is required to determine )(F ,E ). In the following, we will try to simplify this problem by 1) fully utilizing the dynamic range of the transformed image F ' in order to achieve the minimum TFT-LCD power consumption P(F ',E ) and 2) by constraining the pixel transformation function to the family of piecewise linear functions (because these piecewise linear functions are desirable from implementation point of view; cf. section 4.2.)Intuitively, to reduce the dynamic range of a given image one can discard the pixels corresponding to the grayscale levels with low population. This in turn minimizes the number of discarded pixels and hence minimizes the image distortion. On the other hand, for an image with a histogram which is uniformly populated with pixels in different grayscale levels, every level is as important as the other and discarding any grayscale level can cause a significant image distortion. Therefore, a “good” transformation which solves the DBS problem i.e., minimizes P(F ',E ), is the one which transforms the original image histogram into a uniform intensity histogram with a minimal dynamic range. H aving made this selection, we can now deal with the complexity of the distortion function as follows. We set the dynamic range of a benchmark image to some target value and plot the distortion value of thetransformed image as a function of this target range We repeat this process for a number of different target ranges per image and for a largenumber of images in the database. Next, resorting to standard regression analysis techniques, we will calculate the “best” global fit to these distortion values. The result will be an empirical curve which maps the observed distortion function values to target dynamic range of transformed images. We call this curve the distortion characteristic curve (cf. section 5.1c)4 Global Histogram Equalization We propose a global histogram equalization scheme in which the intensity values in the image are altered such that the resulting image has the uniform intensity histogram, with the desired minimum (g min )and maximum (g max ) grayscale limits. This transformation may be accomplished by the use of the cumulative distribution function of the pixel intensities to generate the intensity remapping function. In this approach the resulting image will utilize the available display levels very well, because the transformation function is based on the statistics of the entire image.More precisely, let us denote the cumulative distribution histogram of the original image by H , and the different grayscale values of the image pixels by x which are selected from a finite set of values G , (e.g. G =[0..1]). Transformation function :G G )o is a monotonic function, which maps the original pixel values x into a new pixel values x’ and as a result equalizes cumulative histogram H to become the cumulative uniform histogram, U . (that is a sloped line going from 0 to N, where N represents the number of pixels over which the histogram has been calculated, i.e. number of pixels in the image.)Global Hist ogram Equalizat ion (GHE) Problem: Given the original image cumulative histogram H, find a monotonic transformation:G G )o where G =[0..1] such that a. identity b. grayscale shiftc. grayscale spreadingd. single-bandgrayscale spreadingFigure 2. Pixel transformation functions(())()GU x H x dx) ³(4)is minimized.If the targeted histogram is a uniform distribution between upper and lower limits, g min and g max ,3 then to minimize (4) the transformation function ) can be calculated as:1()(()(()))min max min H x g g g Nx U H x) (5)In actual implementation, it is common to have a discrete version of histogram instead of the cumulative histogram. To convert (5) into a histogram based formulation one can differentiate both sides of (5) to get,()()()max min d x h x g g dx N)(6)where h(x) denotes the marginal distribution histogram. We can then use the first order difference approximation for the differentiation operator, to calculate the discrete transfer function as11()()()i k min min k k k k ki max x g x h x g g x Nx x ) ' '¦(7)where x i G are the center points for the histogram buckets and h (x k ) are the histogram value.4.1 HEBS Implementation Figure 4 depicts the flow of the H EBS algorithm. A user-specified maximum tolerable image distortion is first read as input and is subsequently used to look up the minimum admissible dynamic range for the image from the distortion characteristic curve (cf. Section 3.) Using this minimum admissible dynamic range and the transmissivity characteristics of the TFT display, a maximum backlight scaling factor,E , is calculated and used to scale down the CCFL backlight. Moreover,this minimum dynamic range along with the original image histogram will be used by the GH E problem solver to calculate the pixeltransformation function )(F ,E ). Next, the transformation function is approximated by a piecewise linear function, /(F ,E ), which is in turn used to determine the reference grayscale voltages, and to transform the original pixel values to new ones for the displayed image. In [5], Cheng et al proposed an elegant and highly efficient approach for backlight scaling implementation based on the reference voltages of the source drivers. Figure 5a shows the schematic of the built-in reference voltage driver of a typical LCD panel (recall that the voltage divider provides the required reference voltages for the source driver buffers.) For example in [11], an Analog Devices input LCD reference driver [12] is used with a 10-way voltage divider. The proposed approach in [5] adds controllable switches to both ends of the voltage driver to be able to clamp the low and high gray scale levels to the minimum and maximum gray scale values, respectively, thereby reducing the dynamic range of the image and increasing the slope of the grayscale-voltage function (cf. Sec. 2) to realize contrast compensation. H owever, this solution is limited in two ways; 1) It can only generate a transfer function of single-band grayscale spreading form (cf. Figure 2d), and 2) the linear regionof the transfer function can only have a single slope.Figure 3. k -window grayscale spreading function3U(x)=0 for x < g min ;U(x)=N.(x-g min ) / (g max -g min ) for g min x g max ;and U(x)=N for x>g max.Figure 4. Histogram Equalization for Backlight Scaling flowTo implement H EBS, we use a hierarchical structure for the referencevoltage dividers as shown in figure 5b. This structure provides more flexibility in creating different slopes for multiple linear regions of the grayscale-voltage transfer function. Moreover, adding switches between different grayscale levels enables one to provide flat-bands notonly at the two ends of the image histogram, but also in the middle range of the gray scale levels.To achieve multiple output slopes for the grayscale-voltage transfer function, k different controllable voltage sources, V i , are needed (cf. Figure 5b.) These voltage sources are normally set to voltage levelsi dd V i V kwith i=1…k, creating a transfer function with slope of one.Here V dd denotes the supply voltage, and i and k denote the voltage source number and total number of available voltage sources. To create different slopes for different regions of the grayscale values, one can change the voltage levels of controllable sources to create a k -band grayscale spreading function as described below. In particular we describe a procedure for approximating the pixel transformation function )(F ,E ) with a piecewise linear function /(F ,E ), and then explain how to determine the voltage levels V i , to implement this approximated function. TFT-LCD displays are only capable of displaying a finite number of different grayscale levels, therefore, the input and output values of the transformation function )(F ,E ) are discrete. This observation implies that even the exact form of the transformation function )(F ,E ) as given by (7) is a piecewise linear function. H owever, the number of linear segments of )(F ,E ) is()O G , which is too large for efficient hardwareimplementation. We therefore approximate )(F ,E ) with anotherpiecewise linear function that has a small number of linear segments. Let P ={p 1, …, p n }={(x 1,y 1), …, (x n ,y n )} denote the ordered set of endpoints of each linear segment in exact form of )(F ,E ) starting from x 1=0 for the darkest to x n =255 for the brightest grayscale level. Moreover, let Q ={q 1, …, q m },denote the ordered set of the endpoints of linear segments in /(F ,E ), which is the approximation of )(F ,E ).Clearly, we have the following:111;m n i j i k jQ Pq p and q p q p and q p where k !(8)Piecewise Linear Coarsening (PLC) Problem: Given a piecewise linear curve P, approximate it by another piecewise linear curve Qwith a given number of line segments m so that the mean squared error between )(F ,E )and /(F ,E )is minimized.The PLC problem can be solved using a dynamic programming technique. Let E(n,m) denote the mean squared error between the original curve with n points and its best approximation with m n points. Then,1 (1)(,)((,1)())(1,0)0,(,0),(1,)0,j m n E n m minE j m e j E E n and E m m nf (9)where e(j) denotes the mean squared error incurred by approximating all segments between p j and p n by a single line connecting p j to p n . The time complexity of this algorithm is 2()O mn .Using the solution for the PLC problem, the voltage levels V i , are calculated as.iq i ddY V V E(10)where iq Y denotes the y-component of point q i . Note that the backlight dimming factor E is present in denominator to spread the grayscale level of the resulting image, and hence, compensating the loss of brightness due to backlight dimming. 5 Experimental results5.1 Characterization a)Cold Cathode Fluorescent Lamp (CCFL)The CCFL illumination is a complex function of the driving current,ambient temperature, warm-up time, lamp age, driving waveform, lampdimensions, and reflector design [13]. In the transmissive TFT-LCDapplication, only the driving current is controllable. Therefore, wemodel the CCFL illumination as a function of the driving current onlyand ignore the other parameters. Accounting for the saturationphenomenon in the CCFL light source [10], we use a two-piece linearfunction to characterize the power consumption of CCFL as a functionof the backlight factor:.0().1lin lin sbacklight sat sat s A C C P A C C E E E E E d d d ­®¯(11)Relationship between the CCFL illumination (i.e., luminous flux incident on a surface per unit area) and the driver’s power dissipation for the CCFL in LG Philips transmissive TFT-LCD LP064V1 [11] is shown in Figure 6a. The CCFL illumination increases monotonically as the driving power increases from 0 to 80% of the full driving power. For values of driving power higher than this threshold, the CCFL illumination starts to saturate. The saturation phenomenon is due to the fact that the increased temperature and pressure inside the tube adversely impact the efficiency of emitting visible light [14].After interpolation, we obtain the following coefficient values for the CCFL in LG Philips transmissive TFT-LCD LP064V1:C s =0.8234,A lin =1.9600,C lin =-0.2372, A sat =6.9440, and C sat = 4.3240.b)TFT-LCD display matrixThe hydrogenated amorphous silicon (a-Si:H ) is commonly used tofabricate the TFT in display applications. For a TFT-LCD panel, the a-Si:H TFT power consumption can be modeled by a quadratic function ofpixel value x [0,1] [15]2()..TFT Panel P x a x b x c (12)We performed the current and power measurements on the LGPhilips, LP064V1 LCD. The measurement data are shown in Figure 6b. The regression coefficients are thus determined as:a =0.02449,b = 0.04984, andc =0.993.The power consumption of a normally white TFT-LCD panel decreases slightly as its global transmittance increases. In other words, while maintaining the same luminance, the power consumption of the TFT-LCD decreases when dimming the backlight (which diminishes the power savings of a backlight dimming approach.).However, the change in the TFT-LCD power consumption is quite small compared to the change in CCFL power consumption. In contrast, power consumption of normally black TFT-LCD panel increases slightly as its global transmittance increases (which increases power savings of a backlight dimming approach.) However, for either type of the TFT-LCD, the change in power consumption as a function of the transmittance is so small that it can be ignored.c)Distortion characteristic curveTo characterize the distortion of different images with respect to change in their dynamic range, we set the dynamic range of a benchmark image to some target value and plot the distortion value of the transformedimage as a function of this target range. We adopted the universal image quality index proposed in [8] as our distortion measure and used a set ofbenchmark images from the USC SIPI Image Database (USID) [16].The USID is considered the de facto benchmark suite in the signal andimage processing research field [9]. Figure 7 depicts the resultingdistortion values for these images when the dynamic range of thetransformed image is set to ten different values. Figure 8 is a subset ofbenchmarks reported to provide a visual reference for the distortionmeasure. Next, we used standard curve fitting tools provided inMATLAB version 7 release 14 to find the best “average” and “worst-case” global fits to these distortion values. The result is an empiricalcurve depicted in Figure 7, which maps the observed distortion functionvalues to target dynamic range of transformed images.5.2 Backlight scaling results To show the effectiveness of H EBS approach the power saving for different images from USC SIPI database is reported in table 1. These power savings are generated for three different values of distortion levels. Clearly, by increasing the maximum tolerable distortion level the power saving should increase, which is also confirmed with listed results. Please note that the average power saving of 58% is achieved only for mere distortion level of 10%. This is about 15% increase in power saving comparing to the results reported in [4] and [5]. 6 Conclusions In this paper, histogram equalization for backlight scaling with pre-specified image distortion level was proposed. The proposed approach was based on an accurate definition of the image distortion. Experimental results showed the effectiveness of H EBS method. Infuture work alternative distortion measures and histograms equalizationmethods will be evaluated.0.50.60.70.80.9Figure 6.a. CCFL illuminance (i.e., backlight factor b) versus driver’spower consumption for a CCFL source a. Conventional circuit b. Proposed circuitFigure 5. Reference voltage divider schematics。

相关文档
最新文档