基于计算机视觉的障碍物检测与测量(IJISA-V2-N2-3)

基于计算机视觉的障碍物检测与测量(IJISA-V2-N2-3)
基于计算机视觉的障碍物检测与测量(IJISA-V2-N2-3)

I.J. Intelligent Systems and Applications, 2010, 2, 17-24

Published Online December 2010 in MECS (https://www.360docs.net/doc/9b6074730.html,/)

The Obstacle Detection and Measurement Based

on Machine Vision

Xitao Zheng1, Shiming Wang2, Yongwei Zhang1

1College of IT, Shanghai Ocean University, Shanghai 201306, China

2College of Engineering Science and Technology, Shanghai Ocean University, Shanghai 201306, China xtzheng@https://www.360docs.net/doc/9b6074730.html,, smwang@https://www.360docs.net/doc/9b6074730.html,, zhangyongwei_108@https://www.360docs.net/doc/9b6074730.html,

Abstract - To develop a quick obstacle detection and measurement algorithm for the image-based autonomous vehicle (AV) or computer assisted driving system, this paper utilize the previous work of object detection to get the position of an obstacle and refocus windows on the selected target. Further calculation based on single camera will give the detailed measurement of the object, like the height, the distance to the vehicle, and possibly the width. It adopts a two camera system with different pitch angles, which can perform real-time monitoring for the front area of the vehicle with different coverage. This paper assumes that the vehicle will move at an even speed on a flat road, cameras will sample images at a given rate and the images will be analyzed simultaneously. Focus will be on the virtual window area of the image which is proved to be related to the distance to the object and speed of the vehicle. Counting of the blackened virtual sub-area can quickly find the existence of an obstacle and the obstacle area will be cut to get the interested parameter measurements for the object evaluation.

Index Terms - obstacle detection; object measurement, ALV, virtual window.

I.I NTRODUCTION

Autonomous land vehicles (ALVs) are useful for many automation applications that serve for both indoor and outdoor environments. Vision or digital image based obstacle detection for ALV navigation in outdoor road environments is a difficult and challenging task because of the great variety of object and road conditions, like irregular and unstable features on objects, moving objects, changes of illumination, and even rain. Successful ALV navigation requires the integration of the many techniques, like the environmental sensoring and learning, image processing and feature extraction, ALV location retrieving, travel path planning, vehicle wheel controlling, and so on.

Vehicle intelligent drive system is also an important part of the intelligent transport system. Based on our driving experience, more than 90% of our information is from our vision and hearing, so using computer vision technique to solve this problem is a challenging work for every researcher in this field. The lane keeping and distance measurement based on the computer vision

system.

In reference [1], a real-time distance detection method is proposed to get the real-time distance between the camera car and the obstacle in the front. The paper uses the geometrical reasoning method to obtain the depth information. The shoot angle of the camera is the key factor that affects the accuracy of output. The work uses the two lane boundaries as a constraint to get the pitch angle. The problem is that the results are for static car tests, and algorithm needs to be combined with others to get all the required parameters.

Reference [2] presents an obstacle detection method for AGV under indoor or outdoor environment. Corner or turn points are detected and tracked through an image sequencing camera and grouped into ground region using a method that is called co-planarity checking algorithm. The algorithm is initialized by 5-point planar projective and contour cue which is added to segment the ground region. The method can be applied to some ideal environment and the contour matching can be a problem in a lot of environments.

Reference [3] is the previous work of this paper and it provides the quick detection of the existence of an obstacle, this is a key feature for ALV’s moving where safety is the highest requirement. The frames can be analyzed in a fragment of milliseconds and alert can be sent if an obstacle is found. But sometime the vehicle does not need to stop to wait for a clear road, it need to further analyze the situation and the surrounding environment to make decision whether to slow down or to stop, this part of job is not included in [3].

Reference [4] proposes an effective approach to obstacle detection and avoidance for ALV navigation in outdoor road environments using computer vision and image sequence techniques. To judge whether an object newly appearing in the image is an obstacle or not, the object shape boundary is first extracted from the image. After the translation from the ALV location in the current cycle to that in the next cycle, the position of the object shape in the image is predicted, using coordinate transformation techniques based on the assumption that the height of the object is zero. The predicted object shape is then matched with the extracted shape of the object in the image of the next cycle to decide whether the object is an obstacle. A reasonable distance measure is used to compute the correlation measure between two shapes for shape matching. Finally, a safe navigation point is determined, and a turn angle is computed to

18The Obstacle Detection and Measurement Based on Machine Vision

guide the ALV toward the navigation point for obstacle avoidance. This is one camera based obstacle detection system and the moving speed is relatively slow, the work on the corner sensing can be applied in this paper.

Reference [5] tries to establish restriction functions for lane lines tracks and reconstructs the lane line to get the necessary parameters. There is binocular camera to guide the vehicle moving and another camera on the roadside to trajectory comparison. Reference [6] proposes

a function that can use a monocular camera to detect the lane edge and keep the vehicle in the lane. Reference [7] works on stereo color image with one camera and a lens. The jo

b is based on the assumption that all the brake lights are working and can be clearly pictured. It is a good paper for car-following model which is sensitive to the emerging of the brake signal, i.e., when the traffi

c is slowing down. It will lose target if one brake light is out.

There are a lot of papers that involve the lane detection and obstacle avoidance for ALV navigation in indoor or outdoor road environments using computer vision and image sequence techniques [4, 5, 8]. These methods need to search the whole pictured areas and feature matching can be unreliable.

Reference [9] addresses the problem of visual object class recognition and localization in natural images. Field experiences show how histogram-based image descriptors can be combined with a boosting classi?er to provide a state of the art object detector. Among the improvements the paper introduces a weak learner for multi-valued histogram features and shows how to overcome problems of limited training sets.

Reference [10] summarizes the past work on object detection of feature extraction and classi1cation, proposes that feature selection is an important problem in object detection and demonstrates that genetic algorithms (GAs) provide a simple, general, and powerful framework for selecting good subsets of features, leading to improved detection rates.

Reference [11] proposes an object detection approach using spatial histogram features. As spatial histograms consist of marginal distributions of an image over local patches, they can preserve texture and shape information of an object simultaneously. Their work employs Fisher criterion and mutual information to measure the discriminability and features correlation of spatial histogram features. They further train a hierarchical classi?er by combining cascade histogram matching and support vector machine. The cascade histogram matching is trained via automatically selected discriminative features. A forward sequential selection method is presented to construct uncorrelated and discriminative feature sets for support vector machine classi?cation.

Reference [12] establishes a system that can detect moving object in the 3-D space and propose a background reduction method. The camera can be repositioned to track the moving of a target.

As we know all these works are confined in a specific environment and the methods proposed above can be used in a certain circumstances.

In this paper we will utilize a virtual area-based object detection rule to detect the height of an obstacle and the combination of the two camera images will be able to confirm the movement of the object over the picture; then use the location to retrieve the object area for further analysis. So the work is correlated to reference [3] and some problems of reference [3] are resolved. The system can be very efficient because the object area can be very small compared with the whole image and the image center is nearly found when the obstacle detection is done. The further analysis will tell the driver whether to wind around or to apply brake action.

.

II.P ROPOSED O BJECT H EIGHT D ETECTION M ETHOD A.

System Architecture

Fig 1 The structure diagram of target detection

The system hardware architecture can be shown in Fig 1. The speed sensor is used to get the vehicle speed to do frame differentiation. The speed will be used to determine the height of the smaller object, like a cone or a deer. Lower camera can be viewed as collision prevention precaution and precise distance calculation camera; upper one can be used as lane detection, object pre-alerting, object height analysis, and observation camera.

From Fig 2, we can see that each camera has different coverage and they will overlap in a considerable range. In Fig 3, we can see that the same virtual area will be in different position if we merge the two pictures. This will offer ID verification and approaching information of the object.

.

Fig 2 The covering ranges of the two cameras

Fig 3 Overlapping of virtual area between two cameras B. Virtual Window(s)

As we discussed above, the virtual windows are the only areas that we will be interested, which can be seen in Fig 9-11 as the gridded area. These windows can be dynamically added or deleted depending on the detected road condition. In this model, we will assign the area from 6-11 meters in front of the vehicle as the virtual grid, which is usually used for low-speed drive. Usually multiple windows should be assigned to lower window for smaller object detection and lane detection and larger object detection windows should be assigned to the upper camera. Grid shown in Fig 4 can be an example for slow country road driving window setting.

Fig 4 Small object detection: multiple windows

Virtual windows can be placed at different positions to

serve as the information window, attention window, and action window. Here for the obstacle height detection purpose, we will use the same number and size of windows for both cameras. The accurate mapping and

quick resolution of the virtual windows on the picture re

resolved by our previous work.

Fig 5 The projected plane X-Y

Fig 6 The original rectangle on the road

In reference [1], the relationship between the actual road coordinate and the projected coordinate, which is the picture, can be shown in Fig 5 and 6, the conversion formula between the two coordinate can be seen in formula (1, 2). Detailed definitions of the subscripts can be found in [1].

()()22121341222

34111p p P P p p p

P P p

P k Y h k y k k y U G Y X U G k x k Y k y h h k Y k U G X x k k U G Y μ

?+=???

??????+=

?????

?

??=

?+?+??

??

=???+?

()()()()()()()()()()102

030400000002cos 2cos cos cos k tg H k tg k h k tg W h tg y tg y UG ??

=?α??=γ

??

=?γ??

=β??????αγ?α?=?γ?α?γ?

0(2) Using formula (1) and (2), we can convert a pre-designated area in front of the car to a virtual area in the picture and vise versa.

C. Object Height and Width Analysis

Each virtual window will be counted for the pixels occupied, if the occupied percentage is more than 50%, this window will be assigned black, otherwise white. The black and white windows will be compared against the other camera, the current frame situation will compared again with the following frame, the one with changing black spots will be the obstacle and the numbers of the windows will be the picture height, which can be converted to real height using formula (1) and (2). As seen in Fig 7, the object with a height FS is projected to the ground segment St, ST can be further projected to image size as shown in Fig 8. OE is the height of the camera.

In order to implement our detection algorithm, we find that most road image areas have similar RBG values, so the pixels with this characteristics and the area with even color range are reset to white color. This road information needs to be inputted if different road base color is present. The painting and signs on the road will not affect road color selection. In our samples the asphalt road and stone road are tested and they are a good match

to the RGB rule.

Fig 7 Projection of image length ST to obstacle

height FS

Fig 8 Image formation on camera

1

θFig 9 The view of the observer from different positions

???

?

?

??

???=?=+=?+=??=Δ21120010102121tan tan θθθθctg S h ctg S h S S h S S h S S S I I I I (3)

Applying Fig 7 to a moving condition, Fig 9 can be drew to show the height of an object in the observer’s eyes, i.e., the shadow length. In formula (3), ΔS is the distance moved in the given time interval, S 0 is the initial distance to the object, S 1 is the distance to the object at the next time interval, h 0 is the height of the cameras,θ1 and θ2 are the pitch angles at the two time intervals, h is the height of the object, S I1, S I2 are the viewed object lengths, need to be converted to the picture length using Fig 8, which is very typical for optical camera imaging system.

When an obstacle is found and so an object is located, formula (3) can be used to retrieve the interested height data. There are five variables in formula (3), not include θ1 and θ2. Where ΔS is known if the speed and the time interval is known, and the height of the camera, which is h 0, is also known. So formula will be reduced to an equation of θ1 and θ2, which are unknown because of our moving positions. Note that we have two cameras that are positioned to the same level and symmetrically to the center line, so all the parameters in formula (3) are shared for two cameras, this triangle will give another equation for θ1 and θ2 to get their solution, therefore the solution of all the parameters. Note that we can only get pixels from the picture so the image height and width will be counted from black white small picture to get their pixels, which

can be converted to actual shadow distances S I1, S I2 using Fig8. We have data redundancy so we can cross verify the effectiveness of image edge detection algorithms.

III. E XPERIMENT

The experiment is done with a vehicle moving at 25km/h, the two cameras are horizontally positioned at the front of the vehicle and closely together, so in our model calculation the two cameras are considered at the same vertical location, the upper camera (which has a bigger pitch angle, or have a farther view, not actually higher in position) serves as the information and alert window and the lower one, which has a smaller view area but better focused, serves as the verification window. Both cameras are sampling at 10frame/s, calculation of 5 sample frames pictures are presented here from Fig 10 to Fig 19. Each frame is only processed against its sister frame and the following one or two frame. Calculations are focused on virtual windows and the half processed windows are shown in Fig 20-23, the grids are added to show the virtual window arrangement.

As we already know, the virtual window will be positioned to interested area, where a possible object will appear. If the object is confirmed, which is done by the counting of blackened boxes (data shown in Tab 1) and compared with the previous frames, an object window will appear to cut the object off from the original picture and to do image processing, like the image format conversion and edge detection, pixel data can be found and compared with previous frame for stableness

evaluation. Tab 2 is the data for this experiment.

Fig 10 Frame 1 image from camera 1

Fig11 Frame 1 image from camera 2

Fig 12 Frame 2 image from camera 1

Fig13 Frame 2 image from camera 2

Fig 14 Grid image of frame1 from camera 1

Fig 15 Grid image of frame1 from camera 2

Fig 16 Grid image of frame2 from camera 1

Fig 17 Grid image of frame2 from camera 2

Fig 18 Virtual windows isolated of frame1 (camera1)

Fig 19 Virtual windows isolated of frame1 (camera2)

It can be seen from Fig 10 to Fig 17 that the upper camera (camera 1) views farther and the lower one(camera 2) views closer. As the vehicle approaches the obstacles become bigger in the pictures. Figure 14 to Figure 17 are applied to intelligent grid. Fig 18 and Fig

19 are the separated virtual windows.

Fig 20 Processed windows isolated of frame1 (camera1)

Fig 21 Processed windows isolated of frame3 (camera1)

Fig 22 Processed windows isolated of frame3 (camera2) Fig 20 to Fig 22 are the processed virtual windows to count black boxes for obstacle detection. Fig 23-25 are the height and width analysis figure where the width and height of the object are measured.

Fig 23 Middle object binary images of frame1 and frame2 (camera1)

Fig 24 Difference image of middle object of frame1

and frame2 (camera1)

Fig 25 Middle object binary images of frame1 and frame2 (camera2)

We can see the difference of the two objects in Fig 23-25, their height, width, the pixels occupied are all very different, this is another verification of the effectiveness of obstacle detection algorithm.

So by counting the vertical continuous black spots, the difference prompts an obstacle, and by counting the pixels to the top and bottom of the object, we get the height of the object. If we apply the counting the left and right edge of the object window, we get the width. Applying formula (3), we have Tab 2. We will not elaborate the process to get the pixels length of height and width, because this is the known way to process the digital image.

This quick estimation and multiple verification mechanism will avoid the miscalculation and ensure the driving safety.

Tab 1 is the result of obstacle detection, where the black box numbers are presented for each frames and for each cameras. Tab 2 is the data from each object windows, and then converted by formula (3). The test run proves the two step method is effective. We can also see that the closer image intends to give bigger width which is still within our error tolerance level.

The current method can not resolve the case when the object has similar color to the road. It also proved that when the weather condition is bad, like in the rain or in the night, the system will not be effective. This reason is that we do not apply a good background reduction algorithm, and the complicated illumination condition will considerably affect the detection and analysis. If possible, we will try to research these issues in our later project.

Table 1 Camera 1 and Camera 2 frames

900 pitch angle Fr1 Fr2 Fr3 Fr4 Fr5 camera height(m) 1.035 1.035 1.035 1.035 1.035 Black Boxes 2 4 4 5 7

77.80 pitch angle Fr1 Fr2 Fr3 Fr4 Fr5 camera height(m) 1.035 1.035 1.035 1.035 1.035 Black Boxes 2 4 4 5 8

Table 2 Object Height and Width

Fr1 pitch angle (87.40) Fr2 pitch angle(86.90) camera height(m) 1.035 1.035

Obj dist(m) 11.83 9.75

Obj Ht (pxl) 33 44

Obj Ht(m) 0.48 0.50

Obj Width(pxl) 8 11

Obj Width(m) 0.12 0.13

IV.C ONCLUSION

This method presents a method to detector locate an obstacle, retireve the window that cover the obstacle and then use digital image processing technique to get the detailed measurement of the object. The measurement project the object to a shadow on the background and

then project the shadow to the those in the camera. This is an important effort to facilitate the ALV for its decision making process. Usually the two cameras have the same pitch angle and are symmetric to the center, which can not benefit from our obstacle detection with far view and close view. The structure and algorithm need to be improved in later work for easier projection calculation and enable full object detection, not just the two dimensions in this paper. Pattern recognition and neural network need to be introduced to help the learning process. But because of the hardware architecture relative feature of the system, the learning experience and knowledge base may not be able to be shared like these geometrical algorithms. Also, most of the current works are in the day light, the algorithm needs to be tested modified on different weather conditions to ensure the robustness for actual applications.

R EFERENCES

[1]Lei Guo, Youchun Xu, Keqiang Li, Xiaomin Lian.

“Study on Real-time Distance Detection Based on

Monocular Vision Technique”. Journal of Image

and Graphics. Vol 11, No.1, Jan, 2006.(in

Chinese)

[2]Yong Zhou, Chao Chen, Qingtai Ye. “Obstacle

Detection for AGVS Based on Computer Vision”.

Machine Design and Research. Vol.21, No. 5, Oct,

2005.(in Chinese)

[3]Xitao Zheng, Yongwei Zhang. “An Obstacle

Detection Method Based on Machine Vision”. The

Chinese Symposium on Information Science and

Technology. 2010, page993-996.

[4]Kuang-Hsiung Chen, Wen-Hsiang Tsai. “Vision

based obstacle detection and avoidance for

autonomous land vehicle navigation in outdoor

roads”. Automation in Construction, Volume 10,

Issue 1, November 2000, Pages 1-25.

[5]H.-H. Nagel, F. Heimes, K. Fleischer, M. Haag,

H. Leuck, S. Noltemeier. “Quantitative

comparison between trajectory estimates obtained

from a binocular camera setup within a moving

road vehicle and from the outside by a stationary

monocular camera”. Image and Vision Computing,

Volume 18, Issue 5, April 2000, Pages 435-444. [6]Xin Zhou, Xiyue Huang, Yu Li. “Lane Keeping

and Distance Measurement Based on Monocular

Vision”. Journal of Image and Graphics. Vol.

8(A), No.5. May, 2003. (in Chinese)

[7]Joaquín Ferruz, Aníbal Ollero. “Integrated real-

time vision system for vehicle control in non-

structured environments”. Engineering

Applications of Artificial Intelligence, Volume 13,

Issue 3, 1 June 2000, pages 215-236.

[8]T Kato and Y Ninomiya. “An approach to vehicle

recognition using supervised learning”. Proc. of

Machine Vision Applications [C]. 1998, 77~80. [9]Ivan Laptev. “Improving object detection with

boosted histograms”. Image and Vision

Computing . 27 (2009), 535–544.

[10]Zehang Sun, George Bebis, Ronald Miller.

“Object detection using feature subset selection”.

Pattern Recognition. 37 (2004), 2165 – 2176. [11]Hongming Zhang, Wen Gao, Xilin Chen, Debin

Zhao. “Object detection using spatial histogram

features”. Image and Vision Computing. 24 (2006)

327–341.

[12]Kun Zhu,Tangwen Yang,Qiuqi Ruan,

Hongbo Wang,Jianda Han. “Real-Time Tracking and Measuring of Moving Objects Based OR Binocular Vision”. Robot. Vol, 31. No.

4. July,2009. (in chinese)

西电计算机视觉大作业

数字水印技术 一、引言 随着互联网广泛普及的应用,各种各样的数据资源包括文本、图片、音频、视频等放在网络服务器上供用户访问。但是这种网络资源的幵放也带了许多弊端,比如一些用户非法下载、非法拷贝、恶意篡改等,因此数字媒体内容的安全和因特网上的侵权问题成为一个急需解决的问题。数字水印作为一项很有潜力的解决手段,正是在这种情况下应运而生。 数字水印(技术是将一些代表性的标识信息,一般需要经过某种适合的变换,变换后的秘密信息(即数字水印),通过某种方式嵌入数字载体(包括文档、音频、软件等)当中,但不影响原载体的使用价值,也不容易被人的知觉系统(如视觉或听觉系统)觉察或注意到。通过这些隐藏在载体中的信息,可以达到确认内容创建者、购买者、传送隐秘信息或者判断载体是否被篡改等目的。在发生产权和内容纠纷时,通过相应的算法可以提取该早已潜入的数字水印,从而验证版权的归属和内容的真伪。 二.算法原理 2.1、灰度图像水印 2.1.1基本原理 处理灰度图像数字水印,采用了LSB(最低有效位)、DCT变换域、DWT变换域三种算法来处理数字水印。在此过程中,处理水印首先将其预处理转化为二值图像,简化算法。 (1)LSB算法原理:最低有效位算法(Least Sig nificant Bit , LSB)是很常见的空间域信息隐藏算法, 该算法就是通过改变图像像素最不重要位来达到嵌入隐秘信息的效果, 该方法隐藏的信息在人的肉眼不能发现的情况下, 其嵌入方法简单、隐藏信息量大、提取方法简单等而获得广泛应用。LSB 信息嵌入过程如下: S′=S+f S ,M 其中,S 和S′分别代表载体信息和嵌入秘密信息后的载密信息;M为待嵌入的秘密信息, 而隐写分析则是从S′中检测出M以至提取M 。 (2)DCT算法原理:DCT 变换在图像压缩中有很多应用,它是JPEG,MPEG 等数据

计算机视觉的应用

运动目标检测 目录 基于统计背景模型的运动目标检测方法 背景模型提取 运动目标检测 后处理 基于统计背景模型的运动目标检测方法 问题:(1)背景获取:需要在场景存在运动目标的情况下获得背景图像(2)背景扰动:背景中可以含有轻微扰动的对象,如树枝、树叶的摇动,扰动部分不应该被看做是前景运动目标(3)外界光照变化:一天中不同时间段光线、天气等的变化对检测结果的影响(4)背景中固定对象的移动:背景里的固定对象可能移动,如场景中的一辆车开走、一把椅子移走,对象移走后的区域在一段时间内可能被误认为是运动目标,但不应该永远被看做是前景运动目标(5)背景的更新:背景中固定对象的移动和外界光照条件的变化会使背景图像发生变化,需要及时对背景模型进行更新,以适应这种变化(6)阴影的影响:通常前景目标的阴影也被检测为运动目标的一部分,这样讲影响对运动目标的进一步处理和分析首先利用统计的方法得到背景模型,并实时地对背景模型进行更新以适应光线变化和场景本身的变化,用形态学方法和检测连通域面积进行后处理,消除噪声和背景扰动带来的影响,在HSV色度空间下检测阴影,得到准确的运动目标。 背景模型提取 前提假设在背景模型提取阶段,运动目标在场景区域中运动,不会长时间停留在某一位置视频流中某一像素点只有在前景运动目标通过时,它的亮度值才发生大的变化,在一段时间内,亮度值主要集中在很小的一个区域中,可以用这个区域内的平均值作为该点的背景值。具体实现过程:在YUV颜色空间下,Y值的变化范围为0~255,将该范围划分成若干区间[0,T][T,2T]…[Nt,255],n=255/T,对于每个像素点,统计一段时间内每个区间内亮度值的出现的次数。找出出现次数最多的那个区间,将该区间内所有值的平均值作为背景模型在该点的亮度值。这种方法不受前景运动目标的影响。 运动目标检测 检测当前图像和背景图像中对应像素点的差异,如果差值大于一定阈值,则判定该像素为前景运动目标

计算机视觉检测技术及其在机械零件检测中的应用_张文景

第33卷第5期 1999年5月 上海交通大学学报 JO U RN A L O F SHA N GHA I JIA O T O NG U N IV ERSIT Y Vol.33No.5 M ay 1999  收稿日期:1998-03-16 基金项目:上海市科技发展基金资助项目(951111052)作者简介:张文景(1971~),男,博士生. 文章编号:1006-2467(1999)05-0635-04 计算机视觉检测技术及其在机械零件检测中的应用 张文景, 张文渊, 苏键锋, 许晓鸣 (上海交通大学自动化系,上海200030) 摘 要:基于计算机视觉检测(A VI)技术在检测系统的智能化、柔性、快速性等方面较接触式检测方法具有更大的优越性,综述了其基本原理和分类以及近年来此方面的研究成果.分析了利用AVI 技术对机械零件进行检测的一般方法和系统构成,并探讨了基于CAD 的AV I 的关键环节.最后对AVI 技术的发展趋势提出了作者的观点. 关键词:计算机视觉;计算机视觉检测;计算机辅助设计;机械零件中图分类号:TP 391 文献标识码:A Automated Visual Inspection and Its Application on Inspection of Machined Parts ZH A N G Wen -j ing , ZH AN G W en -y uan , SU J ian -f eng , X U X iao -ming Dept.of Automation,Shang hai Jiaotong Univ.,Shanghai 200030,China Abstract :With development of co mputer vision ,automated visual inspection (AV I )has mor e applications in a variety of fields in industry.AVI takes more advantages in intellig ence,flex ibility and speed o f inspec-tio n sy stem than contacted inspection.T he recent achievemens r esearched by internatio nal scholars in the field of AVI w ere surveyed .Principle and classificatio n o f inspection techniques w ere intro duced .T he anal-ysis for g ener al methods and sy stem of AVI o f machined par ts w as presented.Key technolo gy o f AVI based on CAD w as also discussed.Some o pinio ns about development of AVI w ere proposed at last. Key words :com puter vision;autom ated visual inspection(AVI);com puter aided design (CAD);m achine parts 随着CIM S 的推广应用,企业在向柔性化、自动化发展的进程中,提出了对计算机辅助质量(CAQ )系统的需求.目前,计算机辅助检测计划(CAIP)系统已成为CAQ 系统的重要组成部分,它的核心问题是解决如何检测零件.在柔性制造中坐标测量机(CM M )是重要的检测手段,在制造企业中得到广泛的应用.随着柔性制造系统(FM S)的推广,人们对 检测系统的智能化、柔性、快速性等方面提出了更高的要求,以适应多品种、小批量生产的需要.计算机视觉技术得到不断发展,由于计算机视觉系统可以快速获取大量信息,且易于同设计信息及加工控制信息集成,用于工况监视、现场监控的计算机视觉系统已广泛地应用在工业、商业等领域[1,2].计算机视觉作为一种检测手段已经越来越引起人们的重视,逐步形成一种新的检测技术——计算机视觉检测(Autom ated Visual Inspectio n ,AVI )技术. 本文综合了国内外在AVI 方面的研究成果,总结了利用AVI 技术检测机械零件的一般方法和系

计算机视觉与图像理解

计算机视觉与图像理解 摘要 精确的特征跟踪是计算机视觉中的许多高层次的任务,如三维建模及运动分析奠定了基础。虽然有许多特征跟踪算法,他们大多对被跟踪的数据没有错误信息。但是,由于困难和空间局部性的问题,现有的方法会产生非常不正确的对应方式,造成剔除了基本的后处理步骤。我们提出了一个新的通用框架,使用Unscented转换,以增加任意变换特征跟踪算法,并使用高斯随机变量来表示位置的不确定性。我们运用和验证了金出武雄,卢卡斯- Tomasi 的跟踪功能框架,并将其命名为Unscented康莱特(UKLT)。UKLT能跟踪并拒绝不正确的应对措施。并证明对真假序列的方法真确性,并演示UKLT能做出正确不误的判断出物体的位置。 1.简介 在计算机视觉,对问题反映的准确性取决于于图像的准确测定。特征跟踪会随时间变化对变化的图像进行处理,并更新每个功能的变化作为图像的位置判断。重要的是所选择图像的功能,有足够的信息来跟踪,而且不遭受光圈问题的影响。[1] 在金出武雄,卢卡斯- Tomasi(康莱特)是最知名的跟踪和研究方法之一。它采用一对匹配准则刚性平移模型,它是相当于窗口强度的平方差之和最小化的基础。特征点的正确选择,可大大提高算法的性能。[3] Shi与Tomasi 将初始算法考虑仿射模型,并提出了技术监测的功能对质量进行跟踪。如果第一场比赛中的图像区域之间和当前帧残留超过阈值时,该功能将被拒绝。在随后的工作中,对模型进行了扩展且考虑了光照和反射的变化。 不幸的是,这些算法没有考虑在跟踪的不确定性,和估计的可靠性。如果我们能够考虑到这些问题,我们将能从混乱的数据中提取出更准确的数据。在没有不确定性特设技术条件下,有些研究员试图从中提取有用的数据但是结果都不能令人满意。但是理论上有声音的不确定性为特征跟踪,是可以应用于不同的功能类型的方法。 在一个闭塞,模糊,光照变化的环境中,即使是最复杂的特征跟踪算法一败涂地无法准确跟踪。这些问题导致错误的匹配,就是离群值。虽然有几种方法来减轻异常值的影响,但是其计算成本通常较高[7] [8]。[9]采用随机抽样一致性[10]的方法来消除图像序列异常值。Fusiello提出的康莱特,增加了一种自动拒绝规则功能,所谓的X84。虽然有许多离群排斥的方法,但没有一个单一的算法,尽管该算法在所有情况下都表现良好。 在本文中我们将研究范围扩大,运用高斯随机变量(GRVs)与Unscented变换(SUT 的),计算在一个非线性变换的分布传播,运用标准康莱特算法。采用随机变量来描述图像特征的位置和它们的不确定性既提高了精度又提高了鲁棒性的跟踪过程。虽然我们不知道什么是真正的分布,被测系统为我们提供了理论保证,前两个时刻的估计是正确的。另外,使用异常检测被测样品确定性使我们没有增加任何额外费用。 2.不确定度表示 我们现在引入一个新的通用框架,增强了任意特征跟踪算法,以代表和跟踪高斯随机变量(GRVs)功能的位置。然后,我们说明它可以被应用到最常用的方法,康莱特之一[1]。 GRVs是一种用于图像的特征定位概率分布函数描述的不错选择。他们有一个简单易懂的数学公式(平均向量和协方差矩阵)和紧凑的计算实施。他们也有一个确切的封闭使用的线性代数运算的代数线性变换的制定,并以此作为其参数表示的两个分布的第一时刻。Haralick [13]虽然提出了在计算机视觉中使用协方差传递,但他只考虑一阶线性化。 易用性外,还出现了一些有效的文献,它质疑从本地的图像灰度信息测量协方差是否可以代表的功能位置的不确定性[6]。

北邮模式识别课堂作业答案(参考)

第一次课堂作业 1.人在识别事物时是否可以避免错识 2.如果错识不可避免,那么你是否怀疑你所看到的、听到的、嗅 到的到底是真是的,还是虚假的 3.如果不是,那么你依靠的是什么呢用学术语言该如何表示。 4.我们是以统计学为基础分析模式识别问题,采用的是错误概率 评价分类器性能。如果不采用统计学,你是否能想到还有什么合理地分类 器性能评价指标来替代错误率 1.知觉的特性为选择性、整体性、理解性、恒常性。错觉是错误的知觉,是在特定条件下产生的对客观事物歪曲的知觉。认知是一个过程,需要大脑的参与.人的认知并不神秘,也符合一定的规律,也会产生错误 2.不是 3.辨别事物的最基本方法是计算 . 从不同事物所具有的不同属性为出发点认识事物. 一种是对事物的属性进行度量,属于定量的表示方法(向量表示法 )。另一种则是对事务所包含的成分进行分析,称为定性的描述(结构性描述方法)。 4.风险 第二次课堂作业 作为学生,你需要判断今天的课是否点名。结合该问题(或者其它你熟悉的识别问题,如”天气预报”),说明: 先验概率、后验概率和类条件概率 按照最小错误率如何决策 按照最小风险如何决策 ωi为老师点名的事件,x为判断老师点名的概率 1.先验概率: 指根据以往经验和分析得到的该老师点名的概率,即为先验概率 P(ωi ) 后验概率: 在收到某个消息之后,接收端所了解到的该消息发送的概率称为后验概率。 在上过课之后,了解到的老师点名的概率为后验概率P(ωi|x) 类条件概率:在老师点名这个事件发生的条件下,学生判断老师点名的概率p(x| ωi ) 2. 如果P(ω1|X)>P(ω2|X),则X归为ω1类别 如果P(ω1|X)≤P(ω2|X),则X归为ω2类别 3.1)计算出后验概率 已知P(ωi)和P(X|ωi),i=1,…,c,获得观测到的特征向量X 根据贝叶斯公式计算 j=1,…,x

《图像理解与计算机视觉》习题

《图像理解与计算机视觉》习题 1. 一幅图像的象素灰度级为256、大小为1024×1024的图象的数据量多少MB?假设网络的平均传输速率为1Mbit/s,需要多少秒才能传送完毕?每个像素可以用8比特表示。 解答:图像的数据量是1024×1024×8/8=1 MB,需要8 秒才能传送完毕。 2. 通过你对生活的观察,举出一个书本描述之外的图像处理应用的领域和例子。 解答: 医院CT检查,等。 I f x y z t的各个参数的具体含义,反映的图像类型有多 3. 请说明图像亮度函数(,,,,) 少? 解答: I f x y z t中,(x,y,z)是空间坐标,λ是波长,t 是时间,I 是 图像数学表达式(,,,,) 光点(x,y,z)的强度(幅度)。上式表示一幅运动(t) 的、彩色/多光谱(λ) 的、立体(x,y,z)图像。 I f x y z t表示一幅运动(t) 的、彩色/多光谱(λ) 的、立体(x,y,z)图像。对 (,,,,) 于静止图像,则与时间t 无关;对于单色图像(也称灰度图像),则波长λ为一常数;对于平面图像,则与坐标z 无关,故f(x,y)表示平面上的静止灰度图像,它是一般图像I f x y z t的一个特例。 (,,,,) 4. 色彩具有哪几个基本属性,表述这些属性的含义。 解答: 色相(H):色相是与颜色主波长有关的颜色物理和心理特性,它们就是所有的色相,有时色相也称为色调。 饱和度(S):饱和度指颜色的强度或纯度,表示色相中灰色成分所占的比例,用0%-100%(纯色)来表示。 亮度(B):亮度是颜色的相对明暗程度,通常用0%(黑)-100%(白)来度量。 5. 请解释马赫带效应。 解答: 所谓“马赫带效应(Mach band effect)”是指视觉的主观感受在亮度有变化的地方

计算机视觉检测技术及其在机械零件检测中的应用

泰山职业技术学院2006级毕业论文 目录 摘要 (1) 1自动检测的原理和方法 (2) 1.1检测和检验的不同意义 (2) 1.1.1质量控制中的检测 (2) 1.1.2质量控制中的检验 (3) 1.2统计质量控制基础 (3) 1.2.1统计质量控制的推断 (3) 1.2.2制造中的偏差 (4) 2计算机视觉检测技 (4) 2.1 AVI技术 (4) 2.1.1视觉检测分类 (5) 2.1.2视觉检测方法 (5) 2.1.3视觉检验方法 (5) 2.2机械零件的AVI (5) 2.3基于CAD的AVI (6) 2.4 AVI技术展望 (7) 参考文献 (8) 致谢 (9)

计算机视觉检测技术及其在机械零件检测中的应用 张文景,张文渊,苏键锋,许晓鸣 摘要:基于计算机视觉检测(AVI)技术在检测系统的智能化、柔性、快速性等方面较接触式检测方法具有更大的优越性,综述了其基本原理和分类以及近年来此方面的研究成果.分析了利用AVI技术对机械零件进行检测的一般方法和系统构成,并探讨了基于CAD的AVI的关键环节.最后对AVI技术的发展趋势提出了作者的观点. 关键词:计算机视觉;计算机视觉检测;计算机辅助设计;机械零件 中图分类号:TP 391 文献标识码:A Automated Visual Inspection and Its Application on Inspection of Machined Parts ZHANG Wen-jing,ZHANG Wen-yuan,SU Jian-feng, XU Xiao-ming Dept. of Automation, Shanghai Jiaotong Univ., Shanghai 200030, China Abstract:With development of computer vision,automated visual inspection (AVI) has more applications in a variety of fields in industry.AVI takes more advantages in intelligence,flexibility and speed of inspection system than contacted inspection.The recent achievemens researched by international scholars in the field of AVI were surveyed.Principle and classification of inspection techniques were introduced.The analysis for general methods and system of AVI of machined parts was presented.Key technology of AVI based on CAD was also discussed.Some opinions about development of AVI were proposed at last. Key words:computer vision;automated visual inspection(AVI);computer aided design(CAD);machine parts

计算机视觉期末复习

一、 1.什么是计算机视觉?理解计算机视觉问题的产生原理。 研究用计算机来模拟生物视觉功能的技术学科。具体来说,就是让计算机具有对周围世界的空间物体进行 传感、抽象、分析判断、决策的能力,从而达到识别、理解的目的。 2.直方图的均衡化 处理的“中心思想”是把原始图像的灰度直方图从比较集中的某个灰度区间变成在全部灰度范围内的均匀分布。直方 图均衡化就是对图像进行非线性拉伸,重新分配图像像素值,使一定灰度范围内的像素数量大致相同。直方图均衡化就是 把给定图像的直方图分布改变成“均匀”分布直方图分布。 是将原图像通过某种变换,得到一幅灰度直方图更为均匀分布的新图像的方法。设图像均衡化处理后,图像的直方图 是平直的,即各灰度级具有相同的出现频数,那么由于灰度级具有均匀的概率分布,图像看起来就更清晰了。 二、 1.常见的几何变换:平移T x为点(x,y)在x方向要平移的量。 旋转 变尺度:x轴变大a倍,y轴变大b倍。 2.卷积掩膜技术:(,) (,)(,)(,) m n f i j h i m j n g m n =-- ∑∑ 对应相乘再相加掩膜的有效应用——去噪问题 3. 均值滤波器(低通):抑制噪声 主要用于抑制噪声,对每一个目标像素值用其局部邻域内所有像素值的加权均值置换。con命令高斯滤波器:一个朴素的道理,距离目标像素越近的点,往往相关性越大,越远则越不相干。所以,高斯 滤波器根据高斯函数选择邻域内各像素的权值 medfilt1 。 区别方法是:高通滤波器模板的和为0,低通滤波器模板的和为1 常用的非线性滤波器:中值滤波;双边滤波;非局部滤波 4.边缘检测算子:通过一组定义好的函数,定位图像中局部变换剧烈的部分(寻找图像边缘)。主要方法有:Robert 交叉梯度,Sobel梯度,拉普拉斯算子,高提升滤波,高斯-拉普拉斯变换(都是高通滤波器) 1100 cos sin0 [1][1]sin cos0 001 x y x y θθ θθ - ?? ? = ? ? ?? 1100 00 [1][1]00 00 a x y x y b ab ?? ? = ? ? ?? (,) 1 [,][,] k l N h i j f k l M∈ =∑ ? ? ? ? ? ? ? = 1 1 1 ]1 [ ]1 [ 1 1 y x T T y x y x

计算机视觉复习题

《计算机视觉》复习题 1、利用MFC及OpenCV 库函数编写对话框程序,添加按钮实现图像读入、图像阈值分割、边缘提取等功能(至少实现三个以上功能)。(考前做好并用A4纸打印,考试当天带来) 为旋转不变算子,即当图像()v,u f旋转后,计算值在对应点保持不变。 2、证明Laplace算子 理论 3、计算机视觉研究的目的是什么?它和图像处理及计算机图形学的区别和联系是什么? 从20世纪50年代末开始,计算机开始被作为实现人类智能和人类感知的工具,借助计算机人类第一次可以象借助机械实现对体力的延伸一样实现对脑力和感知能力的延伸。对人类视觉感知能力的计算机模拟导致了计算机视觉的产生。计算机视觉就是用各种成像系统代替视觉器官作为输入敏感手段,由计算机来替代大脑完成处理和解释。计算机视觉使用的理论方法主要是基于几何、概率和运动学计算与三维重构的视觉计算理论。 具体地讲,计算机视觉要达到的基本目的有以下几个: 根据一幅或者多幅二维图像计算出观测点到目标物体的距离; 根据一幅或者多幅二维图像计算出观测点到目标物体的运动参数; 根据一幅或者多幅二维图像计算出观测点到目标物体的表面物理特征; 根据多幅二维投影图像恢复出更大空间区域的投影图像。 简单来说,计算机视觉要达到的最终目的是实现利用计算机对三维景物世界的理解,即实现人的视觉系统的某些功能。从本质上来讲,计算机视觉研究就是利用二维投影图像来重构三维物体的可视部分。 计算机视觉和图像处理及计算机图形学的区别和联系: 区别: 图像处理(image processing)通常是把一幅图像变换为另外一幅图像。它输入的是图像,输出的也是图像。Photoshop中对一幅图像应用滤镜就是典型的一种图像处理。常见操作有模糊、灰度化、增强对比度。 计算机图形学(Computer Graphics)是借助计算机来研究图形表达、处理图像、显示生成的学科。,主要通过几何基元,如线、圆和自由曲面等,来生成图像,属于图像综合。输入的是对虚拟场景的描述,通常为多边形数组,输出的是图像,即二维像素数组。

计算机视觉简介

人们常说:眼睛是心灵的窗户,通过眼睛人们可以轻易地交流情感,眼睛也是与外界交流的窗口,这些都是通过“看”来完成的。 人们可以很容易“看到”一幅画,但这一“简单”过程并不如此简单,大致上它可以分为以下几个阶段:首先是通过眼睛将图成像在视网膜上;其次大脑对图像进行理解;最后根据处理的结果做出反应。用比较专业一点的语言来描述,该过程包括了识别、描述与理解三个层次;这其中还隐含了边缘检测(各物体的轮廓等)、图像的分割(各物体区域的划分)等阶段。以上实际上概述了视觉系统的三个层次,即低层阶段:基于图像特征提取及分割阶段;中层阶段:基于物体的几何模型与图像特性表达阶段;高层阶段:基于景物知识的描述、识别与理解阶段,这是根据先验知识介入的程度划分的,且实现起来也越来越困难。 毫无疑问,如何人工实现这一过程是极具挑战性和应用前景的一项工作,计算机视觉也因此而应运而生。计算机视觉是研究用计算机和成像设备来模拟人和生物视觉系统功能的技术学科,其目标是从图像或图像序列中获取对外部世界的认知和理解,即利用二维图像恢复三维环境中物体的几何信息,比如形状、位置、姿态、运动等,并能描述、识别与理解。 计算机视觉的基础是各种成像设备,例如CCD(Charge Coupled Device )摄像机(数码相机属于此类型)、红外摄像机、医学上常用的核磁共振成像、X射线成像等,这些设备不仅可以成像,还可以获取比人眼更丰富的图像,人们可以形象地把摄像机看成计算机视觉的视网膜部分。可以说从人类拍摄出第一幅图像开始,就为计算机视觉的诞生奠定了基础。 而计算机视觉的核心是数字电子计算机,其发展可谓突飞猛进,在计算和存储能力上,人脑已经无法与之相比,人们的目标就是利用计算机非凡的计算处理能力来代替人脑实现对图像的理解,而计算机日新月异的发展也使得这一愿望越来越成为可能。 用于指导“计算机”这个大脑运作的核心是计算机视觉的理论方法,计算机视觉使用的理论方法主要基于几何、概率和运动学计算与三维重构的视觉计算理论,它的基础包括射影几何学、刚体运动力学、概率论与随机过程、图像处理、人工智能等理论。在20世纪70年代,视觉研究大多采用模式识别的方法;80年代,开始采用空间几何的方法以及物理知识进行视觉研究;90年代以后,随着智能机器人视觉研究的发展,引入了许多新的理论与技术如主动视觉理论、不变量理论、融合技术等,并应用于许多计算机视觉系统中。 研究计算机视觉,不得不提的是英国已故科学家戴维·马尔(David Marr),他在计算机视觉发展史上可谓写下了浓重的一笔。在20世纪70年代末,他提出了第一个

北邮模式识别课堂作业答案(参考)

第一次课堂作业 ? 1.人在识别事物时是否可以避免错识? ? 2.如果错识不可避免,那么你是否怀疑你所看到的、听到的、嗅到的到底 是真是的,还是虚假的? ? 3.如果不是,那么你依靠的是什么呢?用学术语言该如何表示。 ? 4.我们是以统计学为基础分析模式识别问题,采用的是错误概率评价分类 器性能。如果不采用统计学,你是否能想到还有什么合理地分类器性能评价指标来替代错误率? 1.知觉的特性为选择性、整体性、理解性、恒常性。错觉是错误的知觉,是在特定条件下产生的对客观事物歪曲的知觉。认知是一个过程,需要大脑的参与.人的认知并不神秘,也符合一定的规律,也会产生错误 2.不是 3.辨别事物的最基本方法是计算.从不同事物所具有的不同属性为出发点认识事物.一种是对事物的属性进行度量,属于定量的表示方法(向量表示法)。另一种则是对事务所包含的成分进行分析,称为定性的描述(结构性描述方法)。 4.风险 第二次课堂作业 ?作为学生,你需要判断今天的课是否点名。结合该问题(或者其它你熟悉的识别问题, 如”天气预报”),说明: ?先验概率、后验概率和类条件概率? ?按照最小错误率如何决策? ?按照最小风险如何决策? ωi为老师点名的事件,x为判断老师点名的概率 1.先验概率:指根据以往经验和分析得到的该老师点名的概率,即为先验概率P(ωi ) 后验概率:在收到某个消息之后,接收端所了解到的该消息发送的概率称为后验概率。 在上过课之后,了解到的老师点名的概率为后验概率P(ωi|x) 类条件概率:在老师点名这个事件发生的条件下,学生判断老师点名的概率p(x| ωi ) 2. 如果P(ω1|X)>P(ω2|X),则X归为ω1类别 如果P(ω1|X)≤P(ω2|X),则X归为ω2类别 3.1)计算出后验概率 已知P(ωi)和P(X|ωi),i=1,…,c,获得观测到的特征向量X 根据贝叶斯公式计算 j=1,…,x 2)计算条件风险

计算机视觉各种方法

第33卷第1期自动化学报Vol.33,No.1 2007年1月ACTA AUTOMATICA SINICA January,2007 车辆辅助驾驶系统中基于计算机视觉的 行人检测研究综述 贾慧星1章毓晋1 摘要基于计算机视觉的行人检测由于其在车辆辅助驾驶系统中的重要应用价值成为当前计算机视觉和智能车辆领域最为活跃的研究课题之一.其核心是利用安装在运动车辆上的摄像机检测行人,从而估计出潜在的危险以便采取策略保护行人.本文在对这一问题存在的困难进行分析的基础上,对相关文献进行综述.基于视觉的行人检测系统一般包括两个模块:感兴趣区分割和目标识别,本文介绍了这两个模块所采用的一些典型方法,分析了每种方法的原理和优缺点.最后对性能评估和未来的研究方向等一系列关键问题给予了介绍. 关键词行人检测,车辆辅助驾驶系统,感兴趣区分割,目标识别 中图分类号TP391.41 A Survey of Computer Vision Based Pedestrian Detection for Driver Assistance Systems JIA Hui-Xing ZHANG Yu-Jin Abstract Computer vision based pedestrian detection has become one of the hottest topics in the domain of computer vision and intelligent vehicle because of its potential applications in driver assistance systems.It aims at detecting pedestrians appearing ahead of the vehicle using a vehicle-mounted camera,so as to assess the danger and take actions to protect pedestrians in case of danger.In this paper,we give detailed analysis of the di?culties lying in the problem and review most of the literature.A typical pedestrian detection system includes two modules:regions of interest(ROIs) segmentation and object recognition.This paper introduces the principle of typical methods of the two modules and analyzes their respective pros and cons.Finally,we give detailed analysis of performance evaluation and propose some research directions. Key words Pedestrian detection,driver assistance system,ROIs segmentation,object recognition 1引言 车辆辅助驾驶系统中基于计算机视觉的行人检测是指利用安装在运动车辆上的摄像机获取车辆前面的视频信息,然后从视频序列中检测出行人的位置.由于它在行人安全方面的巨大应用前景,成为智能车辆、计算机视觉和模式识别领域的前沿研究课题.欧盟从2000年到2005年连续资助了PROTECTOR[1]和SAVE-U[2]项目,开发了两个以计算机视觉为核心的行人检测系统;意大利Parma[3]大学开发的ARGO智能车也包括一个行人检测模块;以色列的MobilEye[4]公司开发了芯 收稿日期2006-3-14收修改稿日期2006-6-17 Received March14,2006;in revised form June17,2006 国家自然科学基金(60573148),教育部高等学校博士学科点专项科研基金(20060003102)资助 Supported by National Natural Science Foundation of P.R.China(60573148),Specialized Research Fund for the Doc-toral Program of Higher Education(20060003102) 1.清华大学电子工程系北京100084 1.Department of Electronic Engineering,Tsinghua University, Beijing100084 DOI:10.1360/aas-007-0084片级的行人检测系统;日本本田汽车公司[5]开发了基于红外摄像机的行人检测系统;国外的大学如CMU[6]、MIT[7,8]和国内的西安交通大学[9]、清华大学[10]也在该领域做了许多研究工作. 车辆辅助驾驶系统中基于计算机视觉的行人检测属于计算机视觉中人体运动分析的研究范畴,其主要任务是在运动摄像机下快速准确地检测行人.本文主要针对这一特定领域对相关的文献进行综述,重点分析常用方法的原理和优缺点,以期对相关的科技人员起到指导作用.对监控系统和体育运动分析领域中人体检测感兴趣的读者可以参考综述文献[11~14]. 行人检测除了具有一般人体检测具有的服饰变化、姿态变化等难点外,由于其特定的应用领域还具有以下难点:摄像机是运动的,这样广泛应用于智能监控领域中检测动态目标的方法便不能直接使用;行人检测面临的是一个开放的环境,要考虑不同的路况、天气和光线变化,对算法的鲁棒性提出了很高的要求;实时性是系统必须满足的要求,这 c 2007by Acta Automatica Sinica.All rights reserved.

计算机视觉

计算机视觉 计算机视觉是一门研究如何使机器“看”的科学,更进一步的说,就是是指用摄影机和电脑代替人眼对目标进行识别、跟踪和测量等机器视觉,并进一步做图形处理,用电脑处理成为更适合人眼观察或传送给仪器检测的图像。作为一个科学学科,计算机视觉研究相关的理论和技术,试图建立能够从图像或者多维数据中获取‘信息’的人工智能系统。这里所指的信息指Shannon定义的,可以用来帮助做一个“决定”的信息。因为感知可以看作是从感官信号中提取信息,所以计算机视觉也可以看作是研究如何使人工系统从图像或多维数据中“感知”的科学。 目录 1定义 2解析 3原理 4相关 5现状 6用途 7异同 8问题

9系统 10要件 11会议 12期刊 1定义 计算机视觉是使用计算机及相关设备对生物视觉的一种模拟。它的主要任务就是通过对采集的图片或视频进行处理以获得相应场景的三维信息,就像人类和许多其他类生物每天所做的那样。 计算机视觉是一门关于如何运用照相机和计算机来获取我们所需的,被拍摄对象的数据与信息的学问。形象地说,就是给计算机安装上眼睛(照相机)和大脑(算法),让计算机能够感知环境。我们中国人的成语"眼见为实"和西方人常说的"One picture is worth ten thousand words"表达了视觉对人类的重要性。不难想象,具有视觉的机器的应用前景能有多么地宽广。 计算机视觉既是工程领域,也是科学领域中的一个富有挑战性重要研究领域。计算机视觉是一门综合性的学科,它已经吸引了来自各个学科的研究者参加到对它

的研究之中。其中包括计算机科学和工程、信号处理、物理学、应用数学和统计学,神经生理学和认知科学等。 2解析 视觉是各个应用领域,如制造业、检验、文档分析、医疗诊断,和军事等领域中各种智能/自主系统中不可分割的一部分。由于它的重要性,一些先进国家,例如美国把对计算机视觉的 计算机视觉与其他领域的关系 研究列为对经济和科学有广泛影响的科学和工程中的重大基本问题,即所谓的重大挑战(grand challenge)。计算机视觉的挑战是要为计算机和机器人开发具有与人类水平相当的视觉能力。机器视觉需要图象信号,纹理和颜色建模,几何处理和推理,以及物体建模。一个有能力的视觉系统应该把所有这些处理都紧密地集成在一起。作为一门学科,计算机视觉开始于60年代初,但在计算机视觉的基本研究中的许多重要进展是在80年代取得的。计算机视觉与人类视觉密切相关,对人类视觉有一个正确的认识将对计算机视觉的研究非常有益。为此我们将先介绍人类视觉。 3原理 计算机视觉就是用各种成象系统代替视觉器官作为输入敏感手段,由计算机来代替大脑完成处理和解释。计算机视觉的最终研究目标就是使计算机能象人那样通过视觉观察和理解世界,具有自主适应环境的能力。要经过长期的努力才能达到的目标。因此,在实现最终目标以前,人们努力的中期目标是建立一种视觉系统,这个系统能依据视觉敏感和反馈的某种程度的智能完成一定的任务。例如,计算机视觉的一个重要应用领域就是自主车辆的视觉导航,还没有条件实现象人那样能识别和理解任何环境,完成自主导航的系统。因此,人们努力的研究目标是实现在高速公路上具有道路跟踪能力,可避免与前方车辆碰撞的视觉辅助驾驶系统。这里要指出的一点是在计算机视觉系统中计算机起代替人脑的作用,但并不意味

《计算机图形学基础》第一次作业参考答案

《计算机图形学基础》第一次作业参考答案 一、名词解释 1、计算机图形学:是使用计算机通过算法和程序在显示设备上构造出图形来,是真实物体或虚构物体的图形综合技术。 2、图像处理:是景物或图象的分析技术,它所研究的是计算机图形学的逆过程。包括图象增强、模式识别、景物分析、计算机视觉等,并研究如何从图象中提取二维或三维物体的模型。 3、逐点比较法:这是绘图仪经常采用的一种方法,就是在绘图过程中,绘图笔每画一笔,就与规定图形进行比较,然后决定下一步的走向,用步步逼近的方法画出规定的图形。 4、造型技术:要在计算机中构造三维物体的一幅图像,首先必须在计算机中构造出该物体的模型,这一模型是由一批几何数据及数据之间的拓朴关系来表示,这就是造型技术。 二、简答 1、计算机图形系统的组成包括哪些? 答:计算机图形系统由硬件和软件两部分组成,硬件包括:主计算机、图形显示器以及I/O交互工具和存储设备;软件包括操作系统、高级语言、图形软件和应用软件。 现代计算机图形系统与一般计算机系统最主要的差别是具有图形的输入、输出设备以及必要的交互工具,在速度和存储容量上具有较高的要求。另外,人也是这个系统的组成部分。 2、简述图形软件的组成。 答:图形软件系统应该具有良好的结构,要有合理的层次结构的模块结构,以便于设计、维护和调试。 1、零级图形软件:是最底层的软件,主要解决图形设备与主机的通讯、接口等问题,又称设备驱动程序,是一些最基本的输入、输出子程序,要求程序质量高,它是面向系统的,而不是面向用户的。 2、一级图形软件:又称基本子程序,包括生成基本图形元素,对设备进行管理的各程序模块,既面向系统又面向用户。 3、二级图形软件:也称功能子程序,是在一级图形软件基础上编制的,其主要任务是建立图形数据结构,定义、修改和输出图形;以及建立各图形设备之间的联系,要具有较强的交互功能,它是面向用户的。(以上三级通常称为支撑软件) 4、三级图形软件:是为解决某种应用问题的图形软件,是整个应用软件的一部分,通常由用户编写或与设计者一起编写。 三、算法实现 写出Bresenham方法生成直线的C语言算法。 答案: int bres_line(int x1,int y1,int x2,int y2,int color) { int oldcolor,itag; int dx,dy,tx,ty,inc1,inc2,d,curx,cury; oldcolor=setcolor(color);

《计算机视觉与图象处理》.

视觉检测技术基础》课程教学大纲 一、课程基本信息 1、课程代码:MI420 2 、课程名称(中/ 英文):视觉检测技术基础/ Foundation of visual measurement technique 3、学时/ 学分:27/1.5 4、先修课程:高等数学,大学物理 5、面向对象:电子信息类专业本科生 6、开课院(系)、教研室:电子信息与电气工程学院仪器系自动检测技术研究所 7、教材、教学参考书:自编讲义 《机器视觉》,贾云得著,科学出版社,2000 《计算机视 觉》,马颂德著,科学出版社,1997 《图像工程》,章毓晋 著,清华大学出版社,2002 二、本课程的性质和任务 《视觉检测基础》是电子信息学院仪器系四年级本科生的选修课,通过本课程的学习,使学生初步了解视觉检测系统的构成及基本原理,每个组成部分如何选择设计,掌握相应的图像处理方法,增加学生的专业知识。通过上机实践提高学生的实际编程能力,增强感性认识,为以后科研、工作中遇到的相关问题提供一个解决的思想,并能实际运用。 三、本课程教学内容和基本要求

1. 基本要求 《视觉检测基础》作为本科生的选修课,应当主要立足于对学生知识的普及,主要讲述计算机视觉系统的组成、设计、处理等方面的基本知识,以课堂讲述为主,讲述中应结合日常生活实际,提高学生的学习兴趣,让学生掌握基本的处理过程及算法,并辅以实验手段进一步增强学生对视觉检测技术的了解,增加感性认识, 2. 教学内容 (1) 课堂教学部分 第一讲计算机视觉概述 一、什么是计算机视觉 二、计算机视觉的应用 三、计算机视觉的研究内容 1 、主要研究内容 2 、与其它学科的关系 第二讲成像原理与系统 一、成像几何基础 1、透视投影 2、正交投影 二、输入设备 1 、镜头 2 、摄像机

基于计算机视觉的障碍物检测与测量(IJISA-V2-N2-3)

I.J. Intelligent Systems and Applications, 2010, 2, 17-24 Published Online December 2010 in MECS (https://www.360docs.net/doc/9b6074730.html,/) The Obstacle Detection and Measurement Based on Machine Vision Xitao Zheng1, Shiming Wang2, Yongwei Zhang1 1College of IT, Shanghai Ocean University, Shanghai 201306, China 2College of Engineering Science and Technology, Shanghai Ocean University, Shanghai 201306, China xtzheng@https://www.360docs.net/doc/9b6074730.html,, smwang@https://www.360docs.net/doc/9b6074730.html,, zhangyongwei_108@https://www.360docs.net/doc/9b6074730.html, Abstract - To develop a quick obstacle detection and measurement algorithm for the image-based autonomous vehicle (AV) or computer assisted driving system, this paper utilize the previous work of object detection to get the position of an obstacle and refocus windows on the selected target. Further calculation based on single camera will give the detailed measurement of the object, like the height, the distance to the vehicle, and possibly the width. It adopts a two camera system with different pitch angles, which can perform real-time monitoring for the front area of the vehicle with different coverage. This paper assumes that the vehicle will move at an even speed on a flat road, cameras will sample images at a given rate and the images will be analyzed simultaneously. Focus will be on the virtual window area of the image which is proved to be related to the distance to the object and speed of the vehicle. Counting of the blackened virtual sub-area can quickly find the existence of an obstacle and the obstacle area will be cut to get the interested parameter measurements for the object evaluation. Index Terms - obstacle detection; object measurement, ALV, virtual window. I.I NTRODUCTION Autonomous land vehicles (ALVs) are useful for many automation applications that serve for both indoor and outdoor environments. Vision or digital image based obstacle detection for ALV navigation in outdoor road environments is a difficult and challenging task because of the great variety of object and road conditions, like irregular and unstable features on objects, moving objects, changes of illumination, and even rain. Successful ALV navigation requires the integration of the many techniques, like the environmental sensoring and learning, image processing and feature extraction, ALV location retrieving, travel path planning, vehicle wheel controlling, and so on. Vehicle intelligent drive system is also an important part of the intelligent transport system. Based on our driving experience, more than 90% of our information is from our vision and hearing, so using computer vision technique to solve this problem is a challenging work for every researcher in this field. The lane keeping and distance measurement based on the computer vision system. In reference [1], a real-time distance detection method is proposed to get the real-time distance between the camera car and the obstacle in the front. The paper uses the geometrical reasoning method to obtain the depth information. The shoot angle of the camera is the key factor that affects the accuracy of output. The work uses the two lane boundaries as a constraint to get the pitch angle. The problem is that the results are for static car tests, and algorithm needs to be combined with others to get all the required parameters. Reference [2] presents an obstacle detection method for AGV under indoor or outdoor environment. Corner or turn points are detected and tracked through an image sequencing camera and grouped into ground region using a method that is called co-planarity checking algorithm. The algorithm is initialized by 5-point planar projective and contour cue which is added to segment the ground region. The method can be applied to some ideal environment and the contour matching can be a problem in a lot of environments. Reference [3] is the previous work of this paper and it provides the quick detection of the existence of an obstacle, this is a key feature for ALV’s moving where safety is the highest requirement. The frames can be analyzed in a fragment of milliseconds and alert can be sent if an obstacle is found. But sometime the vehicle does not need to stop to wait for a clear road, it need to further analyze the situation and the surrounding environment to make decision whether to slow down or to stop, this part of job is not included in [3]. Reference [4] proposes an effective approach to obstacle detection and avoidance for ALV navigation in outdoor road environments using computer vision and image sequence techniques. To judge whether an object newly appearing in the image is an obstacle or not, the object shape boundary is first extracted from the image. After the translation from the ALV location in the current cycle to that in the next cycle, the position of the object shape in the image is predicted, using coordinate transformation techniques based on the assumption that the height of the object is zero. The predicted object shape is then matched with the extracted shape of the object in the image of the next cycle to decide whether the object is an obstacle. A reasonable distance measure is used to compute the correlation measure between two shapes for shape matching. Finally, a safe navigation point is determined, and a turn angle is computed to

相关文档
最新文档