图像处理外文翻译.doc_第1页
图像处理外文翻译.doc_第2页
图像处理外文翻译.doc_第3页
图像处理外文翻译.doc_第4页
图像处理外文翻译.doc_第5页
已阅读5页,还剩6页未读 继续免费阅读

下载本文档

版权说明:本文档由用户提供并上传,收益归属内容提供方,若内容存在侵权,请进行举报或认领

文档简介

附录a3 image enhancement in the spatial domainthe principal objective of enhancement is to process an image so that the result is more suitable than the original image for a specific application. the word specific is important, because it establishes at the outset than the techniques discussed in this chapter are very much problem oriented. thus, for example, a method that is quite useful for enhancing x-ray images may not necessarily be the best approach for enhancing pictures of mars transmitted by a space probe. regardless of the method used .however, image enhancement is one of the most interesting and visually appealing areas of image processing.image enhancement approaches fall into two broad categories: spatial domain methods and frequency domain methods. the term spatial domain refers to the image plane itself, and approaches in this category are based on direct manipulation of pixels in an image. fourier transform of an image. spatial methods are covered in this chapter, and frequency domain enhancement is discussed in chapter 4.enhancement techniques based on various combinations of methods from these two categories are not unusual. we note also that many of the fundamental techniques introduced in this chapter in the context of enhancement are used in subsequent chapters for a variety of other image processing applications.there is no general theory of image enhancement. when an image is processed for visual interpretation, the viewer is the ultimate judge of how well a particular method works. visual evaluation of image quality is a highly is highly subjective process, thus making the definition of a “good image” an elusive standard by which to compare algorithm performance. when the problem is one of processing images for machine perception, the evaluation task is somewhat easier. for example, in dealing with a character recognition application, and leaving aside other issues such as computational requirements, the best image processing method would be the one yielding the best machine recognition results. however, even in situations when a clear-cut criterion of performance can be imposed on the problem, a certain amount of trial and error usually is required before a particular image enhancement approach is selected.3.1 backgroundas indicated previously, the term spatial domain refers to the aggregate of pixels composing an image. spatial domain methods are procedures that operate directly on these pixels. spatial domain processes will be denotes by the expression (3.1-1)where f(x, y) is the input image, g(x, y) is the processed image, and t is an operator on f, defined over some neighborhood of (x, y). in addition, t can operate on a set of input images, such as performing the pixel-by-pixel sum of k images for noise reduction, as discussed in section 3.4.2.the principal approach in defining a neighborhood about a point (x, y) is to use a square or rectangular subimage area centered at (x, y).the center of the subimage is moved from pixel to starting, say, at the top left corner. the operator t is applied at each location (x, y) to yield the output, g, at that location. the process utilizes only the pixels in the area of the image spanned by the neighborhood. although other neighborhood shapes, such as approximations to a circle, sometimes are used, square and rectangular arrays are by far the most predominant because of their ease of implementation.the simplest from of t is when the neighborhood is of size 11 (that is, a single pixel). in this case, g depends only on the value of f at (x, y), and t becomes a gray-level (also called an intensity or mapping) transformation function of the form (3.1-2)where, for simplicity in notation, r and s are variables denoting, respectively, the grey level of f(x, y) and g(x, y)at any point (x, y).some fairly simple, yet powerful, processing approaches can be formulates with gray-level transformations. because enhancement at any point in an image depends only on the grey level at that point, techniques in this category often are referred to as point processing.larger neighborhoods allow considerably more flexibility. the general approach is to use a function of the values of f in a predefined neighborhood of (x, y) to determine the value of g at (x, y). one of the principal approaches in this formulation is based on the use of so-called masks (also referred to as filters, kernels, templates, or windows). basically, a mask is a small (say, 33) 2-darray, in which the values of the mask coefficients determine the nature of the type of approach often are referred to as mask processing or filtering. these concepts are discussed in section 3.5.3.2 some basic gray level transformationswe begin the study of image enhancement techniques by discussing gray-level transformation functions. these are among the simplest of all image enhancement techniques. the values of pixels, before and after processing, will be denoted by r and s, respectively. as indicated in the previous section, these values are related by an expression of the from s = t(r), where t is a transformation that maps a pixel value r into a pixel value s. since we are dealing with digital quantities, values of the transformation function typically are stored in a one-dimensional array and the mappings from r to s are implemented via table lookups. for an 8-bit environment, a lookup table containing the values of t will have 256 entries.as an introduction to gray-level transformations, which shows three basic types of functions used frequently for image enhancement: linear (negative and identity transformations), logarithmic (log and inverse-log transformations), and power-law (nth power and nth root transformations). the identity function is the trivial case in which out put intensities are identical to input intensities. it is included in the graph only for completeness.3.2.1 image negativesthe negative of an image with gray levels in the range 0, l-1is obtained by using the negative transformation show shown, which is given by the expression (3.2-1) reversing the intensity levels of an image in this manner produces the equivalent of a photographic negative. this type of processing is particularly suited for enhancing white or grey detail embedded in dark regions of an image, especially when the black areas are dominant in size. 3.2.2 log transformationsthe general from of the log transformation is (3.2-2) where c is a constant, and it is assumed that r 0 .the shape of the log curve transformation maps a narrow range of low gray-level values in the input image into a wider range of output levels. the opposite is true of higher values of input levels. we would use a transformation of this type to expand the values of dark pixels in an image while compressing the higher-level values. the opposite is true of the inverse log transformation.any curve having the general shape of the log functions would accomplish this spreading/compressing of gray levels in an image. in fact, the power-law transformations discussed in the next section are much more versatile for this purpose than the log transformation. however, the log function has the important characteristic that it compresses the dynamic range of image characteristics of spectra. it is not unusual to encounter spectrum values that range from 0 to 106 or higher. while processing numbers such as these presents no problems for a computer, image display systems generally will not be able to reproduce faithfully such a wide range of intensity values .the net effect is that a significant degree of detail will be lost in the display of a typical fourier spectrum.3.2.3 power-law transformationspower-law transformations have the basic from (3.2-3) where c and y are positive constants .sometimes eq. (3.2-3) is written as to account for an offset (that is, a measurable output when the input is zero). however, offsets typically are an issue of display calibration and as a result they are normally ignored in eq. (3.2-3). plots of s versus r for various values of y are shown in fig.3.6. as in the case of the log transformation, power-law curves with fractional values of y map a narrow range of dark input values into a wider range of output values, with the opposite being true for higher values of input levels. unlike the log function, however, we notice here a family of possible transformation curves obtained simply by varying y. as expected, we see in fig.3.6 that curves generated with values of y1 have exactly the opposite effect as those generated with values of y1. finally, we note that eq.(3.2-3) reduces to the identity transformation when c = y = 1.a variety of devices used for image capture, printing, and display respond according to as gammahence our use of this symbol in eq.(3.2-3).the process used to correct this power-law response phenomena is called gamma correction.gamma correction is important if displaying an image accurately on a computer screen is of concern. images that are not corrected properly can look either bleached out, or, what is more likely, too dark. trying to reproduce colors accurately also requires some knowledge of gamma correction because varying the value of gamma correcting changes not only the brightness, but also the ratios of red to green to blue. gamma correction has become increasingly important in the past few years, as use of digital images for commercial purposes over the internet has increased. it is not internet has increased. it is not unusual that images created for a popular web site will be viewed by millions of people, the majority of whom will have different monitors and/or monitor settings. some computer systems even have partial gamma correction built in. also, current image standards do not contain the value of gamma with which an image was created, thus complicating the issue further. given these constraints, a reasonable approach when storing images in a web site is to preprocess the images with a gamma that represents in a web site is to preprocess the images with a gamma that represents an “average” of the types of monitors and computer systems that one expects in the open market at any given point in time.3.2.4 piecewise-linear transformation functionsa complementary approach to the methods discussed in the previous three sections is to use piecewise linear functions. the principal advantage of piecewise linear functions over the types of functions we have discussed thus far is that the form of piecewise functions can be arbitrarily complex. in fact, as we will see shortly, a practical implementation of some important transformations can be formulated only as piecewise functions. the principal disadvantage of piecewise functions is that their specification requires considerably more user input. contrast stretchingone of the simplest piecewise linear functions is a contrast-stretching transformation. low-contrast images can result from poor illumination, lack of dynamic range in the imaging sensor, or even wrong setting of a lens aperture during image acquisition. the idea behind contrast stretching is to increase the dynamic range of the gray levels in the image being processed.gray-level slicinghighlighting a specific range of gray levels in an image often is desired. applications include enhancing features such as masses of water in satellite imagery and enhancing flaws in x-ray images. there are several ways of doing level slicing, but most of them are variations of two basic themes. one approach is to display a high value for all gray levels in the range of interest and a low value for all other gray levels. bit-plane slicinginstead of highlighting gray-level ranges, highlighting the contribution made to total image appearance by specific bits might be desired. suppose that each pixel in an image is represented by 8 bits. imagine that the image is composed of eight 1-bit planes, ranging from bit-plane 0 for the least significant bit to bit-plane 7 for the most significant bit. in terms of 8-bit bytes, plane 0 contains all the lowest order bits in the bytes comprising the pixels in the image and plane 7 contains all the high-order bits. 3.3 histogram processingthe histogram of a digital image with gray levels in the range 0, l-1 is a discrete function , where is the kth gray level and is the number of pixels in the image having gray level . it is common practice to pixels in the image, denoted by n. thus, a normalized histogram is given by , for , loosely speaking, gives an estimate of the probability of occurrence of gray level . note that the sum of all components of a normalized histogram is equal to 1.histograms are the basis for numerous spatial domain processing techniques. histogram manipulation can be used effectively for image enhancement, as shown in this section. in addition to providing useful image statistics, we shall see in subsequent chapters that the information inherent in histograms also is quite useful in other image processing applications, such as image compression and segmentation. histograms are simple to calculate in software and also lend themselves to economic hardware implementations, thus making them a popular tool for real-time image processing.附录b第三章 空间域图像增强增强的首要目标是处理图像,使其比原始图像格式和特定应用。这里的“特定”很重要,因为它一开始就确立了本章多讨论技术是面向问题的。例如,一种很合适增强x射线图像的方法,不一定是增强有空间探测器发回的火星图像的最好方法。暂且不谈所用方法,图像增强本身就是图像处理中最具有吸引力的领域之一。图像增强的方法分为两大类:空间域方法和频域方法。“空间域”一次是指图像平面自身,这类方法是以对图像的像素直接处理恩基础的。“频域”处理技术足以修改图像的傅氏变换为基础的。空间域方法在这一章讲述,频域增强将在第四章讨论。以这两类方法的各种结合为基础的增强技术是不常见的。我们也注意到,本章关于增强的许多基本技术在后续章节里的其他图像处理应用中也会用到。图像增强的通用理论是不存在的。当图像为视觉解释而进行处理时,有观察者最后判断特定方法的效果。图像质量的视觉评价是一种高度主观的过程,因此,定义一个“理想图像”标准没通过这个标准去比较算法的性能。当为机器感知而处理图像时,这个评价任务就会容易一些。例如,在一个特征识别的应用中,不考虑像计算要求这些问题,最好的图像处理方法是一种能得到最好的机器可识别结果的方法。无论怎样,甚至在把一个明确的性能标准加于这个问题的情况下,在选择特定的图像增强方法之前,常常需要一个实验和误差的特定量。3.1背景知识如前所述,“空间域增强”是指增强构成图像的像素。空间域方法是直接对这些像素操作的过程。空间域处理可由下式定义: (3.1-1)其中f(x, y)是输入图像,g(x, y)是处理后的图像,t是对f的一种操作,其定义在(x, y)的邻域。另外,t能对输入图像集进行操作,例如,为减少噪音而对k幅图像进行逐像素的求和操作,如3.4.2节所讨论的。定义一个点(x, y)邻域的主要方法是利用中心在(x, y)点的正方形货矩形子图像。子图像的中心从一个像素向另一个像素移动,比如说,可以从左上角开始。t操作应用到每一个(x, y)位置得到该店的输出g。这个过程仅仅用在小范围邻域里的图像像素。尽管像近似于圆的其他邻域形状有时也用,但正方形和矩形列阵因其容易执行操作而占主导地位。t操作最简单的形式是邻域为11的尺度(即单个像素)。在这种情况下,g仅仅依赖于f在(x, y)点的值,t操作成为灰度级变换函数(也叫做强度映射),形成为: (3.1-2)这里,为简便起见,令r和s是所定义的变量,分别是f(x, y)和g(x, y)在任一点(x, y)的灰度级。有的相当简单,却有很大作用,处理方法可以用灰度变换加以公式化。因为在图像任意点的增强仅仅依赖于该点的灰度,这类技术常常是指点处理。更大的邻域会有更多的灵活性。一般的方法是,利用点(x, y)事先定义的邻域里的一个f值的函数来决定g在(x, y)的值,其公式化的一个主要方法是以利用所谓的模板(也指滤波器、核、掩模或窗口)为基础的。从根本上说,模板是一个小的(即33)二维阵列,模板的系数值决定了处理的性质,如图像尖锐化等。以这种方法为基础的增强技术通常是指模板处理或滤波。这些概念将在3.5节讨论。3.2某些基本灰度变换以讨论灰度变换函数开始研究图像增强技术,这些都属于所有图像增强技术最简单的一类。处理前后的像素的值用r和s分别定义。如前节所述,这些值与s = t(r)表达式的形式有关,这里的t是把像素的值r映射到值s的一种变换。由于处理的是数字量,变换函数的值通常储存在一个一维阵列中,并且从r到s的映射通过查表得到。对于8比特环境,一个包含t值的可查阅的表需要有256个记录。正如对灰度变换介绍的那样,它显示了图像增强常用的三个基本类型函数:线性的(正比和反比)、对数的(对数和反对数变换)、幂次的(n次幂和n次方根变换)。正比函数式最一般的,其输出亮度与输入亮度可互换,唯有它完全包括在图形中。3.2.1 图像反转灰度级范围为0,l-1的图像反转可由反转变换获得,表示为: (3.2-1)用这种方式倒转图像的前度产生图像反转的对等图像。这种处理尤其适用于增强嵌入于图像暗色区域的白色或灰色细节,特别是当黑色面积占主导地位时。3.2.2 对数变换对数变换的一般表达式为: (3.2-2)其中c是一个常数,并假设r0。此种变换使窄带低灰度输入图像值映射为一宽带输出值。相对的是输入灰度的高调整值。可以利用这种变换来扩展被压缩的高值图像中的按像素。相对的是反对数变换的调整值。一般对数函数的所有曲线都能完成图像灰度的扩散/压缩。事实上飞,就此

温馨提示

  • 1. 本站所有资源如无特殊说明,都需要本地电脑安装OFFICE2007和PDF阅读器。图纸软件为CAD,CAXA,PROE,UG,SolidWorks等.压缩文件请下载最新的WinRAR软件解压。
  • 2. 本站的文档不包含任何第三方提供的附件图纸等,如果需要附件,请联系上传者。文件的所有权益归上传用户所有。
  • 3. 本站RAR压缩包中若带图纸,网页内容里面会有图纸预览,若没有图纸预览就没有图纸。
  • 4. 未经权益所有人同意不得将文件中的内容挪作商业或盈利用途。
  • 5. 人人文库网仅提供信息存储空间,仅对用户上传内容的表现方式做保护处理,对用户上传分享的文档内容本身不做任何修改或编辑,并不能对任何下载内容负责。
  • 6. 下载文件中如有侵权或不适当内容,请与我们联系,我们立即纠正。
  • 7. 本站不保证下载资源的准确性、安全性和完整性, 同时也不承担用户因使用这些下载资源对自己和他人造成任何形式的伤害或损失。

评论

0/150

提交评论