Tài liệu Xử lý hình ảnh kỹ thuật số P15 docx

66 273 0
Tài liệu Xử lý hình ảnh kỹ thuật số P15 docx

Đang tải... (xem toàn văn)

Tài liệu hạn chế xem trước, để xem đầy đủ mời bạn chọn Tải xuống

Thông tin tài liệu

443 15 EDGE DETECTION Changes or discontinuities in an image amplitude attribute such as luminance or tri- stimulus value are fundamentally important primitive characteristics of an image because they often provide an indication of the physical extent of objects within the image. Local discontinuities in image luminance from one level to another are called luminance edges. Global luminance discontinuities, called luminance boundary seg- ments, are considered in Section 17.4. In this chapter the definition of a luminance edge is limited to image amplitude discontinuities between reasonably smooth regions. Discontinuity detection between textured regions is considered in Section 17.5. This chapter also considers edge detection in color images, as well as the detection of lines and spots within an image. 15.1. EDGE, LINE, AND SPOT MODELS Figure 15.1-1a is a sketch of a continuous domain, one-dimensional ramp edge modeled as a ramp increase in image amplitude from a low to a high level, or vice versa. The edge is characterized by its height, slope angle, and horizontal coordinate of the slope midpoint. An edge exists if the edge height is greater than a specified value. An ideal edge detector should produce an edge indication localized to a single pixel located at the midpoint of the slope. If the slope angle of Figure 15.1-1a is 90°, the resultant edge is called a step edge, as shown in Figure 15.1-1b. In a digital imaging system, step edges usually exist only for artificially generated images such as test patterns and bilevel graphics data. Digital images, resulting from digitization of optical images of real scenes, generally do not possess step edges because the anti aliasing low-pass filtering prior to digitization reduces the edge slope in the digital image caused by any sudden luminance change in the scene. The one-dimensional Digital Image Processing: PIKS Inside, Third Edition. William K. Pratt Copyright © 2001 John Wiley & Sons, Inc. ISBNs: 0-471-37407-5 (Hardback); 0-471-22132-5 (Electronic) 444 EDGE DETECTION profile of a line is shown in Figure 15.1-1c. In the limit, as the line width w approaches zero, the resultant amplitude discontinuity is called a roof edge. Continuous domain, two-dimensional models of edges and lines assume that the amplitude discontinuity remains constant in a small neighborhood orthogonal to the edge or line profile. Figure 15.1-2a is a sketch of a two-dimensional edge. In addi- tion to the edge parameters of a one-dimensional edge, the orientation of the edge slope with respect to a reference axis is also important. Figure 15.1-2b defines the edge orientation nomenclature for edges of an octagonally shaped object whose amplitude is higher than its background. Figure 15.1-3 contains step and unit width ramp edge models in the discrete domain. The vertical ramp edge model in the figure contains a single transition pixel whose amplitude is at the midvalue of its neighbors. This edge model can be obtained by performing a pixel moving window average on the vertical step edge FIGURE 15.1-1. One-dimensional, continuous domain edge and line models. 22× EDGE, LINE, AND SPOT MODELS 445 model. The figure also contains two versions of a diagonal ramp edge. The single- pixel transition model contains a single midvalue transition pixel between the regions of high and low amplitude; the smoothed transition model is generated by a pixel moving window average of the diagonal step edge model. Figure 15.1-3 also presents models for a discrete step and ramp corner edge. The edge location for discrete step edges is usually marked at the higher-amplitude side of an edge transi- tion. For the single-pixel transition model and the smoothed transition vertical and corner edge models, the proper edge location is at the transition pixel. The smoothed transition diagonal ramp edge model has a pair of adjacent pixels in its transition zone. The edge is usually marked at the higher-amplitude pixel of the pair. In Figure 15.1-3 the edge pixels are italicized. Discrete two-dimensional single-pixel line models are presented in Figure 15.1-4 for step lines and unit width ramp lines. The single-pixel transition model has a mid- value transition pixel inserted between the high value of the line plateau and the low-value background. The smoothed transition model is obtained by performing a pixel moving window average on the step line model. FIGURE 15.1-2. Two-dimensional, continuous domain edge model. 22× 22× 446 EDGE DETECTION A spot, which can only be defined in two dimensions, consists of a plateau of high amplitude against a lower amplitude background, or vice versa. Figure 15.1-5 presents single-pixel spot models in the discrete domain. There are two generic approaches to the detection of edges, lines, and spots in a luminance image: differential detection and model fitting. With the differential detection approach, as illustrated in Figure 15.1-6, spatial processing is performed on an original image to produce a differential image with accentu- ated spatial amplitude changes. Next, a differential detection operation is executed to determine the pixel locations of significant differentials. The second general approach to edge, line, or spot detection involves fitting of a local region of pixel values to a model of the edge, line, or spot, as represented in Figures 15.1-1 to 15.1-5. If the fit is sufficiently close, an edge, line, or spot is said to exist, and its assigned parameters are those of the appropriate model. A binary indicator map is often generated to indicate the position of edges, lines, or spots within an FIGURE 15.1-3. Two-dimensional, discrete domain edge models. Fjk,() Gjk,() Ejk,() EDGE, LINE, AND SPOT MODELS 447 image. Typically, edge, line, and spot locations are specified by black pixels against a white background. There are two major classes of differential edge detection: first- and second-order derivative. For the first-order class, some form of spatial first-order differentiation is performed, and the resulting edge gradient is compared to a threshold value. An edge is judged present if the gradient exceeds the threshold. For the second-order derivative class of differential edge detection, an edge is judged present if there is a significant spatial change in the polarity of the second derivative. Sections 15.2 and 15.3 discuss the first- and second-order derivative forms of edge detection, respectively. Edge fitting methods of edge detection are considered in Section 15.4. FIGURE 15.1-4. Two-dimensional, discrete domain line models. 448 EDGE DETECTION 15.2. FIRST-ORDER DERIVATIVE EDGE DETECTION There are two fundamental methods for generating first-order derivative edge gradi- ents. One method involves generation of gradients in two orthogonal directions in an image; the second utilizes a set of directional derivatives. FIGURE 15.1-5. Two-dimensional, discrete domain single pixel spot models. FIRST-ORDER DERIVATIVE EDGE DETECTION 449 15.2.1. Orthogonal Gradient Generation An edge in a continuous domain edge segment such as the one depicted in Figure 15.1-2a can be detected by forming the continuous one-dimensional gradient along a line normal to the edge slope, which is at an angle with respect to the horizontal axis. If the gradient is sufficiently large (i.e., above some threshold value), an edge is deemed present. The gradient along the line normal to the edge slope can be computed in terms of the derivatives along orthogonal axes according to the following (1, p. 106) (15.2-1) Figure 15.2-1 describes the generation of an edge gradient in the discrete domain in terms of a row gradient and a column gradient . The spatial gradient amplitude is given by (15.2-2) For computational efficiency, the gradient amplitude is sometimes approximated by the magnitude combination (15.2-3) FIGURE 15.1-6. Differential edge, line, and spot detection. FIGURE 15.2-1. Orthogonal gradient generation. Fxy,() Gxy,() θ Gxy,() Fxy,()∂ x∂ θcos Fxy,()∂ y∂ θsin+= Gxy,() G R jk,() G C jk,() Gjk,() G R jk,()[] 2 G C jk,()[] 2 +[] 12⁄ = Gjk,() G R jk,() G C jk,()+= 450 EDGE DETECTION The orientation of the spatial gradient with respect to the row axis is (15.2-4) The remaining issue for discrete domain orthogonal gradient generation is to choose a good discrete approximation to the continuous differentials of Eq. 15.2-1. The simplest method of discrete gradient generation is to form the running differ- ence of pixels along rows and columns of the image. The row gradient is defined as (15.2-5a) and the column gradient is (15.2-5b) These definitions of row and column gradients, and subsequent extensions, are cho- sen such that G R and G C are positive for an edge that increases in amplitude from left to right and from bottom to top in an image. As an example of the response of a pixel difference edge detector, the following is the row gradient along the center row of the vertical step edge model of Figure 15.1-3: In this sequence, h = b – a is the step edge height. The row gradient for the vertical ramp edge model is For ramp edges, the running difference edge detector cannot localize the edge to a single pixel. Figure 15.2-2 provides examples of horizontal and vertical differencing gradients of the monochrome peppers image. In this and subsequent gradient display photographs, the gradient range has been scaled over the full contrast range of the photograph. It is visually apparent from the photograph that the running difference technique is highly susceptible to small fluctuations in image luminance and that the object boundaries are not well delineated. θ jk,() arc G C jk,() G R jk,()    tan= G R jk,() Fjk,()Fjk 1–,()–= G C jk,() Fjk,()Fj 1+ k,()–= 0000h 0000 0000 h 2 h 2 000 FIRST-ORDER DERIVATIVE EDGE DETECTION 451 Diagonal edge gradients can be obtained by forming running differences of diag- onal pairs of pixels. This is the basis of the Roberts (2) cross-difference operator, which is defined in magnitude form as (15.2-6a) and in square-root form as (15.2-6b) FIGURE 15.2-2. Horizontal and vertical differencing gradients of the peppers_mon image. ( b ) Horizontal magnitude ( c ) Vertical magnitude ( a ) Original Gjk,() G 1 jk,() G 2 jk,()+= Gjk,() G 1 jk,()[] 2 G 2 jk,()[] 2 +[] 12⁄ = 452 EDGE DETECTION where (15.2-6c) (15.2-6d) The edge orientation with respect to the row axis is (15.2-7) Figure 15.2-3 presents the edge gradients of the peppers image for the Roberts oper- ators. Visually, the objects in the image appear to be slightly better distinguished with the Roberts square-root gradient than with the magnitude gradient. In Section 15.5, a quantitative evaluation of edge detectors confirms the superiority of the square-root combination technique. The pixel difference method of gradient generation can be modified to localize the edge center of the ramp edge model of Figure 15.1-3 by forming the pixel differ- ence separated by a null value. The row and column gradients then become (15.2-8a) (15.2-8b) The row gradient response for a vertical ramp edge model is then FIGURE 15.2-3. Roberts gradients of the peppers_mon image. G 1 jk,() Fjk,()Fj 1+ k 1+,()–= G 2 jk,() Fjk 1+,()Fj 1+ k,()–= θ jk,() π 4 arc G 2 jk,() G 1 jk,()    tan+= G R jk,()Fjk 1+,()Fjk 1–,()–= G C jk,()Fj 1– k,()Fj 1+ k,()–= 00 h 2 h h 2 00 ( a ) Magnitude ( b ) Square root

Ngày đăng: 21/01/2014, 15:20

Tài liệu cùng người dùng

Tài liệu liên quan