Sie sind auf Seite 1von 44

Overview Of Digital Signal/Image Processing

Mohammed A. Hasan Department of Electrical & Computer Engineering University of Minnesota-Duluth Email:mhasan@d.umn.edu

Related ECE Courses and Software: 1. ECE 2111: Signals and Systems 2. ECE 5741: Digital Signal Processing 3. ECE 8741: Digital Image Processing 4. Matlab Working knowledge in Statistics, Calculus, and Dierential Equations is very helpful in understanding Image Processing.

History Many of the techniques of digital image processing, or digital picture processing as it was often called, were developed in the 1960s at the Jet Propulsion Laboratory, MIT, Bell Labs, University of Maryland, and a few other places, with application to satellite imagery, wirephoto standards conversion, medical imaging, videophone, character recognition, and photo enhancement. But the cost of processing was fairly high with the computing equipment of that era. In the 1970s, digital image processing proliferated, when cheaper computers and dedicated hardware became available. Images could then be processed in real time, for some dedicated problems such as television standards conversion. As general-purpose computers became faster, they started to take over the role of dedicated hardware for all but the most specialized and compute-intensive operations. With the fast computers and signal processors available in the 2000s, digital image processing has become the most common form of image processing, and is generally used because it is not only the most versatile method, but also the cheapest.

What is an Image
1. An image f (x, y) is 2-dimensional light intensity function, where f measures brightness at position (x, y). 2. A digital image is a representation of an image by a 2-D array of discrete samples. 3. The amplitude of each sample is represented by a nite number of bits. 4. Each element of the array is called a pixel.

Terminology Images: An image is a two-dimensional signal whose intensity at any point is a function of two spatial variables. Examples are photographs, still video images, radar and sonar signals, chest and dental X-rays. An image sequence such as that seen in a television is a three dimensional signal for which the image intensity at any point is a function of three variables: two spatial variables and time. 1. Digital image processing is a term used to describe the manipulation of image data by a computer. 2. The process of transforming an image to a set of numbers, which a computer can utilized, is called digitization.

3. Digitization is to divide an image up into several picture elements called pixels. A pixel is the smallest resolvable unit of an image which the computer handles.

4. The value of a pixel is referred to as its gray level and can be thought of as the intensity or brightness (or darkness) of the pixel. 5. The number of dierent gray-levels a pixel can have varies from system to system, and is determined by the hardware that produces or displays the image.

Why do we process images


Images (and videos) are everywhere. This includes different imaging modalities such as visual, X-ray, ultrasound, etc. Multimedia information will be the wave of the future. Diverse applications in astronomy, biology, geology, geography, medicine, law enforcement, defense, Industrial inspection, require processing of images.

Applications 1. Editing Digital photos (a) Image Enhancements (b) Contrast Stretching (c) Color Correction (d) White-Balancing (e) Red-Eye Reduction/ Removal (f) Shrinking/Zooming Software (a) Adobe Photoshop (b) ImageMagick (c) GIMP (GNU Image Manipulation Program) 2. Medical Image Processing (a) Chest X-rays (b) Computed Tomography (c) UltraSound Imaging (d) MRI, PET, etc. 3. Machine Vision (a) Inspection (b) Measurement (c) Robotic Vision

4. Remote Sensing (a) Land use (b) Crops (c) Weather Forecasting 5. Defense (a) Target recognition (b) Thermal Imaging 6. Image Storage & Transmission (a) JPEG,MPEG, etc. (b) Eciently store an image in a digital camera (c) Send an image from Mars to Earth 7. OCR, Document Image Processing 8. Prepare for display or printing (a) Adjust image size (b) Halftoning 9. Enhance and restore images (a) Remove scratches from an old movie (b) Improve visibility of tumor in a radiograph 10. Extract information from images (a) Read the ZIP code on a letter (b) Measure water polution from aerial images 11. To hide information in them or watermarking

Image Sensors/Input Devices 1. Film Cameras 2. Digital Cameras (CCD/CMOS) 3. Video Cameras 4. Scanners Image Output Devices 1. Printers 2. TV/Computer Monitors Image Storage 1. CD-ROM 2. Hard disk Processing Hardware 1. PCs, 2. Microprocessors, boards, 3. workstations, etc. Processing Software: Many commercial and non-commercial

Grayscale and Color Images 1. For grayscale image, 256 levels or 8 bits/pixel is sucient for most applications 2. For color image, each component (R, G, B) needs 256 levels or 8 bits/pixel 3. Storage for typical images (a) 512 512, 8 bits grayscale image: 262,144B (b) 1024768, 24 bits true color image: 2,359,296B

Digital B/W Videos


XR (l, m, n) 1. l= vertical position 2. m= horizontal position 3. n= frame number

Grayscale Image

Color Images XR (n, m), XG (n, m), XB (n, m)

F (x, y) = F (m, n),

0 m M 1, 0 n N 1

A digital image can be written as a matrix x(0, 0) x(1, 0) F = . . . x(M 1, 0)

x(0, 1) x(1, 1) . . . x(M 1, 1)

. . .

x(0, N 1) x(1, N 1) . . . x(M 1, N 1)

Popular Image Size: 1. N = 64, M = 64 (26 ), # of pixels = 4,096. 2. N = 512, M = 512 (29), # of pixels = 262,144. 3. N = 1280, M = 1024, # of pixels = 1,310,720.

Image Operations 1. Point Operations 2. Local Operations 3. Global Operations

Image Operations can be classied as Linear and non-linear Operations: H is a linear operator if if satises the superposition principle: H(af + bg) = aH(f ) + bH(g) for all images f and g and all constants a and b. 1. Mean ltering: Linear 2. Median ltering: Non-linear

Simple Operations On Images Digital Negative: Given an image F , the Digital Negative of F is dened as FN egative(m, n) = 255 F (m, n)

Feature Enhancement by Subtraction

A Brief History of Lena (Lenna) Anil K Kandangath Anyone familiar with digital image processing will surely recognize the image of Lena. While going through some old usenet discussions, I got to know that Lena has a history worth all the attention that has been paid to her over the years by countless image processing researchers. Lena Sjblom, (also spelled Lenna by many publications) was the Playboy playmate in November 1972 and rose to fame in the computer world when researchers at the University of Southern California scanned and digitized her image in June 1973. (Lena herself never know of her fame until she was interviewed by a computer magazine in Sweden where she lives with her husband and children). According to the IEEE PCS Newsletter of May/June 2001, they were hurriedly searching for a glossy image which they could scan and use for a conference paper when someone walked in with a copy of Playboy. The engineers tore o the top third of the centerfold and scanned it with a Muirhead wire photo scanner (a distant cry from the atbed scanners of today) by wrapping it around the drum of the scanner. (Now you know why the image shows only a small part of the entire picture.. discounting of course, the fact that the complete picture would raise quite a few eyebrows.

Also, heres the poem dedicated to Lina, written by an anonymous admirer: "0 dear Lena, your beauty is so vast It is hard sometimes to describe it fast. I thought the entire world I would impress If only your portrait I could compress. Alas! First when I tried to use VQ I found that your cheeks belong to only you. Your silky hair contains a thousand lines Hard to match with sums of discrete cosines. And for your lips, sensual and tactual Thirteen Crays found not the proper fractal. And while these setbacks are all quite severe I might have fixed them with hacks here or there But when wavelets took sparkle from your eyes I said, "Skip this stuff. Ill just digitize."

Linear Stretching 1. Enhance the dynamic range by stretching the original gray levels to the range of 0 to 2. Example (a) The original gray levels are [100, 150]. (b) The target gray levels are [0, 255]. (c) The transformation function g(f ) = ((f 100)/50) 255 f or100 f 150

Illustration of Linear Stretching

Image/video Processing Methods


1. Image Enhancement 2. Image Restoration 3. Compression 4. Image reconstruction 5. Morphological image processing 6. Feature extraction and recognition, computer vision

Other Image Operations Image algebra includes mathematical comparisons, altering values of pixels, thresholding, edge detection and noise reduction. 1. Neighborhood averaging is to avoid extreme uctuations in gray level from pixel to pixel. It is also very eective tool for noise reduction. 2. Image Scaling is a means of reducing or expanding the size of an image using existing image data. 3. Histogram Equalization is an adjustment of gray scale based on gray-level histogram. This is eective in enhancing the contrast of an image. 4. Edge Detection is an operation of measuring and analyzing the features in an image by detecting and enhancing the edges of the features. The most common edge detection method is gradient detection.

5. Image Restoration: Given a noisy image y(m, n) y(m, n) = x(m, n) + v(m, n) where x(m, n) is the original image and v(m, n) is noise. The objective is to recover x(m, n) from y(m, n). Color Restoration

Photo Restoration

6. Contrast Enhancement: how to enhance the contrast of an image? 1. Low contrast image values concentrated near narrow range (mostly dark, or mostly bright, or mostly medium values) 2. Contrast enhancement change the image value distribution to cover a wide range 3. Contrast of an image can be revealed by its histogram Histogram The histogram of an image with L possible gray levels, f = 0, 1, , L 1 is dened as: nl P (l) = n where nl is the number of pixels with gray level l. n is the total number of pixels in the image.

Examples of Histograms

Applications Astronomy: Hubble Space Telescope : This telescope has limitation in resolution due to atmospheric turbulence.

Optical problem in a telescope results in blurred, out of focus image. Digital image processing is normally used to recover the desired information from these images.

Medical Imaging: Most of advanced medical imaging tools are based on DSP tools. X-Ray computerized Tomography (X-ray CT) is capable of generating a cross-sectional display of the body. This involves X-ray generation, detection, digitization, processing and computer image reconstruction. Similarly, NMRCT (nuclear magnetic resonance). MRI

Ultrasound

Biometrics and Security:


Biometric recognition refers to the use of distinctive characteristics (biometric identiers) for automatically recognition individuals. These characteristics may be 1. Physiological (e.g., ngerprints, face, retina, iris) 2. a Behavioral (e.g., gait, signature, keystroke) Biometric identiers are actually a combination of physiological and behavioral characteristics, and they should not be exclusively identied into either class. (For example, speech is determined partly by the physiology and partly by the way a person speaks.)

1. Signature Verication 2. Fingerprint identication 3. Face recognition

Signature Verication

Fingerprint: In 1684, an English plant morphologist published the rst scientic paper reporting his systematic study on the ridge and pore structure in ngerprints.

A ngerprint image may be classied as: (a) Oine: Inked impression of the ngertip on a paper is scanned (b) Live-scan: Optical sensor, capacitive sensors, ultrasound sensors, ...

At the local level, there are dierent local ridge characteristics. The two most prominent ridge characteristics, called minutiae, are: (a) Ridge termination (b) Ridge bifurcation At the very-ne level, intra-ridge details (sweat pores) can be detected. They are very distinctive; however, very high-resolution images are required.

Face Recognition

Face Recognition Methods (a) Template matching using minimum-distance classiers metrics (b) Linear discriminants (c) Bayesian approach

1. Watermarking: The World Wide Web and the progress in multimedia storage and transmission technology expanded the possibility of illegal copying and reproducing of digital data. Digital watermarking represents a valid solution to the above problem, since it makes possible to identify the source, author, creator, owner, distributor or authorized consumer of digitized images, video recordings or audio recordings. A digital watermark is an identication code, permanently embedded into digital data, carrying information pertaining to copyright protection and data authentication.

(a) Copyright protection and authentication 2. Data hiding 3. Steganography: Secret communication Steganography is the art and science of communicating in a way which hides the existence of the communication. In contrast to cryptography, where the enemy is allowed to detect, intercept and modify messages without being able to violate certain security premises guaranteed by a cryptosystem, the goal of steganography is to hide messages inside other harmless messages in a way that does not allow any enemy to even detect that there is a second secret message present [Markus Kuhn 1995-07-03]

Entertainment
1. Digital camcorders 2. HDTV 3. DVDs: High quality image/video compression(MPEG2: about 5-10 Million bits/second) 4. Digital Cinema (a) New compression technologies are needed i. Consider a 2 hour movie: 1920 x 1080 x 30 bits/pixel x 24 frames/second 1.5 billion bits/second ! 1.3 terra bytes / 2 hr program

Compression HDTV (high resolution TV broadcasted with the same existing TV channel; requires digital compression technology). Video-phone using existing telephone cable. The amount of intermission to be transmitted for videophone is much bigger than speech signal. Requires Compression Technology. Streaming video over wireless 1. Video is high bandwidth data 2. Wireless, at present, has limited bandwidth 3. Needs ecient and eective compression 4. New coding techniques such as MPEG-4 have been developed. Satellite broadcasting and Satellite Imaging Image Compression Techniques 1. JPEG 2000 standard is based on wavelets 2. JPEG (original) is based on the Discrete Cosine

An Example of Image Compression

Das könnte Ihnen auch gefallen