Sie sind auf Seite 1von 54

|  

  

 
In high resolution field, in addition to the usual
preprocessing functions (offset, dark and flat
corrections), the usefulness of image processing
can be divided into two main functions:
increasing the contrast of planetary details and
reducing the noise.

  
  

 Increasing the contrast of small details is
the aim of many processing algorithms
which all act in the same way: they
amplify the high frequencies in the image.
This is the reason why they are called
Y Y 
, and probably the most
Y 
,
famous of them is Y  . . This
technique is well-
well-known but hard to use in
astrophotography. In digital image
processing the general principle of
unsharp masking is
| 



 a fuzzy image (blue curve) is made
from the initial image (red curve) by
application of a low-
low-pass filter
(gaussian) whose strenght is
adjustable; the high frequencies are
suppressed,
 this fuzzy image is substracted from
the initial image; the result (green
curve) contains only the small details
(high frequencies) but its appearance
is very strange and unaesthetic
(unfortunately, this image also
contains noise),

|   
 m  is choosing which points you
want to have represent a given image.
Given an analog image, sampling
represents a mapping of the image from a
continuum of points in space (and possibly
time, if it is a moving image) to a discrete
set. Given a digital image, sampling
represents a mapping from one discrete
set of points to another (smaller) set.
 

 
 
d  ! d! "
dow pass filters

dow pass filtering, otherwise known as


"smoothing", is employed to remove high
spatial frequency noise from a digital
image. Noise is often introduced during
the analog-
analog-to-
to-digital conversion process
as a side-
side-effect of the physical conversion
of patterns of light energy into electrical
patterns
 

 
     
 If several copies of an image have
been obtained from the source, some
static image, then it may be possible
to sum the values for each pixel from
each image and compute an
average. This is not possible,
however, if the image is from a
moving source or there are other
time or size restrictions.
 #  $ %  
6  & 
 If such averaging is not possible, or if it is
insufficient, some form of R 
RR
 may be required. There
are two main types:
 R
 , where an


  R
 ,
image is restored based on some
knowledge of of the type of degradation it
has undergone. Filters that do this are
often called "optimal filters"
   R
 , which
  R
 ,
attempts to improve the
(subjectively measured) quality of an
image for human or machine
interpretability. Enhancement filters
are generally heuristic and problem
oriented
  &  &  
 The form that low-
low-pass filters usually
take is as some sort of   
   

.
. The operator
usually affects one pixel of the image
at a time, changing its value by
some function of a "local" region of
pixels ("covered" by the window).
The operator "moves" over the
image to affect all the pixels in the
image.
 

 ×  
  
  R

   R

   R

× 

    
 These replace the value of each
pixel, by a weighted-
weighted-average of the
pixels in some neighborhood around
it, i.e. a weighted sum of the weights
are non-
non-negative. If all the weights
are equal then this is a 
 filter.
"linear"
î   
 This replaces each pixel value by the
median of its neighbors, i.e. the
value such that 50% of the values in
the neighborhood are above, and
50% are below. This can be difficult
and costly to implement due to the
need for sorting of the values.
However, this method is generally
very good at preserving edges.
î  
 Each pixel value is replaced by its
most common neighbor. This is a
particularly useful filter for
    procedures where each
pixel corresponds to an object which
must be placed into a class; in
remote sensing, for example, each
class could be some type of terrain,
crop type, water, etc..
These are all   
  in that
the same operation is applied to each
pixel location.
 ¦ non-
non-space invariant filtering, using
the above filters, can be obtained by
changing the type of filter or the
weightings used for the pixels for
different parts of the image.
 ×  R 
filters also exist which
are not space invariant; these
attempt to locate edges in the noisy
image before applying smoothing, a
difficult task at best, in order to
reduce the blurring of edges due to
smoothing.
Ô    
 ¦ high pass filter is used in digital image
processing to remove or suppress the low
frequency component, resulting in a
sharpened image. High pass filters are
often used in conjunction with low pass
filters. For example, the image may be
smoothed using a low pass filter, then a
high pass filter can be applied to sharpen
the image, therefore preserving boundary
detail.
|   
 ¦n edge may be regarded as a
boundary between two dissimilar
regions in an image.
 These may be different surfaces of
the object, or perhaps a boundary
between light and shadow falling on
a single surface.
 ' 
 edges have been loosely defined as pixel
intensity discontinuities within an
image. While two experimenters
processing the same image for the same
purpose may not see the same edge pixels
in the image, two for different applications
may never agree.
 In a word, edge detection is usually a
subjective task.
 In principle an edge is easy to find
since differences in pixel values
between regions are relatively easy
to calculate by considering gradients.
 (
 
)

 '' *   & 



 
 Finding pixels in the image where
edges are likely to occur by looking
for discontinuities in gradients.
 Candidate points for edges in the
image are usually referred to as
›, › › ›
,
› ›, › ›
, or

.
› ›
.
 dinking these edge points in some
way to produce descriptions of edges
in terms of lines, curves ›
a       

 ¦n edge point can be regarded as a


point in an image where a
discontinuity (in gradient) occurs
across some line. ¦ discontinuity
may be classified as one of three
types
 
a     
 -- where the gradient of the pixel
values changes across a line. This
type of discontinuity can be classed
as
 ` edges
 `  edges
 › edges
  › edges
--by
--by noting the sign of the component
of the gradient perpendicular to the
edge on either side of the edge.
 Ramp edges have the same signs in
the gradient components on either
side of the discontinuity, while roof
edges have opposite signs in the
gradient components.
¦       
 -- where pixel values themselves
change suddenly across some line.
¦     
 -- where pixel values rapidly increase
then decrease again (or  ››` )
across some line.
 ( + (
 +

 jump discontinuities occur where one


object occludes another (or another
part of itself).
 Gradient discontinuities usually occur
between adjacent faces of the same
object.
 (  +

 a bar discontinuity would represent


cases like a thin black line on a white
piece of paper.
 Step edges may separate different
objects, or may occur where a
shadow falls across an object.
G    

 ,
 Since First derivative operators
exaggerate the effects of noise,
Second derivatives exaggerate noise
twice as much.
 No directional information about the
edge is given.
x   

 Edge detectors yield pixels in an


image lie on edges.
 Next collect these pixels together
into a set of edges.
 Replace many points on edges with a
few edges themselves.
 ' -
 Small pieces of edges may be
missing,
 Small edge segments may appear to
be present due to noise where there
is no real edge, ›
› ..
  x   
 -- where edge points are grouped to
form edges by considering each
point's relationship to any
neighbouring edge points.
a  x   
 -- where all edge points in the image
plane are considered at the same
time and sets of edge points are
sought according to some similarity
constraint, such as points which
share the same edge equation.
d d

   

 îost edge detectors yield


information about the magnitude of
the gradient at an edge point and,
more importantly, the direction of
the edge in the locality of the point.
’  ¦ 

 In many machine vision and image


processing algorithms, simplifying
assumptions are made about the
uniformity of intensities in local
image regions. However, images of
real objects often do not exhibit
regions of uniform intensities.
 Image texture, defined as a function
of the spatial variation in pixel
intensities (gray values), is useful in
a variety of applications and has
been a subject of intense study by
many researchers. One immediate
application of image texture is the
recognition of image regions using
texture properties.
(  
Texture boundaries can be found even
if the texture surfaces cannot be
classified. The goal of texture
segmentation is to obtain the
boundary map separating the
differently textured regions in an
image.
(  
 ’

 Y
  is often used for
image compression applications. It is
also important in computer graphics
where the goal is to render object
surfaces which are as realistic
looking as possible.
  (
 The Y
 

problem is
one instance of a general class of
vision problems known as ``shape
from X.'' The goal is to extract three-
three-
dimensional surface shape from
variations in textural properties in
the image. The texture features are
distorted due to the imaging process
and the perspective projection which
provide information about surface
orientation and shape.

Das könnte Ihnen auch gefallen