0 Bewertungen0% fanden dieses Dokument nützlich (0 Abstimmungen)
10 Ansichten12 Seiten
This document discusses image classification using the Maximum Likelihood Classifier. It explains that the classifier uses Bayes' theorem to calculate the posterior probability that an object belongs to a particular class, given its features. It provides an example where the classifier distinguishes between nuts and bolts based on their circularity. It outlines calculating the prior probabilities of each class, as well as the probability distributions of each class's features. The Maximum Likelihood Classifier assigns an object to the class with the highest posterior probability based on Bayes' theorem.
This document discusses image classification using the Maximum Likelihood Classifier. It explains that the classifier uses Bayes' theorem to calculate the posterior probability that an object belongs to a particular class, given its features. It provides an example where the classifier distinguishes between nuts and bolts based on their circularity. It outlines calculating the prior probabilities of each class, as well as the probability distributions of each class's features. The Maximum Likelihood Classifier assigns an object to the class with the highest posterior probability based on Bayes' theorem.
This document discusses image classification using the Maximum Likelihood Classifier. It explains that the classifier uses Bayes' theorem to calculate the posterior probability that an object belongs to a particular class, given its features. It provides an example where the classifier distinguishes between nuts and bolts based on their circularity. It outlines calculating the prior probabilities of each class, as well as the probability distributions of each class's features. The Maximum Likelihood Classifier assigns an object to the class with the highest posterior probability based on Bayes' theorem.
Muhammad Rzi Abbas Department of Mechatronics and Control Engineering
muhammadrziabbas@uet.edu.pk Lecturer, Mechatronics Dept. University of Engineering and Technology, Lahore Classification Module • There are however more advanced classifiers, of which one based on Bayes’ theorem from statistical decision theory is called the Maximum-Liklihood Classifier. • We will change our example a bit so that we want to classify two objects using one feature only. • Suppose we wish to distinguish between nuts and bolts. • Circularity measure will suffice this time and we will have one- dimensional feature space with two classes of objects. Classification Module • Let Cn = class of nuts, Cb = class of bolts, and x = circularity measure. • The first thing that is required is the PDF for each of these two classes, i.e. a measure of the probabilities that an object from a particular class will have a given feature value. • PDF can be estimated using a large set of nuts and bolts. • PDF is nothing but the normalized histograms of the feature values Classification Module • Using a-priori knowledge we may assume that the class of nuts is likely to occur twice as often as the class of bolts. • In this case we say that the a-priori probabilities of the two classes are: • P(Cn) = 0.666 • P(Cb) = 0.333 • But in this case it is more likely that they will have the same a-priori probabilities (0.5) Classification Module • The PDFs tell us the probability that the circularity x will occur, given that the object belongs to the class of nuts Cn in the first instance and to the class of bolts Cb in the second instance. • This is called ‘Conditional Probability’ of an object having a certain feature value, given that we know that it belongs to a particular class. • P(x|Cb) [It gives the probability that a circularity x will occur, given that the object is a bolt] Classification Module Classification Module • This is not what we wished for. • We want to determine the probability that an object belongs to a particular class, given that a particular value of x has occurred. • This is called the a-posteriori probability P(Ci|x) and is given by Bayes’ Theorem: Classification Module • In effect, this Bayes’ theorem allows us to use the a-priori probability of objects occurring in the first place, the conditional probability of an object having a particular feature value given that it belongs to a particular class, and the actual measurement of a feature value to estimate the probability that the measured object belongs to a given class. Classification Module • Once we can estimate the probability that, for a given measurement, the object is a nut and the probability that it is a bolt, we can make a decision as to its identity, choosing the class with the higher probability. • This is why it is called the Maximum Liklihood Classifier. Classification Module • The Maximum Liklihood Classifier’s test may be modified as follows: Classification Module • Advantage of using The Maximum Liklihood Classifier • Machine Vision by David Vernon, Published in 1991 • Chapter 6 (Section 6.3)
Katja Kruckeberg, Wolfgang Amann, Mike Green-Leadership and Personal Development - A Toolbox For The 21st Century Professional-Information Age Publishing (2011)