Beruflich Dokumente
Kultur Dokumente
Machine learning is a subfield of computer science that evolved from the study of pattern
recognition and computational learning theory in artificial intelligence. In 1959, Arthur Samuel
defined machine learning as a "Field of study that gives computers the ability to learn without being
explicitly programmed". Machine learning explores the study and construction of algorithms that
can learn from and make predictions on data. Such algorithms operate by building a model from
example inputs in order to make data-driven predictions or decisions.
Types of machine learning?
Supervised learning: The computer is presented with example inputs and their desired outputs,
given by a "teacher", and the goal is to learn a general rule that maps inputs to outputs.
Unsupervised learning: No labels are given to the learning algorithm, leaving it on its own to find
structure in its input. Unsupervised learning can be a goal in itself (discovering hidden patterns in
data) or a means towards an end (feature learning).
Machine Learning Algorithms:
Artificial neural networks
An artificial neural network (ANN) learning algorithm, usually called "neural network" (NN), is a
learning algorithm that is inspired by the structure and functional aspects of biological neural
networks. Computations are structured in terms of an interconnected group of artificial neurons,
processing information using a connectionist approach to computation. Modern neural networks are
non-linear statistical data modeling tools. They are usually used to model complex relationships
between inputs and outputs, to find patterns in data, or to capture the statistical structure in an
unknown joint probability distribution between observed variables.
How do ANNs work?
An artificial neuron is an imitation of a human neuron
The output is a function of the input, that is affected by the weights, and the transfer
functions.
Three types of layers: Input, Hidden, and Output:
Problem Statement:
Semantic Segmentation Using convolutional neural
network.
"segmentation" is a partition of an image into several "coherent" parts, but without any attempt at
understanding what these parts represents. One of the most famous works (but definitely not the
first) is Shi and Malik "Normalized Cuts and Image Segmentation" PAMI 2000. These works
attempt to define "coherence" in terms of low-level cues such as color, texture and smoothness of
boundary. You can trace back these works to the Gestalt theory.
On the other have "semantic segmentation" attempts to partition the image into semantically
meaningful parts, and to classify each part into one of pre-determined classes. You can also achieve
the same goal by classifying each pixel (rather than the entire image/segment). In that case you are
doing pixel-wise classification, which leads to the same end result but in a slightly different path...
So, I suppose you can say that "semantic segmentation", "scene labeling" and "pixelwise
classification" are basically trying to achieve the same goal: semantically understanding the role of
each pixel in the image.
*For different methods of semantic segmentation follow survey paper attached in mail:
Some time delay neural networks also use a very similar architecture to convolutional neural
networks, especially those for image recognition and/or classification tasks, since the tiling of
neuron outputs can be done in timed stages, in a manner useful for analysis of images.
Compared to other image classification algorithms, convolutional neural networks use relatively
little pre-processing. This means that the network is responsible for learning the filters that in
traditional algorithms were hand-engineered. The lack of dependence on prior knowledge and
human effort in designing features is a major advantage for CNNs.