Sunteți pe pagina 1din 20

OBJECT DETECTION AND

IDENTIFICATION
COMPUTER SCIENCE ENGINEERING
Sainagesh.V 160031460
Sindhusha.R 160031283
Rohith sri sai.M 160030891
PROJECT GUIDE
Mohammad Gouse Assistant professor
INTRODUCTION
• In this project we will be using highly accurate object detection-algorithms
and methods such as R-CNN, Fast-RCNN, Faster-RCNN, RetinaNet and fast
yet highly accurate ones like SSD and YOLO.
• Using these methods and algorithms, based on deep learning which is also
based on machine learning require lots of mathematical and deep learning
frameworks understanding by using dependencies such as
tensorflow,opencv,imageai..etc., we can detect each object in image by the
area object in an highlighted rectangular boxes and identify each object and
assign its tag to the object.
• this also includes accuracy of each method for identifying objects.
OBJECT DETECTION
• It is a critical part of many applications such as image search, image auto-
annotation and scene understanding, object tracking. Moving object tracking
of video image sequences was one of the most important subjects in
computer vision.
• Object recognition is a general term to describe a collection of related
computer vision tasks that involve identifying objects in digital photographs.
• Image classification involves predicting the class of one object in an image.
Object localization refers to identifying the location of one or more objects in
an image and drawing abounding box around their extent.
• Object detection combines these two tasks and localizes and classifies one
or more objects in an image.
BACKGROUND STEPS IN OBJECT DETECTION

• Object detection systems always construct a model for an object class from a
set of training examples.
• Let’s assume the size of the input image to be 16 × 16 × 3. If we are using
the sliding window approach, then we would have passed this image to the
above ConvNet four times, where each time the sliding window crops the part
of the input image matrix of size 14 × 14 × 3 and pass it through the
ConvNet.
• This results will give an output matrix of shape 2 × 2 × 4. Each cell in the
output matrix represents the result of the possible crop and the classified
value of the cropped image.
BACKGROUND STEPS IN OBJECT DETECTION
Before Detection After Detection
METHODS
• R-CNN
• Fast R-CNN
• Faster R-CNN
• ResNet
• YOLO
• SSD
• MANet
R-CNN

• Generate initial sub-segmentation, we generate many candidate regions


• Use greedy algorithm to recursively combine similar regions into larger
ones
• Use the generated regions to produce the final candidate region proposals
FAST R-CNN

• The same author of the previous paper(R-CNN) solved some of the drawbacks of R-
CNN to build a faster object detection algorithm and it was called Fast R-CNN.
• The approach is similar to the R-CNN algorithm. But, instead of feeding the region
proposals to the CNN, we feed the input image to the CNN to generate a
convolutional feature map.
FASTER R-CNN

• The image is provided as an input to a convolutional network which provides a convolutional feature map.
• Instead of using selective search algorithm on the feature map to identify the region proposals, a separate
network is used to predict the region proposals.
• The predicted region proposals are then reshaped using a RoI pooling layer which is then used to classify the
image within the proposed region and predict the offset values for the bounding boxes.
COMPARISION OF THREE R-CNN
ALGORITHMS
ResNet
• To train the network model in a more effective manner, we herein adopt the
same strategy as that used for DSSD (the performance of the residual
network is better than that of the VGG network). The goal is to improve
accuracy.
• The first implemented for the modification was the replacement of the VGG
network which is used in the original SSD with ResNet. We will also add a
series of convolution feature layers at the end of the underlying network.
• Although the ResNet–101 layer is deeper than the VGG–16 layer, it is
experimentally known that it replaces the SSD’s underlying convolution
network with a residual network, and it does not improve its accuracy but
rather decreases it.
YOLO

• we take an image and split it into an SxS grid, within each of the grid we take m
bounding boxes. For each of the bounding box, the network outputs a class
probability and offset values for the bounding box.
• The bounding boxes having the class probability above a threshold value is selected
and used to locate the object within the image.
• YOLO is orders of magnitude faster(45 frames per second) than other object
detection algorithms. The limitation of YOLO algorithm is that it struggles with small
objects within the image, for example it might have difficulties in detecting a flock of
birds.
SSD(SINGLE SHOT DETECTION)
• The SSD object detection composes of 2
parts:Extract feature maps, and Apply
convolution filters to detect objects.
• Each prediction composes of a boundary
box and 21 scores for each class and we
pick the highest score as the class for the
bounded object. Conv4_3 makes total of
38 × 38 × 4 predictions: four predictions
per cell regardless of the depth of
featuremaps. A expected, many
predictions contain no object. SSD
reserves a class “0” to indicate
INCEPTION V3
• Inception v3 is widely used as image recognition model that has showed to
obtain accuracy of greater than 78.1% on the ImageNet dataset. The model
is the culmination of many ideas developed by researchers over years. It is
based on “Rethinking the Inception Architecture Computer Vision” by
Szegedy
• The model is made of symmetric and asymmetric building blocks, including
convolutions, average pooling, max pooling, concats, dropouts, and fully
connected layers. Batchnorm is used more throughout the model and applied
to activation inputs. Loss is computed via Softmax.
INCEPTION V3
DenseNet
• DenseNet stands for Densely Connected Convolutional Networks
it is one of the latest neural networks for visual object recognition.
It is similar to ResNet but has some fundamental differences.
• Usually Convolutionnetworks work such a way that
• We have an initial image, say having a shape of (29, 34, 31). After
we apply set of convolution or pooling filters on it, squeezing
dimensions of width and height and increasing features
dimension.
• So the output from the Li layer is input to the Li+1 layer.
DenseNet
DenseNet
CONCLUSION
By the end of this project we are able to detect and
identify the objects in the picture and assign them the
tags that they belong to more accurately.We will be able
to plot the coordinates of the object identified along with
its tag from the image.

S-ar putea să vă placă și