Professional Documents
Culture Documents
ISSN No:-2456-2165
We use a YOLOv3 module only for detection of traffic large-scale datasets from ImageNet, by transferring its
signs, instead of training the model to classify traffic signs learned image representations and reuse them to the
across 43 different classes, we trained the model to detect classification task with limited training data. The main idea
traffic signs in images across 4 types based on color and is based on designing a method which reuses a part of training
shape. The detected sign is cropped and passed on to a layers of AlexNet. But this approach comes with limitations
separate 26 layered CNN model which classifies the traffic such as the model depends very much on the accuracy, fitting
sign across 43 different classes. With this approach we can and availability of pre-trained model.
ensure that accuracy and efficiency of detection as well as
classification is maintained even with limited training data. B. ‘Traffic Sign Detection and Recognition using a CNN
Ensemble’ by Aashrith Vennelakanti, Smriti Shreya,
II. LITERATURE REVIEW Resmi Rajendran, Debasis Sarkar, Deepak
Muddegowda, Phanish Hanagal
A. ‘Convolutional Neural Networks for image The System is divided into two phases: detection and
classification’ by Nadia Jmour, Sehla Zayen, Afef recognition. Detection is done based on shape and color of
Abdelkrim the traffic sign, followed by the sign validation. Once the
The proposed approach involves special case of transfer sign is detected, the portion of the image is cropped and is
learning on CNN variation called AlexNet applied on the feed to a CNN model for recognition. The use of CNN
A. Object Detection
The object detector is basically a YOLOv3 model which is 106 Layer Convolutional Neural Network, based on the darknet
YOLOv3 model trained on COCO dataset for image classification for 80 classes.
The cropped image is converted to grayscale so that function which is best suited for multi-class classification.
single intensity value for each pixel is retrieved instead to 3.
Then we apply Histogram equalization the grayscale image,
it increases the global contrast of the image to help
distinguish between a background and foreground. Now we
normalize the image based on intensity which changes the
range of pixel intensity values. Since the image is in
grayscale, only one channel needs to be normalized. Then
we apply dimensionality reduction by collapsing the image
into single channel and the processed image has dimensions Fig. 5:- 26-Layer CNN Architecture
as 32×32×1. The image is then forwarded to CNN classifier.
The input to the CNN is an image of 32X32X1
C. Classifier dimension. The filters move around this 2D matrix
The CNN classifier used is a 26 layered architecture producing feature maps by performing dot product. These
which includes convolutional, max pooling, batch are then feed to the dense or the fully connected layers of the
normalization, dense, dropout and flatten layers. The filter CNN to perform matrix vector multiplication for
size used is 3X3 and the number of filters in first layer is 16 classification and the last layer implements softmax function
which increases to 64 at the end of the feature extraction on the layer inputs to provide list of classes and its
segment. To prevent the over-fitting of the model we add probability. The one with highest probability is considered
dropout layer. There are 4 dense layer which perform the as the label for traffic sign.
classification task, the last of which uses softmax classifier
As mentioned, in our approach we used the YOLOv3 Prohibitory – Circular, White background, Red border
as a detector rather than a classifier, we divided the dataset Danger – Triangular, White background, Red border
across 4 categories instead of considering 43 classes. The 4 Mandatory – Circular, Blue background
categories were based on following characteristics of shape, Others – Remaining classes
background color and border color:
After the updated categorization of dataset, the
distribution was comparatively balanced:
V. RESULTS