Traffic Sign Recognition Using CNN
Traffic Sign Recognition Using CNN
Traffic Sign Recognition Using CNN
PREFACE
Subject to limitation of time effort and resource every possible attempt has
been made to study the problem deeply. The whole project is measure with
the help of deep learning and CNN model, the data further analyzed and
interpreted and the result are obtained.
Page | 1
\
DECLARATION
I/ We certify that,
------------------------------------
Harshit Agarwal (10300320068)
------------------------------------
Pragati Priya (10300320098)
------------------------------------
Anshu Kumari (10300320031)
------------------------------------
Page | 2
\
Page | 3
\
ACKNOWLEDGEMENT
Finally, we would like to thank all other teaching staff and non-
teaching staff for allowing us to carry out the project work.
Page | 4
\
Contents
Preface……………………………………………………………1
Declaration………………………………………………………..2
Certificate …………….………………………………………,,,,.3
Acknowledgement………………………………………………..4
Contents…………………………………………………………..5
Chapter-1 : Introduction…………………………………..7
1.1 Background Motivation .............................................8
1.2 Research Questions ................................................... 9
1.3 Purpose of Project ................................................... 10
1.4 Definitions .............................................................. 10
1.5 Software Requirement Specification………………. 52
1.5.1 Functional Requirements .................................. 52
1.5.2 Non-Functional Requirements .......................... 54
1.6 Literature Review………………………………...…. 59
Page | 5
\
Chapter – 6 : Results………………………………………137
6.1 Data Description………………………………….138
6.2 Outputs of Traffic Sign Recognition…………….139
Chapter – 8 : Bibliography………………………………...178
Page | 6
\
CHAPTER – 1
INTRODUCTION
Page | 7
\
Page | 8
\
(1) Which deep learning algorithms are suitable for recognizing various
sizes of real-world traffic signs?
(2) What detection and classification algorithm systems are suitable for
our project?
Page | 9
\
1.4 Definitions:
Page | 10
\
Page | 11
\
Page | 12
\
2. Error Function: The error function, also known as the loss function,
evaluates the performance of the model by quantifying the discrepancy
between the predicted outcomes and the actual outcomes. If labeled
data is available, the error function compares the predicted values to
the ground truth labels to determine the accuracy of the model's
predictions. The goal of the error function is to minimize this
discrepancy, ensuring that the model's predictions align as closely as
possible with the true outcomes.
Page | 13
\
Page | 14
\
Page | 15
\
Page | 16
\
Page | 17
\
• Semi-supervised learning:
Page | 18
\
data often proves to be a formidable bottleneck due to the need for manual
annotation by domain experts. In domains such as sentiment analysis,
document classification, and entity recognition, semi-supervised learning
algorithms can leverage a modest corpus of labeled text data, comprising
categorized news articles, customer reviews, or social media posts,
alongside an extensive collection of unlabeled text data. By discerning
subtle linguistic patterns, semantic relationships, and contextual cues
embedded within the unlabeled corpus, these algorithms can enhance their
understanding of language dynamics, sentiment nuances, and topical
coherence, thereby enriching their predictive capabilities and
generalization performance.
Page | 19
\
Page | 20
\
Page | 21
\
the difference between the predicted and actual prices, thus enabling
accurate price predictions for new instances.
Page | 22
\
Page | 23
\
Page | 24
\
• What is Python?
Page | 25
\
• Deep learning:
At the heart of deep learning lies the concept of artificial neural networks
(ANNs), sophisticated constructs inspired by the biological neural
networks that underpin human cognition. Comprising interconnected
nodes or neurons organized into layers, these neural networks possess
Page | 26
\
Page | 27
\
Page | 28
\
Page | 29
\
At the crux of CNNs lie convolutional layers, which serve as the bedrock
for local feature extraction and representation learning. These
convolutional layers convolve learnable filters or kernels over input
images, systematically scanning and capturing local patterns, edges, and
textures. Through successive convolutions and non-linear activations,
CNNs adeptly discern salient features and spatial hierarchies within
images, enabling them to discriminate between objects, scenes, and
textures with remarkable accuracy and efficiency.
Page | 30
\
Page | 31
\
Page | 32
\
Advantages:
• Very High accuracy in image recognition problems.
• Automatically detects the important features without any human
supervision
Disadvantages:
• CNN do not encode the position and orientation of object.
• Lack of ability to be spatially invariant to the input data.
• Lots of training data is required.
Page | 33
\
Page | 34
\
Alex Net:
AlexNet was the champion CNN model in the most difficult ImageNet
challenge named the ImageNet Large Scale Visual Recognition Challenge
(ILSVRC) in 2012 (Krizhevsky et al., 2012). This model proposed by
Alex and others were deeper and wider than the previous neural network
(LeNet), and it achieved the astonishing recognition accuracy against all
the traditional approaches. The appearance of AlexNet could be seen as
the turning point of the development of using machine learning and
computer vision for object detection and classification tasks.
where 𝑎𝑥,𝑦 𝑖 denotes the value yields by the number of 𝑖 convolution at the
position (𝑥, 𝑦) and the result of outputting by the ReLU activation function.
𝑛 is the number of neigh boring convolution kernels, and N is the total
number of convolution kernels in this layer. The rest of variants are
parameters, which are obtained in the experimental validation set.
Page | 35
\
Page | 36
\
VGG net:
Page | 37
\
Page | 38
\
Page | 39
\
classification tasks.
While VGG Net has been surpassed by more recent architectures in terms
of computational efficiency and parameter optimization, its simplicity and
effectiveness have made it a popular choice for educational purposes,
benchmarking, and as a baseline for comparison in research and
development. The principles underlying VGG Net—deep, homogeneous
architectures with small convolutional filters—have also influenced
subsequent CNN architectures, contributing to the ongoing evolution of
deep learning in computer vision and beyond.
Res Net:
Page | 40
\
Page | 41
\
Page | 42
\
• Faster R-CNN:
Faster R-CNN represents a significant advancement in object detection
within the realm of computer vision. Building upon the frameworks of
Fast R-CNN and R-CNN, Faster R-CNN introduces a novel architecture
that significantly improves the speed and accuracy of object detection
tasks.
Page | 43
\
The use of anchor boxes enables the RPN to efficiently generate region
proposals by eliminating the need for exhaustive search across different
scales and aspect ratios. Instead, the RPN focuses on refining the anchor
boxes based on the learned features extracted from the shared convolutional
layers. This approach significantly reduces computational overhead,
allowing for real-time object detection.
Page | 44
\
Page | 45
\
Page | 46
\
the raw data. The main aim of this step is to reduce the volume of data so
that it can be easily used and managed for data modelling. Feature
extraction methods include cluster analysis, text analytics, edge detection
algorithms, and principal components analysis (PCA).
Feature extraction can be accomplished manually or automatically
• Extracting features manually involves identifying and describing
pertinent attributes for a specific problem and devising methods to extract
them. A deep understanding of the domain aids in discerning which
features might be beneficial. Across decades of study, experts have
devised techniques for feature extraction from images, signals, and text.
For instance, a basic feature could be the average value within a signal
window.
• Automated feature extraction employs specialized algorithms or deep
neural networks to autonomously extract features from signals or images,
eliminating the need for manual intervention. This method proves
invaluable when aiming for swift progression from raw data to machine
learning algorithm development. Wavelet scattering exemplifies
automated feature extraction, showcasing its efficacy in swiftly and
accurately identifying salient features within datasets.
With the ascent of deep learning, feature extraction has been largely
replaced by the first layers of deep networks – but mostly for image data. For
signal and time-series applications, feature extraction remains the first
challenge that requires significant expertise before one can build effective
predictive models.
• Feature Selection:
Feature selection is a critical preprocessing step in machine learning
pipelines, aimed at identifying and retaining the most informative
features while discarding redundant, irrelevant, or noisy ones. The
process of feature selection plays a pivotal role in enhancing model
performance, reducing overfitting, and improving the interpretability of
machine learning models.
Page | 47
\
Page | 48
\
Page | 49
\
On the other hand, average pooling computes the average value of all
activations within each pooling region, leading to a smoother
downsampling process that reduces dimensionalities and suppresses noises.
While average pooling is computationally less intensive than max pooling,
it may not be as effective in preserving the most discriminative features,
especially in tasks where spatial localization is crucial.
In practice, max pooling is often favored over average pooling due to its
superior performance in tasks requiring feature localization, noise
reduction, and dimensionality reduction. By selectively retaining the most
significant activations, max pooling facilitates effective feature extraction
and enhances the discriminative power of CNNs.
Page | 50
\
Fig: Max and average polling with a filter of size 2x2 and stride 2
• ReLU Layer: -
The allure of ReLU lies not only in its computational efficiency but also
in its efficacy in mitigating the vanishing gradient problem encountered
in deeper networks. By allowing only positive activations to pass through
unchanged, ReLU facilitates smoother and faster gradient propagation
during backpropagation, thereby accelerating the convergence of the
training process. This characteristic has propelled ReLU to the forefront
of activation functions, rendering it a staple choice in modern deep
learning architectures.While ReLU's dominance is indisputable, it is
worthwhile to explore alternative activation functions that offer distinct
nonlinear properties. One such contender is the saturating hyperbolic
Page | 51
\
tangent function, \( f(x) = \tanh(x) \), which compresses input values into
the range \((-1, 1)\), thereby inducing saturation at extreme values.
Variants such as \( f(x) = |\tanh(x)| \) alleviate this saturation effect by
enforcing non-negativity, akin to ReLU. Additionally, the sigmoid
function \( \sigma(x) = \frac{1}{1 + e^{-x}} \) provides a smooth
transition from zero to one, lending itself well to binary classification
tasks.
Page | 52
\
4. Data Pre-processing:
Converting the raw image data into a format suitable for machine learning
is necessary. Each image is transformed into a matrix, and class labels are
encoded using one-hot encoding, essential for categorical data. This step
prepares the dataset for training by ensuring compatibility with machine
learning algorithms.
Page | 53
\
both training and validation datasets provides insights into the model's
performance and helps identify potential overfitting.
Page | 54
\
Adam Optimizer:
• Adam stands for Adaptive Moment Estimation. It's an optimization
algorithm that combines ideas from RMSprop and Momentum.
• It maintains two moving averages: the first moment (the mean) of the
gradients and the second moment (the uncentered variance) of the
gradients.
• Adam computes adaptive learning rates for each parameter. It updates
the parameters with the moving averages instead of raw gradients.
• The algorithm is known for its efficiency and good performance
across a wide range of deep learning architectures and tasks.
Benefits:
• Adaptive learning rate: Adam computes individual learning rates for
each parameter, which allows for faster convergence and better
generalization.
• Robustness: It performs well across a wide range of tasks and
architectures without requiring manual tuning of learning rates.
• Efficiency: Adam combines the benefits of adaptive learning rates
with the computational efficiency of stochastic gradient descent.
Adamax Optimizer:
• Adamax is a variant of the Adam optimizer, which is specifically
tailored to address certain limitations of Adam.
• While Adam calculates the second moment with a moving average
Page | 55
\
Scratch CNN :
1.Data Preprocessing:
• Load and preprocess your dataset. Typical preprocessing steps
include resizing images to a uniform size, normalization (scaling
pixel values to a range like [0, 1]), and splitting the dataset into
training, validation, and test sets.
2.Convolutional Layers:
• The convolutional layer is the fundamental building block of a
CNN. It applies a set of learnable filters (kernels) to the input data
to extract features. Each filter slides over the input data,
performing element-wise multiplication and summation to
produce feature maps.
Page | 56
\
Page | 57
\
Page | 58
\
Page | 59
\
Page | 60
\
Findings:
Convolutional Neural Networks (CNNs): CNNs are a specialized type of
deep learning architecture particularly adept at image recognition tasks.
Their ability to learn from vast datasets allows them to excel at identifying
patterns and features within images. By training a CNN on a comprehensive
dataset of labeled traffic signs, the system learns to distinguish between
different signs with high accuracy.
Dataset Selection: The quality and size of the training dataset are
paramount for the system's performance. We opted to utilize the German
Traffic Sign Benchmarks (GTSRB) dataset, known for its extensive
library of over 51,900 traffic sign images categorized into 43 distinct
classes. This rich dataset provided a robust foundation for training the
CNN to achieve high accuracy in recognizing various traffic signs.
Conclusion:
Real-time traffic sign recognition and voice alert systems represent a
significant leap forward in road safety technology. By leveraging the
power
Page | 61
\
Page | 62
\
Page | 63
\
Conclusion:
Deep learning-based traffic sign recognition systems hold immense
potential to enhance road safety by providing real-time assistance to
drivers. This research has demonstrated the effectiveness of CNNs in
accurately recognizing and classifying circular traffic signs. Further
research can explore the application of this approach to a wider range of
traffic signs and investigate integration with existing driver assistance
systems for a more comprehensive solution.
Page | 64
\
Page | 65
\
7. This research delves into the application of deep learning for the
automatic identification and recognition of traffic signs specific to
India. The proposed approach utilizes two powerful deep learning
architectures:
This research presents a practical deep learning approach for detecting and
recognizing Indian traffic signs. The proposed RMR-CNN model
demonstrates exceptional performance across various real-world
challenges:
Page | 66
\
Conclusion:
This research demonstrates the potential of deep learning for automated
traffic sign recognition in India. The proposed RMR-CNN model exhibits
promising results in accurately detecting and classifying a wide range of
Indian traffic signs under diverse real-world conditions. This research
contributes to the advancement of Intelligent Transportation Systems
(ITS) and paves the way for enhanced road safety in India for both human
drivers and autonomous vehicles.
Page | 67
\
Traffic signs play a vital role in ensuring the safety and order of our
roadways. These signs communicate vital information to drivers regarding
speed limits, road closures, directions, and warnings. However, several
factors can hinder the effectiveness of traffic signs:[4]
1. Image Preprocessing:
Page | 68
\
2. Feature Extraction:
3. Classification:
Findings:
Page | 69
\
Conclusion:
Comparative of Survey
Page | 70
\
2. Fast R-CNN:
3. Faster R-CNN:
Page | 71
\
Chapter -2
METHODOLOGY
Page | 72
\
Furthermore, benchmark datasets used for training and evaluating traffic sign
recognition algorithms often exhibit an uneven distribution of data
categories. For example, the German Traffic Sign Recognition Benchmark
(GTSRB) contains 43 classes of traffic signs, with some classes having
significantly lower frequencies than others. This imbalance in class
frequencies can pose challenges for machine learning algorithms, as they
may struggle to effectively learn from and generalize to the entire dataset.
Page | 73
\
Page | 74
\
Page | 75
\
and redundancy.
Among these, the most common dataset is the GTSRB (German Traffic
Sign Recognition Benchmark) dataset. The reason for its popularity is:
Page | 76
\
Page | 77
\
The loss layer serves a crucial role in neural networks by quantifying the
disparity between predicted outputs and ground truth labels. Various loss
functions are tailored to specific tasks and output types. For instance,
softmax cross-entropy loss is commonly used for multi-class
classification tasks, where it computes the cross-entropy between
predicted probabilities and true labels. Contrastingly, sigmoid cross -
entropy loss is employed in binary classification scenarios to evaluate
independent probabilities between 0 and 1. Additionally, mean squared
error (MSE) loss is prevalent in regression tasks, measuring the average
squared difference between predicted and actual values. Proper selection
of the loss function is pivotal for optimizing the neural network's
performance, ensuring accurate training updates and convergence
towards the desired output distribution.
Page | 78
\
Page | 79
\
5. The dropout layer is used to avoid over fitting by setting some of the
input neurons to 0 during the training process. The dense layer, on the
other hand, feeds all the outputs from the preceding layer to all its
neurons and perform the matrix- vector multiplication (the row vector
of the output from the preceding layer should be equal to the column
vector of the dense layer), to generate a m-dimensional vector.
5. The dropout layer is used to avoid over fitting by setting some of the
input neurons to 0 during the training process. The dense layer, on
the other hand, feeds all the outputs from the preceding layer to all
its neurons and perform the matrix- vector multiplication (the row
vector of the output from the preceding layer should be equal to the
column vector of the dense layer), to generate a m-dimensional
vector.
7. Next, the model is trained using the training dataset, by passing the
pre-processed images from the training dataset.
Page | 80
\
8. Finally, the predictions on the test data are done using the trained
model and the traffic sign name along with the class Id is shown as
an output.
Page | 81
\
Gray Scale:
Before Gray scaling, image data shape (of training data set) is: (34799,
32, 32, 3). This means that the images were of 32x32 size and coloured
in RGB format (3 channels). After Gray scaling, image data shape (of
training data set) becomes: (34799, 32, 32, 1). After Gray scaling the
size of the image remains the same (32x32) but the number of channels
is reduced to 1.
Page | 82
\
4. Data Augmentation:
Augment the training dataset with variations like rotation,
scaling, and changes in brightness to improve the model’s
robustness.
5. Transfer Learning:
Page | 83
\
8. Real-time Processing:
Optimize the system for real-time processing to ensure
timely recognition of traffic signs.
Page | 84
\
2. CNN Variants:
AlexNet: Introduced by Alex Krizhevsky et al., AlexNet was the
pioneering CNN architecture that achieved significant
performance improvements in the ImageNet Large Scale Visual
Recognition Challenge (ILSVRC) in 2012. It popularized the use
of deep CNNs for image classification tasks.
VGG (Visual Geometry Group)Net: VGGNet, developed by
the Visual Geometry Group at the University of Oxford, is
known for its simplicity and uniform architecture, consisting of
multiple convolutional layers followed by max-pooling layers.
It achieved high accuracy on various image classification tasks.
3. Object Detection Libraries:
Detectron2: A powerful object detection library built on
PyTorch, often used for CNN.
4. Image Preprocessing:
OpenCV (Open Source Computer Vision Library): Used for
image processing tasks, including resizing, normalization, and
augmentation.
Page | 85
\
9. Communication Protocols:
MQTT (Message Queuing Telemetry Transport): Used for
lightweight communication between the traffic sign recognition
system and other components.
12. Database:
SQLite, MySQL, or MongoDB: Storage for maintaining records
and data related to the recognized traffic signs.
Page | 86
\
Dataset Selection:
3 Identify existing datasets: Look for publicly available
datasets that contain images of traffic signboards along
with corresponding annotations. Datasets like German
Traffic Sign Recognition Benchmark (GTSRB) or
LISA Traffic Sign Dataset can be useful.
4 Custom dataset creation: If existing datasets do not
meet your requirements, you may need to create a
custom dataset by capturing images of traffic sign-
boards in the target environment. Ensure diversity in
lighting conditions, weather, and traffic scenarios.
1.Image Acquisition:
2.Annotation Process:
Page | 87
\
6. Data Preprocessing:
• Resize images to a consistent size: Standardize the image
dimensions to ensure uniformity during training.
• Normalize pixel values: Scale pixel values to a standard
range (e.g., 0 to 1) to facilitate training convergence.
Page | 88
\
Chapter- 3
FUNCTIONAL MODULES
Page | 89
\
3.1 Datasets:
The following five new images have been added to the model from
the internet:
A total of 39209 images are used in this model of which 31367 images
are taken for training the model and 7842 images were taken to test
the detection phase of the models. These traffic images were taken
from online sources with traffic sign having different viewing angle
and position on image. We have taken these dataset from German
traffic sign recognition benchmark.
Page | 90
\
Page | 91
\
4. Slippery road
Page | 92
\
6. Go straight or right
Page | 93
\
7. Go straight or left
8. Keep left
Page | 94
\
Page | 95
\
Page | 96
\
Page | 97
\
16. Yield
Page | 98
\
Page | 99
\
Page | 100
\
In India, road traffic signals are fundamental for ensuring the safety of
all road users and minimizing potential property damage. These
signals are categorized into three main types, each serving distinct
purposes:
Effective signage design and placement are essential for ensuring that
road users can quickly and easily interpret and respond to traffic
signals. Standardized sign shapes, colors, and symbols help create
consistency and clarity across road networks, facilitating safer and
more efficient traffic management. Moreover, periodic maintenance
Page | 101
\
Page | 102
\
The detection process begins with the original traffic image, where red
pixels are segmented using color thresholding. Subsequently, edge
detection is applied to the resulting mask. Following this, the Hough
transform is employed to detect circular shapes, identifying the center
and radius information crucial for cropping the candidate traffic sign.
The cropped image, containing the entire traffic sign along with the
red circular boundary, proceeds to the classification phase. A traffic
sign is deemed detected if the cropped image fully encompasses the
entirety of the sign. This methodical approach ensures accurate
detection and classification of traffic signs, contributing to enhanced
road safety and efficient traffic management. Additionally, by
incorporating multiple detection methods, the system improves
robustness and reliability in diverse environmental conditions, such as
varying lighting and weather conditions.
Page | 103
\
Page | 104
\
Page | 105
\
Page | 106
\
Page | 107
\
Page | 108
\
Chapter-4
ANALYSIS MODELING
Page | 109
\
1.Dynamic Beahviour
Page | 110
\
3. State Variables:
2. Economics:
Page | 111
\
1. Mathematical Models:
Page | 112
\
2. Agent-Based Models:
5. Simulation Models:
Page | 113
\
2. Collect Data:
3. Develop Relationships:
Page | 114
\
over time.
1. Actors:
Types:
2. Use Cases:
Page | 115
\
3. System Boundary:
4. Relationships:
Identify Actors:
3 Determine all the external entities that will interact with the
system. Consider both primary and secondary actors.
Page | 116
\
• Identify and list the key functionalities that the system must
provide. Each use case should represent a goal that an actor
wants to achieve.
1. Establish Relationships:
1. Clear Communication:
2. Requirement Clarification:
3. Scope Definition:
Page | 117
\
expectations.
4. System Design:
5. Documentation:
Page | 118
\
Page | 119
\
Define the specific scenario or use case you want to model. Clearly
describe the process and the goal of the interaction.
2. Identify Participants:
List all the objects and actors involved in the scenario. Each
participant will have a lifeline in the diagram.
3. Draw Lifelines:
4. Add Messages:
5. Include Activations:
Page | 120
\
Page | 121
\
Page | 122
\
• Identify Activities:
5.Add the Final Node: Place the final node at the end of
the process.
Page | 123
\
Page | 124
\
Page | 125
\
Page | 126
\
Page | 127
\
Chapter 5
SYSTEM DESIGN
Page | 128
\
Page | 129
\
2. Dataset Splitting:
3. CNN Model:
4. Transfer Learning:
Page | 130
\
5. Training:
7. Post-processing:
8. Alert Generation:
Page | 131
\
Page | 132
\
2. Preprocessing Module:
Page | 133
\
Page | 134
\
and identification.
Page | 135
\
Page | 136
\
Chapter-6
RESULTS
Page | 137
\
Page | 138
\
Epoch:
After running the Adam optimiser on 50 epochs here are the results:
Page | 139
\
Graph:
Here we are going to draw Accuracy graph (accuracy vs epochs graph)
and Loss graph (loss vs epochs graph), after using adam optimizer.
Accuracy graph:
Page | 140
\
Loss Graph:
Page | 141
\
Confusion matrix:-
Page | 142
\
PREDICTED LABELS
Page | 143
\
Class: No passing
TP: 280
FP: 2
TN: 7559
FN: 1
Page | 144
\
Class: Yield
TP: 424
FP: 4
TN: 7408
FN: 6
Class: Stop
TP: 164
FP: 0
TN: 7676
FN: 2
Class: No vehicles
TP: 116
FP: 17
TN: 7708
FN: 1
Class: No entry
TP: 226
FP: 2
TN: 7614
FN: 0
Page | 145
\
Page | 146
\
Class: Pedestrians
TP: 44
FP: 2
TN: 7793
FN: 3
Page | 147
\
Page | 148
\
Page | 149
\
Page | 150
\
The next change we did was using a different value for optimiser
(Adamax) and the results we observed are as follows:
Result:
Page | 151
\
Graph:
Here we are going to draw Accuracy graph (accuracy vs epochs graph)
and Loss graph (loss vs epochs graph), after using adamax optimizer.
Accuracy graph:
Page | 152
\
Loss graph:
Page | 153
\
Confusion matrix:
Page | 154
\
TP: 451
FP: 3
TN: 7388
FN: 0
Class: No passing
TP: 281
Page | 155
\
FP: 2
TN: 7559
FN: 0
Class: Yield
TP: 429
FP: 1
TN: 7411
FN: 1
Class: Stop
TP: 166
FP: 1
TN: 7675
FN: 0
Class: No vehicles
TP: 116
FP: 1
TN: 7724
FN: 1
Page | 156
\
TN: 7751
FN: 1
Class: No entry
TP: 226
FP: 1
TN: 7615
FN: 0
Page | 157
\
FN: 1
Class: Pedestrians
TP: 47
FP: 0
TN: 7795
FN: 0
Page | 158
\
Page | 159
\
Page | 160
\
Page | 161
\
Input - C:\Users\harsh\Desktop\traffic-signboard-
main\test\01958.png
Output 1:-
Page | 162
\
Input - C:\Users\harsh\Desktop\traffic-signboard-
main\test\02001.png
Output 2:-
Page | 163
\
Input - C:\Users\harsh\Desktop\traffic-signboard-
main\test\00058.png
Output 3:-
Page | 164
\
Input - C:\Users\harsh\Desktop\traffic-signboard-
main\test\00003.png
Output 4:-
Page | 165
\
Input - C:\Users\harsh\Desktop\traffic-signboard-
main\test\00100.png
Output 5:-
Page | 166
\
Page | 167
\
Page | 168
\
Page | 169
\
Page | 170
\
Page | 171
\
Chapter -7
Page | 172
\
Page | 173
\
Page | 174
\
7.1 Advantages:
Page | 175
\
7.2 Limitations:
• Fixed Input Size: CNN models typically have fixed input sizes,
which might pose challenges when dealing with images of
varying resolutions. This limitation can affect the system’s
adaptability to different camera setups or environments.
Page | 176
\
Future work will focus on several key areas to enhance and extend
the capabilities of our model. We plan to implement our model on
various other datasets to ensure its robustness and generalizability.
Additionally, we will aim to achieve high accuracy with our
lightweight model across these different input datasets, reinforcing
its effectiveness and adaptability. Moreover, we will focus on
improving our model's ability to capture real-time video streams,
providing both voice and display alerts to drivers. This will involve
enhancing the model's real-time processing capabilities and
ensuring that the alerts are timely and accurate, further contributing
to road safety and the practicality of our solution in real-world
driving scenarios. Ability to recognize a wide range of traffic signs
accurately.
Page | 177
\
CHAPTER-8
BIBLIOGRAPHY
Page | 178
\
Books:
Page | 179
\
Articles:
Page | 180
\
Page | 181
\
APPENDIX
Page | 182