Article citation information:
Parse, M., Pramod, D. Edge detection technique based on
bilateral filtering and iterative threshold selection algorithm and transfer
learning for traffic sign recognition. Scientific Journal of Silesian
University of Technology. Series Transport. 2023, 119, 199-222. ISSN: 0209-3324. DOI: https://doi.org/10.20858/sjsutst.2023.119.12.
Milind PARSE[1],
Dhanya PRAMOD[2]
EDGE DETECTION
TECHNIQUE BASED ON BILATERAL FILTERING AND ITERATIVE THRESHOLD SELECTION
ALGORITHM AND TRANSFER LEARNING FOR TRAFFIC SIGN RECOGNITION
Summary. The traffic
sign identification and recognition system (TSIRS) is
an essential component for autonomous vehicles to succeed. The TSIRS helps to collect and provide helpful information for
autonomous driving systems. The information may include limits on speed,
directions for driving, signs to stop or lower the speed, and many more
essential things for safe driving. Recently, incidents have been reported
regarding autonomous vehicle crashes due to traffic sign identification and
recognition system failures. The TSIRS fails to
recognize the traffic signs in challenging conditions such as skewed
signboards, scratches on traffic symbols, discontinuous or damaged traffic
symbols, etc. These challenging conditions are presented for various reasons,
such as accidents, storms, artificial damage, etc. Such traffic signs contain
an ample amount of noise, because of which traffic sign identification and
recognition become a challenging task for automated TSIRS
systems. The proposed method in this paper addresses these challenges. The sign
edge is a helpful feature for the recognition of traffic signs. A novel traffic
sign edge detection algorithm is introduced based on bilateral filtering with
adaptive thresholding and varying aperture size that
effectively detects the edges from such noisy images. The proposed edge
detection algorithm and transfer learning is used to train the Convolutional
Neural Network (CNN) models and recognize the traffic signs. The performance of
the proposed method is evaluated and compared with existing edge detection
methods. The results show that the proposed algorithm achieves optimal Mean
Square Error (MSE) and Root Mean Square Error (RMSE) error rates and has a better Signal to Noise Ratio
(SNR) and Peak Signal to Noise Ratio (PSNR) ratio
than the traditional edge detection algorithms. Furthermore, the precision
rate, recall rate, and F1 scores are evaluated for
the CNN models. With the German Traffic Sign Benchmark database (GTSRB), the proposed algorithm and Inception V3 CNN model gives promising results when it receives the
edge-detected images for training and testing.
Keywords: bilateral
filtering, edge detection, transfer learning, traffic sign identification and
recognition
1. INTRODUCTION
Traffic signs play a significant role in
creating a secure travel environment and controlling the flow of vehicles
running on the road worldwide. Traffic signs are the visual representation of
regulations, warnings, and information, which is beneficial for drivers to
drive cars safely on roads [1]. Traffic signs have different shapes and provide
some vital information to drivers. The commonly used forms are circles,
triangles, hexagons, squares containing numbers, text or images, or icons [1,
2]. Over the past decade, powering the automotive industry with artificial
intelligence-enabled technologies has been gaining the attention of
academicians and industries. Increased awareness of road safety and the global
inclination towards adapting more intelligent technology have encouraged
researchers to focus on automated solutions for recognizing traffic signs under
different challenging conditions and developing appropriate and timely
responses according to the situation at hand to avoid road accidents and save
people's lives. Even future smart cars can take the input from TSRIS and adjust the driving to prevent fatalities on the
road [3-5]. Image processing involves several steps that are carried out to
extract meaningful information from the digital images for analysis. Typically,
these steps are acquiring, importing, manipulating, and analyzing the image.
When an image is processed, its results could be helpful information about parameters,
features, or a new resulting image. The resulting image or information helps
build specific practical industrial applications [6]. The edge in an image is a
continuous set of pixels. It is also denoted using the change in pixel
intensity across two disjoint regions. Edge detection is a segmentation method.
With the help of brightness, sharpness, or pixel intensity changes, one can
classify or identify continuous or discontinuous regions in the image [7]. Over
the past years, researchers have developed several edge detection
algorithms/operators/methods like Robert, Sobel, Prewitt, and Canny. These
methods fall into two categories: Gradient and Gaussian [8-10]. An image may
contain noise introduced for various reasons such as lighting conditions, dust
on the camera, images taken during rain, objects in the background, etc. The
different types of noise in an image are Gaussian, poison, impulse, salt and
pepper, and speckle noise. Thus, edge detection in images becomes a more
demanding and challenging job [4]. However, images can be denoised
using filters and a few other techniques. Nevertheless, excessive filtering for
denoising images can reduce the edge strength [11,
12]. In the literature, many
methods are available that accurately detect edges in the image. A fast edge
computing algorithm introduced for the first time was Robert's algorithm. To
detect the sharp edges, Canny algorithm is developed
and most widely used in image processing research applications [8-10, 13, 14].
In the literature, numerous studies are available that employ neural
network-based classifiers to recognize the traffic sign. The proposed method
can be used for feature extraction in conjunction with neural network-based
traffic sign detection. Edge images are proved to produce better results as
compared to the scenario where unprocessed images are provided as input to the
classification model. The popular machine learning library, Keras
has various pretrained models which are trained on
ImageNet dataset having 1000 classes of images. In this study researchers have
used two such pretrained models, VGG-16
and Inception V3 for the performance evaluation.
Weights of VGG-16 and Inception V3
models are fine-tuned for traffic sign recognition task using transfer learning
technique. Transfer learning has gained the attention of researchers in
recently as it saves the time required to train and validate the dataset. The
significant contributions of this study include identifying the suitability of
existing techniques to detect the edges of images that contain skewed
signboards, scratches on traffic symbols, and deformed or discontinuous traffic
signs in Indian traffic sign images. An image dataset is created by capturing
traffic sign images using a mobile phone camera. A new technique is implemented
based on bilateral filtering, adaptive thresholding, and varying aperture size.
Then the proposed algorithm is combined with two different CNN models, and the
results obtained on the GTSRB dataset are presented and evaluated. Also, the
proposed edge detection method’s performance is evaluated and compared
with existing techniques on quality metric parameters.
2.
LITERATURE REVIEW
This section describes the popular
edge detection techniques and relevant work available in the literature.
2.1. Edge detection technique
The edge detection techniques
identify the true and false edges along the boundary of the image. The Gradient
and Second derivative operators are the two types in which all the edge
detection algorithms are classified [8-10]. Some of the extensively used edge
detection are as follows:
2.1.1. Roberts edge detection operator
Lawrence Robert, in 1963 introduced
Roberts cross operator for edge detection. It is a differential operator that
approximates an image's gradient via discrete differentiation, which is
accomplished by calculating the sum of the squares of the differences between
diagonally adjacent pixels. The Roberts edge detection operator quickly
calculates a 2-D spatial gradient of an input image in a straightforward. It
highlights the strong spatial gradient zones that represent the corresponding
edges. This operator takes grayscale images as input and produces grayscale
edge image output. The magnitude of the input image's spatial gradient at a
particular position is represented by the pixel value at each place in the
output image. Unfortunately, it is susceptible to noise and inaccurate edge
detection [8-10, 13].
2.1.2. Sobel edge detection
The Sobel edge detection method is a
widely used technique where image processing is done separately in the X and Y
direction. Then a new image is formed that represents the sum of the edges in X
and Y directions in the processed image. Commonly, this method encounters an
issue of noise in the final processed image. To reduce the noise in the image,
one can use an averaging filter for smoothening the image and then apply the
Sobel operator again and compare the differences [8-10, 13].
2.1.3. The Prewitt edge detection
Judith M. S. Prewitt developed the Prewitt
operator to detect an object's edge in an image. Prewitt operator detects
horizontal edges along the x-axis and vertical edges on the y-axis. An edge is
detected whenever there is a change in the intensity of the pixel; hence,
differentiation is used to calculate the edge. The local maxima or minima
represent the edge in the Prewitt mask result. The first order and second order
derivatives in the Prewitt masks have the following properties:
- Maximum weight means more chances of the detection of
edge.
- Both (+ and -) signs should be present in the mask.
- The sum of the mask values must be equal to zero.
Prewitt operator provides us with
two masks, one for detecting edges in the horizontal direction and another for
detecting edges in a vertical order [8-10, 13].
2.1.4. Canny edge detection
It is a popular edge-detection
method today because it is robust and flexible. The algorithm itself follows a
three-stage process for extracting edges from an image. Then, add image
blurring, a necessary pre-processing step to reduce noise. This makes it a
four-stage process, including Noise Reduction, Calculating the Intensity
Gradient of the Image, Suppression of False Edges, and Hysteresis Thresholding
[8-10, 13].
2.2. Relevant works
Yu Z., Feng C., Liu M.Y., and Ramalingam S. (2017)
addressed the category awareness problem in semantic edge detection. Unlike
traditional frameworks, which use multiple classes, the authors developed a
learning framework based on multiple labels that improve edge detection. The
proposed CASENet architecture improves DSN architecture. The feature extraction layer replaces the
bottom-side classification module. Supervision is imposed at the top layer in
the network. Finally, sliced concatenation is replaced by shared concatenation
[15].
Liu Y. and et al., (2017) presented
a CNN-based richer convolutional features (RCF)
framework that uses a hierarchy of image features for edge detection. Here, any
size image is taken as input, and the method produces the same size output
image with detected edges. This framework utilizes fine details, semantic
features, and fine details of images. The RCF
framework is also suitable for other computer vision problems, as it provides
efficient and accurate results. Furthermore, when applied to classical
segmentation, the RCF offers promising results [16].
Lakhani K., Minocha
B., & Gugnani N. (2016) analyzed
and evaluated edge extraction techniques on dental X-ray images. The authors
addressed the problem of misaligned images. Sobel and Prewitt's results showed
a notable difference between the original and processed images when the authors
applied a Gaussian filter. Similarly, smoothening and sharpening images help
diagnose dental diseases and medical X-ray images [17].
Srujana P., Priyanka J., Patnaikuni V.S., & Vejendla N. (2021) in their research work, used Sobel,
Prewitt, Roberts, Log, and Canny operators to find the edges of images. The
authors used building and flower images for edge detection using all methods.
The authors found the Canny edge detector as a suitable method for detection
when the results of all of these methods were compared using the parameters
like SNR, PSNR, and MSE [18].
Halder A., Bhattacharya P., Kundu A. (2019) used Richardson's extrapolation technique
to detect the edge and quantify the edge strength of neighbouring pixels. The
extrapolation is used for numerical analysis and to compute the convergence
rate. This technique is also used to identify the respective edges of a binary
image [19].
Srinithyee S.K., Srivarsha E., Priyadharsini R.,
& Beulah A. (2021) came up with an augmented edge detection technique that
skipped the smoothing process to reduce the computational time. In the proposed
method, Sobel operator is modified, and the Kernal is
made up of a fractional calculus-based mask instead of the classical one. As a
result, the algorithm can work with any image and resist noise. The order of
the fractional calculus decides the quality of the output image. It is measured
in Pratt's Figure of merit [20].
Carmelo M., Pierce S.G., Summan R. (2019) introduced
a novel filtering method based on Spatial FFT (Fast
Fourier Transform) for edge reconstruction and a boundary point detection
algorithm. The proposed FFT-based edge reconstruction
helps eliminate the polynomial curve fitting problem. Independent of threshold
values, the Boundary Points Detection (BPD) algorithm detects boundary points
using the maximum number of points on the region of interest in boundary points
and internal points. The proposed algorithm provides satisfactory results when
less than 75% noise [21].
Zheng Z., Zha
B., Yuan H., Xuchen Y., Gao
Y., and Zhang H. (2020) have implemented an improved model based on grey pixel
values in neighboring pixels and a grey prediction
method to remove noise and edge localization problems. The model employs 24
directions on the image edge to extract texture features of the edge in the
image. To obtain the enhanced edge, the model replaces the values of pixels in
the original image with the predicted value of the grey model. Through the
global iteration method, the model selects the adaptive threshold and then
removes the unwanted points and blur in the image [22].
Shi J., Jin
H., Xiao Z. (2020) proposed a fusion of 'polarimetric
constant false alarm rate' and 'weighted gradient-based detector to remove
speckle noises and detect weak or strong edges in synthetic aperture radar
(SAR) images. The Constant False Alarm Rate (CFAR)
detector based on KK distribution is proposed to
detect weak images, and the weighted gradient-based detector detects the
changes in the intensity of strong pixels across heterogeneous areas [23].
Xizhen S., Wei Z., Yiling
G., and Shengyang Y. (2021) came up with a solution
to the problems associated with the Canny operator,
such as poor adaptive ability and image edge determination in noisy images.
First, the authors de-noise the images by improving the filtering method and
computing the gradient amplitude using a 4-direction template. Then, Otsu's
inter-class algorithm combines the image blocks to obtain the low and high
threshold values. Finally, the experiment uses leaf images with complex
backgrounds [24].
Bausys R., Kazakeviciute-Januskeviciene
G., Cavallaro F., Usovaite
A. (2020) worked on solving the task of detecting and monitoring the region of
interest from satellite images. The proposed method adaptively selects an edge
detection algorithm based on a decision matrix. Also, the authors used the
multi-criteria decision-making method. Finally, the authors compared the
results of the several edge detection operators, and the Canny
edge detector was more accurate and applicable for most of the images. However,
it is difficult to predict which edge detector is more appropriate for a
particular image content [25].
Dhillon D., and Chouhan,
R. (2022) discussed a unique way of finding the best-connected edges and noise
removal from the input image. The authors modified the Canny
operator that accepts the same parameters but provides better-connected edges
using the Stochastic Resonance threshold and window mapping. The proposed
method has a limitation. It cannot work without a threshold value; however, the
authors suggest using automatic thresholding techniques like Otsu's method. The
algorithm performs well on the Barcelona Images for Perceptual Edge Detection
Dataset (BIPED) benchmarking dataset [26].
How D.N.T.,
Sahari K.S.M., Hou Y.C., and Basubeit
O.G.S. (2019) applied transfer learning to the
Malaysian Traffic Sign dataset, which contained five different classes using
the different deep CNN pre-trained model. For most deep CNN models, the authors
got an accuracy above 90%. Among all the models, the highest accuracy was
obtained by the DenseNet169 pre-trained model, which
is 98.33% [30].
Palavanchu S. (2021) has explored a few
transfer learning models in the Keras library to
detect, recognize and classify the GTSRB traffic
signs. The author achieved 95.04% accuracy and 0.2311 loss value for Xception network compared to InceptionV3
Networks, Residual Networks ResNet50, VGG-16, and EfficientNetB0 models
[31].
2.3. Limitations of existing edge detection
techniques
Roberts's edge detection method is
sensitive to noise because it uses a small kernel, and unless the edges are
sharp, its detection of edges is very poor. The Soble
operator produces inaccurate results with an increase in noise, and hence
degrades the edge magnitude. The Prewitt method is also noise sensitive and
cannot accurately detect edges with thin and smooth. However, most research
proves that Canny edge detection is better than the
other three methods. Still, it has greater computational complexity, higher
time consumption, and is sometimes inaccurate in edge detection [8-10, 13].
3. PROPOSED WORK
The primary objective was to find
the suitability of existing edge detection techniques on the images that
present challenges in detecting the traffic symbol, such as skewed signboards,
scratches on traffic symbols, and discontinuous or damaged traffic symbols.
Initially, the four popular edge detection techniques, viz. Roberts, Sobel,
Prewitt, and Canny are implemented and tested [8-10, 13]. A novel Edge
detection technique based on bilateral filtering [27-29] with adaptive
thresholding [6] and aperture size is proposed. The algorithm contains several
steps which are discussed in subsequent sections. First, the algorithm makes
the slide adjustment to select the aperture size, the maximum, and the lower
threshold to get the object edges. The aperture size and threshold values are
iterated until the continuous edge of the symbol is obtained. The results are
presented, and the performance of the proposed algorithm with the four existing
algorithms is evaluated using SNR, PSNR, MSE, and RMSE parameters [14].
Then the proposed algorithm is used to pre-process the German Traffic Sign Benchmark
dataset and using transfer learning with Inception V3
and VGG-16 the traffic sign recognition task is
carried out. The image classification results are compared against F1 score, Precision and Recall metrics.
3.1. Dataset and methods
For this research, an image dataset
was created that contains 1000 images of traffic signboards captured using a
mobile phone. These images are randomly captured using a mobile phone on roads,
presenting many challenges in recognizing the traffic symbol [4]. Some challenges
include varying lighting conditions, color variance,
contrast, occlusion by objects like poles, electricity
lines, trees, leaves, improper shape, etc. [4, 29]. This dataset is mainly used
to test the proposed bilateral filtering-based edge detection algorithm.
Secondly, the proposed algorithm and
transfer learning on VGG-16 and InceptionV3
applied to the German Traffic Sign Benchmark dataset. The German Traffic Sign
Benchmark dataset has 50,000 images consist of 43 categories of traffic symbols
and is widely used by researchers worldwide for traffic sign recognition tasks.
The German Traffic Sign Benchmark dataset is used to find how the proposed edge
detection algorithm works on such a standard dataset.
3.2. Proposed edge detection method
The proposed method is based on
bilateral filtering, adaptive thresholding, and aperture size. The explanation
of important concepts involved in the proposed method is discussed below.
3.2.1. Reducing noise using bilateral filtering
The edge detection process is noise
sensitive and requires the image to be smoothened. Image convolution and
Gaussian Kernel are widely used for smoothening the images [27]. The size of
the Kernel is responsible for deciding blur visibility [28-29]. The blur
visibility decreases with lesser kernel size. The Gaussian filter with kernel
size (2k+1)×(2k+1) is given by
equation (1).
Hij =
Using Gaussian filters helps to
reduce noise to some extent, but they still present a challenge in detecting
object edges accurately. Whereas, the bilateral filter has a spatial kernel and
range kernel, which are more effective in noise removal [24]. In bilateral
filtering, the, spatial kernel uses the Gaussian function given in equation 1
to smoothen the image. The spatial kernel describes the distance dependence
between the image pixels (Euclidean distance). In contrast, the Range Kernel
gives the intensity similarity between two image pixels [26-27].
Thus, the spatial kernel defines the
spatial proximity measurement. The spatial kernel (P𝜎𝑠) is given by equation (2).
P𝜎𝑠 = exp (−(‖a−b‖2)2𝜎2𝑠)
(2)
While the Range Kernel (K𝜎𝑟) defines the weights of pixels with the
intensity difference in size pixel under consideration with the image's centre
pixel, each pixel range kernel is given by the equation (3).
K𝜎𝑟 = exp (−(|𝐼(a)−𝐼(b)|2)/2𝜎2𝑠 )
(3)
Thus, the Bilateral Filter BFb
can be defined using equation (4) and the normalization factor in equation (5)
BFb=
Fn = Σ b ∈
𝑁 (𝑝) K𝜎𝑠 ((‖a − b‖) K(|𝐼a −
𝐼b|)
(5)
Where,
I is a Grey level image
N(p) represents neighbouring pixel p
Ia
and Ib give
the pixel intensities
Fn
is the factor of normalization that ensures pixel weights sum is 1.
3.2.2. L2 gradient
The L2
Gradient is calculated to find pixel intensity and edge direction in the image.
Edge connectivity is decided by the change in the intensity of pixels. The
bilateral filter is applied to highlight the pixel intensity changes along the
horizontal X-axis and vertical Y-axis [27]. Image smoothing is achieved using
the Spatial kernel and Range kernel as per equation
(2) and equation (3). The magnitude of the Gradient is given by equation (6),
and slope θ is given by equation (7).
The resulting image may contain a
few thick edges, non-max suppression is used to make them thinner. In addition,
the gradient intensity levels are non-uniform, meaning that they may be between
0 and 255. Therefore, the resulting image should have the same intensity across
the edges to form a connected edge. To overcome the non-uniformity of intensity
levels, double thresholding is used [27-29].
3.2.3. Double thresholding
Some level of intensity for an image
pixel is defined using thresholding.
Mainly, it is used to separate the edge pixels from the background
pixels. Several types of thresholding techniques are available in image
processing [6]. To overcome the problem of the non-uniform pixel intensity,
double thresholding technique is more effective, and hence it is used in this
experiment. In double thresholding, strong, weak, and irrelevant image pixels
are found:
- Strong pixels have higher intensity and they
contribute to edge formation.
- The pixels whose intensity values are not high enough
to consider as strong and not too small enough to consider as non-relevant to
form the edge are regarded as weak pixels.
- Pixels that are not strong or not weak are considered
non-relevant pixels, and these do not contribute to edge formation.
Thus, in double thresholding, the
following are determined:
- Decide the value for the high threshold to determine
the strong pixels, i.e., the pixels having a higher intensity value than the
high threshold value.
- Decide the value for the low threshold to determine
the non-relevant pixels, i.e., the pixels having a lower intensity value than
the low threshold value.
- The pixels having intensity values between high
threshold and low threshold are considered as weak pixels.
Thus, there are two threshold
characters ST= High threshold and WT=Low threshold. To get the edge region
when the pixel's grey intensity value is greater than ST, i.e., it represents a strong edge pixel. A non-edge area is
obtained when the pixel's grey intensity level is less than WT, i.e., it represents the weak edge
pixel. The neighbouring pixel’s grey intensities are considered to
determine the strong edge or weak edge pixel when the grey intensity value lies
between ST and WT. The result of double thresholding produces an image with
two-pixel intensity values, i.e., strong and weak.
3.2.4. Iterating threshold values
To ensure the intensity levels,
threshold values are iterated. Then the L2 Gradient
is calculated for the iterated values of the lower threshold and upper
threshold each time. For example, the lower threshold from 100 pixels to 700
pixels is repeated and the value of the upper threshold from 200 pixels to 500
pixels with an iteration of 10 pixels, concerning aperture, sizes 3, 5, and 7
each time. In general, determining the optimal threshold value becomes
difficult with images containing noise. To reduce the noise interference,
threshold is selected by iterating the values for the aperture sizes [33].
Suppose the original image with noise is given by equation (8)
I (p,q) = f(p,q)
(8)
then the algorithm has the following specific
steps.
3.3. Algorithm
The specific steps in the algorithm
are listed below.
(1)
Apply the
bilateral filter on the input image to smoothen the image so that sharp edges
are detected and preserved, which reduces the false edge detection probability.
(2)
Find the L2 Gradient and slope using equations (6) and (7).
(3)
Segment the image
into two parts, I1 and I2,
using a set threshold; to get the image parts as shown in equations (9) and
(10).
I_1 (p,q)=f_1 (p,q) (9)
I_2 (p,q)=f_2 (p,q) (10)
(4)
Obtain the grey
values for I1 and I2 using
equations (11) and (12).
E{I_1
(p,q)}=E{f_1 (p,q)+e(p,q)}
(11)
E{I_2
(p,q)}=E{f_2 (p,q)+e(p,q)}
(12)
(5)
Calculate the
iterative loop. The initial threshold value is given by equation (13).
Where,
T represents the initial value of threshold
Gmax is maximum grey level value
Gmin is minimum grey level value
(6)
For 'N' iterations, calculate ST and WT. Divide the image into ST
and WT such that ST is greater than the initial threshold and WT is less than the initial value of the threshold, which is given
by equations (14) and (15).
ST={f(p,q)|f(p,q)>T}
(14)
WT={f(p,q)|f(p,q)<T}
(15)
(7)
Calculate the
upper (Tu)
and lower (Tl) threshold values using
equations (16) and (17).
Where f(p,j) gives the Grey level value of
image I(i,j)
for the pixel points N_0(i,j) and N_1(i,j) that
satisfies the following conditions given in equations (18) and (19).
N0(i,j)=
N1(i,j)=
(8)
The iteration
continues until the conditions are satisfied and achieve the initial value of
the threshold. Once the iteration terminates, the optimal double threshold
values Tu
and Tl are obtained.
4. RESULTS
The effectiveness of the proposed
edge detection algorithm is evaluated by comparing its output with that of
other popular edge detection techniques, namely Prewitt ,
Sobel, Roberts and Canny. The proposed edge detection algorithm is then
combined with VGG-16 and Inception V3 models to evaluate the improvement resulting in final
traffic sign classification activity.
4.1. Edge detection using the proposed
algorithm
A dataset of 1000 images was created
which were captured using a mobile phone camera. The major focus of this work
is to detect the edges of traffic symbols on signboard images in which the
traffic signboard is skewed, scratched and the traffic sign or board is
deformed. Also, most of these images contain trees or other objects in the
background. Figure 1. (a), (b), (c), and (d) shows a few images that are used
as input for the proposed algorithm.
Figure 2 (a-d) to Figure 6 (a-d)
shows the results of the existing edge detection algorithms and the proposed
traffic sign edge detection algorithm. The Prewitt and Sobel methods behave
differently with the curved edges and noise. Both methods are less effective in
handling the noise, background objects, and curved edges. Comparatively,
Roberts and Canny Edge detectors can effectively detect the traffic symbol
edges. Also, both methods do not deal with background noise effectively. The
proposed method is more effective in handling the background noise and
detecting the edges of traffic symbols in traffic sign images of signboards
having scratches on the symbol, deformed and Skewed signboards.
(a)Scratches on Sign Board |
(b)Skewed Sign
Board |
(c)Skewed Sign
Board |
(d) Deformed Sign Board |
|
|
|
|
Fig 1. Input Image – Indian
Traffic Signs (a-d) |
|||
|
|
|
|
Fig 2. The output of the Prewitt Method
(a-d) |
|
|
|
|
Fig 3. The output of the Sobel Method
(a-d) |
|
|
|
|
Fig 4. The output of Roberts Method
(a-d) |
|
|
|
|
Fig 5. The output of the Canny Method
(a-d) |
|
|
|
|
Fig 6. The output of the proposed method (a-d) |
4.2. Combining
proposed edge detection technique with CNN for traffic sign recognition
The task of traffic sign recognition
is carried out using the transfer learning method. Transfer learning allows
using a pre-trained model to solve a similar problem by certain fine-tuning
parameters across the model layers and learning from new data. The images in GTSRB are categorized across 43 classes and resized to 224
x 224 pixels. All data is randomly divided into the training set and testing
with the ratio of 80:20, i.e., 80% for training and 20% for testing. The
training set is used for model training and parameter learning. The model is
tested in the training process, and the network fine-tuning is performed
according to the model test results. The test set is used to test the model's
identification capability and the model's generalization ability. In this work,
the two pre-trained CNN models VGG-16 and Inception V3 are used. The fully-connected layers are fine-tuned, and
the output layer is reconfigured for the traffic sign recognition task on the
same CNN models.
Tab. 1
Results before edge detection
Training
accuracy |
Validation
accuracy |
Testing
accuracy |
|
VGG-16 |
93.04% |
93.53% |
91.78% |
Inception
V3 |
96.81% |
98.81% |
98.67% |
Later, the input images are
pre-processed using a proposed edge detection algorithm to transform the
original images into edged images, the pre-processed images are provided as
input to the CNN models. Weights of the convolutional layers are pre-trained on
ImageNet. The fully-connected layers are fine-tuned, and the output layer is
reconfigured for the traffic sign recognition task on the same CNN models; the
following results are obtained.
Tab. 2
Results after edge detection
Training accuracy |
Validation accuracy |
Testing accuracy |
|
VGG-16 |
96.36% |
97.21% |
94.32% |
Inception V3 |
99.40% |
99.03% |
99.49% |
Comparing the results in Table 1 and
Table 2 shows that the same model gets higher training and testing accuracy if
the original images are replaced with the edge images. Though edge images are
obtained from original images, they give different information to the Deep CNNs
model. More specifically, it is observed that the edge images have more than
99% accuracy because they focus on the inner edge and shape of the object,
which is more distinguishable. The below section describes the results of the
Inception V3 model in detail. Even in the case of VGG-16 model, when the edge detection algorithm is applied
in pre-processing the model performs better and achieves higher accuracy as
compared to the model which takes the GTSRB images
directly without pre-processing.
4.2.1. Inception V3
implementation without edge detection
As shown in Figure 7, there are 43
different classes of traffic signs in the German traffic sign benchmark
dataset. As shown in Figure 8, it is observed that during the training, the
loss continues to decrease, and accuracy continues to increase. Similarly, the
loss continues to fall after each epoch, and validation accuracy continues to grow after each epoch when the Inception
V3 model is trained and validated using the images in
the GSTB dataset. The training accuracy of 96.81% is
achieved after 10 epochs, with a loss of 0.66. at the
same time, the validation accuracy is 98.81%, with a loss of 0.16. Figure
9(a-b) shows the respective graphs for Training and Validation loss, Training
and Validation Accuracy on Inception V3 model without
edge detection.
Comparing the results in Table 1 and Table 2
shows that the same model gets higher training and testing accuracy if the
original images are replaced with the edge images. Though edge images are
obtained from original images, they give different information to the Deep CNNs
model. More specifically, it is observed that the edge images have more than
99% accuracy because they focus on the inner edge and shape of the object,
which is more distinguishable.
The below section describes the results of the
Inception V3 model in detail. Even in the case of VGG-16 model, when the edge detection algorithm is applied
in pre-processing the model performs better and achieves higher accuracy as
compared to the model which takes the GTSRB images
directly without pre-processing.
Fig. 7. Input image classes of GSTB dataset |
Fig. 8. Training and validation loss, training and validation accuracy on
Inception V3 model without edge detection |
Fig. 9. Training
and validation loss, training and validation accuracy on
inception V3 model without edge detection
4.2.2. Inception V3 implementation
with edge detection
Figure 10 shows the edge detected
images. All the images of the German traffic sign benchmark dataset are first
pre-processed using the proposed edge detection algorithm. As shown in Figure
11, it is observed that during the training, the loss continues to decrease,
and accuracy continues to increase. Similarly, the loss continues to fall after
each epoch, and validation accuracy continues to grow after each epoch during
training and validation of the Inception V3 model using
the images in the GSTB dataset. The training accuracy
achieved is 99.40% after 10 epochs, with a loss of 0.34. at
the same time, the validation accuracy is 99.03%, with a loss of 0.16. Figure
12(a-b) shows the respective graphs. For the test dataset, the optimal test
loss is 0.11 and test accuracy is 99.49% as shown in Figure 13. Figure 14(a-b)
shows the graphs for training loss, test loss, training accuracy and test
accuracy. The model accuracy improved when the edge detected images were input
to the model during training, validation, and testing.
Fig. 10. Input image classes of GSTB dataset with
edge detected |
Fig. 11. Training and validation loss, training and validation accuracy of
inception V3 model with edge detection |
Fig. 12. Training and validation loss,
training and validation accuracy of
inception V3 model with edge detection
Fig. 13. Train and test loss, train test
accuracy of inception V3 model with edge detection
Fig. 14. Train and test loss, train accuracy
and testing accuracy of
inception V3 model with edge detection
5. PERFORMANCE EVALUATION
The performance efficiency of edge
detection of the proposed Traffic Sign Edge Detection algorithm is evaluated on
parameters like Signal to Noise Ratio, Peak Signal to Noise Ratio, and Mean Square Error.
5.1. Mean square error (MSE)
MSE is one of the popular quality assessment
metrics used to compare the image edge detection quality. The smaller value of MSE indicates
that the edge detection method performs better.
It can be calculated by evaluating the difference between the original image
and the edge detected image, as given in equation (20).
MSE =[f1(x,y)-f2(x,y)]
(20)
Where f1 is the original image and f2 is the edge
detected image.
5.2. Root mean square error (RMSE)
RMSE is also a quality assessment metric similar to
MSE. The
following formula shown in equation (21) is used to calculate the value of RMSE.
RMSE =
5.3. Signal to noise ratio (SNR)
SNR is a
quality metric used to calculate the false switching of pixels while estimating
an edge. It is widely used to compare the performance of different methods of
segmentation and edge detection. It represents the level of noise in detected
edges [35, 36]. The higher SNR value
indicates the sharpness of the edge. To evaluate the proposed method, SNR is calculated using equation (22).
SNR =
Where f1 is the original image and f2 is the edge
detected image. The image size is M×N. The rows and columns are given by x and y respectively.
5.4. Peak signal to noise ratio (PSNR)
PSNR describes how the noise affects the pixel
quality. It calculates the ratio between the max value of a pixel and the
noise. It is given on a logarithmic decibel scale. The higher value of PSNR generally
indicates that the error is less but, in some cases, like edge detection, the
lesser value of PSNR
shows the accurate edge detection [36, 37]. Equation (23) gives the value of PSNR.
PSNR = 10 log
Table 3 to Table 6 shows the
statistical comparison of the proposed method with existing traditional methods
using the noise and edge detection quality assessment metrics. The proposed
method is more accurate and has a good quality of edge detection as compared to
traditional methods for the input image that contains scratches on the
signboard (Table 3), skewed signboard images (Table 4, and Table 5), and the
deformed signboard image (Table 6).
Tab. 3
Performance Comparison for Input Image in Fig. 1(a)
Criteria |
Prewitt |
Sobel |
Robert |
Canny |
Proposed (TSED) |
SNR |
3.8874 |
3.2169 |
4.7832 |
8.3726 |
9.0142 |
PSNR |
2.9054 |
2.9843 |
2.7987 |
2.4568 |
2.9122 |
MSE |
3.8352 |
4.4786 |
3.3303 |
3.0215 |
2.2113 |
RMSE |
1.9583 |
2.1162 |
1.8249 |
1.7382 |
1.4870 |
Tab. 4
Performance comparison for
input image in Fig. 1(b)
Criteria |
Prewitt |
Sobel |
Robert |
Canny |
Proposed (TSED) |
SNR |
3.3832 |
3.6432 |
4.8934 |
8.8735 |
9.1823 |
PSNR |
2.9364 |
2.8954 |
2.9987 |
3.6943 |
4.2342 |
MSE |
3.4023 |
3.9237 |
3.0012 |
2.7926 |
2.4134 |
RMSE |
1.8445 |
1.9808 |
1.7323 |
1.6711 |
1.5535 |
Tab. 5
Performance
comparison for input image in Fig. 1(c)
Criteria |
Prewitt |
Sobel |
Robert |
Canny |
Proposed (TSED) |
SNR |
3.2834 |
3.5792 |
4.4892 |
8.6253 |
9.4844 |
PSNR |
2.8247 |
2.7962 |
2.8246 |
3.4956 |
4.0122 |
MSE |
3.0624 |
3.4642 |
3.3389 |
4.0062 |
1.9834 |
RMSE |
1.7499 |
1.8612 |
1.8272 |
2.0015 |
1.4083 |
Tab. 6
Performance
comparison for input image in Fig. 1(d)
Criteria |
Prewitt |
Sobel |
Robert |
Canny |
Proposed (TSED) |
SNR |
3.7832 |
3.7653 |
4.6946 |
8.7673 |
9.3472 |
PSNR |
3.0313 |
3.2372 |
3.1034 |
3.5532 |
4.3032 |
MSE |
3.1023 |
4.9145 |
3.4312 |
4.0216 |
2.7134 |
RMSE |
1.7613 |
2.2168 |
1.8523 |
2.0053 |
1.6472 |
As shown in Table 3 to Table 6 for
an input image, the parameter values help determine the edge's quality detected
in the image. The proposed traffic sign edge detection algorithm achieved a
higher quality of edge detection than the other algorithms. By this comparative
analysis of the result, it can be concluded that the proposed traffic sign edge
detection algorithm is the best and most optimal algorithm concerning all other
existing algorithms.
Impact of proposed novel edge
detection technique on VGG-16 and Inception V3 model’s performance are compared using Precision
Rate, Recall Rate, and F1 Score metrics.
5.5. Precision rate
Precision is used to compute the
model’s classification accuracy in terms of positive samples classified
either correctly or incorrectly. It is the ratio of the number of samples
classified as True positive to the total number of positively classified
samples, as given in equation (24). Higher
the precision, better the model performance. Ideal model would have a perfect
score 1.0
P = TP/(TP+FP)
(24)
5.6. Recall rate
The model’s ability to detect
positive samples is called the recall rate. The higher recall rate indicates
that the model can detect more positive samples. It is given as the ratio
between the number of Positive samples correctly
classified as Positive to the total number of Positive samples. Equation (25)
is used to calculate the recall rate.
R
= TP/(TP+FN)
(25)
5.7. F1 score
The F1
Score combines the precision and recall rate in one single metric. It is given
by the equation (26). F1 score is the harmonic mean
of precision and recall. F1 score 1 represents that
model perfectly classified all observations correctly.
F1 Score = 2 * { (P*R)
/ (P+R) }
(26)
Figure 15-17 show that the models
perform better when the input image is edge detected. It is evident that for
the classes 16th, 17th, 22nd, 29th, and 30th, the Recall rate is apparent, and
the Precision rate is higher.
|
|
Fig. 15.
Precision rate
without edge detection (a); precision
rate with edge detection (b) |
|
|
|
Fig. 16. Recall rate without edge
detection (a); recall rate
with edge detection (b) |
|
|
|
Fig. 17. F1
Score without edge detection (a); F1 score with
edge detection (b) |
6. COMPARATIVE ANALYSIS
The proposed approach combines edge
detection with transfer learning for traffic sign recognition and
identification. Table 7 below summarizes some similar works. It is observed
that the proposed edge detection algorithm helps to improve the accuracy of
traffic sign recognition.
Tab. 7
Performance comparison of previous
similar studies
Reference paper |
Dataset |
Classes |
Edge detection |
Model |
Training set (%) |
Validation set (%) |
Testing set (%) |
[31] |
GTSRB |
43 |
No |
InceptionV3 |
- |
70.74 |
- |
[42] |
GTSRB |
43 |
No |
InceptionV3 |
- |
- |
96.03 |
[31] |
GTSRB |
43 |
No |
VGG-16 |
- |
30.61 |
- |
[43] |
GTSRB |
43 |
No |
VGG-16 |
- |
- |
74.5 |
[44] |
GTSRB |
43 |
No |
IVGG-16 |
- |
- |
99.00 |
This study |
GTSRB |
43 |
No |
VGG-16 |
93.04 |
93.53 |
91.78 |
This study |
GTSRB |
43 |
No |
InceptionV3 |
96.81 |
98.81 |
98.67 |
This study |
GTSRB |
43 |
Yes |
VGG-16 |
96.36 |
97.21 |
94.32 |
This study |
GTSRB |
43 |
Yes |
InceptionV3 |
99.40 |
99.03 |
99.49 |
7. CONCLUSION AND FUTURE SCOPE
This paper discussed why traffic
sign recognition and classification are crucial today. A novel edge detection
technique is implemented, and the proposed algorithm results are presented. The
comparative analysis of the performance metric parameters shows that the
proposed traffic sign edge detection algorithm provides promising results for
edge detection in challenging conditions. Thresholding is the critical factor
for the extraction of visual features in images. The proposed method defines
three levels of thresholding: fixed, floating, and ideal. In fixed and ideal
thresholding conditions, the Canny edge detector
performs well compared to Robert, Prewitt, and Sobel edge detection algorithms.
However, the proposed traffic sign edge detection algorithm, derived from canny
edge detection, produces more promising results with a floating thresholding
range from 100 pixels to 600 pixels. The results clearly show that the proposed
algorithm considers the actual images for the symbol. When integrated with
pre-trained CNN models, the proposed traffic sign edge detection algorithm
gives better training, validation, and testing accuracy. Two pre-trained
models, VGG-16 and Inception V3,
were used in the study, out of which Inception V3 is
a top performer with 99.49% test accuracy. The results show that the proposed
algorithm achieves optimal MSE and RMSE error rates and has a better SNR and PSNR
ratio than the traditional edge detection algorithms. There is a scope to
address a few challenges discussed in section two in the future. In the future,
the researcher will aim to implement the proposed algorithm to detect and
recognize the traffic signs from skewed, scratched, and deformed traffic
symbols in real-time using the other CNN models.
References
1.
Liu Chunsheng, Shuang
Li, Faliang Chang, Yinhai
Wang. 2019. „Machine vision
based traffic sign detection methods: Review, analyses and perspectives.”
IEEE Access 7: 86578-86596. ISSN: 2169-3536. DOI:
10.1109/ACCESS.2019.2924947.
2.
Alam Altaf, Zainul Abdin Jaffery. 2020. „Indian
traffic sign detection and recognition. International Journal of Intelligent
Transportation Systems Research 18(1): 98-112. DOI:
https://doi.org/10.1007/s13177-019-00178-1.
3.
Winkle Thomas. 2022. “Findings from Traffic Accident Analysis”.
Product Development within Artificial Intelligence, Ethics and Legal Risk.
Springer Vieweg, Wiesbaden. P. 7-43. ISBN: 978-3-658-34293-7. DOI:
https://doi.org/10.1007/978-3-658-34293-7
4.
Temel Dogancan,
Min-Hung Chen, Ghassan AlRegib.
2019. „Traffic sign detection under challenging conditions: A deeper look
into performance variations and spectral characteristics.” IEEE
Transactions on Intelligent Transportation Systems 21(9): 3663-3673. DOI: https://doi.org/10.1109/TITS.2019.2931429.
5.
Barodi Anass, Abdrrahim Bajit, Mohammed Benbrahim, Ahmed Tamtaoui. 2020.
„An Enhanced Approach in Detecting Object Applied to Automotive Traffic
Roads Signs.” In: 2020 IEEE 6th International Conference on
Optimization and Applications (ICOA). P. 1-6.
IEEE 2020. 20-21 April 2020. Beni Mellal,
Morocco. ISBN: 978-1-7281-6655-1. DOI: https://doi.org/10.1109/ICOA49421.2020.9094457.
6.
Sundararajan Duraisamy. 2017. Digital
image processing: a signal processing and algorithmic approach. Springer
International Publishing. ISBN: 978-981-10-6113-4.
7.
Ghosh Chinmoy, Suman Majumder,
Sangram Ray, Shrayasi Datta, Satyendra Nath Mandal. 2020. „Different EDGE detection
techniques: a review.” Electronic Systems and Intelligent Computing:
885-898. DOI: https://doi.org/10.1007/978-981-15-7031-5_84.
8.
Amer Ghassan
Mahmoud Husien, Ahmed Mohamed Abushaala.
2015. „Edge detection methods.” In: 2015 2nd World Symposium on
Web Applications and Networking (WSWAN). P. 1-7.
IEEE, 2015. 21-23 March 2015. Sousse, Tunisia. ISBN: 978-1-4799-8172-4. DOI: https://doi.org/10.1109/WSWAN.2015.7210349.
9.
Yousaf Rehan Mehmood, Hafiz Adnan Habib, Hussain Dawood, and Sidra Shafiq. 2018. „A comparative study of various edge
detection methods.” In: 2018 14th International Conference on
Computational Intelligence and Security (CIS). P. 96-99. IEEE, 2018. 16-19
November 2018. Hangzhou, China. ISBN: 978-1-7281-0170-5. DOI: https://doi.org/10.1109/CIS2018.2018.00029.
10.
Shah Bickey Kumar, Vansh
Kedia, Rohan Raut, Sakil Ansari, Anshul Shroff.
2020. „Evaluation and Comparative Study of Edge Detection Techniques.”
IOSR Journal of Computer Engineering 22(5):
6-15. DOI: https://doi.org/10.9790/0661-2205030615. ISSN: 2278-8727.
11.
Deshpande Abhinav V., Ramani
Kannan, M. Monica Subashini. 2018. „Study of
Various Image De-Noising Methods Used for the Purpose of Traffic Sign Board
Recognition in an Intelligent Advanced Driver Assistance System.” In: 2018
International Conference on Intelligent and Advanced System (ICIAS). P. 1-6. IEEE, 2018. 13-14 August 2018. Kuala
Lumpur, Malaysia. ISBN: 978-1-5386-7270-9. DOI: https://doi.org/10.1109/ICIAS.2018.8540630.
12.
Kumar Arvind, Sartaj Singh Sodhi.
2020. „Comparative analysis of gaussian filter,
median filter and denoise autoenocoder.”
In: 2020 7th International Conference on Computing for Sustainable Global
Development (INDIACom). P. 45-51. IEEE, 2020.
12-14 March 2020. New Delhi, India. DOI: 10.23919/INDIACom49435.2020.9083712.
13.
Kumar Sunil, Amit Kumar Upadhyay, Preeti Dubey, and Sudeep Varshney. 2021. “Comparative analysis for Edge
Detection Techniques.” In: 2021 International Conference on Computing,
Communication, and Intelligent Systems (ICCCIS).
P. 675-681. IEEE, 2021. 19-20 February 2021. Greater Noida,
India. ISBN: 978-1-7281-8530-9. DOI: 10.1109/ICCCIS51004.2021.9397225.
14. Tariq Nazish, Rostam Affendi Hamzah, Theam Foo Ng, Shir Li Wang, Haidi Ibrahim. 2021 „Quality assessment methods to
evaluate the performance of edge detection algorithms for digital image: A systematic
literature review.” IEEE Access. ISSN: 2169-3536. DOI:
10.1109/ACCESS.2021.3089210.
15.
Yu Zhiding, Chen Feng, Ming-Yu Liu, Srikumar Ramalingam. 2017 „Casenet: Deep category-aware semantic edge detection.”
In: Proceedings of the IEEE conference on computer vision and pattern recognition(CVPR). P.
5964-5973. July 2017.
16.
Liu Yun, Ming-Ming Cheng, Xiaowei Hu, Kai
Wang, and Xiang Bai. 2017. „Richer convolutional features for edge
detection.” In: Proceedings of the IEEE conference on computer vision
and pattern recognition(CVPR).
P. 3000-3009. July 2017.
17. Lakhani
Kanika, Bhawna Minocha, Neeraj Gugnani. 2016. „Analyzing
edge detection techniques for feature extraction in dental radiographs.” Perspectives in Science 8:
395-398.
DOI: https://doi.org/10.1016/j.pisc.2016.04.087.
18.
Srujana P., J. Priyanka, Vyss Sudir Patnaikuni,
Nancharaiah Vejendla. 2021. „Edge
Detection with different Parameters in Digital Image Processing using GUI.”
In: 2021 5th International Conference on Computing Methodologies and Communication
(ICCMC). P. 795-802. IEEE, 2021. 08-10
April 2021. Erode, India. ISBN: 978-1-6654-4775-1. DOI:
10.1109/ICCMC51019.2021.9418327.
19.
Halder Amiya, Pritam
Bhattacharya, Aritra Kundu. 2019. „Edge detection method using Richardson’s
extrapolation formula.” In: Soft Computing in Data Analytics, Advances in Intelligent Systems and
Computing 758. 727-733. Springer, Singapore. ISBN: 978-981-13-0513-9. DOI: 10.1007/978-981-13-0514-6_69.
20.
Srinithyee S.K.,
E. Srivarsha, R. Priyadharsini,
A. Beulah. 2021. „Optimized Image Edge Detection Approach using
Fractional Order Calculus.” In: 2021 6th International Conference on
Communication and Electronics Systems (ICCES). P.
1179-1183. IEEE. 08-10 July 2021. Coimbatre, India.
ISBN: 978-1-6654-1182-0. DOI: 10.1109/ICCES51350.2021.9489066.
21.
Mineo Carmelo, Stephen Gareth Pierce,
Rahul Summan. 2019 „Novel algorithms for 3D
surface point cloud boundary detection and edge reconstruction.” Journal
of Computational Design and Engineering 6(1): 81-91. DOI: https://doi.org/10.1016/j.jcde.2018.02.001.
22.
Zheng Zhen, Bingting Zha,
Hailu Yuan, Youshi Xuchen, Yanliang Gao, He Zhang. 2020. „Adaptive edge detection
algorithm based on improved grey prediction model.” IEEE Access 8:
102165-102176. ISSN: 2169-3536. DOI: 10.1109/ACCESS.2020.2999071.
23. Shi Junfei,
Haiyan Jin, Zhaolin Xiao.
2020. „A novel hybrid edge detection method for polarimetric
SAR images.” IEEE Access 8: 8974-8991.
ISSN: 2169-3536. DOI: 10.1109/ACCESS.2020.2963989
24.
Xizhen Shen, Zeng Wei, Guo
Yiling, Yin Shengyang.
2021. „Edge detection algorithm of plant leaf image based on improved
Canny.” In: 2021 6th International Conference on Intelligent Computing
and Signal Processing (ICSP). P. 342-345. IEEE.
09-11 April 2021. Xi'an, China. ISBN: 978-1-6654-4730-0. DOI:
10.1109/ICSP51882.2021.9408929.
25.
Bausys Romualdas, Giruta Kazakeviciute-Januskeviciene, Fausto Cavallaro,
Ana Usovaite. 2020. „Algorithm
selection for edge detection in satellite images by neutrosophic
WASPAS method.” Sustainability 12(2): 548. DOI: https://doi.org/10.3390/su12020548.
26.
Dhillon Deepak, Rajlaxmi Chouhan.
2022. „Enhanced Edge Detection Using SR-Guided Threshold Maneuvering and
Window Mapping: Handling Broken Edges and Noisy Structures in Canny Edges.”
IEEE Access 10: 11191-11205.
ISSN: 2169-3536. DOI: 10.1109/ACCESS.2022.3145428.
27.
Fawwaz I., M. Zarlis,
R.F. Rahmat. 2018.
„The edge detection enhancement on satellite image using bilateral
filter.” In: IOP Conference Series:
Materials Science and Engineering 308(1): 012052. IOP
Publishing. 24-25 August 2017, Banda Aceh, Indonesia.
28. Ai Jiaqiu,
Ruiming Liu, Bo Tang, Lu Jia,
Jinling Zhao, Fang Zhou. 2019. „A refined
bilateral filtering algorithm based on adaptively-trimmed-statistics for
speckle reduction in SAR imagery.” IEEE Access 7: 103443-103455. ISSN: 2169-3536. DOI: 10.1109/ACCESS.2019.2931572.
29.
Galetto Fernando J., Guang Deng, Mukhalad Al-Nasrawi, Waseem Waheed. “Edge-aware
filter based on adaptive patch variance weighted average. 2021.” IEEE
Access 9: 118291-118306. ISSN: 2169-3536. DOI:
10.1109/ACCESS.2021.3106907.
30.
How Dickson Neoh Tze,
Khairul Salleh Mohamed Sahari, Yew Cheong Hou, Omar Gumaan Saleh Basubeit. 2019.
„Recognizing Malaysia traffic signs with pre-trained deep convolutional
neural networks.” In: 2019 4th International Conference on Control,
Robotics and Cybernetics (CRC). P. 109-113. IEEE. 27-30 September 2019.
Tokyo, Japan. ISBN: 978-1-7281-4621-8. DOI: 10.1109/CRC.2019.00030.
31.
Palavanchu Sowjanya.
2021. „Transfer Learning Models for Traffic Sign Recognition System.”
Annals of the Romanian Society for Cell Biology: 3477-3489.
32.
Rajagopal Balaji
Ganesh. 2020. „Intelligent traffic analysis system for Indian road
conditions.” International Journal of Information Technology:
1-13. DOI: https://doi.org/10.1007/s41870-020-00447-3.
33.
Ghedia Navneet
S., C. H. Vithalani. 2021. „Outdoor object
detection for surveillance based on modified GMM and
Adaptive Thresholding.” International Journal of Information
Technology 13(1): 185-193. DOI:
https://doi.org/10.1007/s41870-020-00522-9.
34.
Pamuji Fandi Yulian, Moch Arief
Soeleman. 2020. „Improved Number Detection For Low Resolution Image Using the Canny Algorithm.”
In: 2020 International Seminar on Application for Technology of Information
and Communication (iSemantic). P. 638-642. IEEE.
9-20 September 2020. Semarang, Indonesia. ISBN: 978-1-7281-9069-3. DOI: 10.1109/iSemantic50169.2020.9234190.
35.
Nadipally Manasa.
2019. „Optimization of methods for image-texture segmentation using ant
colony optimization.” In: Intelligent data analysis for biomedical
applications. P. 21-47. Academic Press. ISBN: 9780128155530. DOI: https://doi.org/10.1016/B978-0-12-815553-0.00002-1.
36.
Malarvizhi C., P. Balamurugan. 2019. „Qualitative Analysis Of Various
Edge Detection Techniques Applied On Cervical Herniated Spine Images.” ICTACT Journal of Image and Video Processing
9(04): 1986-1991. DOI: 10.21917/ijivp.2019.0282.
37.
Paul Eben Sophia, J. Anitha.
2019. „Analysis of Transform-Based Compression Techniques for MRI and CT
Images.” In: Intelligent Data Analysis for Biomedical Applications.
P. 103-120. Academic Press. ISBN: 9780128155530. DOI: https://doi.org/10.1016/B978-0-12-815553-0.00005-7.
38.
Fitriyah Hurriyatul,
Edita Rosana Widasari, Gembong Edhi Setyawan. 2017.
„Traffic sign recognition using edge detection and eigen-face:
Comparison between with and without color pre-classification based on Hue.”
In: 2017 International Conference on Sustainable Information Engineering and
Technology (SIET). P. 155-158. IEEE. 24-25
November 2017. Malang, Indonesia. ISBN: 978-1-5386-2183-7. DOI: 10.1109/SIET.2017.8304127.
39.
Zhu Yanzhao, Wei Qi Yan. 2022. „Traffic
sign recognition based on deep learning.” Multimedia Tools and
Applications 81(13): 17779-17791. DOI: https://doi.org/10.1007/s11042-022-12163-0.
40. Wan Haifeng,
Lei Gao, Manman Su, Qinglong You, Hui Qu, Qirun Sun.
2021. „A novel neural network model for traffic sign detection and
recognition under extreme conditions.” Journal of Sensors. DOI: https://doi.org/10.1155/2021/9984787.
41.
Lin Chunmian, Lin Li, Wenting
Luo, Kelvin C.P. Wang, Jiangang
Guo. 2019. „Transfer learning based traffic
sign recognition using inception-v3 model.” Periodica Polytechnica
Transportation Engineering 47(3): 242-250. DOI: https://doi.org/10.3311/PPtr.11480.
42.
Atif Muhammad, Tommaso
Zoppi, Mohamad Gharib,
Andrea Bondavalli. 2022. „Towards enhancing
traffic sign recognition through sliding windows.” Sensors 22(7):
2683. DOI: https://doi.org/10.3390/s22072683.
43.
Persson Siri. 2018.
„Application of the German Traffic Sign Recognition Benchmark on the VGG16 network using transfer learning and bottleneck
features in Keras.” Thesis. Uppsala universitet, Datalogi. Available at: http://urn.kb.se/resolve?urn=urn:nbn:se:uu:diva-344672.
44.
Kheder
Mohammed Qader. 2022. „Improved Traffic Sign
Recognition System (Itsrs) for Autonomous Vehicle
Based on Deep Convolutional Neural Network.” SSRN. DOI: http://dx.doi.org/10.2139/ssrn.4135313.
Available at: https://ssrn.com/abstract=4135313.
Received 07.12.2022; accepted in
revised form 29.03.2023
Scientific Journal of Silesian University of Technology. Series Transport is licensed under a Creative Commons
Attribution 4.0 International License
[1] Symbiosis International
(Deemed University) (SIU), Pune, India. Email: parsemv@gmail.com. ORCID: https://orcid.org/0000-0002-3946-2291
[2]
Symbiosis Centre for Information Technology (SCIT),
Symbiosis International (Deemed University), Pune, India. Email: director@scit.edu. ORCID:
https://orcid.org/0000-0003-3451-9794