Article
citation information:
Czapla, Z. Description of vehicle
passage through a multisegment detection field. Scientific Journal of Silesian University of
Technology. Series Transport. 2020, 106,
41-50. ISSN: 0209-3324. DOI: https://doi.org/10.20858/sjsutst.2020.106.3.
Zbigniew CZAPLA[1]
DESCRIPTION
OF VEHICLE PASSAGE THROUGH A MULTISEGMENT
DETECTION FIELD
Summary. This paper presents the video-based description
method for vehicles passing a detection field. A sequence of source images is
created by consecutive frames of the input video stream. The source images are
converted into binary target images using the analysis of small gradients.
Binary values of the target images represent edges and surfaces comprised in
the source images. For all images, the same detection field composed of
segments is defined. Inside each segment of the detection field, the sum of
edge values is calculated. For the entire detection field, an adjusted sum of
the edge values is determined. A vehicle passing the detection field changes
the number of edge values within individual segments and the adjusted sum of
the edge values for the entire detection field. Vehicle passage through the
detection field is described by a discrete function that associates the
adjusted sum of the edge values determined for the entire detection field in
the current binary image to the ordinal number of the current image in the
sequence of source images.
Keywords: vehicle passage
description, multisegment detection field, image
processing
1. INTRODUCTION
Present-day road traffic systems employ vision data
for the determination of traffic parameters and traffic surveillance [6]. The
main aim defined for road traffic systems is vehicle detection [7,10]. Aside from vehicle detection, processing and analysis
of vision date allow determination of a number of vehicles moving along a road
during an assumed time period, vehicle classification into type categories,
tracking of vehicles, and vehicle speed estimation. Description of vehicles
passing through detection fields allows vehicle detection and determination of
traffic parameters.
In [1], the traffic system using corner features for
vehicle detection is presented. The regions with brightness changing in more
than one direction are analysed, then vehicle trajectories are described and
traffic parameters are determined. In [4], vehicles are separated through the
determination of the difference between the current image and the background
image, thresholding and binarization of the
difference image, and updating of the background image. In [5], the difference
image is calculated, then lane-dividing lines are determined,
shadow-elimination is performed and Kalman filtering
is used for vehicle tracking. In [2], high-level and low-level modules are
applied. The low-level modules are utilised for the extraction of image
objects, and the high-level module is employed for vehicle tracking. In [3],
captured images are segmented, then static and dynamic analysis is performed
for traffic monitoring. The static analysis is intended for single processed
images and the dynamic analysis encompasses previous images. In [9], vehicles
are detected with the use of time-spatial images created on the basis of
virtual lines. The virtual lines are defined for the frames from the input
video data. In [8], vehicle tracking at intersections is performed. For vehicle
tracking, a stochastic algorithm is applied. In the applied algorithm, image
division into square blocks is required. In [11], background modelling based on
application of a mixture of Gaussian distributions is presented. Background
modelling allows detaching vehicles from the background.
The proposed description method of vehicle passage through
the multisegment detection field assumes the
conversion of source images into a binary form on the basis of the small
gradient analysis. Description of vehicle passage allows determination of the
state of the multisegment detection field and
efficient vehicle detection. The increase of the number of the detection fields
makes possible estimation of vehicle speed and classification of vehicles into
type categories.
2. IMAGE DATA
Image data are obtained at the
measuring station. The measuring station includes the stationary video camera
mounted over a road and directed towards the approaching vehicles. An input
greyscale video stream is obtained from the video camera. Consecutive frames
taken from the input video stream make the sequence of the source images. Size
of each source image is M columns by N rows. Image coordinates m and n point the position of the individual pixels in columns and rows,
respectively. The ordinal number of each individual source image is denoted by i and indicates
the image position in the sequence of the source images.
Subsequent source images are
processed separately. The single source image is represented by source matrix A = [an,m].
Each source
image is converted into a binary target image represented by target matrix B = [bn,m]. Pixels of the source
images are processed by rows. For each processed pixel, the magnitude of small
gradients in rows, in columns and in two diagonal directions are successively
compared to the preset threshold value.
For magnitudes greater than the
threshold value, binary values corresponding to the pixels employed for the
gradient magnitude determination are set to the logical value 1, otherwise,
these binary values are appropriate set to the logical value 0. Elements of
binary target matrix B equal to 1
corresponds to edges in the source image unlike elements equal to 0 which represent
smooth surfaces. Binary elements of the target matrix B which are equal to 1, are called the edge values. The location of
edge values in target images is consistent with the content of input images.
3. MULTISEGMENT
DETECTION FIELD
Individual detection
fields are defined for analysed road lanes. A single detection field is a
rectangle specified by four vertices. The location of the vertices is indicated
by the coordinates of the column (left mL or right mR) and the row (upper nU
or bottom nB), respectively. Thus, the detection field is described by the set of
coordinates {mL, mR, nU, nB}.
The width w of the detection field is given by w = mR – mL + 1,
and the height h of the detection field is expressed by h = nB – nU + 1.
The multisegment
detection field consists of K > 1
segments that cover the entire detection field. Each segment k is
described by the set of segment coordinates {mLk,
mRk, nUk,
nBk}. The segment coordinates
satisfy
(1)
The
width wk and the height hk of the segments are given by wk = mRk – mLk + 1 and hk = nBk – nUk + 1,
respectively.
For the current image
denoted by i,
the features of the segments of the multisegment
detection field are calculated. The relative arithmetic sums of the edge values inside the individual
segments of the multisegment detection field are
given by
(2)
The maximum value of the
relative arithmetic sums of
the edge values within the individual segments of the multisegment
detection field is taken as the adjusted relative sum of the edge values for the
entire multisegment detection field
(3)
Taking into account
uneven distribution of the edge values, the averaging adjusted relative sum of
the edge values inside the entire multisegment detection
field is calculated (a form of lowpass filtering), on
the basis of the current image i and the preset number P of previous images, using the equation
(4)
For comparison to the
uniform detection field of the same entire size, the relative arithmetic sum of the edge values inside the
uniform detection field is given by
(5)
and the averaging relative sum of the edge values is expressed by the
equation
(6)
Application of the multisegment detection field improves the quality of the
description of vehicles passing the detection field and allows increasing
efficiency of proper vehicle detection.
4. EXPERIMENTAL RESULTS
Measurements have been taken at the
measuring station in daytimes and under good light and weather conditions.
Traffic conditions during measurements were normal off-peak and without
congestions. The result of measurements is a registered video stream containing
traffic scenes. The video stream was registered with the use of a typical video
camera. Video stream frame rate is 30 frames per second. Analysed sequences of
input images are created on the basis of the video stream and present
particular traffic scenes. Sequences of input images consist of consecutive
greyscale images with intensity resolution of 8 bits per pixel and of size
384 x 384 pixels. The multisegment
detection field consists of 3 segments of size 52 x 5 pixels. The
segments horizontal overlap with one another and they are shifted in relation
to one another by half the length.
Inside particular segments, the arithmetic
sums of the edge values are calculated relative to the maximum number of the
edge values within the segment. On the basis of the segment sums, the adjusted
relative sum of the edge values is calculated for the entire multisegment detection field. On the basis of the adjusted
relative sums, the averaging adjusted relative sums of edge values are
calculated in the setting P = 0
(without consideration of previous images) and P = 4 (with consideration of 4 previous images).
The test sequences of the source images
were considered. These sequences of source images present varying types of
vehicles moving through the measuring station.
4.1. Description of Passenger Car
Passage
The examples of source and binary target
images that present the passenger car approaching the multisegment
detection field are shown in Fig. 1. The source image is placed at the
left side in the figure and the binary target image on the right side. The multisegment detection field is indicated by the black
rectangles. The black points in the binary image signify the edge values.
Fig. 1. Passenger car by the multisegment detection field
Descriptions of the passenger car passing the multisegment detection field in the form of dependence of the
relative sum of the edge values on the image ordinal number in the sequence of
images show Fig. 2 (the description without consideration of previous
images) and Fig. 3 (the description with consideration of 4 previous images).
Fig. 2. Passenger car passage
through the multisegment detection field (P = 0)
Fig. 3. Passenger car passage
through the multisegment detection field (P = 4)
Description of the passenger car passing
the multisegment detection field is dense. Passenger
cars usually do not have large surfaces without edges. Application of the multisegment detection field gives the proper results which
allow efficient vehicle detection.
4.2. Description of Van Passage
Figure 4 shows the examples of source and
binary target images (at the left and right side, respectively) that present
the van approaching the multisegment detection field.
Similar to images of the passenger car, the multisegment
detection field is pointed by the black rectangles, and the black points in the
binary image denote the edge values.
Van passage through the multisegment detection field is described by the
appropriate relative adjusted sums of edge values. Figure
5 shows the description of van passage through the multisegment
detection field without consideration of the previous images (P = 0). The description of van
passage with consideration of 4 previous images (P = 4) is shown in Fig. 6.
Fig. 4. Van by
the multisegment detection field
Fig. 5. Van passage through
the multisegment detection field (P = 0)
Fig. 6. Van passage through
the multisegment detection field (P = 4)
Vans usually have large surface without
edges, which is visible in the van passage description. Application of the multisegment detection field increases the number of edge
values taken into account, thus improves the possibility of proper vehicle
detection.
4.3. Description of Truck Passage
The examples of images of the truck moving
into the multisegment detection field are shown in
Fig. 7. As in the previous examples, the source image is placed at the left
side in the figure, and the binary target image at the right side, also in the
binary target image, the black rectangles indicates the multisegment
detection field, and the black points signify the edge values.
Fig. 7. Truck by
the multisegment detection field
Figures 8 and 9 show a description of the
truck passage through the multisegment detection
field. Similar to the previous examples, the description presents the dependence of the
relative adjusted sums of the edge values, inside the multisegment
detection field, on the image ordinal number in the image sequence.
Fig. 8. Truck passage through
the multisegment detection field (P = 0)
Fig. 9. Truck passage through
the multisegment detection field (P = 4)
Description of the trucks passing the multisegment detection field can have varying density, of
which degree depends on the number and size of surfaces without edges. As in
the previous examples, the application of multisegment
detection field increases the number of the edge values taken into account.
4.4. Comparison to the Uniform
Detection Field
For efficiency estimation of the multisegment detection field, the description of vehicle
passage through the multisegment detection field was
compared to the description of vehicle passage through the uniform detection
field of the same size. The examples of images of the passenger car moving to
the uniform detection field show Fig. 10. Differences of the relative
adjusted sums of the edge values between descriptions of passage through the
uniform and multisegment detection fields, with and
without consideration of the previous images, show Fig. 11
and 12, respectively.
Fig. 10. Passenger car by the uniform detection field
Fig. 11. Passenger car passage
through the detection fields (P = 0)
Application of the multisegment
detection field instead of the uniform detection field significantly increases
the number of the edge values in the descriptions of vehicle passage through
the detection field, both with and without consideration of the previous
images. The increase of the edge values taken into account for the descriptions
concerns all types of vehicles.
Fig. 12. Passenger car passage
through the detection fields (P = 4)
5. CONCLUSIONS
Vehicle passage through the detection
field can be described on the basis of image gradients. Source images are
converted into binary target images using the comparison of gradient magnitudes
to the preset threshold value. The obtained binary
target images contain edge values which are consistent with edges in the source
images. Vehicles passing the detection field are described through the dependence of the
relative sum of the edge values inside the detection field on the image number
in the sequence of source images. The replacement of the uniform detection
field by the multisegment detection field caused the
increase of the edge values in the vehicle passage description.
Application of the multisegment detection field
facilitates effective vehicle detection.
The proposed method of the description of
vehicles passing the multisegment detection field is
computationally simple and allows effective vehicle detection. This method can
be useful for determination of traffic parameters. The fundamental
advantage of the proposed method is the small number of required operations. As
other methods based on image data, the proposed method of vehicle description
is sensitive to changes in weather and illumination. The
descriptions of vehicle passages through the multisegment
detection field are intended for systems of traffic measurements and
surveillance.
References
1.
Coifman Benjamin, David Beymer,
Philip McLauchlan, Jitendra
Malik. 1998. “A real-time computer vision system for vehicle tracking and
traffic surveillance”. Transportation Research Part C 6(4):
271-288. DOI: 10.1016/s0968-090x(98)00019-9.
2.
Cucchiara Rita, Piccardi Massimo,
Mello Paola. 2000. “Image Analysis and Rule-Based Reasoning for a Traffic
Monitoring System”. IEEE Transaction on Intelligent Transaction
Systems 1(2): 119-130. DOI:
10.1109/6979.8809.69.
3.
Fernandez-Caballero
Antonio, Francisco J. Gomez, Juan Lopez-Lopez. 2008. “Road traffic
monitoring by knowledge-driven static and dynamic image analysis”. Expert Systems with Applications 35(3): 701-719. DOI:
10.1016/j.eswa.2007.07.017.
4.
Gupte
Surendra, Osama Masoud,
Robert F.K. Martin, Nikolaos P. Papanikolopoulos.
2002. “Detection and Classification
of Vehicles”. IEEE Transaction on Intelligent Transportation Systems
3(1): 37-47. DOI: 10.1109/6979.994794.
5.
Hsieh Jun Wei,
Shih-Hao Yu, Jung-Sheng
Chen, Wen-Fong Hu. 2006. „Automatic Traffic Surveillance System for
Vehicle Tracking and Classification”. IEEE Transaction on Intelligent
Transportation Systems 7(2): 175-187. DOI: 10.1109/tits.2006.874722.
6.
Jarašūniene Aldona. 2007.
“Research into intelligent transport systems (ITS) technologies and
efficiency”. Transport 22(2):
61-67. DOI: 10.1080/16484142.2007.9638100.
7.
Joonho
Ko, Daejin
Kim, Heung Gweon Sin & Seungjae
Lee. 2014. “The efficiency of vehicle monitoring locations for a
voluntary travel demand management program”. Transport 29(3): 326-333. DOI:
10.3846/16484142.2014.953206.
8.
Kamijo
Shunsuke, Yasuyuki
Matsushita, Katsushi Ikeuchi, Masao Sakauchi. 2000. “Traffic Monitoring and Accident
Detection at Intersections”. IEEE Transactions on Intelligence
Transportation Systems 1(2): 108-118. DOI: 10.1109/6979.880968.
9.
Mithun
Niluthpol Chowdhury, Nafi
Ur Rashid, S.M. Mahbubur
Rahman. 2012. “Detection and Classification of Vehicles From Video Using Multiple Time-Spatial Images”. IEEE
Transaction on Intelligent Transportation Systems 13(3): 1215-1225. DOI: 10.1109/tits.2012.2186128.
10. Niu Hongxia, Tao Hou. 2018. “Fast detection study of foreign object
intrusion on railway track”. Archives of Transport
47(3): 79-89. ISSN 0866-9546.
11. Stauffer Chris, W.E.L.
Grimson. 1999. “Adaptive background mixture
models for real time tracking”. In IEEE
Computer Society Conference on Computer Vision and Pattern Recognition 2:
246-252. IEEE. 23-25 June 1999, Fort Collins, CO, USA. ISBN: 0-7695-0149-4. DOI: 10.1109/cvpr.1999.784637.
Received 02.11.2019; accepted in revised form 28.12.2019
Scientific
Journal of Silesian University of Technology. Series Transport is licensed
under a Creative Commons Attribution 4.0 International License
[1] Faculty of Transport, The Silesian University of Technology, Krasińskiego
8 Street, 40-019 Katowice, Poland. Email: zbigniew.czapla@polsl.pl