Next Article in Journal
Multi-Scale Earthquake Damaged Building Feature Set
Previous Article in Journal
Tuning Data Mining Models to Predict Secondary School Academic Performance
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Data Descriptor

A Point Cloud Dataset of Vehicles Passing through a Toll Station for Use in Training Classification Algorithms

by
Alexander Campo-Ramírez
*,
Eduardo F. Caicedo-Bravo
and
Eval B. Bacca-Cortes
School of Electrical and Electronic Engineering, Faculty of Engineering, Universidad del Valle, Cali 760032, Colombia
*
Author to whom correspondence should be addressed.
Submission received: 3 May 2024 / Revised: 6 June 2024 / Accepted: 19 June 2024 / Published: 27 June 2024

Abstract

:
This work presents a point cloud dataset of vehicles passing through a toll station in Colombia to be used to train artificial vision and computational intelligence algorithms. This article details the process of creating the dataset, covering initial data acquisition, range information preprocessing, point cloud validation, and vehicle labeling. Additionally, a detailed description of the structure and content of the dataset is provided, along with some potential applications of its use. The dataset consists of 36,026 total objects divided into 6 classes: 31,432 cars, campers, vans and 2-axle trucks with a single tire on the rear axle, 452 minibuses with a single tire on the rear axle, 1158 buses, 1179 2-axle small trucks, 797 2-axle large trucks, and 1008 trucks with 3 or more axles. The point clouds were captured using a LiDAR sensor and Doppler effect speed sensors. The dataset can be used to train and evaluate algorithms for range data processing, vehicle classification, vehicle counting, and traffic flow analysis. The dataset can also be used to develop new applications for intelligent transportation systems.
Dataset: The data presented in this study are openly available at: https://doi.org/10.5281/zenodo.10974361
Dataset License: Creative Commons Attribution 4.0 International

1. Introduction

Modern countries face an increasingly complex panorama: efficient traffic management and road safety. The rapid increase in the number of vehicles, traffic congestion, traffic accidents and environmental pollution are some of the problems that plague most countries [1,2]. These challenges not only negatively impact the quality of the life of citizens but also generate significant economic and environmental costs [3].
Responding to these urgent needs, intelligent transportation systems (ITS) emerge as an innovative technological solution that is changing the field of transportation [4]. These systems are a wide range of technological solutions applied to transportation, which integrate a network of subsystems and distributed sensors to collect data in real time about the road environment [5]. By processing this big data using machine learning algorithms and artificial intelligence techniques, ITS can obtain crucial information to optimize traffic flow, improve road safety and reduce the environmental impact of transportation [6].
ITS are fundamental to the continuous transformation of mobility and offer solutions to improve safety, efficiency and sustainability in transportation. At the core of these systems, one essential need is focused on real-time object detection and classification, especially as it relates to vehicles traversing roads and infrastructure points such as toll stations [7].
In recent years, the field of vehicle detection and classification has been revolutionized by the arrival of new sensor technologies and the increase in data processing algorithms. In particular, LiDAR (light detection and ranging) sensors have become a valuable tool for capturing high-resolution three-dimensional point cloud data, offering information on the dynamics of vehicular movement [8]. Processing these point clouds using specialized algorithms allows ITS to detect, track and recognize objects.
Accurate vehicle identification and classification are essential for numerous ITS applications, such as dynamic lane assignment, differentiated toll collection, traffic restrictions, implementation of automated tolling systems, emergency vehicle prioritization, and enforcement of traffic rules [9,10,11,12]. However, the development and validation of robust classification algorithms depend on access to high-quality datasets that faithfully represent real-world traffic scenarios. These datasets are the fundamental basis for training, testing and benchmarking many algorithms, allowing researchers to evaluate their performance under different environmental conditions and operational constraints [13].
In response to this demand, this work presents a novel point cloud dataset designed from the capture vehicles passing through a toll station in Colombia. Leveraging LiDAR and speed sensors, this dataset provides a detailed three-dimensional representation of vehicle geometry. Considering the richness of point cloud data, our dataset offers valuable insights into the various shapes and sizes of vehicles encountered in real-world traffic scenarios.
This article clarifies the methodologies used to collect, process and label the dataset, highlighting its suitability for training and validating classification algorithms adapted to vehicle recognition tasks in intelligent transportation systems. Additionally, this work mentions potential applications of this dataset, including but not limited to vehicle counting, size estimation, vehicle type classification and 3D modeling, thereby facilitating advances in traffic management, road safety, and overall efficiency of ITS solutions.
Figure 1 shows some examples of vehicle point clouds from the dataset; on the left are the raw 3D point clouds, and on the right are the 2D point clouds after processing to extract only the side profile of the vehicles, excluding background, ground surface and noise information. The point clouds are presented on Cartesian planes where the Z axis corresponds to the height of the vehicle, the X axis to the length of the vehicle and the Y axis to the width. The units of the Cartesian plane in the 2D point clouds for the X and Z axes are meters and allow the size of the vehicles to be observed and compared.
Technology is advancing rapidly, and with access to datasets like the one presented in this work, ITS are poised to revolutionize mobility. By sharing this dataset, the goal is to foster collaboration between researchers to improve classification algorithms, drive innovation in ITS, and create safer and more efficient transportation networks for the future.
This paper is structured as follows: Section 1.1 reviews previous work related to vehicle point cloud datasets. Section 2 describes the structure and format of the dataset, along with a detailed description of the hardware used. Section 3 details the data acquisition process, including the processing stages and tools used for the construction and filtering of the point clouds from the dataset files. In addition, the methodology and tools used for labeling and validation are explained. Section 4 presents examples of applications of the data set. Finally, Section 5 presents the conclusions derived from this work.

1.1. Related Work

In recent years, the domain of ITS has witnessed remarkable progress in the development of vehicle detection and classification algorithms, fueled by advancements in sensor technologies, machine learning techniques and computational capabilities. These advancements have been instrumental in addressing the increasing demand for safer, more efficient, and sustainable transportation solutions [5]. In this context, point cloud datasets have become a fundamental tool for the development and evaluation of real-time vehicle classification algorithms.
Some of the widely used point cloud datasets for vehicle classification are the KITTI [14], Argoverse [15], nuScenes [16], Waymo Open [17], A2D2 [18], ApolloScape [19] and PandaSet [20] datasets, which feature traffic scenes captured from moving vehicles equipped with multiple sensors, including cameras and LiDAR. However, due to their generic nature, these datasets mainly focus on autonomous driving, which limits their direct applicability in specific scenarios such as the classification of vehicles traversing roads and infrastructure points such as toll stations.
Table 1 provides an overview of the characteristics of these datasets. The “Dataset” column specifies the name of the dataset, “Ann. fr.” indicates the number of annotation frames present and “3D box.” reflects the number of objects detected in the frames and represented by 3D point clouds. “LiDAR” indicates how many 3D LiDAR sensors were used to acquire the data, while “Classes” shows the number of categories into which the objects in the dataset have been labeled. The “Location” column indicates the city or cities where the data acquisition was performed, “Distance” indicates the distance traveled, “Night/Rain” indicates whether the data was taken under these conditions and, finally, “Duration” indicates the time duration of data collection.
The Waymo dataset stands out, containing 12 million identified objects in 230,000 frames. However, it only classifies these objects into four different categories. In contrast, NuScenes includes 1.4 million objects classified into 23 different categories. It is relevant to note that these categories cover a variety of objects, not only vehicles but also pedestrians, seated people, cyclists, trams and others.
Although these datasets are valuable resources for research and development of perception systems in autonomous driving environments, they also present some specific limitations regarding vehicle classification [21]:
  • Vehicle variety: Although these datasets contain a wide variety of vehicles common in urban environments, they may not fully represent the diversity of vehicle types found in other geographic regions. Examples of less common vehicles may be missing.
  • Resolution and data quality: Although datasets typically have high resolution, data quality can vary depending on acquisition conditions, such as lighting, weather, or the presence of obstacles that may partially obstruct objects. This can make it difficult to accurately classify vehicles in some situations.
  • Point cloud information: The density of point cloud data in sets such as these may be insufficient to provide a detailed description of each vehicle. The abundance of information in the point cloud depends on factors such as the distance between objects, weather conditions, and acquisition equipment. Consequently, vehicle point cloud data is often sparse, resulting in limited representation of different vehicle parts.
Considering these topics, several researchers have addressed these limitations by developing their datasets. These sets are generated from data acquisition systems strategically located at fixed points, both overhead and to the side of traffic routes. This strategy allows a wide variety of traffic scenarios to be captured from different perspectives, enriching the diversity and quality of data available for vehicle classification research.
For example, in Refs. [11,22], a dataset including 4955 vehicles was created using a LiDAR mounted overhead of a road. Five different classes were generated, ranging from motorcycles to trucks and buses. In Ref. [23], two LiDARs were installed on the side of a toll station road, generating a dataset with 206 vehicles for axle detection and counting in four different classes. In works [24,25], the authors placed three LiDARs on a gantry in a three-lane road, one scanner per lane, obtaining a dataset with 30,000 vehicles distributed in six different classes. Authors in Ref. [26] used two laser distance detectors on a gantry, one on each side of a gantry, along with a third sensor on a support pillar in front of the gantry, to collect data from 270 vehicles, including saloon cars, passenger cars and trucks. In Ref. [27], the authors collected data on an entrance ramp to a truck scale, using a LiDAR located on the side and collecting 10,024 vehicles in 11 different classes. Finally, the authors in [21] employed three LiDARs in a similar configuration to Ref. [26], achieving a dataset with 800 vehicles and high-density point clouds, covering 11 different vehicle types, from vans to fuel tank trucks.
Table 2 presents a summary of the main characteristics of these datasets, designed specifically for vehicle classification. The “Veh.” column reflects the size of the dataset, that is, the number of vehicles detected, “Type” indicates the type of laser sensors used, and “Sensor Position” specifies the location of the sensors on the road. The other columns are similar to Table 1. The work in Refs. [24,25] stands out, which has 30,000 available vehicles classified into six categories, including passenger vehicles, passenger vehicles with trailers, trucks, trucks with trailers, trucks with two trailers and motorcycles.
However, although these works show significant advances and important results in research on vehicle classification from point clouds, it is still possible to identify some limitations related to the data sets:
  • The data sets are not publicly accessible.
  • The number and variety of classes are limited. In Refs. [11,22,24,25], sensors mounted overhead of a road only allow classification into a few classes: motorcycle, passenger car, bus, and truck, with or without trailers. In Refs. [21,23,26], the number of objects is small, less than 1000. In Ref. [23], although the sensors were installed in a toll plaza, the same four vehicles were used for collection of data, one for each class, which does not reflect a real traffic environment. In Refs. [21,27], the focus was mainly on large vehicles.
In contrast, the main advantages of the data set presented in this work are as follows:
  • The data set is publicly accessible.
  • It contains a significant number of objects, 36,026, divided into 6 classes: 31,432 cars, campers, vans, and 2-axle trucks with a single tire on the rear axle; 452 minibuses with a single tire on the rear axle; 1158 buses; 1179 2-axle small trucks; 797 2-axle large trucks; and 1008 trucks with 3 or more axles.
  • The data were acquired in a real traffic environment 24 h a day for 12 days, allowing a wide variety of vehicles to be obtained. Although the data are labeled and classified into six classes according to Colombian regulations, it is possible to relabel it and find a wider range of classes according to specific needs. These classes include cars, pickups, campers or vans with trailer, 3-axle buses, single unit 3-axle trucks, single unit 4 or more-axle trucks, single trailer 3 or 4-axle trucks, single trailer 5-axle trucks, single trailer 6 or more-axle trucks, multi-trailer 5 or less-axle trucks, multi-trailer 6-axle trucks, multi-trailer 7 or more-axle trucks, barn trucks, fence trucks, crane trucks, semi-trailer tractors, garbage trucks, watering trucks, fuel tank trucks, among others.
As a result, while current datasets have been very useful for autonomous driving research, it is evident that there is a growing demand for specialized datasets designed for specific applications in ITS. These datasets are essential for training classification algorithms that can be applied in both ITS and autonomous driving tasks, capturing information relevant to the particular environment and specific tasks at hand. For example, a dataset focused on vehicles passing through a toll station would benefit from annotations that cover not only the type of vehicle, but also details such as the lane used, the traffic flow rate, and a more detailed description of the vehicles, which would facilitate a more precise classification and with a greater diversity of classes.
The dataset presented in this paper offers significant advantages compared to other works in terms of the number of vehicle samples and the diversity of vehicle classes. Consequently, it becomes a valuable tool for the development and research of intelligent transportation systems. This dataset has the potential to improve the accuracy of vehicle classification algorithms, as well as point cloud and range data processing, along with other crucial traffic analysis tasks.

2. Dataset Description

This section begins by offering a thorough description of the hardware used, covering its components, configuration, operating conditions and installation in a toll station in Colombia. This information lays the foundation for understanding the structure of the data. Next, a detailed description of the dataset is provided, including the files that comprise it and the format in which they are organized.

2.1. Hardware Description

The dataset was obtained using a Hokuyo UTM-30LX scanning laser rangefinder [28], two Stalker stationary speed sensors [29] and an industrial fanless minicomputer. The main technical characteristics of the sensors are shown in Table 3 and Table 4, and the main technical characteristics of the computer are shown in Table 5. The Hokuyo UTM-30LX scanning laser rangefinder is a 2D and single layer sensor.
This configuration makes it possible to monitor two adjacent lanes at a toll, as illustrated in Figure 2. Each lane is equipped with a speed radar, and given the range of the laser sensor, a single device is used to monitor both lanes simultaneously.
The sensors were installed in a toll station, as shown in Figure 2, Figure 3 and Figure 4. This station has 5 lanes, and data recording was carried out in the center lane and an adjacent one. The separators between lanes have a width of 2 m, while each lane has a width of 3.5 m.
Figure 3 presents a picture of the toll with the sensors installed. In it, the lanes designated as right and left are identified. The laser is installed in a structure at 1 m high, visible on the left side of the picture, while on the right side, we can see the structure that holds the speed sensors and the computer. The laser rangefinder was installed in the middle of two lanes with its light beams pointing toward the side of the vehicles to create 3D models of the side faces. Considering that the angular range of the sensor is 270°, it was installed with the connection cables pointing perpendicularly toward the ground as seen in Figure 4. The speed sensors were installed at a distance of 5 m from the laser rangefinder, with their focus pointing towards the center of the lane at the point through which the light beams of the laser rangefinder pass, as seen in Figure 2.
Since the laser rangefinder has an angular range of 270°, a polar coordinate plane was defined as a reference frame for the scanner reading. In this plane, the 0° angle is oriented towards the right lane in a horizontal direction parallel to the ground, as depicted in Figure 5. Consequently, the sensor makes measurements within an angular range that goes from −45° to 225°, according to the established polar coordinate plane. Figure 5 provides a visual representation of the angular range of the scanner with a front view from the position of the speed sensors.
To ensure accurate data acquisition using the sensors, the following installation and operating conditions were prepared and strictly adhered to during the data collection period:
  • The sensors must be protected from rain and direct sunlight, a protection that the toll station cover adequately provides.
  • It is recommended that vehicles drive between 0.5 m and 2.5 m from the laser rangefinder to ensure enough point resolution. Given that the width of the lane separator is 2 m, it is established that the minimum distance at which vehicles travel is 1 m.
  • There must be no obstructions between the vehicle and the ground to the laser rangefinder. The sensor has been installed without obstacles that prevent proper scanning of vehicles and the ground surface.
  • Vehicles must travel at a maximum speed of 40 km/h to ensure adequate point density (in case of different circumstances, the possibility of changing the range sensor for one with a higher sampling rate should be considered). The lanes are equipped with speed bumps that limit the speed of vehicles to this maximum.
  • Vehicles are required to drive in front of the laser rangefinder without stopping until the scan is completed. Since the sensor is located after the toll booth, vehicles do not normally stop in front of the sensor during normal toll operation.

2.2. Data Format

Vehicles are classified into six different categories, which generally represent the classes of vehicles used for classification at Colombian tolls. Table 6 presents a detailed breakdown of the vehicles obtained, including the classes they have been classified into, a brief description of each class, the number of vehicles per class, and visual examples of each.
The dataset is composed of 36,026 object classes stored in text files with out extension and organized into daily and hourly folders according to the date and time of capture. Figure 6 shows the storage structure of the dataset.
During the data acquisition process, the software automatically divides the laser light beams into two parts: one for the right lane and one for the left lane. This ensures that there are no problems if two vehicles pass simultaneously. Therefore, each file of the dataset corresponds to a single vehicle that passed through the toll station.
There are 12 folders at the root of the dataset, one for each day, and they are named with the corresponding year, month and day, for example, folder 20230401 contains the data for 1 April 2023. Each day’s folder can contain up to 24 subfolders, one for each hour of the day, which are named with the corresponding hour in two-digit format from 00 to 24. The files are named with the date and time of the capture, the lane by which the vehicle passed (D for the right lane and I for the left lane) and the type of vehicle (see Table 6). The following format is used for the file name:
year_month_day_hour_minute_second_lane_type.out
The scanner’s acquisition rate is 40 SPS (Scans Per Second), which means that the sensor performs a full sweep of distance measurements from −45° to 225° every 0.25 ms. The angular resolution of the scanner is 0.25°, which implies that each scan or sample of the scanner contains 1081 values. Of these values, the first 541 are assigned to the right lane and the remaining 540 to the left lane.
Each file in the dataset is made up of a matrix, where each row represents a sample of the scanner. Figure 7 shows the general structure of the rows in each file, the upper array corresponds to the files of vehicles in the right lane and the lower one to the vehicles in the left lane.
The number of rows in each file depends on the size of the vehicle, as shown in Figure 8. Each row is a set of float values separated by whitespace. The first value corresponds to the speed in m/s, measured with the speed sensor of the corresponding lane at the same time of the scan, and is highlighted in red in Figure 8. The second value indicates the relative time in seconds, obtained with a timer that starts with data acquisition, and is represented in blue. The remaining values are measurements from the laser rangefinder scan or sample and are identified in green. Therefore, for vehicles traveling in the right lane, the files will contain rows with 543 values, while for vehicles traveling in the left lane, the files will have 542 elements.
For the 36,026 vehicles, the average speed measured was 14.2 km/h with a standard deviation of 3.7 km/h.
The data set is shared in its raw form, that is, comprising points with polar coordinates, allowing interested researchers to explore a wide range of analysis and processing options, including filtering, interpolation, analysis of sensor response, among others. As detailed in the following sections and following the instructions in Section 3.2.1 and Algorithm 3, a simple script can be created to convert the polar coordinates of the data set files into a Cartesian coordinate matrix. This matrix can be saved in a PCD file, which is compatible with software such as CloudCompare [30] or other similar programs.

3. Methods

This section addresses the software tools used to read and interpret files, allowing the raw point clouds to be viewed. Likewise, the processing algorithms used to generate point clouds with the lateral profiles of the vehicles are detailed, eliminating background information, the ground surface and noise. This provides a clear display of the vehicle type and its distinctive features. In addition, the methodology and tools used for validation and labeling of the dataset are presented.
The data acquisition software was developed in Python 3.69. The interface with the laser rangefinder was made with the HokuyoAIST library [31,32] and the Point Cloud Library (PCL) [33] version 1.10.1 was used to process the point clouds.

3.1. Data Acquisition

Algorithm 1 presents a brief description of the computational steps invoked to acquire the range and speed data of the vehicles, as well as their storage in the files that constitute the dataset. In summary, the algorithm is responsible for the initial configuration, activation and reading of the sensors, the selection of valid samples to determine their belonging to a vehicle, and finally the registration of the valid samples in the corresponding file.
Algorithm 1 The main computational steps to acquire vehicle range data.
1:
r i g h t _ f i n a l i z e d F a l s e         // indicates when capture of a vehicle in right lane ends
2:
l e f t _ f i n a l i z e d F a l s e         // indicates when capture of a vehicle in left lane ends
3:
i n i t i a l _ t i m e read current time         // read the current time
4:
turn on sensors         // turn on laser rangefinder and speed sensors
5:
while  T r u e  do
6:
     s c a n read lidar         // read a LiDAR sample
7:
     r i g h t _ s p e e d read right sensor         // read a value from right lane speed sensor
8:
     l e f t _ s p e e d read left sensor         // read a value from left lane speed sensor
9:
    if  r i g h t _ s p e e d > 0  then         // vehicle detection in right lane if speed > 0
10:
         t i m e read current time i n i t i a l _ t i m e         // estimate relative time
11:
         r i g h t _ a r r a y { r i g h t _ s p e e d t i m e s c a n [ 0 . . 540 ] }         // create array of the sample
12:
        // detects a vehicle and selects valid samples
13:
         r i g h t _ l a s e r _ b l o c k , r i g h t _ f i n a l i z e d SamplesSelection( r i g h t _ a r r a y )
14:
        if  r i g h t _ f i n a l i z e d is T r u e  then
15:
           // save data to file if vehicle detection ends
16:
           save r i g h t _ l a s e r _ b l o c k to file
17:
            r i g h t _ f i n a l i z e d F a l s e
18:
        end if
19:
    end if
20:
    if  l e f t _ s p e e d > 0  then         // vehicle detection in left lane if speed > 0
21:
         t i m e read current time i n i t i a l _ t i m e
22:
         l e f t _ a r r a y { l e f t _ s p e e d t i m e s c a n [ 541 . . 1080 ] }
23:
         l e f t _ l a s e r _ b l o c k , l e f t _ f i n a l i z e d SamplesSelection( l e f t _ a r r a y )
24:
        if  l e f t _ f i n a l i z e d is T r u e  then
25:
           save l e f t _ l a s e r _ b l o c k to file; l e f t _ f i n a l i z e d F a l s e
26:
        end if
27:
    end if
28:
end while
The proper selection of valid samples is essential to determine the precise moment at which a vehicle passes in front of the laser sensor, thus marking the beginning of the creation of a new laser block. This block records all light beams captured during vehicle detection, except those associated with a speed equal to zero. To carry out this process, a region of interest (ROI) is established whose dimensions are adjusted according to the specific location, being activated or deactivated depending on the presence or absence of an object.
Since the laser sensor provides the range measurements in polar format, the region of interest is configured in a cone shape, with parameters such as a minimum radius (minimum distance), a maximum radius (maximum distance), and top and bottom opening angles. These parameters are centered at the point (0, 0) corresponding to the sensor position and are adjusted independently for each lane under system supervision. Figure 9 illustrates the region of interest (represented by red cones) when selecting valid samples. For the right lane, a region of interest has been defined that spans from −30° to 30° in angle, and from 1 m to 3 m in distance. As for the left lane, the region of interest extends from 150° to 210° in angle, maintaining the same distance range, that is, from 1 m to 3 m.
To guarantee good performance of the algorithm, it is essential to ensure that the region of interest does not include light beams that impact the ground or surrounding surfaces, such as walls or objects in the environment, as this could cause erroneous activations in the region of interest.
In Algorithm 1, it begins with the acquisition of data by the laser and speed sensors. Subsequently, a validation is carried out to discard samples in which the recorded speed is equal to 0. Those samples with speeds other than 0 are subjected to the process of selecting valid samples. On the other hand, Algorithm 2 offers a synthesis of the computational steps necessary for this selection. Once the region of interest (ROI) has been defined by adjusting the parameters, and in the absence of vehicle detection, the activity in that region is monitored. Detection of activity in the ROI indicates the creation of a new laser block. Once a vehicle is detected by activating the ROI, its continued presence is verified. Otherwise, a countdown is started to stop the block and send it to the storage stage.
Algorithm 2 The main computational steps for the selection of valid samples.
1:
R O I         // set region of interest
2:
c o u n t 0         // counter continuous samples that are outside ROI
3:
s t a t u s w a i t i n g         // algorithm status: waiting or collecting a vehicle
4:
declare b u f f e r [ 10 , 543 ]         // temporary buffer that stores last 10 samples
5:
l a s e r _ b l o c k [ ]         // laser block with all samples of the vehicle
6:
function SamplesSelection( a r r a y )
7:
    if  b u f f e r is f u l l  then
8:
        rotate b u f f e r         // rotates buffer and remove oldest sample
9:
        append a r r a y to b u f f e r         // append the sample to buffer end
10:
       if  s t a t u s is w a i t i n g  then
11:
            l a s e r _ b l o c k [ ]         // cleans laser block
12:
           // If 10% of sample points belong to ROI, vehicle data recording begins
13:
           if  10 % of a r r a y R O I  then  s t a t u s c o l l e c t i n g
14:
           end if
15:
        else
16:
           append b u f f e r [ 0 ] [ a l l ] to l a s e r _ b l o c k         // append the buffer oldest sample to laser block
17:
           // counts continuous samples with 90% of points outside ROI (deactivation) to detect object
18:
           // completion
19:
           if  90 % of a r r a y R O I  then  c o u n t c o u n t + 1
20:
           else  c o u n t 0
21:
           end if
22:
           if  c o u n t > 20  then
23:
               // if more than 20 continuous samples have 90% of points outside ROI, object detection ends
24:
                s t a t u s w a i t i n g ; c o u n t 0
25:
               // if the laser block has more than 30 samples, it is considered a valid vehicle
26:
               if size of l a s e r _ b l o c k > 30  then return  l a s e r _ b l o c k , T r u e
27:
               end if
28:
           end if
29:
        end if
30:
    else append a r r a y to b u f f e r
31:
    end if
32:
    return  l a s e r _ b l o c k , F a l s e
33:
end function

3.2. Data Processing

The data files must be processed to obtain a point cloud of the vehicle, in which the stage background, the ground surface and the associated noise have been removed. A general diagram of this process can be seen in Figure 10. The following subsections show each of the four stages presented in the diagram in more detail.

3.2.1. Creation of the Point Cloud

In the initial phase, the process involves reading the vehicle’s range data from a file, to generate a three-dimensional representation of the captured scene. For this representation, a Cartesian coordinate system is established where the Z axis denotes the height of the vehicle, the X axis the length, and the Y axis the width. Each row of the file contains scan data expressed in polar coordinates (r, θ ), where r indicates the distance measured in mm and θ represents the corresponding angle. For the right lane, the angles vary from −45° to 90° in 0.25° increments, while for the left lane they range from 90.25° to 225°. The construction of the point cloud requires the conversion of the range data to Cartesian coordinates (Y, Z) and the adaptation of the measurement units to m, using Equations (1) and (2).
Y = r c o s π 180 θ / 1000
Z = r s i n π 180 θ / 1000
The X-axis coordinate is obtained using Equation (3), where X p r e v is the X coordinate of the previous sample (this variable is initialized to 0), s p e e d is the measured speed of the sample in m/s and is the first value in the array of each row in the range data file, t i m e c u r r is the relative time in seconds of the sample and is the second value in the array, and  t i m e p r e v is the relative time in seconds of the previous sample, that is, in the array of the row immediately preceding the sample being processed.
X = X p r e v + s p e e d · ( t i m e c u r r t i m e p r e v )
Algorithm 3 presents a brief description of the computational steps used to construct the point cloud from the dataset files and Figure 11 shows the 3D image of a point cloud built from the data in one of the files. It is possible to observe the entire captured scene, including the vehicle, the ground surface and the background.
Algorithm 3 The main computational steps for the creation of the point cloud.
1:
d a t a read file
2:
a n g u l a r _ r e s o l u t i o n 0.25         // angular resolution of LiDAR
3:
a n g u l a r _ r a n g e 270         // angular range of LiDAR
4:
if  l a n e is right then  a n g l e _ a r r a y { 45 : 0.25 : 90 }         // angle array of right lane
5:
else  a n g l e _ a r r a y { 90.25 : 0.25 : 225 }         // angle array of left lane
6:
end if
7:
s p e e d _ d a t a d a t a [ a l l ] [ 0 ]         // speed array of all samples
8:
t i m e _ d a t a d a t a [ a l l ] [ 1 ]         // time array of all samples
9:
s a m p l e s _ n u m b e r rows of d a t a         // number of samples or scans
10:
p o i n t s _ n u m b e r columns of d a t a 2         // number of values or points per sample
11:
X p r e v 0         // first position of the X axis
12:
X , Y , Z         // point cloud coordinate array
13:
for  i 1 , s a m p l e s _ n u m b e r do
14:
    // calculates the point cloud coordinates for each sample
15:
     s p e e d s p e e d _ d a t a [ i ]
16:
    append { X p r e v + s p e e d ( t i m e _ d a t a [ i ] t i m e _ d a t a [ i 1 ] ) ones( p o i n t s _ n u m e r )} to X
17:
     X p r e v X [ e n d ]
18:
    append { d a t a [ i ] [ 2 . . e n d ] c o s ( a n g l e _ a r r a y π / 180 ) / 1000 } to Y
19:
    append { d a t a [ i ] [ 2 . . e n d ] s i n ( a n g l e _ a r r a y π / 180 ) / 1000 } to Z
20:
end for

3.2.2. Distance Filtering

Once the point cloud is created, the next step is to eliminate non-relevant information, such as pedestrians walking in front of the sensor, background elements, other sensors at the toll, and vehicles traveling in other lanes. All information from these unnecessary areas is removed by distance filtering, that is, thresholding in terms of height (Z axis) and depth (Y axis) in the point cloud.
The distance filtering limits were set between 1 m and 3.5 m in the Y axis and between −1.5 m and 3 m in the Z axis, considering that the origin is the position of the laser rangefinder, as seen in Figure 12. To define these limits, the width of the lane and the maximum height of the vehicles were considered.
Figure 13 shows the point cloud of Figure 11 after applying distance filtering. The 3D representation of the filtered point cloud is shown on the left, while on the right, it is presented in 2D format. The elimination of background elements can be seen, leaving only the information about the vehicle and the surface of the lane through which it travels.

3.2.3. Ground Surface Extraction

One of the most common processes in the segmentation of elements in a point cloud is the estimation of parametric shape models (planes, cylinders, circles, etc.), facilitating the detection of elements in the scene. To identify and manipulate the ground surface present in the point cloud, a plane extraction algorithm was used. The identification of the planar model of the ground surface allows for the elimination of information present on the road surface. For the segmentation of the ground model, the implementation of the techniques present in Refs. [34,35] was used, which uses the RANSAC (RANdom SAmple Consensus) estimation method.
The method is used to estimate the model of a dominant plane perpendicular to the Z axis, that is parallel to the ground. The plane model has the form of Equation (4). First, the model of a plane is identified with the greatest number of points that fit it, which must be at a distance from the plane less than a defined distance threshold of 5 cm and with a normal approximately parallel to that of the plane, that is, with an angle between them less than an angle threshold defined by 10°. Subsequently, the points that fit the found plane are eliminated from the point cloud, which corresponds to the ground surface. Algorithm 4 outlines the basic steps that are used to extract the ground surface.
a x + b y + c z + d = 0
Algorithm 4 The main computational steps for ground surface extraction.
1:
i t e r a t i o n s 5000
2:
d 0.05         // set distance threshold
3:
ϵ 10         // set angle epsilon threshold
4:
P { p 1 p n }         // set of 3D points
5:
P z ��� { p i | n i × z 0 }         // select set P z with normals parallel to z , angle between n i and z ϵ
6:
for  i 1 , i t e r a t i o n s  do
7:
     C i random P z i P z         // select a random subset of P z
8:
    // find the best plane fit using sample consensus
9:
    estimate ( { a , b , c , d } , A = a · p j x + b · p j y + c · p j z + d , p j C i )
10:
    P i { p j P z | p j fits to A }         // select the set that fit to A , distance between p j and A plane ≤ d
11:
end for
12:
P P larger ( P i )         // point cloud without ground surface, ground surface is set with the most points
Figure 14 shows a segmented point cloud, depicting the vehicle information in black and the ground surface in red. The precise identification of the ground surface stands out, which can be successfully eliminated.

3.2.4. Statistical Filtering

Finally, statistical filtering is used to remove noise from the scene by implementing algorithms described in Refs. [34,36]. Noise appears for several reasons such as dust in the environment, uncertainty in the measurement, etc. In point clouds, this noise can be observed as isolated points without any relationship with the real scene.
The method calculates the mean distance d ¯ of each point to its k closest neighbors. It then estimates the mean μ and standard deviation σ of the mean distance space of all points. Points whose mean distance d ¯ to their k closest neighbors is approximate to the mean μ in a range no greater than α times the standard deviation σ are retained. That is, the remaining filtered point cloud P can be estimated by Equation (5) where P is the complete point cloud. The parameter k was set to 50 closest neighbors and the factor α to 2.5. The computational steps of the statistical filtering method are presented in Algorithm 5.
P { p i P | ( μ α · σ ) d i ( μ + α · σ ) }
Algorithm 5 The main computational steps for statistical filtering.
1:
k 50
2:
α 2.5
3:
P { p 1 p n }         // set of 3D points
4:
for all  p i P  do
5:
    estimate P k         // set of k closest neighbors to p i
6:
    for all  p j P k  do
7:
         d i j { | p i p j | }         // compute the distances between p i and its k closest neighbors
8:
    end for
9:
     d ¯ i mean ( d i j for j [ 1 k ] )         // compute mean distance of p i to its k closest neighbors
10:
end for
11:
μ mean ( d ¯ i for i [ 1 n ] )         // compute mean of mean distance space of all points
12:
σ stddev ( d ¯ i for i [ 1 n ] )         // compute standard deviation of mean distance space of all points
13:
P { p i P | ( μ α · σ ) d i ( μ + α · σ ) }         // filtered point cloud
On the left in Figure 15, the filtered point cloud is shown in black, and the removed points are in red. On the right, only the filtered point cloud is shown, where greater homogeneity can be observed in the surfaces and how some scattered points disappear.

3.3. Dataset Validation and Labeling

Once the range data has been obtained and stored in files, it is essential to perform two fundamental steps. First, extensive validation must be performed to ensure the accuracy and reliability of the data. Secondly, it is necessary to label each file with the corresponding vehicle type, which will allow its use in training automatic classification models. In this specific project, this process was carried out visually and manually, with the support of videos from cameras at the toll.
To streamline and systematize this process, a Graphical User Interface (GUI) was developed in Python. This tool makes it easy to view each point cloud individually, giving the user the ability to verify and validate the accuracy, integrity, and consistency of the data. Additionally, it allows the selection of the vehicle type from a predefined list of options. Once the vehicle type is selected, the tool automatically renames the file, adding an identification number corresponding to that vehicle type.
The creation of this GUI considerably simplified the validation and labeling process, while ensuring greater accuracy in assigning vehicle categories and validating the quality of the generated dataset. Figure 16 shows the interface developed for this purpose, providing a clear and friendly view of the labeling workflow.

4. Application Examples

The purpose of this section is to offer a broader view of the possible applications of the dataset, as well as to explore some potential usage scenarios. In particular, we will focus on two key areas: automatic classification and 3D modeling.

4.1. Application for Automatic Classification

Vehicle classification represents a fundamental field of study, given its relevance in a wide range of applications ranging from surveillance and security systems to traffic management and accident prevention, among others. In this context, a variety of algorithms aimed at accurate object classification have been developed and applied, many of which are based on the processing of point clouds obtained by LiDAR sensors.
Some of the techniques for point cloud classification include: neural networks [37,38,39], Bayesian networks [40], K-nearest neighbors KNN [41], support vector machines SVM [42,43], among others.
In this work, to demonstrate that the dataset is well constructed and labeled, a classifier was implemented and trained from a support vector machine (SVM). The input data to the classifier is geometric information of the vehicles. Figure 17 shows the stages of the feature extraction and classification process.
According to the vehicle types in Table 6, the classification depends largely on the number of tires touching the ground and the dimensions of the vehicles. For this reason, a method was implemented to extract geometric characteristics of vehicles from the information contained in the point cloud. In this work, the following five geometric features were used: number of tires touching the ground, length and height of the vehicle, diameter of the first tire, and height of the vehicle on the first axle (see Figure 18).
Finally, a classifier based on an SVM with a polynomial kernel was trained. The dataset is used with a distribution of 80% point clouds for training and 20% for validation. To balance the number of samples of each vehicle type, only 3% of the point clouds from class 1 (cars, campers, vans) were used. The training results are shown in Figure 19 using a confusion matrix.
In general, a very good performance of the classifier is observed. For the “Cars” and “+3-axles Trucks” classes, the performance is above 95%. With the “Minibuses” class a performance of 79% is achieved and some samples are confused with the “Cars” class, this may be due to some minibuses being as small as vehicles. Similarly, the “Buses” class achieves a performance of 84% and some samples are confused with minibuses for similar reasons to the previous case. In the case of the “2-axle Truck” class, although good performance is also achieved, in the labeling process it is sometimes difficult to differentiate between a large truck and a small one, which according to the regulations has to do with the size of the tires.
It is crucial to highlight that the SVM classifier presented in this study is used only for demonstration purposes on the data set. However, employing a more advanced classifier could lead to higher classification accuracy.

4.2. Application for 3D Modeling

The three-dimensional creation of objects and scenarios is a constant challenge and of great interest in the communities dedicated to graphics, computer vision and photogrammetry. Three-dimensional digital models are essential in a variety of applications, including inspection, navigation, object recognition, visualization and animation [44].
In this study, due to the characteristics and installation method of the data acquisition equipment, only one face of each vehicle is captured in the point cloud dataset. To fill in the missing coordinate points in the unscanned planes, 3D surface reconstruction techniques are employed, as described in Refs. [45,46,47,48].
In the experiment, the scanned face was mirrored and separated from the original by a distance approximately the width of the vehicle. Point clouds were then added to the front, back, top and bottom faces, followed by a plane reconstruction using the Poisson method with the CloudCompare tool [30].
Figure 20 exemplifies the result of 3D modeling of the point cloud dataset. Details such as the side-view mirror and the fender can be seen, demonstrating the wealth of detailed vehicle information in this dataset and its usefulness in 3D modeling.

5. Conclusions

This work addresses the creation of a point cloud data set obtained by laterally scanning vehicles at a toll station, using a laser rangefinder and two Doppler speed sensors. Several range image processing stages, such as distance filtering, ground surface removal, and statistical filtering, were applied to obtain vehicle side profiles without noise and irrelevant information.
The manual labeling and validation process used to classify vehicles into six distinct classes is described; although, the structure of the dataset allows for flexibility in relabeling vehicles based on the specific application.
In addition, two examples of applications of the dataset were presented. In the first, a classification method based on geometric features was proposed using a support vector machine. In the second, a 3D modeling application was demonstrated using three-dimensional reconstruction techniques.
Notwithstanding the above, this dataset provides researchers and developers with the opportunity to test and validate range image processing and classification methods and algorithms, such as deep learning and support vector machines, among others.
The main contribution of this dataset is the 36,026 point clouds of side views of a good variety of vehicles. Therefore, the dataset is an important contribution to research in computer vision and computational intelligence.
In future work, the authors promise to acquire more data and label it to expand the data set. In addition, efforts will be devoted to modifying and testing the implemented processing and classification methods in order to improve the classification accuracy. It is also planned to integrate video cameras into the system for license plate detection and thus add more information to the data set.

Author Contributions

Conceptualization, A.C.-R., E.F.C.-B. and E.B.B.-C.; methodology, A.C.-R., E.F.C.-B. and E.B.B.-C.; software, A.C.-R.; validation, A.C.-R.; formal analysis, A.C.-R.; investigation, A.C.-R., E.F.C.-B. and E.B.B.-C.; resources, A.C.-R.; data curation, A.C.-R.; writing—original draft preparation, A.C.-R.; writing—review and editing, A.C.-R., E.F.C.-B. and E.B.B.-C.; visualization, A.C.-R.; supervision, E.F.C.-B. and E.B.B.-C.; project administration, A.C.-R.; funding acquisition, A.C.-R. All authors have read and agreed to the published version of the manuscript.

Funding

This work is part of a beneficiary project from the National Doctorate in Business call number 758-2016, funded by the Ministry of Science, Technology and Innovation of Colombia, the Universidad del Valle, and the company SSI Soluciones y Suministros para Ingenierías SAS.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are openly available at https://doi.org/10.5281/zenodo.10974361.

Acknowledgments

We thank the Autopistas del Café concession for allowing us access to one of their toll stations to acquire the data presented in this work.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ITSIntelligent transportation systems
LiDARLight detection and ranging
PCLPoint Cloud Library
RANSACRANdomized SAmple Consensus
SVMSupport vector machine
ROIRegion of interest
GUIGraphical User Interface

References

  1. National Research Council (U.S.); Transportation Research Board; Committee for a Study for a Future Strategic Highway Research Program. Strategic Highway Research: Saving Lives, Reducing Congestion, Improving Quality of Life; Technical Report Special Report 260; The National Academies Press: Washington, DC, USA, 2002. [CrossRef]
  2. World Health Organization. Global Status Report on Road Safety 2023; Technical report; World Health Organization: Geneva, Switzerland, 2023. [Google Scholar]
  3. Litman, T. Transportation and public health. Annu. Rev. Public Health 2013, 34, 217–233. [Google Scholar] [CrossRef] [PubMed]
  4. Newman-Askins, R.; Ferreira, L.; Bunker, J. Intelligent transport systems evaluation: From theory to practice. In Proceedings of the 21st ARRB and 11th REAAA Conference, Cairns, Australia, 18–23 May 2003. [Google Scholar]
  5. Qi, L. Research on intelligent transportation system technologies and applications. In Proceedings of the 2008 Workshop on Power Electronics and Intelligent Transportation System, PEITS, Guangzhou, China, 2–3 August 2008; pp. 529–531. [Google Scholar] [CrossRef]
  6. Banse, K.; Herrera, L.F.; Nuñez, M.; Navarro, J.C.; Bermudez, D.; Chavarriaga, J. ITS development in Colombia: Challenges and opportunities. In Proceedings of the 2018 ICAI Workshops, ICAIW, Bogotá, Colombia, 1–3 November 2018. [Google Scholar] [CrossRef]
  7. Keerthi Kiran, V.; Parida, P.; Dash, S. Vehicle detection and classification: A review. In Proceedings of the 2019 Innovations in Bio-Inspired Computing and Applications, IBICA, Gunupur, Odisha, India, 16–18 December 2019; pp. 45–56. [Google Scholar] [CrossRef]
  8. Zheng, J.; Yang, S.; Wang, X.; Xia, X.; Xiao, Y.; Li, T. A Decision Tree Based Road Recognition Approach Using Roadside Fixed 3D LiDAR Sensors. IEEE Access 2019, 7, 53878–53890. [Google Scholar] [CrossRef]
  9. Sundaravalli, G.; Krishnaveni, K. A survey on vehicle classification techniques. Int. J. Eng. Res. Comput. Sci. Eng. (IJERCSE) 2018, 5, 268–273. [Google Scholar]
  10. Celso, R.N.; Ting, Z.B.; Del Carmen, D.J.R.; Cajote, R.D. Two-Step Vehicle Classification System for Traffic Monitoring in the Philippines. In Proceedings of the 2018 IEEE Region 10 Annual International Conference (TENCON), Jeju, Republic of Korea, 28–31 October 2018; pp. 2028–2033. [Google Scholar] [CrossRef]
  11. Hussain, K.F.; Moussa, G.S. Automatic Vehicle Classification System using range sensor. In Proceedings of the 2005 International Conference on Information Technology: Coding and Computing, ITCC, Las Vegas, NV, USA, 4–6 April 2005; Volume 2, pp. 107–112. [Google Scholar] [CrossRef]
  12. Singh, V.; Srivastava, A.; Kumar, S.; Ghosh, R. A Structural Feature Based Automatic Vehicle Classification System at Toll Plaza. In Proceedings of the 2019 4th International Conference on Internet of Things and Connected Technologies (ICIoTCT), Jaipur, India, 9–10 May 2019; Volume 1122, pp. 1–10. [Google Scholar] [CrossRef]
  13. Ahmed, W.; Arafat, S.Y.; Gul, N. A Systematic Review on Vehicle Identification and Classification Techniques. In Proceedings of the 2018 21st International Multi Topic Conference, INMIC, Karachi, Pakistan, 1–2 November 2018; pp. 1–6. [Google Scholar] [CrossRef]
  14. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision meets robotics: The KITTI dataset. Int. J. Robot. Res. 2013, 32, 1231–1237. [Google Scholar] [CrossRef]
  15. Chang, M.F.; Lambert, J.; Sangkloy, P.; Singh, J.; Bak, S.; Hartnett, A.; Wang, D.; Carr, P.; Lucey, S.; Ramanan, D.; et al. Argoverse: 3D tracking and forecasting with rich maps. In Proceedings of the 2019 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 8740–8749. [Google Scholar] [CrossRef]
  16. Caesar, H.; Bankiti, V.; Lang, A.H.; Vora, S.; Liong, V.E.; Xu, Q.; Krishnan, A.; Pan, Y.; Baldan, G.; Beijbom, O. Nuscenes: A multimodal dataset for autonomous driving. In Proceedings of the 2020 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11618–11628. [Google Scholar] [CrossRef]
  17. Sun, P.; Kretzschmar, H.; Dotiwalla, X.; Chouard, A.; Patnaik, V.; Tsui, P.; Guo, J.; Zhou, Y.; Chai, Y.; Caine, B.; et al. Scalability in perception for autonomous driving: Waymo open dataset. In Proceedings of the 2020 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2443–2451. [Google Scholar] [CrossRef]
  18. Geyer, J.; Kassahun, Y.; Mahmudi, M.; Ricou, X.; Durgesh, R.; Chung, A.S.; Hauswald, L.; Pham, V.H.; Mühlegg, M.; Dorn, S.; et al. A2D2: Audi Autonomous Driving Dataset. arXiv 2020, arXiv:2004.06320. [Google Scholar]
  19. Huang, X.; Wang, P.; Cheng, X.; Zhou, D.; Geng, Q.; Yang, R. The ApolloScape Open Dataset for Autonomous Driving and Its Application. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 2702–2719. [Google Scholar] [CrossRef] [PubMed]
  20. Xiao, P.; Shao, Z.; Hao, S.; Zhang, Z.; Chai, X.; Jiao, J.; Li, Z.; Wu, J.; Sun, K.; Jiang, K.; et al. PandaSet: Advanced Sensor Suite Dataset for Autonomous Driving. In Proceedings of the 2021 IEEE Conference on Intelligent Transportation Systems, ITSC, Indianapolis, IN, USA, 19–22 September 2021; pp. 3095–3101. [Google Scholar] [CrossRef]
  21. Ye, Z.; Wang, Z.; Chen, X.; Zhou, T.; Yu, C.; Guo, J.; Li, J. ZPVehicles: A dataset of large vehicle 3D point cloud data. In Proceedings of the 2023 IEEE International Workshop on Metrology for Automotive, MetroAutomotive, Modena, Italy, 28–30 June 2023; pp. 234–239. [Google Scholar] [CrossRef]
  22. Abdelbaki, H.M.; Hussain, K.; Gelenbe, E. A laser intensity image based automatic vehicle classification system. In Proceedings of the 2001 IEEE Intelligent Transportation Systems, ITSC, Oakland, CA, USA, 25–29 August 2001; pp. 460–465. [Google Scholar] [CrossRef]
  23. Sato, T.; Aoki, Y.; Takebayashi, Y. Vehicle axle counting using two LIDARs for toll collection systems. In Proceedings of the 2014 21st World Congress on Intelligent Transport Systems, ITSWC: Reinventing Transportation in Our Connected World, Detroit, MI, USA, 7–11 September 2014. [Google Scholar]
  24. Chidlovskii, B.; Csurka, G.; Rodriguez-Serrano, J. Vehicle type classification from laser scans with global alignment kernels. In Proceedings of the 2014 7th IEEE International Conference on Intelligent Transportation Systems, ITSC, Qingdao, China, 8–11 October 2014; pp. 2840–2845. [Google Scholar] [CrossRef]
  25. Sandhawalia, H.; Rodriguez-Serrano, J.A.; Poirier, H.; Csurka, G. Vehicle type classification from laser scanner profiles: A benchmark of feature descriptors. In Proceedings of the 2013 IEEE Conference on Intelligent Transportation Systems, ITSC, The Hague, The Netherlands, 6–9 October 2013; pp. 517–522. [Google Scholar] [CrossRef]
  26. Xu, Z.; Wei, J.; Chen, X. Vehicle recognition and classification method based on laser scanning point cloud data. In Proceedings of the 2015 3rd International Conference on Transportation Information and Safety, ICTIS, Wuhan, China, 25–28 June 2015; pp. 44–49. [Google Scholar] [CrossRef]
  27. Li, Y.; Tok, A.Y.; Sun, Z.; Ritchie, S.G.; Allu, K.R. LiDAR Vehicle Point Cloud Reconstruction Framework for Axle-Based Classification. IEEE Sens. J. 2023, 23, 11168–11180. [Google Scholar] [CrossRef]
  28. Hokuyo Automatic Co. Ltd. Scanning Rangefinder UTM-30LX; Hokuyo Automatic Co. Ltd.: Osaka, Japan, 2021; Available online: https://www.hokuyo-aut.jp/search/single.php?serial=169 (accessed on 17 April 2024).
  29. Applied Concepts Inc. Stalker Stationary Speed Sensor; Applied Concepts Inc.: Richardson, TX, USA, 2019; Available online: https://www.stalkerradar.com/oem/Stationary.html (accessed on 17 April 2024).
  30. CloudCompare (Version 2.13.1). GPL Software. 3D Point Cloud and Mesh Processing Software Open Source Project. Available online: https://www.cloudcompare.org/ (accessed on 17 April 2024).
  31. National Institute of Advanced Industrial Science and Technology. Hokuyoaist: Hokuyo Range Sensor Driver; National Institute of Advanced Industrial Science and Technology: Tokyo, Japan, 2017. Available online: https://github.com/gbiggs/hokuyoaist (accessed on 17 April 2024).
  32. National Institute of Advanced Industrial Science and Technology. HokuyoAIST Range Sensor Driver—HokuyoAIST 3.0.0 Documentation; National Institute of Advanced Industrial Science and Technology: Tokyo, Japan, 2011; Available online: https://gbiggs.github.io/hokuyoaist/ (accessed on 17 April 2024).
  33. Rusu, R.B.; Cousins, S. 3D is here: Point Cloud Library (PCL). In Proceedings of the 2011 IEEE International Conference on Robotics and Automation, Shanghai, China, 9–13 May 2011; pp. 1–4. [Google Scholar] [CrossRef]
  34. Rusu, R.B. Semantic 3D Object Maps for Everyday Manipulation in Human Living Environments. Ph.D. Thesis, Technical University of Munich, Munich, Germany, 2009. [Google Scholar] [CrossRef]
  35. Fischler, M.A.; Bolles, R.C. Random sample consensus: A Paradigm for Model Fitting with Applications to Image Analysis and Automated Cartography. Commun. ACM 1981, 24, 381–395. [Google Scholar] [CrossRef]
  36. Rusu, R.B.; Blodow, N.; Marton, Z.; Soos, A.; Beetz, M. Towards 3D object maps for autonomous household robots. In Proceedings of the 2007 IEEE International Conference on Intelligent Robots and Systems, San Diego, CA, USA, 29 October–2 November 2007; pp. 3191–3198. [Google Scholar] [CrossRef]
  37. Wei, S. Three-dimensional point cloud classification based on multi-scale dynamic graph convolutional network. In Proceedings of the 2021 3rd International Academic Exchange Conference on Science and Technology Innovation, IAECST, Guangzhou, China, 10–12 December 2021; pp. 603–606. [Google Scholar] [CrossRef]
  38. Zhang, Z.; Wang, Q.; Wang, M.; Shen, T. Graph Neural Network with Multi-Kernel Learning for Multispectral Point Cloud Classification. In Proceedings of the 2023 International Geoscience and Remote Sensing Symposium (IGARSS), Pasadena, CA, USA, 16–21 July 2023; pp. 970–973. [Google Scholar] [CrossRef]
  39. Yu, S.; Wang, M.; Zhang, C.; Li, J.; Yan, K.; Liang, Z.; Wei, R. A Dynamic Multi-Branch Neural Network Module for 3D Point Cloud Classification and Segmentation Using Structural Re-parametertization. In Proceedings of the 2023 11th International Conference on Agro-Geoinformatics (Agro-Geoinformatics), Wuhan, China, 25–28 July 2023; pp. 1–6. [Google Scholar] [CrossRef]
  40. Kang, Z.; Yang, J.; Zhong, R. A Bayesian-Network-Based Classification Method Integrating Airborne LiDAR Data with Optical Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 1651–1661. [Google Scholar] [CrossRef]
  41. Wang, L.; Kou, Q.; Wei, R.; Zhou, L.; Fang, T.; Zhang, J. An Improved Substation Equipment Recognition Algorithm by KNN Classification of Subspace Feature Vector. In Proceedings of the 2021 China Automation Congress, CAC, Beijing, China, 22–24 October 2021; pp. 7541–7546. [Google Scholar] [CrossRef]
  42. Zhan, Q.; Yu, L. Objects classification from laser scanning data based on multi-class support vector machine. In Proceedings of the 2011 International Conference on Remote Sensing, Environment and Transportation Engineering, RSETE, Nanjing, China, 24–26 June 2011; pp. 520–523. [Google Scholar] [CrossRef]
  43. Raktrakulthum, P.A.; Netramai, C. Vehicle classification in congested traffic based on 3D point cloud using SVM and KNN. In Proceedings of the 2017 9th International Conference on Information Technology and Electrical Engineering, ICITEE, Phuket, Thailand, 12–13 October 2017; pp. 1–6. [Google Scholar] [CrossRef]
  44. Remondino, F.; El-hakim, S. Image-based 3D modelling: A review. Photogramm. Rec. 2006, 21, 269–291. [Google Scholar] [CrossRef]
  45. Kazhdan, M.; Bolitho, M.; Hoppe, H. Robust poisson surface reconstruction. In Proceedings of the Symposium on Geometry Processing, Cagliari, Italy, 26–28 June 2006. [Google Scholar] [CrossRef]
  46. Yan, S.; Peng, Y.; Wang, G.; Lai, S.; Zhang, M. Weakly Supported Plane Surface Reconstruction via Plane Segmentation Guided Point Cloud Enhancement. IEEE Access 2020, 8, 60491–60504. [Google Scholar] [CrossRef]
  47. Wang, S.; Cai, G.; Cheng, M.; Marcato Junior, J.; Huang, S.; Wang, Z.; Su, S.; Li, J. Robust 3D reconstruction of building surfaces from point clouds based on structural and closed constraints. ISPRS J. Photogramm. Remote Sens. 2020, 170, 29–44. [Google Scholar] [CrossRef]
  48. Lu, C.H.; Chen, X.H. Improved iterative Poisson point cloud surface reconstruction. In Proceedings of the 2023 3rd International Conference on Digital Society and Intelligent Systems (DSInS), Chengdu, China, 10–12 November 2024; pp. 382–385. [Google Scholar] [CrossRef]
Figure 1. Dataset examples.
Figure 1. Dataset examples.
Data 09 00087 g001
Figure 2. Top view of the sensor installation diagram: (1) Scanning laser rangefinder. (2) Speed sensors.
Figure 2. Top view of the sensor installation diagram: (1) Scanning laser rangefinder. (2) Speed sensors.
Data 09 00087 g002
Figure 3. Sensors installed at the toll station: (a) Scanning laser rangefinder. (b,c) Speed sensors for the right and left lanes, respectively.
Figure 3. Sensors installed at the toll station: (a) Scanning laser rangefinder. (b,c) Speed sensors for the right and left lanes, respectively.
Data 09 00087 g003
Figure 4. Front view of the laser rangefinder installation diagram.
Figure 4. Front view of the laser rangefinder installation diagram.
Data 09 00087 g004
Figure 5. Laser rangefinder angular range.
Figure 5. Laser rangefinder angular range.
Data 09 00087 g005
Figure 6. Dataset storage structure.
Figure 6. Dataset storage structure.
Data 09 00087 g006
Figure 7. Data arrays for the right and left lanes, respectively.
Figure 7. Data arrays for the right and left lanes, respectively.
Data 09 00087 g007
Figure 8. Example of file content.
Figure 8. Example of file content.
Data 09 00087 g008
Figure 9. Region of interest for vehicle detection.
Figure 9. Region of interest for vehicle detection.
Data 09 00087 g009
Figure 10. Processing stages for the point cloud.
Figure 10. Processing stages for the point cloud.
Data 09 00087 g010
Figure 11. Three-dimensional image of a raw point cloud.
Figure 11. Three-dimensional image of a raw point cloud.
Data 09 00087 g011
Figure 12. Region for distance filter.
Figure 12. Region for distance filter.
Data 09 00087 g012
Figure 13. Point cloud with distance filtering.
Figure 13. Point cloud with distance filtering.
Data 09 00087 g013
Figure 14. Point cloud with ground surface extraction.
Figure 14. Point cloud with ground surface extraction.
Data 09 00087 g014
Figure 15. Point cloud with statistical filtering.
Figure 15. Point cloud with statistical filtering.
Data 09 00087 g015
Figure 16. GUI for labeling and validation.
Figure 16. GUI for labeling and validation.
Data 09 00087 g016
Figure 17. Stages for feature extraction and classification.
Figure 17. Stages for feature extraction and classification.
Data 09 00087 g017
Figure 18. Geometric features.
Figure 18. Geometric features.
Data 09 00087 g018
Figure 19. Confusion matrix for SVM training.
Figure 19. Confusion matrix for SVM training.
Data 09 00087 g019
Figure 20. Three-dimensional model of a vehicle.
Figure 20. Three-dimensional model of a vehicle.
Data 09 00087 g020
Table 1. Overview of publicly available point cloud datasets.
Table 1. Overview of publicly available point cloud datasets.
DatasetYearAnn. Fr.3D Box.LiDARClassesLocationDistanceNight/RainDuration
KITTI [14]201315 k200 k18Karlsruhe-No/No6 h
Argoverse [15]201922 k993 k215Pittsburgh, Miami300 kmYes/Yes320 h
NuScenes [16]202040 k1.4 M123Boston, Singapore242 kmYes/Yes15 h
Waymo [17]2020230 k12 M543× USA76 km2Yes/Yes6 h
A2D2 [18]202012 k-5143× Germay-Yes/Yes-
ApolloScape [19]2020144 k70 k2354× China-Yes/No-
PandaSet [20]20218 k8 k228San Francisco-Yes/-0.3 h
Table 2. Overview of non-generic datasets.
Table 2. Overview of non-generic datasets.
AuthorsYearVeh.LiDARTypeClass.LocationSensor PositionNight/RainDuration
Abdelbaki et al. [11,22]20015 k12D5USAOverhead-/Yes-
Sato et al. [23]201420622D4JapanSide-/--
Chidlovskii et al. [24,25]201330 k32D6USAOverhead-/--
Xu et al. [26]201527032D3ChinaSide and overhead-/--
Li et al. [27]202310 k13D11USASideYes/Yes19 days
Ye et al. [21]202380032D11ChinaSide and overhead-/--
This work202430 k12D6ColombiaSideYes/Yes12 days
Table 3. Main technical specifications of the Hokuyo UTM-30LX scanning laser range finder [28].
Table 3. Main technical specifications of the Hokuyo UTM-30LX scanning laser range finder [28].
ParameterDetails
Light sourceSemiconductor laser diode ( λ = 905 nm)
Detection range0.1 to 30 m
Accuracy0.1 to 10 m: ±30 mm, 10 to 30 m: ±50 mm
Scan angle270°
Angular resolution0.25° (360°/1440 steps)
Scan time25 ms/scan
InterfaceUSB2.0 (Full Speed)
Command systemExclusively designed command SCIP Ver.2.0
Table 4. Main technical specifications of the Stalker stationary speed sensor [29].
Table 4. Main technical specifications of the Stalker stationary speed sensor [29].
ParameterDetails
TypeStationary Doppler radar speed sensor
Communication interfaceRS-232
Speed range1 MPH to 200 MPH (1.6 to 321 km/h)
Accuracy+/−3%
Table 5. Main technical specifications of the computer.
Table 5. Main technical specifications of the computer.
ParameterDetails
OSUbuntu 18.04.6 LTS
ProcessorIntel(R) Core(TM) i7-8550U CPU @ 1.80 GHz
RAM16 GB
Storage512 GB SSD
Table 6. Vehicle types.
Table 6. Vehicle types.
TypeDescriptionQuantityExamples
1Cars, campers and vans, with or without a trailer, ambulances, small 2-axle trucks with a single tire on the rear axle31,432Data 09 00087 i001
2Minibuses with single tire on the rear axle452Data 09 00087 i002
32-axle buses with double tires on the rear axle1158Data 09 00087 i003
4Buses and trucks with 3 or more axles1008Data 09 00087 i004
5Small 2-axle trucks with double tires on the rear axle 11179Data 09 00087 i005
6Large 2-axle trucks with double tires on the rear axle 2797Data 09 00087 i006
Total36,026
1 Trucks whose gross vehicle weight must be less than 8.5 tons. For example: light trucks, food transport trucks, cranes, etc. 2 Trucks whose gross vehicle weight must be greater than 8.5 tons, for example, heavy trucks, trucks for transporting materials, dump trucks, etc.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Campo-Ramírez, A.; Caicedo-Bravo, E.F.; Bacca-Cortes, E.B. A Point Cloud Dataset of Vehicles Passing through a Toll Station for Use in Training Classification Algorithms. Data 2024, 9, 87. https://doi.org/10.3390/data9070087

AMA Style

Campo-Ramírez A, Caicedo-Bravo EF, Bacca-Cortes EB. A Point Cloud Dataset of Vehicles Passing through a Toll Station for Use in Training Classification Algorithms. Data. 2024; 9(7):87. https://doi.org/10.3390/data9070087

Chicago/Turabian Style

Campo-Ramírez, Alexander, Eduardo F. Caicedo-Bravo, and Eval B. Bacca-Cortes. 2024. "A Point Cloud Dataset of Vehicles Passing through a Toll Station for Use in Training Classification Algorithms" Data 9, no. 7: 87. https://doi.org/10.3390/data9070087

Article Metrics

Back to TopTop