LiDAR-based identification is a recent yet challenging initiation considering various circumstances and precisions. Few comprehensive studies have begun with the parallel of conventional video-based tracking with the LiDAR sensor. But all these studies considered 3D LiDAR as their alternative. We thought differently and used a 2D LiDAR sensor to do so. Considerable outcomes, less processing time, inexpensive installation, and wide practical applications made this research worthy.
4.1: Gait based person identification
We thought of the different data combinations during acquisition to make the application-wide in range and reliability. A homogeneous and heterogeneous LiDAR setup was considered for the collection of data. Though a single LiDAR is sufficient to get a pedestrian's desired data, critically distributed and overlapped data are always dilute the system performance. Ankle level LiDAR setup was our primary focus to track and identify a pedestrian. Besides, the uppermost setup of the sensors does not enhance the system performance when the measurements consider a two-dimensional scenario. We placed two LiDAR sensors on a single tripod; parallelly, another two sensors were placed on the second tripod. These two tripods were kept 2 meters distance in a ninety-degree angle experiment setup. We considered different ages and heights pedestrians. They were unbiased in gender, clothes, shoes, and region also. We substantially found numerous walking styles based on their properties. To plot LiDAR data on an image was our primary challenge, but we cautiously did the experiments and found excellent accuracy there. Our proposed KoLaSU dataset consists of fourteen outdoor sequences where twenty-nine participants were attended with five to ten minutes walking with the experiment setup. We considered a standard forty fps rate to write the data. For cross-validation, we considered the 100 fps rate also. In all our experiments, we kept our datasets into three groups. Sixty percent of the data were considered train data, where the remaining 40 percent were split equally into test and validation sets.
Table 1 shows an overall experimental result of gait-based recognition. As discussed, we considered 14 different conditions data for our KoLaSU person tracking dataset; here, we placed 9 of these. We split out all four LiDAR's data individually as we kept four top rows in table 1. Similarly, in the table, we merged different LiDAR data as the number assigned (i.e., LiDAR 12, merged LiDAR 1 and LiDAR 2's data in MHI with 40 fps). Among 29 participants, we keep 26 pedestrians for our experiment after the filter. Here 60 percent of the total data of a set were kept for training the system, and the remaining 40 percent were divided equally for validation and test dataset. For this research, we used the GIGABYTE BRIX GPU machine for the analysis of our data. The batch size was considered as 38, and the number of epochs varied from 25 to 50 periodically. We used a deep neural network to train our model. Here a pre-trained ResNet 18 network was used to train the dataset. We used ResNet 34 and ResNet 50 also for cross-checking the system performance. We placed some results of these in the next section. We randomly selected our train, test, and validation datasets among all machine-generated data, where all segments were utterly disjoint. Initially, we kept every person's data in three parts: train, test, and validation group. Further, we enhanced our study for unknown test data set. Without prior information, man or machine cannot identify a new individual; thus, the system reacts. From table 1, we see that accuracy in three different segments is very impressive and near about 99 percent correctly identified. Though some data are not accurately captured due to congestion problems (i.e., KoLaSU LiDAR 3 and KoLaSU LiDAR 4), their performance in the test case did not go below 93 percent. Moreover, their combined dataset (KoLaSU LiDAR 34) performed significantly well with 99 percent precession.
Table 1
Gait-based person identification on different parameters
Data
|
Experiments
Type
|
Batch
Size
|
Epoch
Size
|
GPU
|
DNN
Model
|
Train
Accuracy
|
Train
Loss
|
Test Accuracy
|
Test Loss
|
Validation Accuracy
|
Validation Loss
|
KoLaSU LiDAR 1
|
26 Persons Individual (60%,20%,20%)
|
38
|
25
|
Yes
|
Resnet 18
|
0.99421
|
0.02
|
0.9843
|
0.05
|
0.9851
|
0.04718
|
KoLaSU LiDAR 2
|
26 Persons Individual (60%,20%,20%)
|
38
|
25
|
Yes
|
Resnet 18
|
0.99324
|
0.02329
|
0.9846
|
0.0494
|
0.98502
|
0.04515
|
KoLaSU LiDAR 3
|
26 Persons Individual (60%,20%,20%)
|
38
|
25
|
Yes
|
Resnet 18
|
0.97721
|
0.00388
|
0.9336
|
0.2144
|
0.93478
|
0.2078
|
KoLaSU LiDAR 4
|
26 Persons Individual (60%,20%,20%)
|
38
|
25
|
Yes
|
Resnet 18
|
0.98038
|
0.0627
|
0.9479
|
0.1689
|
0.94711
|
0.16857
|
KoLaSU LiDAR 13
|
26 Persons Individual (60%,20%,20%)
|
38
|
25
|
Yes
|
Resnet 18
|
0.9982
|
0.0076
|
0.9962
|
0.0132
|
0.996
|
0.0145
|
KoLaSU LiDAR 24
|
26 Persons Individual (60%,20%,20%)
|
38
|
25
|
Yes
|
Resnet 18
|
0.99831
|
0.00717
|
0.99622
|
0.01315
|
0.99611
|
0.013
|
KoLaSU LiDAR 12
|
26 Persons Individual (60%,20%,20%)
|
38
|
25
|
Yes
|
Resnet 18
|
0.99719
|
0.00033
|
0.9934
|
0.0208
|
0.99305
|
0.0208
|
KoLaSU LiDAR 34
|
26 Persons Individual (60%,20%,20%)
|
38
|
25
|
Yes
|
Resnet 18
|
0.99821
|
0.00772
|
0.99343
|
0.0209
|
0.9942
|
0.0198
|
KoLaSU LiDAR 1234
|
26 Persons Individual (60%,20%,20%)
|
38
|
25
|
Yes
|
Resnet 18
|
0.99869
|
0.00561
|
0.99658
|
0.01118
|
0.99713
|
0.0095
|
4.2 Comparison to different data types
Figure 8 shows three different input datasets' accuracy with a different dataset, also showing designed network is accurately trained with all kinds of data. To reduce the overfitting, we validated the network, and here its accuracy is impressive, and none of the datasets goes below 93 percent. The test accuracy also goes through with validation accuracy and follows its footstep. Accuracies and losses are inversely proportional in a system. Our system is also showing this trend. Most of the loss calculations remain on the borderline of the scale in Fig. 9. The symmetry in the accuracy and loss carves of the three parameters substantially shows the network credibility. This designed network and its performances instinctively develop a logical ground of using two-dimensional LiDAR sensors for person identification in broad.
We performed rigorous testing with different datasets to test the system performance. All fourteen datasets were considered for this cross-testing. We analyzed the results and found an essential symmetry in the performance analysis phase, and those wholly aligned with our theoretical expectations—figure 10 shows these results in detail. Suppose the top four datasets on the figure are KoLaSU LiDAR 24, 13 respectively. We trained and validated the system with the same dataset but changed the test data only in four cases. LiDAR 24 and LiDAR 13 are created with merging sensors 2 and 4 and sensors 1 and 3, respectively. For testing, we used only LiDAR 4 and 2 and LiDAR 3 and LiDAR 1 separately. The bar chart shows that training and validation accuracy is nearly absolute, with a test accuracy below 20 percent. We considered unbiased disjoint data for all cases. As the figure shows, if a person is accurately trained by the neural network and tested differently, the system performance will degrade. The same performance persists for all the cases except the combined dataset LiDAR 1234 is tried with test data 24 and 13. Here the performance reached up to 38 percent but was not impressive. So here, we can conclude that to achieve the best performance of the system; it should be trained and tested with the same types of data; any other biasedness is not necessary there.
Table 2
Performance test with different DNN model
Data
|
KoLaSU LiDAR 1234 TestCross24
|
KoLaSU LiDAR 1234
TestCross24
|
Experiments Type
|
26 Persons Individual (60%,20%,20%)
|
26 Persons Individual (60%,20%,20%)
|
Batch Size
|
38
|
38
|
Epoch
|
25
|
40
|
GPU
|
Yes
|
Yes
|
Model
|
Resnet 18
|
Resnet 50_2
|
Train Accuracy
|
0.99864
|
0.99999
|
Train Loss
|
0.00589
|
0.000354
|
Test Accuracy
|
0.379
|
0.4007
|
Test Loss
|
4.2741
|
3.5852
|
Validation Accuracy
|
0.99721
|
0.99956
|
Validation Loss
|
0.00589
|
0.001578
|
Besides ResNet 18, we analyzed different neural networks to test our data's system performance and effectiveness. We show one of such types of analysis in Table 2. For the same dataset KoLaSU LiDAR 1234, we trained and validated it by ResNet 18 and ResNet 50. The rest of the parameters remained the same except epoch size. We tested the system with different datasets KoLaSU LiDAR 24 and tried to analyze the performance of two separate networks. ResNet 18 gave almost 38 percent accuracy, whereas ResNet 50 performed with 40 percent accuracy. But network size of ResNet 50 is abruptly huge than ResNet 18, and computation time is highly excessive. Thus, we decided to consider ResNet 18 rather than ResNet 50 even though its performance is little improved.
To test the system performance in different ways, we combined different datasets and trained and validated our system. Our achieved accuracies were impressive in all cases. In Fig. 11, we placed some experiments based on combined datasets. Here we put system accuracy and loss together. Six datasets were connected with their aligned ones. Suppose LiDAR 24 data was combined with LiDAR 2 and 4 data for training and validation of the system. Further, we tested the system individually with LiDAR 24, 2, and 4 data. The same scenarios were performed with LiDAR 1234, 13, and 24 also. In all cases, though the system was trained with multiple groups of data and tested with individual one, it performed as a regular system performed previously. From Fig. 11, we see that all six combinations' train, validation, and test accuracies lay around 99 percent, where its loss remains more minor as below 20 percent. Here accuracy and loss curves follow a symmetry as well, which is an indication of network performance.
4.3 Comparison with contemporary studies
To the best of our knowledge, there was no such research performed with 2D LiDAR sensors to identify a person based on gait analysis. Even a few studies were conducted with a 3D LiDAR sensor. Though they used different sensor setups and designed their model, we compared the overall system precisions in Table 3. Benedek et al. [4] initiated the research for lidar-based gait analysis. They prepared their dataset, SZTAK-LGA, with 28 participants. They used CNN (convolutional neural network) and MLP (multi-layer perceptron) for training and testing the system. They used a different number of people in their experiments; an increased number of people degraded the system performance from 92 percent (for five people) to 75 percent (for 28 people). Yamada et al. [3] performed a thorough experiment on lidar-based gait analysis. This research was also conducted with a 3D LiDAR sensor. They also prepared their dataset, PCG (point cloud gait), with 30 participants. A CNN and LSTM (long short-term memory) neural network model was applied for training and testing the system. Though they get different accuracies in different input patterns, here l = 1:8 gave a maximum of 72 percent in general. Compared with rest two, we used our own dataset KoLaSU, with 29 participants, and a two-dimensional LiDAR sensor only. We used a residual deep neural network (ResNet) for training, validation, and testing the system. We randomly used utterly unbiased datasets categorized into three classes (train, test, and validation). The average system performance is greater than 98 percent, emphasizing a wide use of 2D sensors in different applications.
Table 3
Performance comparison with state-of-the-art technologies
Method
|
Dataset
|
Sensor
|
Model
|
Accuracies
|
Benedek et al. [4]
|
SZTAKI-LGA (28 People)
|
3D LiDAR
|
CNN + MLP
|
80% (approx.)
|
Yamada et al. [3]
|
PCG (30 People)
|
3D LiDAR
|
CNN + LSTM
|
72% (approx.)
|
Our
|
KoLaSU (29 People)
|
2D LiDAR
|
ResNet
|
98% (approx.)
|