iBet uBet web content aggregator. Adding the entire web to your favor.
iBet uBet web content aggregator. Adding the entire web to your favor.



Link to original content: https://doi.org/10.3390/s22145270
Tracking Missing Person in Large Crowd Gathering Using Intelligent Video Surveillance
Next Article in Journal
Classification of Three Volatiles Using a Single-Type eNose with Detailed Class-Map Visualization
Next Article in Special Issue
Artificial Intelligence Applications and Self-Learning 6G Networks for Smart Cities Digital Ecosystems: Taxonomy, Challenges, and Future Directions
Previous Article in Journal
Fair Facial Attribute Classification via Causal Graph-Based Attribute Translation
Previous Article in Special Issue
Fatigue Driving Detection Method Based on Combination of BP Neural Network and Time Cumulative Effect
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Tracking Missing Person in Large Crowd Gathering Using Intelligent Video Surveillance

1
Faculty of Computer and Information System, Islamic University of Madinah, Madinah 42351, Saudi Arabia
2
Department of Physics, Federal Urdu University of Arts, Science & Technology, Karachi 75300, Pakistan
3
Department of Computer Science, Federal Urdu University of Arts, Science & Technology, Islamabad 45570, Pakistan
4
Department of Computer Science and Information Technology, Sir Syed University of Engineering and Technology, Karachi 75300, Pakistan
5
James Watt School of Engineering, University of Glasgow, Glasgow G12 8QQ, UK
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(14), 5270; https://doi.org/10.3390/s22145270
Submission received: 30 May 2022 / Revised: 1 July 2022 / Accepted: 11 July 2022 / Published: 14 July 2022
(This article belongs to the Special Issue Artificial Intelligence Methods for Smart Cities)

Abstract

:
Locating a missing child or elderly person in a large gathering through face recognition in videos is still challenging because of various dynamic factors. In this paper, we present an intelligent mechanism for tracking missing persons in an unconstrained large gathering scenario of Al-Nabawi Mosque, Madinah, KSA. The proposed mechanism in this paper is unique in two aspects. First, there are various proposals existing in the literature that deal with face detection and recognition in high-quality images of a large crowd but none of them tested tracking of a missing person in low resolution images of a large gathering scenario. Secondly, our proposed mechanism is unique in the sense that it employs four phases: (a) report missing person online through web and mobile app based on spatio-temporal features; (b) geo fence set estimation for reducing search space; (c) face detection using the fusion of Viola Jones cascades LBP, CART, and HAAR to optimize the results of the localization of face regions; and (d) face recognition to find a missing person based on the profile image of reported missing person. The overall results of our proposed intelligent tracking mechanism suggest good performance when tested on a challenging dataset of 2208 low resolution images of large crowd gathering.

1. Introduction

Tracking and locating a person automatically in an unconstrained large crowd gathering through face detection and recognition is still challenging. Face detection and recognition is challenging due to various dynamic factors such as low resolution, variable crowd distance from installed cameras, mobility of cameras and the crowd. In this paper, we propose an automatic tracking of the registered missing persons in a large Al-Nabawi mosque gathering scenario where millions of pilgrims gather to perform religious rituals in a covered area of approximately 73 acres.
The contribution in this paper is novel in two aspects. First, to the best of our knowledge, this is one of the few proposals for automatic tracking of the missing persons in a large gathering with low-resolution images. There are various proposals in the literature, which apply face recognition algorithms to large crowd images such as [1,2,3,4]; however, tracking a person in a large crowd with low-resolution images is rare. Several state-of-the-art deep learning algorithms are used for face recognition but with high-quality images such as [5]; however, our study of the related work suggested that they do not show good performance on low-resolution images of the unconstrained environment. To the best of our knowledge, dividing the surveillance premises into geofences and estimating set of geofences for probable presence of missing person is used for the first time in such works.Therefore, we believe our proposal in this paper contributes to the existing knowledge.
Second, our proposed mechanism is unique in terms of its methodology. In which we first efficiently reduce the search space to locate the missing person by applying the proposed geofence set estimation algorithm and then employ our face detector algorithm which fuses the three cascades of Viola Jones to optimize the number of localized face regions from each frame. Finally, the proposed mechanism applies the face recognition algorithm using the registered profile images of the missing persons to track them in the crowd.
We consider the unconstrained large gathering scenario of Al-Nabawi mosque, Madinah, Saudi Arabia. We first divided the total area of the mosque in 25 geofences and 20 cameras installed in this covered areas as our source of input data in the form of video sequences. Then, the system extracts frames of low quality images. We developed a mobile and web based system through which a head of the pilgrim’s family or the head of the pilgrims group can report his missing companion with time and location. Then, proposed Geo fence set estimation algorithm will result in suggested set of geofences where the missing person could be found. This will significantly reduce the search space and the system will start tracking from the videos of cameras installed in the suggested geo fences premises. Then, we apply our face detector algorithm which is a fusion of three Viola Jones cascades which produces high number of localized face regions with accuracy which is much higher than applying the viola Jones cascades individually. Then, face recognition algorithms is employed to find the missing person from the detected face regions based on the profile image of reported missing person. The prediction of proposed system in this paper has improved significantly from our previous work [6] where we just employed a single face detector algorithm. Secondly, it is an automated system as it first reduces the search space to increase the efficiency in terms of time and then employ our proposed face detector and recognition algorithm to track missing person in challenging unconstrained large gathering scenario with low resolution data.
The paper is structured as follows. Section 2 presents our brief related work of recent studies including gap analysis. Our proposed methodology is presented in Section 3 including algorithms and technical details. Then, we elaborate the implementation results of training and testing of our proposed methodology on the dataset in Section 4. Finally, the conclusion and future work is presented in Section 5.

2. Related Work

Facial recognition is critical for real-time applications of locating missing persons. Therefore, for our presented scenario it is the matter of an immense importance to identify and recognize human faces in a large crowd having unconstrained environment. Therefore, missing person identification could be attained to find vulnerable group of persons including elderly, children and people with disease (i.e., Dementia, Alzheimer, etc.). We now briefly present our review of recent literature related to the tracking of missing persons in the large crowd scenarios using face detection and recognition on video sequences.
According to Sanchez-Moreno, A.S. et al. [7], some deep neural networks techniques have recently been created to attain state-of-the-art performance on tracking of missing person through face detection and recognition problem. Their work is not for a densely populated environment. They employed the YOLO Face approach for face identification because of its high speed real time detector based on YOLO version 3. Secondly, for classification to recognize faces, a combination of FaceNet and a supervised learning technique, such as the support vector machine (SVM), is proposed. Their experiments are based on unconstrained datasets and show that YOLO-Face performs better when the face under analysis has partial occlusion and position fluctuations. Nonetheless, it can recognise little faces. The Honda/UCSD dataset is used to obtain an accuracy of more than 89.6 percent for face identification. Furthermore, the testing findings showed that the FaceNet+SVM model achieved an accuracy of 99.7 percent when utilizing the LFW dataset. FaceNet+KNN and FaceNet+RF score 99.5 percent and 85.1 percent, respectively, on the identical dataset, while FaceNet achieves 99.6 percent. Finally, when both the face detection and classification phases are active, the suggested system has a recognition accuracy of 99.1 percent and a run-time of 49 ms.
The early work published by Nadeem A. et al. [6] and Nadeem A. et al. [8] using an unique integration of face-recognition algorithms, which employs many recognition algorithms in parallel and combines their predictions using a soft voting mechanism, shown improved accuracy. Based on spatio-temporal constructs, this delivers a more sophisticated forecast. However, the technique was used on low-resolution cropped photos of recognized faces in order to discover missing people in the previously described difficult large crowd gathering. That was explored for scenarios involving enormous crowds at the Al-Nabawi mosque in Madinah. It is a highly unregulated environment with a data collection of low-resolution pictures collected from publicly recorded moving CCTV cameras. The proposed model first detects faces in real time from camera-captured photos, then applies face recognition algorithms with soft voting to get better prediction for identifying the missing persons. A tiny series of consecutive frames reveals the presence of a missing individual.
The method suggested by Li, W., and Siddique, A. A., [9] used the notion of face recognition by utilizing a pre-trained LBPH Face Recognizer to identify the individual in the acquired frame in conjunction with a drone mounted camera to capture the live video stream. An inbuilt Raspberry Pi module analyses the obtained video data, detecting the intended individual with an accuracy of 89.1%.
The authors Ullah, R. et al. [10] proposed a real-time framework for human face detection and recognition in CCTV images over a 40 K images with different environmental condition, background and occlusions. In addition, they performed a comparison analysis between different machine / deep learning algorithms such as decision trees, random forest, K-NN and CNN. The authors claimed that they have achieved 90% overall accuracy with minimum computing time using CNN.
As we noticed in Table 1, the authors in [7] applied state-of-art deep learning techniques for face detection and recognition using conversion of low resolution images to high-quality images, but the technique is not tested in low-resolution images from large gatherings. Moreover, literature in [1,2,3,4,5,11] shows work on recognizing people based on large crowd and low resolution image data, whereas the literature presented in [12] only depicts exploitation of large crowd data and in [13] research carried out only on low resolution data. However, emotional expression of human face have been found in [4] crowded environment showed happy faces are easily by identified. Finally, we found research in [14] that carried out identifying and tracking of pilgrims using CNN and Yolo v4 in unconstrained environment but used high resolution images data to identify smaller sized faces in the crowd.
By analyzing the state-of-the-art, we therefore state that no significant work found with human facial recognition and tracking of missing person based work on low resolution dataset in unconstrained and large crowded environments with above mentioned constraints. However, an ample amount of literature was found in the large crowd domain based on re-identification, tracking and crowd count, etc. Therefore, in this study, we presented our proposed mechanism considering the gap to find missing person by identification and recognition as well in a large crowded gathering of people with diverse age groups having fully unconstrained environment. In this regard, we used dataset built on the pre-processed frames extracted from publicly filmed CCTV videos in Al-Nabawi Mosque, Al Madinah, KSA.

3. Proposed Methodology

This research work is proposed for the automated tracking of reported missing person from live videos of unconstrained large gathering. The proposed mechanism is general but to prove the concept we consider large gathering scenarios of Al-Nabawi mosque (Madinah, KSA) where thousands of pilgrims daily visit. The probability of losing vulnerable companions such as a child or an older person in such large gatherings is high and their automated tracking, using intelligent video surveillance, is an extremely challenging task. The proposed work tries to mitigate this challenging task by dividing the experimented premises into geofences where each geofence is installed with particular cameras.
The proposed tracking mechanism is efficient as it reduces the search space by estimating the probabilistic region of a missing person through a novel geofence set estimation algorithm. This algorithm uses spatio-temporal information of a missing person reported by his/her group head through a mobile application. The query face image of a missing person is fetched from database and, afterwards, it is recognized in videos by cameras that are installed within output set of estimated geofences. This task is accomplished by applying face recognition algorithm on all detected faces which were detected in earlier stage by applying face detection algorithm on video frames. The main tasks of proposed methodology are depicted in Figure 1 and the details about all these tasks are given in following subsections.

3.1. Reporting Missing Person

It is a general practice that pilgrims in Al-Nabawi mosque move in the form of groups. Sometimes each family has a separate group headed by a family member, as shown in Figure 2a, and sometimes it consists of people from different families and headed by a group leader as shown in Figure 2b.
The missing person reporting is conducted by group leader through a mobile application. The reporting includes selecting the missing person from the list of his/her group members. This information is accompanied by the approximate missing time and location (in terms of geofence). This information is passed on to geofence set estimation algorithm given in next section.

3.2. Geofence Set Estimation

The perimeter of Al-Nabawi mosque (including the courtyard) is calculated as 2.166 km. We further divided this premises into a 5 × 5 matrix of square sized geofences. The partitions are shown in Figure 3a and the exact dimensions are shown in Figure 3b.
The whole premise is covered by 20 surveillance cameras that are installed inside the boundary of Al-Nabawi mosque. Each surveillance camera covers a particular set of geofences. The missing person reporting includes spatio-temporal features of missing event. It includes geo-location of missing person that is approximated by mobile based location of group head and it also includes the estimated time laps (in minutes) since the person is missed. Therefore, based upon this information, a set of geofences is obtained by applying geofence set estimation algorithm. This algorithm defines several crowd levels based upon the automated counting score of people. Then, based upon that crowd level score, the maximum possible distance, covered by missing person, is calculated around all four possible directions and finally a set of geofences is calculated where that person can be found. The algorithm’s output reduces the search space and hence the missing person is tracked only in videos of those cameras that are installed within the output set of geofences. The geofence set estimation is given in Algorithm 1.
Algorithm 1 Geo-Fence Set Estimation
1:
Begin
input:t (estimated time laps, in minutes, since person missed), l (mobile based location of group head)
output:G (set of geo-fences / Range for probability of presence of the missing person)
2:
Derive geo-fence G i j of group head (i.e., reporting person) based upon his reporting mobile’s location l
3:
Calculate crowd level (i.e., C L i j ) in all geo-fences G i j ( i , j = 1 5 ) based upon total sum ( S i j ) of automated counting score of persons in placed n camera images in that geo-fence premises. categorized crowd score levels as per following rule:
C L i j = R o u n d ( S i j / 60 )
4:
x = ( t × C L i j ) / n
5:
A = { } , B = { } , C = { } , D = { }
6:
if ( ( i + x ) > 5 ) then
7:
    A = ( ( i + x 5 ) × 110 ) meters outside mosque premises
8:
end if// (i.e., calculating vertically down from current geo-fence)
9:
if  ( ( i x ) < 1 )  then
10:
    B = ( | i x | × 110 ) meters outside mosque premises
11:
end if// (i.e., calculating vertically up from current geo-fence)
12:
if ( ( j + x ) > 5 ) then
13:
    A = ( ( j + x 5 ) × 110 ) meters outside mosque premises
14:
end if// (i.e., calculating horizontally right from current geo-fence)
15:
if  ( ( j x ) < 1 )  then
16:
    B = ( | j x | × 110 ) meters outside mosque premises
17:
end if// (i.e., calculating horizontally left from current geo-fence)
18:
G = G i j a = i i + x b = j + 1 j + x G a b a = i i + x b = j x j 1 G a b a = i i 1 b = j + 1 j + x G a b a = i i 1 b = j x j 1 G a b
19:
if  ( A = { } )   a n d   ( B = { } )   a n d   ( C = { } )   a n d   ( D = { } )  then
20:
   exit
21:
else
22:
   i. G = G { G a b | ( 1 a 5 ) , ( 1 b 5 ) }
   ii. G = G A B C D
23:
end if
24:
end
In Algorithm 1 C L i j is monotonically increasing function and directly proportional to S i j value. This rule is based upon geo-fence length and width (which is nearly equal to 110 m as geo-fence is nearly square shaped) and 0.5 m/s is the observed average walking speed of person when the premises is nearly vacant (which is estimated as under 60–70 persons) and then walking speed reduced as geo-fence premises gets populated. As per the observation, the average speed of walking person reduced to nearly half as it gets double populated (i.e., 90–120 persons in a geo-fence) then further reduced to one-third when it gets populated nearly 150–200 persons and so on.

3.3. Faces Detection in Video Frames

The faces are detected at frame level, whereas the video streams of only those cameras are examined which are installed within the geofences mentioned in output set of estimation algorithm described earlier. A tracking workflow that examines the video streams is presented in Algorithm 2, which is the improved version of our previously proposed tracking workflow in [6]. It samples every 10th frame and detects the face regions on that frame. There exists several face detectors, but no one is capable of detecting all the faces in given image correctly. Therefore, a sampled frame is simultaneously fed to three established face detectors called the Cascaded CART, the Cascaded Haar and the Cascaded LBP face detector, then output from these detectors is merged to improve the face detection process. A new face fusion technique is proposed in Algorithm 3, which not only controls the merger of detected overlapping faces, but also maintain the bounding box for updated face region. The fusion strategy increases the face count at frames level, which may also increase a person face count in temporal domain. Therefore, it will enhance the missing person tracking by reducing the negative errors.
Algorithm 2 Tracking workflow
1:
// create face detector objects
2:
F a c e D e t e c t c a r t v i s i o n . C a s c a d e O b j e c t D e t e c t o r ( Frontal Face CART )
3:
F a c e D e t e c t h a a r v i s i o n . C a s c a d e O b j e c t D e t e c t o r ( Frontal Face Haar )
4:
F a c e D e t e c t l b p v i s i o n . C a s c a d e O b j e c t D e t e c t o r ( Frontal Face LBP )
5:
while  o n  do
6:
    f r t c a m e r a // get video frame t
7:
    B B c a r t s t e p ( F a c e D e t e c t c a r t , f r t ) ; // face detection by CART…Where BB defines set of bounding boxes
8:
    B B h a a r s t e p ( F a c e D e t e c t h a a r , f r t ) ; // face detection by Haar…
9:
    B B l b p s t e p ( F a c e D e t e c t l b p , f r t ) ; // face detection by LBP…
10:
    B B Face Fusion B B c a r t , B B h a a r , B B l b p
11:
   for  b B B  do
12:
      f f r t ( b ) // crop the face region
13:
      f e r i m r e s i z e ( f , i n t e r p o l a t i o n , [ 50 , 50 ] ) ;
14:
     for  j A l o g s  do
15:
         ( I D , S c o r e ) j a l g o j ( f e r ) ; // for jth algorithm
16:
     end for
17:
      ( I D , S c o r e ) V o t i n g ( [ ( I D , S c o r e ) 1 , , ( I D , S c o r e ) 5 ] )
18:
   end for
19:
    T r a c k s T r a c k i n g ( I D s t , I D s t 1 , I D s t 2 , I D s t 3 )
20:
   // go for next frame
21:
end while
Algorithm 3 Proposed Face fusion
1:
for  c B B c a r t do
2:
   for  h B B h a a r  do
3:
      a i o u c h / c h
4:
     if  a i o u > 0.50  then
5:
         B B c a r t c h // over write the box c in  B B c a r t
6:
         B B h a a r B B h a a r h // cut the box h from B B h a a r
7:
     end if
8:
   end for
9:
end for
10:
B B f u s i o n B B c a r t B B h a a r
11:
for  f B B f u s i o n do
12:
   for  l B B l b p  do
13:
      a i o u f l / f l
14:
     if  a i o u > 0.50  then
15:
         B B f u s i o n f l // over write the box f in  B B f u s i o n
16:
         B B l b p B B l b p l // cut the box l from B B l b p
17:
     end if
18:
   end for
19:
end for
20:
B B f u s i o n B B f u s i o n B B l b p
The face regions detected on a sampled frame are shown in Figure 4, where face regions detected by individual face detectors can be observed clearly. Figure 5 shows the comparative analysis of detected faces, where overlapping and non-overlapping face regions can be observed easily. The great extent of overlap recommend fusing these regions to a single face region, where additional efforts may be required to adjust the bounding box over updated face region. The process not only adjust the bounding boxes but also increases the face count on sampled frame.

3.4. Face Recognition

All the face regions detected on a sampled frame are cropped, enhanced and resized to a size of 50 × 50. Then, every face image is simultaneously fed to five recognition algorithms, where each algorithm provides an (ID, Score) pair for that face image. All the five algorithms may or may not predict the same identification result for input face, therefore obtained (ID, Score) pairs are fed to a soft-voting algorithm that produces a mature identification result for input face. The details about the voting scheme can be seen in our previously published paper [6]. An example of recognizing the detected face regions is presented in Figure 6a, where predicted identity of every face region is labeled on box and the associated score is illustrated by bounding box color. The score-color scheme is completely in accordance with [6], where NM-2 stands for no match recommended and NM-1 indicates no match suggested due to the confusion. Face regions with white boxes show no identity, and a tag of NM-1 or NM-2 is mentioned on every white box, which means this face region does not match to any personnel stored in database. The actual identification of faces on sampled frame is presented in Figure 6b. Only 15 faces were detected on sampled frame, where 9 faces find a proper match in database, while 6 faces did not find any valid match. The predicted identity of 9 faces on sampled frame can be confirmed from Figure 6b.

3.5. Missing Person Tracking

The missing persons are tracked in all cameras installed in recommended geofences. For example the tracking of subjects ID-47, 51 and 53 in three different camera views are shown in Figure 7a–c given below.
The presence of every identified personnel in video sequence is recorded temporally. The tracking of above 12 personnel is presented in Figure 8, where actual and predicted presence of 12 personnel is presented with different color. As the subject in video sequence is free to move his face leftward, rightward or downward, the presence may not be recorded at every frame correctly, and the ID track looks rough in temporal domain. To resolve this issue the IDs tracks are smoothed along time domain, which improves the tracking of missing personnel significantly. The roughness of an ID track is minimized by holding the presence record of that ID over multiple consecutive past frames.
The proposed technique of smoothing the presence track over temporal domain is presented in Figure 9.
The smoothed presence tracks of 12 personnel are presented in Figure 10, where false positive and the false negative presence of some of the 12 personnel can be seen easily.

4. Results

The experiments were conducted on a large gathering images dataset. These images were obtained through short videos captured by 20 installed surveillance cameras inside Al-Nabawi mosque as shown in Figure 11.
It consists of 2208 sampled raw video frames, processed face images and the presence tracks of 188 subjects. In the following subsections, the experimentation results for training and testing on our dataset are presented separately for face detection, face recognition and tracking.

4.1. Face Detection

Detecting faces in sampled frames is the first and important step that significantly affects the subsequent processes. There exist several face detectors but Viola Jones is the most frequently used face detector as it can quickly and accurately detect faces in the image. Although it shows good performance, but still some of the face regions are missed by the algorithm. Therefore, first using the cascaded face detectors of CART, HAAR and LBP in parallel and then fusing their output was proposed, which results in more detected faces than individual detection algorithms. The video sequence of 2208 sampled frames was fed to the system and the total number of faces detected over this sequence is presented in Figure 12, where Cascade HAAR detects a total of 7316 faces, Cascade CART a total of 6131 faces, Cascade LBP a total of 3317 faces and their fusion detects a total of 10235 faces on entire video sequence. The frame level detection counts are presented in Figure 13, where faces counts of three cascades fusion are better than individual cascaded algorithms. The qualitative appearance of detected faces is presented in methodology section, where face boxes representing the entire face region looks good. Since every detected face covers the entire face region, and the total number of faces counts increases, it will definitely support and enhance the subsequent recognition and tracking processes.

4.2. Face Recognition

Face recognition at every sampled frame plays a significant role in tracking the personnel in entire video sequence. The individual recognition algorithm may perform poorly, and results an incorrect identification, therefore input faces were fed to five recognition algorithms simultaneously and then resultant identifications were fed to a soft-voting algorithm to mature the input face identification. The recognition process is completely in accordance with [6] algorithm. Since face detection was executed by three face detectors, the matured recognition of detected faces for every face detector is presented in Figure 14, where recognition results over cascade CART and cascade fusion are better than other two detectors. The faces detected on sampled frame were matched to the stored faces in database and identification for every detected face was determined. Few of the faces did not find any match and were tagged “(NM-1 or NM-2)”, while remaining faces found a correct match in database. The tag of NM-1 stands for “No match suggested due to confusion,” and every algorithm assigns this tag to a face if it finds a little match for that face region, on the other hand the tag of NM-2 stands for “No match recommended,” and every algorithm assigns this tag to a face only if it finds a very little match for that face region. All the faces detected by CART on sampled frame were identified correctly. The stored reference faces, which can be found on the current sample frame are presented in Figure 6. The predicted identification identifications for most faces in sampled frame are exactly the same as mentioned in database.

4.3. Missing Personnel Tracking

The main objective of proposed work is to find an efficient tracking methodology for missing personnel, which definitely depend on face detection and recognition results. All the recognition algorithms were fine tuned to perform their optimal, the tracking results of each recognition algorithm against every detector is presented in following figures. The tracking results for PCA are presented in Figure 15, where precision and recall curves present the tracking analysis, the fine-tuned point is highlighted by plotting a circle on drawn curves. To further elaborate the fine tuning process, f1-score and accuracies are plotted against the tuning parameter, and fine-tuned points are highlighted by plotting a circle on f1-score curves. F1-score is an evaluation measure that finds a balance point over precision recall curves for optimal performance, it does not consider the faces which finds no match in recognition process and plays an important role where false positive and false negative errors have different impact. Accuracy curves are also plotted to evaluate the tracking performance. It considers true positive, false positive, false negative and true negative scores, and here considers the faces which find no match in stored database. Since we maintain the presence records of personnel stored in missing personnel database, and mainly focus ourselves over their tracking, the f1-score is given more importance than accuracy measure. Therefore we found the optimal point over f1-score curves and plotted the performance evaluation for those points. According to the evaluation measures presented in Figure 16, Figure 17, Figure 18, Figure 19 and Figure 20, tracking performance for cascades fusion is better than individual face detecting algorithms, which support our claim of using more than one detectors and then fusing their outputs for improving the face detection rate.
The overall f1-score and accuracy rate of cascade fusion is 67.1083% and 72.4924% (before smoothing) and 71.644% and 75.9249% (after smoothing) respectively.

5. Conclusions and Future Work

Large crowd management poses various challenges including tracking or locating a missing person and connecting him/her with his/her head of the family/group. In this paper, we present our work from a funded research project related to the automatic tracking of a missing person in an unconstrained large gathering scenario. We proposed a geofence set estimation to reduce the search space for finding registered missing persons. We first tested three Viola Jones cascades—CART, HAAR and LBP—individually on our unconstrained large gathering dataset for localization of face images. Then, to optimize the results of face detection, we proposed the fusion of these cascades which results in improving both the number of detected faces and their accuracy. This has subsequently helped in better face recognition and identification of the missing person. This work is limited to face recognition for tracking of missing person in videos of large crowd gathering scenarios. In order to cover other dimensions such as detecting missing person when his/her face is hidden, more research is required in other research fields such as “gait recognition”, “person re-identification” and “tracking using wearable devices” that are the part of our planned future work.

Author Contributions

Conceptualization, A.N., M.A. and N.Q.; methodology, M.A., A.N. and N.Q.; validation, A.A., Q.H.A. and A.M.; formal analysis, A.N. and Q.H.A.; investigation, A.N., M.A. and N.Q.; data curation, A.M., K.R. and A.A.; writing—original draft preparation, A.N., M.A., N.Q. and K.R.; writing—review and editing, F.N., N.Q. and Q.H.A.; visualization, M.A., K.R., N.Q. and F.N.; supervision, A.N.; project administration, A.N.; funding acquisition, A.N. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Deanship of Scientific Research, Islamic University of Madinah, Madinah (KSA), under Tammayuz program grant number 1442/505.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The dataset will be available for further research in future.

Conflicts of Interest

Authors declare no conflict of interest.

References

  1. Jurevičius, R.; Goranin, N.; Janulevičius, J.; Nugaras, J.; Suzdalev, I.; Lapusinskij, A. Method for real time face recognition application in unmanned aerial vehicles. Aviation 2019, 23, 65–70. [Google Scholar] [CrossRef] [Green Version]
  2. Kramer, R.S.; Hardy, S.C.; Ritchie, K.L. Searching for faces in crowd chokepoint videos. Appl. Cogn. Psychol. 2020, 34, 343–356. [Google Scholar] [CrossRef]
  3. Best-Rowden, L.; Bisht, S.; Klontz, J.C.; Jain, A.K. Unconstrained face recognition: Establishing baseline human performance via crowdsourcing. In Proceedings of the IEEE International Joint Conference on Biometrics, Clearwater, FL, USA, 29 September–2 October 2014; pp. 1–8. [Google Scholar]
  4. Becker, D.V.; Rheem, H. Searching for a face in the crowd: Pitfalls and unexplored possibilities. Attention Perception Psychophys. 2020, 82, 626–636. [Google Scholar] [CrossRef] [PubMed]
  5. Rajnoha, M.; Mezina, A.; Burget, R. Multi-frame labeled faces database: Towards face super-resolution from realistic video sequences. Appl. Sci. 2020, 10, 7213. [Google Scholar] [CrossRef]
  6. Nadeem, A.; Ashraf, M.; Rizwan, K.; Qadeer, N.; AlZahrani, A.; Mehmood, A.; Abbasi, Q.H. A Novel Integration of Face-Recognition Algorithms with a Soft Voting Scheme for Efficiently Tracking Missing Person in Challenging Large-Gathering Scenarios. Sensors 2022, 22, 1153. [Google Scholar] [CrossRef] [PubMed]
  7. Sanchez-Moreno, A.S.; Olivares-Mercado, J.; Hernandez-Suarez, A.; Toscano-Medina, K.; Sanchez-Perez, G.; Benitez-Garcia, G. Efficient Face Recognition System for Operating in Unconstrained Environments. J. Imaging 2021, 7, 161. [Google Scholar] [CrossRef] [PubMed]
  8. Nadeem, A.; Rizwan, K.; Mehmood, A.; Qadeer, N.; Noor, F.; AlZahrani, A. A Smart City Application Design for Efficiently Tracking Missing Person in Large Gatherings in Madinah Using Emerging IoT Technologies. In Proceedings of the 2021 Mohammad Ali Jinnah University International Conference on Computing (MAJICC), Karachi, Pakistan, 15–17 July 2021; pp. 1–7. [Google Scholar]
  9. Wang, L.; Siddique, A.A. Facial recognition system using LBPH face recognizer for anti-theft and surveillance application based on drone technology. Meas. Control 2020, 53, 1070–1077. [Google Scholar] [CrossRef]
  10. Ullah, R.; Hayat, H.; Siddiqui, A.A.; Siddiqui, U.A.; Khan, J.; Ullah, F.; Hassan, S.; Hasan, L.; Albattah, W.; Islam, M.; et al. A Real-Time Framework for Human Face Detection and Recognition in CCTV Images. Math. Probl. Eng. 2022, 2022, 3276704. [Google Scholar] [CrossRef]
  11. Phillips, P.J.; Yates, A.N.; Hu, Y.; Hahn, C.A.; Noyes, E.; Jackson, K.; Cavazos, J.G.; Jeckeln, G.; Ranjan, R.; Sankaranarayanan, S.; et al. Face recognition accuracy of forensic examiners, superrecognizers, and face recognition algorithms. Proc. Natl. Acad. Sci. USA 2018, 115, 6171–6176. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  12. Preeja, P.; Rahmi, S.N. A Survey on Multiple Face Detection and Tracking in Crowds. Int. J. Innov. Eng. Technol. 2016, 7, 211–216. [Google Scholar]
  13. Zhou, Y.; Liu, D.; Huang, T. Survey of face detection on low-quality images. In Proceedings of the 2018 13th IEEE International Conference on Automatic Face & Gesture Recognition (FG 2018), Xi’an, China, 15–19 May 2018; pp. 769–773. [Google Scholar]
  14. Alharbey, R.; Banjar, A.; Said, Y.; Atri, M.; Alshdadi, A.; Abid, M. Human Faces Detection and Tracking for Crowd Management in Hajj and Umrah. Comput. Mater. Contin. 2022, 71, 6275–6291. [Google Scholar] [CrossRef]
Figure 1. Main phases in the proposed methodology.
Figure 1. Main phases in the proposed methodology.
Sensors 22 05270 g001
Figure 2. Groups with head/leader. (a) Family with head; (b) Group with leader.
Figure 2. Groups with head/leader. (a) Family with head; (b) Group with leader.
Sensors 22 05270 g002
Figure 3. Geofences in Al-Nabawi mosque. (a) Geofences for Haram An-Nabavi; (b) Dimensions of geofences for Haram An-Nabavi.
Figure 3. Geofences in Al-Nabawi mosque. (a) Geofences for Haram An-Nabavi; (b) Dimensions of geofences for Haram An-Nabavi.
Sensors 22 05270 g003
Figure 4. Face detection by three cascaded face detectors.
Figure 4. Face detection by three cascaded face detectors.
Sensors 22 05270 g004aSensors 22 05270 g004b
Figure 5. Cascades fusion by overlapping face regions (detected by three cascades).
Figure 5. Cascades fusion by overlapping face regions (detected by three cascades).
Sensors 22 05270 g005aSensors 22 05270 g005b
Figure 6. Recognized faces on sampled frame. (a) Predicted identification; (b) Actual identification.
Figure 6. Recognized faces on sampled frame. (a) Predicted identification; (b) Actual identification.
Sensors 22 05270 g006aSensors 22 05270 g006b
Figure 7. Tracking of subjects ID-47, 51 and 53 in three different camera views. (a) Camera view 1; (b) Camera view 2; (c) Camera view 3.
Figure 7. Tracking of subjects ID-47, 51 and 53 in three different camera views. (a) Camera view 1; (b) Camera view 2; (c) Camera view 3.
Sensors 22 05270 g007aSensors 22 05270 g007b
Figure 8. IDs tracking before smoothing over time domain.
Figure 8. IDs tracking before smoothing over time domain.
Sensors 22 05270 g008
Figure 9. Smoothing operation to minimize irregularity of matched person presenc.
Figure 9. Smoothing operation to minimize irregularity of matched person presenc.
Sensors 22 05270 g009
Figure 10. IDs tracking after smoothing over time domain.
Figure 10. IDs tracking after smoothing over time domain.
Sensors 22 05270 g010
Figure 11. Multi-Camera video feed.
Figure 11. Multi-Camera video feed.
Sensors 22 05270 g011
Figure 12. Detected faces counts on entire video sequence.
Figure 12. Detected faces counts on entire video sequence.
Sensors 22 05270 g012
Figure 13. Detected faces counts on individual frames.
Figure 13. Detected faces counts on individual frames.
Sensors 22 05270 g013
Figure 14. Facial Recognition using CART, HAAR, LBP and Cascade fusion. (a) Predicted identification for CART faces; (b) Predicted identification for Haar faces; (c) Predicted identification for LBP faces; (d) Predicted identification for Cascades fusion.
Figure 14. Facial Recognition using CART, HAAR, LBP and Cascade fusion. (a) Predicted identification for CART faces; (b) Predicted identification for Haar faces; (c) Predicted identification for LBP faces; (d) Predicted identification for Cascades fusion.
Sensors 22 05270 g014aSensors 22 05270 g014b
Figure 15. Tracking Evaluation for PCA.
Figure 15. Tracking Evaluation for PCA.
Sensors 22 05270 g015
Figure 16. Tracking Evaluation for DCT.
Figure 16. Tracking Evaluation for DCT.
Sensors 22 05270 g016
Figure 17. Tracking Evaluation for LBP.
Figure 17. Tracking Evaluation for LBP.
Sensors 22 05270 g017
Figure 18. Tracking Evaluation for LGBP.
Figure 18. Tracking Evaluation for LGBP.
Sensors 22 05270 g018
Figure 19. Tracking Evaluation for ASR+.
Figure 19. Tracking Evaluation for ASR+.
Sensors 22 05270 g019
Figure 20. Tracking Evaluation for fusion of all recognition cascades through soft-voting scheme.
Figure 20. Tracking Evaluation for fusion of all recognition cascades through soft-voting scheme.
Sensors 22 05270 g020
Table 1. Comparison of various parameters used in the problem domain.
Table 1. Comparison of various parameters used in the problem domain.
ReferenceLow Resolution Image DataHuge Crowd EnvironmentUnconstrained EnvironmentIdentifying and Recognizing for Tracking PersonsUsed Machine Learning Algorithms for Detection (D)/ Recognition (R)Fusion of Face Detection (D) and Recognition (R) Algorithms
[7]YesYesYesNoFaceNet+SVM (D)/ YOLO v3 (R)  No
[6,8]YesYesYesNoViola Jones(D)/ PCA, DCT, LGBP, LBP, ASR+ (R)   Yes (R)
[9]YesYesYesNoLBP(RH )   No
[1,2,3,4,5]YesYesYesNoWell known algorithms for (D) & (R)   No
[14]NoYesYesYesCNN (D) / YOLO v4 (R)   No
[10]NoYesYesYesPCA, CNN (D) / DT, RF, KNN, CNN (R)No
ProposedYesYesYesYesLBP, CAR, HAAR (D) / PCA, DCT, LGBP, LBP, ASR+ (R)Yes (D)&( R)
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Nadeem, A.; Ashraf, M.; Qadeer, N.; Rizwan, K.; Mehmood, A.; AlZahrani, A.; Noor, F.; Abbasi, Q.H. Tracking Missing Person in Large Crowd Gathering Using Intelligent Video Surveillance. Sensors 2022, 22, 5270. https://doi.org/10.3390/s22145270

AMA Style

Nadeem A, Ashraf M, Qadeer N, Rizwan K, Mehmood A, AlZahrani A, Noor F, Abbasi QH. Tracking Missing Person in Large Crowd Gathering Using Intelligent Video Surveillance. Sensors. 2022; 22(14):5270. https://doi.org/10.3390/s22145270

Chicago/Turabian Style

Nadeem, Adnan, Muhammad Ashraf, Nauman Qadeer, Kashif Rizwan, Amir Mehmood, Ali AlZahrani, Fazal Noor, and Qammer H. Abbasi. 2022. "Tracking Missing Person in Large Crowd Gathering Using Intelligent Video Surveillance" Sensors 22, no. 14: 5270. https://doi.org/10.3390/s22145270

APA Style

Nadeem, A., Ashraf, M., Qadeer, N., Rizwan, K., Mehmood, A., AlZahrani, A., Noor, F., & Abbasi, Q. H. (2022). Tracking Missing Person in Large Crowd Gathering Using Intelligent Video Surveillance. Sensors, 22(14), 5270. https://doi.org/10.3390/s22145270

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop