Graph-Powered Interpretable Machine Learning Models for Abnormality Detection in Ego-Things Network
Abstract
:1. Introduction
- A novel method is proposed to learn data-driven ML models representing self-awareness and interactive collective awareness of agents’ networks. The learned model is interpretable, i.e., the model is self-explainable about the results it produces and the decisions taken in different situations. For the inferences, a Markov jump particle filter (MJPF) based on generative DBN models is used and extended to become able to detect local and global abnormalities.
- The system incrementally learns new models when the agents encounter new experiences, i.e., when the model detects abnormal situations. The abnormalities generated at different abstraction levels of the models are presented and compared. In this work, interpretability is related to using anomaly data to modify the existing model used to detect the anomaly itself.
2. State of the Art
3. Interpretable Machine Learning Models: Design and Implementation
3.1. Offline Phase: Model Learning
3.1.1. Estimation of Generalized Errors (GEs)
3.1.2. Transition Probability from Discrete Vocabulary
3.2. Online Phase: Model Testing
3.2.1. Anomaly Detection: D-MJPF
- Local anomaly detection: Self-awarenessThe continuous level of the learned DBN model (refer to Figure 2) is dedicated to predicting future states of ego-things separately and detecting anomalies around a specific ego-thing. The innovation measurement of the D-MJPF is used and can be estimated as below:The innovation measurements (i.e., abnormality signal) produced (see Equations (11) and (12)) from the anomaly detection process is used to learn new models incrementally and to represent the situation. Once detected anomaly at continuous level (i.e., local abnormality), the GEs produced from the anomaly signal have been used to correct the model incrementally. Here the GEs are the predicted states and the anomaly measurements (i.e., the probabilistic distance between predicted states and observed evidence). The new model has to capture the current situation that produced anomalies and predict the states better when similar situations occur in the future.The GEs signal of the anomaly has been firstly clustered by applying the GNG algorithm. These clusters generated from the GEs will mainly differ from previously generated clusters on those state-space regions where the anomaly occurred. Then the information extracted by the clustering is used to learn vocabulary and transition probability matrix. The transition probability (shown as green arrows in Figure 2), which influences the model’s continuous level, will be improved to predict the situation well and consequently to produce low or null GEs.
- Global anomaly detection: Collective awarenessThe DBN models’ discrete level (shaded in orange color) shown in Figure 2 detects the global anomaly, i.e., the anomaly happening anywhere in the network. In co-operative task scenarios, this metric can detect the anomaly happening around other agents in the network. The metric used to estimate anomaly is Kullback–Leibler divergence [27]. The Kullback–Leibler (KL) divergence measures how a probability distribution differs from another probability distribution. This work has used this metric to estimate the difference between the predicted discrete level variables distribution and the discrete state’s distribution estimated from the observed sensory variables. All the variables considered here in the discrete level are calculated jointly by considering both ego-things. The equation below calculates the KL distance:Once a global abnormality is detected, i.e., an anomaly happens anywhere in the ego-thing’s network, firstly list out those ego-things that encountered abnormal situations locally. Then the GEs of the abnormality signals detected by the ego-things at the continuous level will be clustered separately. This clustered information will be used to update the corresponding discrete vocabulary and transition probability matrices. The very next step is to update the joint vocabulary (words) and joint transition probability matrix (represented by the green arrow in Figure 2). Therefore, if a similar joint event occurs in the future, the model can represent the situation to make appropriate decisions.
3.2.2. Continual Model Learning and Interpretability
4. Experimental Study
4.1. Training Datasets
4.2. Test Datasets
- Emergency stop 1 (Task 2): while both vehicles jointly navigate a rectangular trajectory, an unexpected pedestrian appears in front of the follower vehicle (iCab2), which performs an emergency brake operation. Since the leader vehicle is not encountering any obstacles, it continues the navigation task. The follower vehicle continues the navigation once the pedestrian crosses the danger zone.
- Emergency stop 2 (Task 3): the task is similar to the previous one (Emergency stop 1), except the pedestrian appears in front of the leader vehicle (iCab1). As soon the leader detects the presence of the pedestrian, it performs an emergency brake operation. Subsequently the follower (iCab2) vehicle also performs emergency brake. Once the pedestrian crosses, both the vehicles (iCab1 and iCab2) continue the joint perimeter monitoring task.
- Emergency stop 3 (Task 4): the task is similar to previous scenarios, but in this case, pedestrians appear in two locations in the trajectory.
5. Results and Discussion
5.1. Abnormality Detection by D-MJPF
5.1.1. Local Abnormality Detection: Self-Awareness
5.1.2. Global Abnormality Detection: Collective-Awareness
5.2. Interpretability and Comparative Analysis
5.2.1. Continuous Level
- (1)
- 2.
- Next, the model was tested with the rotor velocity data sequence collected from Task 3. This time, the model detecting an anomaly at the continuous level belongs to iCab1 only, and it is plotted in Figure 9a. A new model called will be learned incrementally in this phase. Since both vehicles perform emergency stop operations, the anomaly was detected only for the leader vehicle (iCab) because the follower (iCab2) previously experienced a similar situation. Hence, the model was able to represent well the behaviors of iCab2, and the model correction was required only for the part related to iCab1. The GEs produced from the anomaly data for iCab1 and iCab2 are plotted in Figure 21a and Figure 21b, respectively. The additional node generated is for iCab2 only as iCab 1 experiences similar situation while learning the model . The parameters extracted from the clusters are used to generate discrete vocabulary and transition probability matrix to learn the model .
- 3.
- Finally, we tested the model with the data collected from Task 4. This time, a random pedestrian appeared two times in the entire trajectory of the vehicles’ maneuvering operation. However, the model has not detected anomaly at the continuous level (see Figure 10a) because the existing model was able to represent well the current situation and to predict correctly the states of the vehicles. Therefore, matching the dual maps and joint transition probability matrices between Task 3 test data w.r.t. to the previous case did not produce any error. Thus, the model remains unchanged as the existing model could represent the current situation.
5.2.2. Discrete Level
6. Conclusions and Future Work
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Conflicts of Interest
References
- Deo, R.C. Machine learning in medicine. Circulation 2015, 132, 1920–1930. [Google Scholar] [CrossRef] [Green Version]
- Shen, S.; Jiang, H.; Zhang, T. Stock Market Forecasting Using Machine Learning Algorithms; Department of Electrical Engineering, Stanford University: Stanford, CA, USA, 2012; pp. 1–5. [Google Scholar]
- Berk, R. Criminal Justice Forecasts of Risk: A Machine Learning Approach; Springer Science & Business Media: Cham, Switzerland, 2012. [Google Scholar]
- Kanapram, D.T.; Marin-Plaza, P.; Marcenaro, L.; Martin, D.; de la Escalera, A.; Regazzoni, C. Self-awareness in intelligent vehicles: Feature based dynamic Bayesian models for abnormality detection. Robot. Auton. Syst. 2020, 134, 103652. [Google Scholar] [CrossRef]
- Regazzoni, C.; Marcenaro, L.; Campo, D.; Rinner, B. Multisensorial Generative and Descriptive Self-Awareness Models for Autonomous Systems. Proc. IEEE 2020, 108, 987–1010. [Google Scholar] [CrossRef]
- Murdoch, W.J.; Singh, C.; Kumbier, K.; Abbasi-Asl, R.; Yu, B. Definitions, methods, and applications in interpretable machine learning. Proc. Natl. Acad. Sci. USA 2019, 116, 22071–22080. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Du, M.; Liu, N.; Hu, X. Techniques for interpretable machine learning. Commun. ACM 2019, 63, 68–77. [Google Scholar] [CrossRef] [Green Version]
- Miller, T. Explanation in artificial intelligence: Insights from the social sciences. Artif. Intell. 2019, 267, 1–38. [Google Scholar] [CrossRef]
- Molnar, C. Interpretable Machine Learning; Lulu Press: Morrisville, NC, USA, 2020. [Google Scholar]
- Ahmad, M.A.; Eckert, C.; Teredesai, A. Interpretable machine learning in healthcare. In Proceedings of the 2018 ACM International Conference on Bioinformatics, Computational Biology, and Health Informatics, Washington, DC, USA, 29 August–1 September 2018; pp. 559–560. [Google Scholar]
- Ye, H.; Liang, L.; Li, G.Y.; Kim, J.; Lu, L.; Wu, M. Machine learning for vehicular networks: Recent advances and application examples. IEEE Veh. Technol. Mag. 2018, 13, 94–101. [Google Scholar] [CrossRef]
- Kanapram, D.; Campo, D.; Baydoun, M.; Marcenaro, L.; Bodanese, E.L.; Regazzoni, C.; Marchese, M. Dynamic Bayesian Approach for decision-making in Ego-Things. In Proceedings of the 2019 IEEE 5th World Forum on Internet of Things (WF-IoT), Limerick, Ireland, 15–18 April 2019; pp. 909–914. [Google Scholar] [CrossRef]
- Kanapram, D.; Marin-Plaza, P.; Marcenaro, L.; Martin, D.; de la Escalera, A.; Regazzoni, C. Self-awareness in Intelligent Vehicles: Experience Based Abnormality Detection. In Proceedings of the Iberian Robotics Conference, Porto, Portugal, 20–22 November 2019; Springer: Cham, Switzerland, 2019; pp. 216–228. [Google Scholar]
- Ravanbakhsh, M.; Baydoun, M.; Campo, D.; Marin, P.; Martin, D.; Marcenaro, L.; Regazzoni, C. Learning Self-Awareness for Autonomous Vehicles: Exploring Multisensory Incremental Models. IEEE Trans. Intell. Transp. Syst. 2020, 22, 3372–3386. [Google Scholar] [CrossRef]
- Zhang, Y.; Song, B.; Du, X.; Guizani, M. Vehicle tracking using surveillance with multimodal data fusion. IEEE Trans. Intell. Transp. Syst. 2018, 19, 2353–2361. [Google Scholar] [CrossRef] [Green Version]
- Kanapram, D.T.; Marchese, M.; Bodanese, E.L.; Gomez, D.M.; Marcenaro, L.; Regazzoni, C. Dynamic Bayesian collective awareness models for a network of ego-things. IEEE Internet Things J. 2020, 8, 3224–3241. [Google Scholar] [CrossRef]
- Campo, D.; Betancourt, A.; Marcenaro, L.; Regazzoni, C. Static force field representation of environments based on agents’ nonlinear motions. EURASIP J. Adv. Signal Process. 2017, 2017, 13. [Google Scholar] [CrossRef] [Green Version]
- Kohonen, T. Self-organizing maps: Ophmization approaches. In Artificial Neural Networks; Elsevier: Amsterdam, The Netherlands, 1991; pp. 981–990. [Google Scholar]
- MacQueen, J. Some methods for classification and analysis of multivariate observations. In Proceedings of the Fifth Berkeley Symposium on Mathematical Statistics and Probability, Oakland, CA, USA, 1 January 1967; Volume 1, pp. 281–297. [Google Scholar]
- Martinetz, T.M.; Berkovich, S.G.; Schulten, K.J. ‘Neural-gas’ network for vector quantization and its application to time-series prediction. IEEE Trans. Neural Netw. 1993, 4, 558–569. [Google Scholar] [CrossRef] [PubMed]
- Fritzke, B. A Growing Neural Gas Network Learns Topologies. In Proceedings of the 7th International Conference on Neural Information Processing Systems, Denver, CO, USA, 28 November–1 December 1994; MIT Press: Cambridge, MA, USA, 1994; NIPS’94, pp. 625–632. [Google Scholar]
- Baydoun, M.; Campo, D.; Sanguineti, V.; Marcenaro, L.; Cavallaro, A.; Regazzoni, C. Learning switching models for abnormality detection for autonomous driving. In Proceedings of the 2018 21st International Conference on Information Fusion (FUSION), IEEE, Cambridge, UK, 10–13 July 2018; pp. 2606–2613. [Google Scholar]
- Särkkä, S.; Vehtari, A.; Lampinen, J. Rao-Blackwellized particle filter for multiple target tracking. Inform. Fusion 2007, 8, 2–15. [Google Scholar] [CrossRef]
- Blom, H.A.P. An efficient filter for abruptly changing systems. In Proceedings of the 23rd IEEE Conference on Decision and Control, Las Vegas, NV, USA, 12–14 December 1984; pp. 656–658. [Google Scholar]
- Doucet, A.; Gordon, N.J.; Krishnamurthy, V. Particle filters for state estimation of jump Markov linear systems. IEEE Trans. Signal Process. 2001, 49, 613–624. [Google Scholar] [CrossRef]
- Gordon, N.J.; Salmond, D.J.; Smith, A.F.M. Novel approach to nonlinear/non-Gaussian Bayesian state estimation. IEE Proc. F Radar Signal Process. 1993, 140, 107–113. [Google Scholar] [CrossRef] [Green Version]
- Murphy, K.P. Machine Learning: A Probabilistic Perspective; MIT Press: Cambridge, MA, USA, 2012. [Google Scholar]
- Caetano, T.S.; McAuley, J.J.; Cheng, L.; Le, Q.V.; Smola, A.J. Learning graph matching. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31, 1048–1058. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Riesen, K.; Jiang, X.; Bunke, H. Exact and inexact graph matching: Methodology and applications. In Managing and Mining Graph Data; Springer: Cham, Switzerland, 2010; pp. 217–247. [Google Scholar]
- Gomez, D.; Marin-Plaza, P.; Hussein, A.; Escalera, A.; Armingol, J. Ros-based architecture for autonomous intelligent campus automobile (icab). UNED Plasencia Rev. Investig. Univ. 2016, 12, 257–272. [Google Scholar]
- Friston, K. A free energy principle for biological systems. Entropy 2012, 14, 2100–2121. [Google Scholar]
Activated Node Pairs | Label | Label Values (4D) | |
---|---|---|---|
iCab1 | iCab2 | ||
a | e | [5.353 −0.007 5.327 −0.003] | |
c | e | [4.642 0.001 5.327 −0.003] | |
c | f | [4.642 0.001 4.596 −0.009] | |
b | f | [3.925 −0.005 4.596 −0.009] | |
b | d | [3.925 −0.005 3.801 −0.006] | |
c | d | [4.642 0.001 3.801 −0.006] |
Sl No. | Matching of Vocabulary (Node Pairs) | Label | |
---|---|---|---|
Train: Task 1 | Test: Task 2 | ||
1 | - | a,f | |
2 | - | a,g | |
3 | b,d | c,g | |
4 | c,d | b,g | |
5 | c,f | b,f | |
6 | b,f | c,f | |
7 | - | a,d | |
8 | c,e | c,e | |
9 | - | b,e | |
10 | a,e | a,e |
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2022 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Thekke Kanapram, D.; Marcenaro, L.; Martin Gomez, D.; Regazzoni, C. Graph-Powered Interpretable Machine Learning Models for Abnormality Detection in Ego-Things Network. Sensors 2022, 22, 2260. https://doi.org/10.3390/s22062260
Thekke Kanapram D, Marcenaro L, Martin Gomez D, Regazzoni C. Graph-Powered Interpretable Machine Learning Models for Abnormality Detection in Ego-Things Network. Sensors. 2022; 22(6):2260. https://doi.org/10.3390/s22062260
Chicago/Turabian StyleThekke Kanapram, Divya, Lucio Marcenaro, David Martin Gomez, and Carlo Regazzoni. 2022. "Graph-Powered Interpretable Machine Learning Models for Abnormality Detection in Ego-Things Network" Sensors 22, no. 6: 2260. https://doi.org/10.3390/s22062260