-1- Correlation-aware Deep Generative Model for Unsupervised Anomaly Detection Haoyi Fan 1 , Fengbin Zhang 1 , Ruidong Wang 1 , Liang Xi 1 , Zuoyong Li 2 Harbin University of Science and Technology 1 Minjiang University 2 isfanhy@hrbust.edu.cn
-2- Background Anomaly Anomaly Observed Space Latent Space Normal
-3- Background https://www.explosion.com/135494/5-effective-strategies-of-fraud- https://towardsdatascience.com/building-an-intrusion-detection-system- detection-and-prevention-for-ecommerce/ using-deep-learning-b9488332b321 Fraud Detection Intrusion Detection https://planforgermany.com/switching-private-public-health-insurance- https://blog.exporthub.com/working-with-chinese-manufacturers/ germany/ Disease Detection Fault Detection
-4- Background Unsupervised Anomaly Detection β From the Density Estimation Perspective Data samples: π π’π πππ = π¦ 1 , π¦ 2 , π¦ 3 , β¦ , π¦ π , π¦ π is assumed normal. Latent Space
-5- Background Unsupervised Anomaly Detection β From the Density Estimation Perspective Data samples: π π’π πππ = π¦ 1 , π¦ 2 , π¦ 3 , β¦ , π¦ π , π¦ π is assumed normal. Model: π(π¦) Latent Space
-6- Background Unsupervised Anomaly Detection β From the Density Estimation Perspective Data samples: π π’π πππ = π¦ 1 , π¦ 2 , β¦ , π¦ π , π¦ π is assumed normal. Model: π(π¦) Test samples: π π’ππ‘π’ = π¦ 1 , π¦ 2 , β¦ , π¦ π , π¦ π’ is unknow. if π(π¦ π’ ) < π , π¦ π’ is abnormal . if π(π¦ π’ ) β₯ π , π¦ π’ is normal . Latent Space
-7- Background Unsupervised Anomaly Detection β From the Density Estimation Perspective Data samples: π π’π πππ = π¦ 1 , π¦ 2 , β¦ , π¦ π , π¦ π is assumed normal. Model: π(π¦) Test samples: π π’ππ‘π’ = π¦ 1 , π¦ 2 , β¦ , π¦ π , π¦ π’ is unknow. if π(π¦ π’ ) < π , π¦ π’ is abnormal . if π(π¦ π’ ) β₯ π , π¦ π’ is normal . Latent Space Anomalies reside in the low probability density areas.
-8- Background Correlation among data samples Conventional Anomaly Feature Learning Detection Graph Modeling Feature Space Correlation-aware Anomaly Feature Learning Detection Structure Space How to discover the normal pattern from both the feature level and structural level ?
-9- Problem Statement Anomaly Detection Notations π : Graph. Given a set of input samples π¨ = {π¦ π |π = π¦ : Set of nodes in a graph. 1, . . . , π} , each of which is associated π : Set of edges in a graph. with a πΊ dimension feature π π β β πΊ , we π : Number of nodes. aim to learn a score function π£(π π ): β πΊ β¦ πΊ : Dimension of attribute. π β β πΓπ : Adjacency matrix β , to classify sample π¦ π based on the threshold π : of a network. π β β πΓπΊ : Feature matrix of all nodes. π§ π = {1, ππ π£(π π ) β₯ π, 0, ππ’βππ π₯ππ‘π. where π§ π denotes the label of sample π¦ π , with 0 being the normal class and 1 the anomalous class.
-10- Method CADGMM Feature Dual-Encoder Decoder Graph Estimation Construction network
-11- Method CADGMM K-Nearest Neighbor e.g. K=5 Original feature: π¨ = {π¦ π |π = 1, . . . , π} Find neighbors by K-NN: π π = {π¦ π π |π = 1, . . . , πΏ ΰ΅ Model correlation as graph: π = {π¦, π, π} Graph Construction π¦ = {π€ π = π¦ π |π = 1, . . . , π} π = {π π π = (π€ π , π€ π π )|π€ π π β π π }
-12- Method CADGMM Feature Encoder e.g. MLP , CNN, Feature Decoder LSTM Graph Encoder e.g. GAT
-13- Method CADGMM Gaussian Mixture Model Initial embedding: Z Membership: Z π(π β³ ) = π Z π π β³ β1 W π π β³ β1 + b π π β³ β1 , Z π(0) = Z π = Softmax ( Z π(π β³ ) ) , π β β πΓπ Parameter Estimation: π π π,π ( Z π βπ π )( Z π βπ π ) T π π π,π Z π ΰ· ΰ· π=1 π=1 π π = , π» π = π π π π,π π π,π ΰ· π=1 ΰ· π=1 Estimation network Energy: 2 ( Z βπ π ) T π» π exp (β 1 β1 ( Z βπ π )) π π,π π E Z = β log Ο π=1 π Ο π=1 1 π |2ππ» π | 2
-14- Method Loss and Anomaly Score Loss Function: 2 + π 1 E Z + π 2 Ο π=1 1 β = || X β ΰ·‘ π π 2 X || 2 (π» π ) ππ + π 3 || Z || 2 Ο π=1 Embedding Covariance Rec. Error Energy Penalty Penalty Anomaly Score: ππππ π = E Z Solution for Problem: π§ π = {1, ππ π£(π π ) β₯ π, 0, ππ’βππ π₯ππ‘π. π =Distribution( ππππ π )
-15- Experiment Datasets Baselines Evaluation Metrics Precision OC-SVM Chen et al. 2001 Recall IF Liu et al. 2008 F1-Score DSEBM Zhai et al. 2016 DAGMM Zong et al. 2018 AnoGAN Schlegl et al. 2017 ALAD Zenati et al. 2018
-16- Experiment Results Consistent performance improvement!
-17- Experiment Results Less sensitive to noise data! More robust!
-18- Experiment Results Fig. Impact of different K values of K-NN algorithms in graph construction. Less sensitive to hyper-parameters! Easy to use!
-19- Experiment Results (a). DAGMM (b). CADGMM Fig. Embedding visualization on KDD99 (Blue indicates the normal samples and orange the anomalies). Explainable and Effective!
-20- Conclusion and Future Works Conventional feature learning models cannot β’ effectively capture the correlation among data samples for anomaly detection. We propose a general representation learning β’ framework to model the complex correlation among data samples for unsupervised anomaly detection. We plan to explore the correlation among samples β’ for extremely high-dimensional data sources like image or video. We plan to develop an adaptive and learnable graph β’ construction module for a more reasonable correlation modeling.
-21- Reference [OC-SVM] Chen, Y., Zhou, X.S., Huang, T.S.: One-class svm for learning in image β’ retrieval. ICIP . 2001 [IF] 8. Liu, F.T., Ting, K.M., Zhou, Z.H.: Isolation forest. ICDM . 2008. β’ [DSEBM] Zhai, S., Cheng, Y., Lu, W., Zhang, Z.: Deep structured energy based β’ models for anomaly detection. ICML . 2016. [DAGMM] Zong, B., Song, Q., Min, M.R., Cheng, W., Lumezanu, C., Cho, D., Chen, β’ H.: Deep autoencoding gaussian mixture model for unsupervised anomaly detection. ICLR . 2018. [AnoGAN] Schlegl, T., Seebβ’ ock, P ., Waldstein, S.M., Schmidt-Erfurth, U., Langs, β’ G.: Unsupervised anomaly detection with generative adversarial networks to guide marker discovery. IPMI . 2017. [ALAD] Zenati, H., Romain, M., Foo, C.S., Lecouat, B., Chandrasekhar, V.: β’ Adversarially learned anomaly detection. ICDM . 2018.
-22- Thanks Thanks for listening! Contact: isfanhy@hrbust.edu.cn Home Page: https://haoyfan.github.io/
Recommend
More recommend