Abstract
Anomaly detection (AD) involves detecting abnormality from normality and has a wide spectrum of applications in reality. Kernel-based methods for AD have been proven robust with diverse data distributions and offering good generalization ability. Stochastic gradient descent (SGD) method has recently emerged as a promising framework to devise ultra-fast learning methods. In this paper, we conjoin the advantages of Kernel-based method and SGD-based method to propose fast learning methods for anomaly detection. We validate the proposed methods on 8 benchmark datasets in UCI repository and KDD cup 1999 dataset. The experimental results show that the proposed methods offer a comparable one-class classification accuracy while simultaneously achieving a significantly computational speed-up.
Original language | English |
---|---|
Title of host publication | 2016 International Joint Conference on Neural Networks (IJCNN) |
Subtitle of host publication | San Francisco, CA, USA — February 27 - March 02, 2016 |
Editors | Pablo A. Estevez |
Place of Publication | New York NY USA |
Publisher | IEEE, Institute of Electrical and Electronics Engineers |
Pages | 3211-3217 |
Number of pages | 7 |
ISBN (Electronic) | 9781509006199, 9781509006205 |
ISBN (Print) | 9781509006212 |
DOIs | |
Publication status | Published - 2016 |
Externally published | Yes |
Event | IEEE International Joint Conference on Neural Networks 2016 - Vancouver, Canada Duration: 24 Jul 2016 → 29 Jul 2016 https://ewh.ieee.org/conf/wcci/2016/ https://ieeexplore.ieee.org/xpl/conhome/7593175/proceeding (Proceedings) |
Conference
Conference | IEEE International Joint Conference on Neural Networks 2016 |
---|---|
Abbreviated title | IJCNN 2016 |
Country/Territory | Canada |
City | Vancouver |
Period | 24/07/16 → 29/07/16 |
Internet address |
Keywords
- Anomaly detection
- Kernel method
- Stochastic algorithm