/RAMScrapersAttentionDetectors

Attention Models Implementation to test anomaly detection for POS RAM Scraper Data exfiltration methods.

Primary LanguagePython

RAMScrapersAttentionDetectors

This repo contains:

  • Transaction network traffic pcap format in "TransactionTracks" folder;
  • Malware network traffic in pcap format "POSmalwareTracks" folder;
  • Models implementation in "models" folder

Models implemented:

  • LSTM Autoencoder

  • LSTM Autoencoder with attention

    • Bahdanau attention [1]
    • Luong attention with several score function [2]
      • Dot Score
      • General Score
      • Concat Score
      • LATTE Score [3]
  • Transformer [4]

References

[1] Bahdanau, Dzmitry, Kyunghyun Cho, and Yoshua Bengio. "Neural machine translation by jointly learning to align and translate." arXiv preprint arXiv:1409.0473 (2014).

[2] Luong, Minh-Thang, Hieu Pham, and Christopher D. Manning. "Effective approaches to attention-based neural machine translation." arXiv preprint arXiv:1508.04025 (2015).

[3] Kukkala, Vipin Kumar, Sooryaa Vignesh Thiruloga, and Sudeep Pasricha. "LATTE: L STM Self-Att ention based Anomaly Detection in E mbedded Automotive Platforms." ACM Transactions on Embedded Computing Systems (TECS) 20.5s (2021): 1-23.

[4] Vaswani, Ashish, et al. "Attention is all you need." Advances in neural information processing systems 30 (2017).