Dual-Stream Time-Series Transformer-Based Encrypted Traffic Data Augmentation Framework

Research output: Contribution to journalArticlepeer-review

Abstract

We propose a Transformer-based data augmentation framework with a time-series dual-stream architecture to address performance degradation in encrypted network traffic classification caused by class imbalance between attack and benign traffic. The proposed framework independently processes the complete flow’s sequential packet information and statistical characteristics by extracting and normalizing a local channel (comprising packet size, inter-arrival time, and direction) and a set of six global flow-level statistical features. These are used to generate a fixed-length multivariate sequence and an auxiliary vector. The sequence and vector are then fed into an encoder-only Transformer that integrates learnable positional embeddings with a FiLM + context token-based injection mechanism, enabling complementary representation of sequential patterns and global statistical distributions. Large-scale experiments demonstrate that the proposed method reduces reconstruction RMSE and additional feature restoration MSE by over 50%, while improving accuracy, F1-Score, and AUC by 5–7%p compared to classification on the original imbalanced datasets. Furthermore, the augmentation process achieves practical levels of processing time and memory overhead. These results show that the proposed approach effectively mitigates class imbalance in encrypted traffic classification and offers a promising pathway to achieving more robust model generalization in real-world deployment scenarios.

Original languageEnglish
Article number9879
JournalApplied Sciences (Switzerland)
Volume15
Issue number18
DOIs
StatePublished - Sep 2025

Keywords

  • attack traffic
  • augmentation
  • dual-stream
  • encrypted traffic
  • transformer

Fingerprint

Dive into the research topics of 'Dual-Stream Time-Series Transformer-Based Encrypted Traffic Data Augmentation Framework'. Together they form a unique fingerprint.

Cite this