Skip to content

Commit

Permalink
* update 2024-09-27 06:19:36
Browse files Browse the repository at this point in the history
  • Loading branch information
actions-user committed Sep 26, 2024
1 parent 14ecba8 commit d0e5028
Show file tree
Hide file tree
Showing 2 changed files with 13 additions and 1 deletion.
12 changes: 12 additions & 0 deletions arXiv_db/Malware/2024.md
Original file line number Diff line number Diff line change
Expand Up @@ -2798,3 +2798,15 @@

</details>

<details>

<summary>2024-09-25 05:38:09 - Examining the Rat in the Tunnel: Interpretable Multi-Label Classification of Tor-based Malware</summary>

- *Ishan Karunanayake, Mashael AlSabah, Nadeem Ahmed, Sanjay Jha*

- `2409.16639v1` - [abs](http://arxiv.org/abs/2409.16639v1) - [pdf](http://arxiv.org/pdf/2409.16639v1)

> Despite being the most popular privacy-enhancing network, Tor is increasingly adopted by cybercriminals to obfuscate malicious traffic, hindering the identification of malware-related communications between compromised devices and Command and Control (C&C) servers. This malicious traffic can induce congestion and reduce Tor's performance, while encouraging network administrators to block Tor traffic. Recent research, however, demonstrates the potential for accurately classifying captured Tor traffic as malicious or benign. While existing efforts have addressed malware class identification, their performance remains limited, with micro-average precision and recall values around 70%. Accurately classifying specific malware classes is crucial for effective attack prevention and mitigation. Furthermore, understanding the unique patterns and attack vectors employed by different malware classes helps the development of robust and adaptable defence mechanisms. We utilise a multi-label classification technique based on Message-Passing Neural Networks, demonstrating its superiority over previous approaches such as Binary Relevance, Classifier Chains, and Label Powerset, by achieving micro-average precision (MAP) and recall (MAR) exceeding 90%. Compared to previous work, we significantly improve performance by 19.98%, 10.15%, and 59.21% in MAP, MAR, and Hamming Loss, respectively. Next, we employ Explainable Artificial Intelligence (XAI) techniques to interpret the decision-making process within these models. Finally, we assess the robustness of all techniques by crafting adversarial perturbations capable of manipulating classifier predictions and generating false positives and negatives.

</details>

Loading

0 comments on commit d0e5028

Please sign in to comment.