Improving Deep Learning Based Password Guessing Models Using Pre-processing

计算机科学 人工智能 密码 深度学习 字典攻击 预处理器 人工神经网络 自编码 机器学习 计算机网络
作者
Yuxuan Wu,Ding Wang,Yunkai Zou,Ziyi Huang
出处
期刊:Lecture Notes in Computer Science 卷期号:: 163-183
标识
DOI:10.1007/978-3-031-15777-6_10
摘要

Passwords are the most widely used authentication method and play an important role in users' digital lives. Password guessing models are generally used to understand password security, yet statistic-based password models (like the Markov model and probabilistic context-free grammars (PCFG)) are subject to the inherent limitations of overfitting and sparsity. With the improvement of computing power, deep-learning based models with higher crack rates are emerging. Since neural networks are generally used as black boxes for learning password features, a key challenge for deep-learning based password guessing models is to choose the appropriate preprocessing methods to learn more effective features. To fill the gap, this paper explores three new preprocessing methods and makes an attempt to apply them to two promising deep-learning networks, i.e., Long Short-Term Memory (LSTM) neural networks and Generative Adversarial Networks (GAN). First, we propose a character-feature based method for encoding to replace the canonical one-hot encoding. Second, we add so far the most comprehensive recognition rules of words, keyboard patterns, years, and website names into the basic PCFG, and find that the frequency distribution of extracted segments follows the Zipf's law. Third, we adopt Xu et al.'s PCFG improvement with chunk segmentation at CCS'21, and study the performance of the Chunk+PCFG preprocessing method when applied to LSTM and GAN. Extensive experiments on six large real-world password datasets show the effectiveness of our preprocessing methods. Results show that within 50 million guesses: 1) When we apply the PCFG preprocessing method to PassGAN (a GAN-based password model proposed by Hitja et al. at ACNS'19), 13.83%–38.81% (26.79% on average) more passwords can be cracked; 2) Our LSTM based model using PCFG for preprocessing (short for PL) outperforms Wang et al.'s original PL model by 0.35%–3.94% (1.36% on average). Overall, our preprocessing methods can improve the attacking rates in four over seven tested cases. We believe this work provides new feasible directions for guessing optimization, and contributes to a better understanding of deep-learning based models.

科研通智能强力驱动
Strongly Powered by AbleSci AI
更新
大幅提高文件上传限制,最高150M (2024-4-1)

科研通是完全免费的文献互助平台,具备全网最快的应助速度,最高的求助完成率。 对每一个文献求助,科研通都将尽心尽力,给求助人一个满意的交代。
实时播报
1秒前
顾右完成签到,获得积分10
1秒前
跳跃文轩发布了新的文献求助20
2秒前
盖伊福克斯完成签到,获得积分10
2秒前
3秒前
3秒前
3秒前
kokocrl完成签到,获得积分10
4秒前
FODCOC发布了新的文献求助200
4秒前
AbOO发布了新的文献求助10
5秒前
ste56完成签到,获得积分10
5秒前
万能图书馆应助邪恶板凳采纳,获得10
6秒前
李爱国应助喵喵采纳,获得10
6秒前
7秒前
跳跃文轩完成签到,获得积分10
8秒前
情怀应助ccj采纳,获得30
8秒前
喵喵喵发布了新的文献求助10
9秒前
9秒前
wxr发布了新的文献求助10
10秒前
耍酷花卷完成签到,获得积分10
11秒前
Crisp完成签到,获得积分10
11秒前
黄同学发布了新的文献求助10
11秒前
12秒前
Ajax完成签到,获得积分10
13秒前
白白发布了新的文献求助10
14秒前
风光旖旎发布了新的文献求助10
14秒前
15秒前
科研通AI2S应助雪1采纳,获得10
15秒前
666完成签到 ,获得积分10
18秒前
18秒前
喵喵发布了新的文献求助10
19秒前
谦让的紫蓝完成签到,获得积分10
19秒前
20秒前
机智ss完成签到,获得积分10
23秒前
喵喵完成签到,获得积分20
26秒前
星辰大海应助腼腆的修杰采纳,获得10
27秒前
Umwandlung完成签到,获得积分10
28秒前
领导范儿应助健忘的寄瑶采纳,获得10
29秒前
半江完成签到,获得积分10
31秒前
陶醉的夏菡完成签到,获得积分10
32秒前
高分求助中
rhetoric, logic and argumentation: a guide to student writers 1000
Cambridge introduction to intercultural communication 1000
One Man Talking: Selected Essays of Shao Xunmei, 1929–1939 1000
A Chronicle of Small Beer: The Memoirs of Nan Green 1000
Understanding Autism and Autistic Functioning 950
From Rural China to the Ivy League: Reminiscences of Transformations in Modern Chinese History 900
Eric Dunning and the Sociology of Sport 850
热门求助领域 (近24小时)
化学 医学 材料科学 生物 工程类 有机化学 生物化学 物理 内科学 纳米技术 计算机科学 化学工程 复合材料 基因 遗传学 物理化学 催化作用 免疫学 细胞生物学 电极
热门帖子
关注 科研通微信公众号,转发送积分 2915614
求助须知:如何正确求助?哪些是违规求助? 2554443
关于积分的说明 6910937
捐赠科研通 2215813
什么是DOI,文献DOI怎么找? 1177869
版权声明 588353
科研通“疑难数据库(出版商)”最低求助积分说明 576535