Biblio

Filters: Author is Ni, Huanhuan  [Clear All Filters]
2023-08-03
Duan, Xiaowei, Han, Yiliang, Wang, Chao, Ni, Huanhuan.  2022.  Optimization of Encrypted Communication Model Based on Generative Adversarial Network. 2022 International Conference on Blockchain Technology and Information Security (ICBCTIS). :20–24.
With the progress of cryptography computer science, designing cryptographic algorithms using deep learning is a very innovative research direction. Google Brain designed a communication model using generation adversarial network and explored the encrypted communication algorithm based on machine learning. However, the encrypted communication model it designed lacks quantitative evaluation. When some plaintexts and keys are leaked at the same time, the security of communication cannot be guaranteed. This model is optimized to enhance the security by adjusting the optimizer, modifying the activation function, and increasing batch normalization to improve communication speed of optimization. Experiments were performed on 16 bits and 64 bits plaintexts communication. With plaintext and key leak rate of 0.75, the decryption error rate of the decryptor is 0.01 and the attacker can't guess any valid information about the communication.
2022-07-12
Duan, Xiaowei, Han, Yiliang, Wang, Chao, Ni, Huanhuan.  2021.  Optimization of Encrypted Communication Length Based on Generative Adversarial Network. 2021 IEEE 4th International Conference on Big Data and Artificial Intelligence (BDAI). :165—170.
With the development of artificial intelligence and cryptography, intelligent cryptography will be the trend of encrypted communications in the future. Abadi designed an encrypted communication model based on a generative adversarial network, which can communicate securely when the adversary knows the ciphertext. The communication party and the adversary fight against each other to continuously improve their own capabilities to achieve a state of secure communication. However, this model can only have a better communication effect under the 16 bits communication length, and cannot adapt to the length of modern encrypted communication. Combine the neural network structure in DCGAN to optimize the neural network of the original model, and at the same time increase the batch normalization process, and optimize the loss function in the original model. Experiments show that under the condition of the maximum 2048-bit communication length, the decryption success rate of communication reaches about 0.97, while ensuring that the adversary’s guess error rate is about 0.95, and the training speed is greatly increased to keep it below 5000 steps, ensuring safety and efficiency Communication.