privacy leakage in long short term memorybrief overview of long short term memory. recurrent neural...

18
Privacy Leakage in Long Short Term Memory Lun Wang [email protected] Peking University Advisor: Chang Liu, Dawn Song Thanks: Nicholas Carlini

Upload: others

Post on 24-Mar-2021

0 views

Category:

Documents


0 download

TRANSCRIPT

Page 1: Privacy Leakage in Long Short Term MemoryBrief Overview of Long Short Term Memory. Recurrent neural networks \⠀刀一一猀尩 are proposed to address the issue. As we can see in

Privacy Leakage in Long Short Term Memory

Lun [email protected]

Peking UniversityAdvisor: Chang Liu, Dawn Song

Thanks: Nicholas Carlini

Presenter
Presentation Notes
Hi everyone! I am Lun Wang from Peking University. And I will talk about the privacy leakage phenomenon in Long Short Term Memory. The work is finished when visiting UC Berkeley under the instruction of Chang Liu and Dawn Song. The idea is proposed by Nicholas Carlini and this is just a imcomplete presentation of the whole project.
Page 2: Privacy Leakage in Long Short Term MemoryBrief Overview of Long Short Term Memory. Recurrent neural networks \⠀刀一一猀尩 are proposed to address the issue. As we can see in

• Brief Overview of Long Short Term Memory

• Privacy Leakage in Long Short Term Memory

• Targeted Attack on Privacy Leakage

• Conclusion

• References

Presenter
Presentation Notes
So what’s long short term memory? In machine learning, there is a bunch of tasks concerning language, like language modeling, speech recognition, and machine translation. These tasks require the neural network to understand each word based on its understanding of previous words. However, traditional neural networks can’t do that.
Page 3: Privacy Leakage in Long Short Term MemoryBrief Overview of Long Short Term Memory. Recurrent neural networks \⠀刀一一猀尩 are proposed to address the issue. As we can see in

Brief Overview of Long Short Term Memory

Presenter
Presentation Notes
Recurrent neural networks (RNNs) are proposed to address the issue. As we can see in the Figure, RNNs introduce loops into neural networks, allowing previous information to persist.
Page 4: Privacy Leakage in Long Short Term MemoryBrief Overview of Long Short Term Memory. Recurrent neural networks \⠀刀一一猀尩 are proposed to address the issue. As we can see in

Brief Overview of Long Short Term Memory

Presenter
Presentation Notes
However, as the gap between dependencies grows, RNNS become unable to learn to connect the information. The problem was explored in depth by Hochreiter (1991) [German] and Bengio, et al. (1994), who found some pretty fundamental reasons why it might be difficult. LSTM, a variant of RNN, is proposed to solve the problem. As shown in the Figure, LSTM introduces more structures to support the long-term dependency. The key idea of LSTM is cell state, represented by the horizontal line running through the top of the diagram. It allows information to flow through LSTM. In order to remove or add information to the cell state, LSTM uses structures called gates to regulate the information flow. A gate is composed of a sigmoid layer and a pointwise multiplication operation. The sigmoid layer outputs numbers between 0 and 1, representing how much of each component will be let through. The first gate in a cell is called "forget gate". It decides how much information of the previous cell state to forget. The second gate is called "input gate". It decides how much information of the new words to add to cell state. The last gate filters the cell state to give the prediction of next word or character. One cell in LSTM can be represented in mathematical format as following.
Page 5: Privacy Leakage in Long Short Term MemoryBrief Overview of Long Short Term Memory. Recurrent neural networks \⠀刀一一猀尩 are proposed to address the issue. As we can see in

Brief Overview of Long Short Term Memory

• Language modeling

• Speech Recognition

• Machine Translation

The Unreasonable Effectiveness of Recurrent Neural Networks. Andrej Karpathy.

Presenter
Presentation Notes
This is an example of the application of LSTM from Andrej Karpathy’s well-known blog: The unreasonable effectiveness of recurrent neural networks. They downloaded the raw Latex source file and trained a multilayer LSTM. Amazingly, the resulting sampled Latex almost compiles
Page 6: Privacy Leakage in Long Short Term MemoryBrief Overview of Long Short Term Memory. Recurrent neural networks \⠀刀一一猀尩 are proposed to address the issue. As we can see in

• Brief Overview of Long Short Term Memory

• Privacy Leakage in Long Short Term Memory

• Targeted Attack on Privacy Leakage

• Conclusion

• References

Presenter
Presentation Notes
However, the concern is that its overwhelming performance will lead to the leakage of private information in the training text, because it can accurately copy the pattern of the language in training set. For example, if the training text contains many social security numbers, will the LSTM give these information out during sampling? According to our results, the answer is ‘yes’.
Page 7: Privacy Leakage in Long Short Term MemoryBrief Overview of Long Short Term Memory. Recurrent neural networks \⠀刀一一猀尩 are proposed to address the issue. As we can see in

I. Training Text

• Penn Treebank dataset

• Insertion Rate

• Cardinality

Presenter
Presentation Notes
First, in order to generate the training text, we insert different kinds of private information into Penn Treebank dataset. One million words of 1989 Wall Street Journal material. The insertion rate means the frequency we insert private information. Cardinality means the number of different private information inserted. We insert three kinds of private information, social security number, credit card number, and address, as shown in the Table.
Page 8: Privacy Leakage in Long Short Term MemoryBrief Overview of Long Short Term Memory. Recurrent neural networks \⠀刀一一猀尩 are proposed to address the issue. As we can see in

II. Configuration

• Ubuntu 16.04.2 LTS

• GeForce GTX TITAN X

Presenter
Presentation Notes
We run the experiments on a Ubuntu 16.04 LTS system with a TITAN X GPU. The hyper-parameters are set as in the Table, and are got via undergraduate student descent.
Page 9: Privacy Leakage in Long Short Term MemoryBrief Overview of Long Short Term Memory. Recurrent neural networks \⠀刀一一猀尩 are proposed to address the issue. As we can see in

III. Experimental Results

Presenter
Presentation Notes
In Figure 3, we evaluate the relation between the number of private-information-like patterns in the sampling texts and cardinality of private information. We can tell that (1) With the increase of training iterations, the model becomes more over-fitted, so the leakage number increases. (2) The number of privacy leakage will converge and oscillate around some certain value for fixed insertion rate. (3) The influence of cardinality is subtle. For rigorous patterns like SSN and CCN, the influence seems small. For flexible patterns, like address, the higher the cardinality is, it will take more iterations to converge. In Figure 4, we evaluate the relation between the entropy of private-information-like patterns and cardinality of private information. We can tell that (1) The entropy of the patterns will first increase with training iterations, and then drop down after a threshold of training iterations. (2) The entropy will finally converge around the expectation. (1) and (2) mean that the leaked private information is inaccurate first, so as the leakage number increases, the entropy increases. Then the leaked information become more and more accurate, so even the leakage number still increases, the entropy will drop until converge around expectation. Results in Figure 3 and 4 show that even training text containsinterfered private information, LSTM can still give out accurate private information.
Page 10: Privacy Leakage in Long Short Term MemoryBrief Overview of Long Short Term Memory. Recurrent neural networks \⠀刀一一猀尩 are proposed to address the issue. As we can see in

III. Experimental Results

Presenter
Presentation Notes
In Figure 5, we evaluate the relation between the leakage number and the density of private information in training texts. We can tell that (1) The number of leakage number will increase with the training iterations first. (2) After some threshold, it oscillates around some certain value. (3) The higher the density of private information, the more the number of privacy leakage will occur. Results in Figure 5 show that even with low density of private information, LSTM can still give out private information.
Page 11: Privacy Leakage in Long Short Term MemoryBrief Overview of Long Short Term Memory. Recurrent neural networks \⠀刀一一猀尩 are proposed to address the issue. As we can see in

• Brief Overview of Long Short Term Memory

• Privacy Leakage in Long Short Term Memory

• Targeted Attack on Privacy Leakage

• Conclusion

• References

Presenter
Presentation Notes
After validating the existence of privacy, we propose a targeted attack on the privacy leakage phenomenon, which can reveal specific private information from a LSTM.
Page 12: Privacy Leakage in Long Short Term MemoryBrief Overview of Long Short Term Memory. Recurrent neural networks \⠀刀一一猀尩 are proposed to address the issue. As we can see in

I. Targeted Attack

Attacking Seed Private Information

CCN 0000 CCN 0000-0000-0000-0000

Mr.Harry Mr.Harry Potter, Ginger Street, Maple City, CA 90230, U.S.A

SSN 123 SSN 123-45-6789……..……..……..

Presenter
Presentation Notes
The attack mechanism is shown in the figure. In order to launch the attack, the adversary needs some side information about the target private information such as the pattern of the private information or part of the private information. For example, if you know the first 3 digits of a social security number, you can feed that into LSTM and get the complete private information.
Page 13: Privacy Leakage in Long Short Term MemoryBrief Overview of Long Short Term Memory. Recurrent neural networks \⠀刀一一猀尩 are proposed to address the issue. As we can see in

II. Experimental Results

Presenter
Presentation Notes
In Figure 6, we evaluate the relation between success rate of attack and training iterations. We can tell that (1) For most cases, the success rate will increase with more iterations. (2) In some cases, the upper relation may not hold because during the process of converging, the direction towards global minimum does not always necessarily conform with the direction increasing success rate (although in many cases they do conform). These results show that we can not always promise the success of attack even for long-trained models. However, the following experiments show that even the attack result is not accurate private information, it is extremely similar to the original private information. In Figure 7, we evaluate the cumulative distribution function of hamming distance between attack result and private information. We can tell that most of the attack results (nearly 60% in SSN, over 95% in CCN, nearly 50% in address except 1000 iterations) have hamming distance of no more than 1 from original private information. These results show that although the attack success rate does not always go up with training iterations, the hamminng distance from original private information is not large. If the character set of private information is limited, this indicates a secondround attack by replacing each character in the attack result with all the possible characters. If the hamming distance is 1, the complexity will only be O(l||C||), where l is the length of attack result, and C is the character set.
Page 14: Privacy Leakage in Long Short Term MemoryBrief Overview of Long Short Term Memory. Recurrent neural networks \⠀刀一一猀尩 are proposed to address the issue. As we can see in

• Brief Overview of Long Short Term Memory

• Privacy Leakage in Long Short Term Memory

• Targeted Attack on Privacy Leakage

• Conclusion

• References

Page 15: Privacy Leakage in Long Short Term MemoryBrief Overview of Long Short Term Memory. Recurrent neural networks \⠀刀一一猀尩 are proposed to address the issue. As we can see in

Conclusion

In this paper, we explored privacy leakage and targeted attack in LSTM

model. The experimental results show that LSTM is likely to give out

private information in training. The attack against privacy leakage is

also simple and feasible. All the results show that the training set

should be properly anonymized to protect privacy of data providers.

Page 16: Privacy Leakage in Long Short Term MemoryBrief Overview of Long Short Term Memory. Recurrent neural networks \⠀刀一一猀尩 are proposed to address the issue. As we can see in

• Brief Overview of Long Short Term Memory

• Privacy Leakage in Long Short Term Memory

• Targeted Attack on Privacy Leakage

• Conclusion

• References

Page 17: Privacy Leakage in Long Short Term MemoryBrief Overview of Long Short Term Memory. Recurrent neural networks \⠀刀一一猀尩 are proposed to address the issue. As we can see in

References

[1] D. Bahdanau, K. Cho, and Y. Bengio. Neural machine translation by jointly learning to align and translate. arXiv preprint arXiv:1409.0473, 2014.[2] Y. Bengio, P. Simard, and P. Frasconi. Learning long-term dependencies with gradient descent is difficult. IEEE transactions on neural networks,5(2):157–166, 1994.[3] A. Graves, A.-r. Mohamed, and G. Hinton. Speech recognition with deep recurrent neural networks. In Acoustics, speech and signal processing (icassp), 2013 ieeeinternational conference on, pages 6645–6649. IEEE, 2013.[4] S. Hochreiter and J. Schmidhuber. Long short-term memory. Neural computation, 9(8):1735–1780, 1997.[5] M. P. Marcus, M. A. Marcinkiewicz, and B. Santorini. Building a large annotated corpus of english: The penntreebank. Computational linguistics, 19(2):313–330, 1993.[6] M. Sundermeyer, R. Schluter, and H. Ney. Lstm neural networks for language modeling. In Thirteenth AnnualConference of the International Speech Communication Association, 2012.

Page 18: Privacy Leakage in Long Short Term MemoryBrief Overview of Long Short Term Memory. Recurrent neural networks \⠀刀一一猀尩 are proposed to address the issue. As we can see in

Q & A