obfuscation using
play

Obfuscation Using Distributional Features Bachelors Thesis Defense - PowerPoint PPT Presentation

Authorship Verification and Obfuscation Using Distributional Features Bachelors Thesis Defense by Janek Bevendorff Date: 27. October 2016 Referees: Prof. Dr. Benno Stein PD Dr. Andreas Jakoby What Is Authorship Verification? Authorship


  1. Authorship Verification and Obfuscation Using Distributional Features Bachelor’s Thesis Defense by Janek Bevendorff Date: 27. October 2016 Referees: Prof. Dr. Benno Stein PD Dr. Andreas Jakoby

  2. What Is Authorship Verification? Authorship Identification Reference Texts Verification Attribution ? ? May solve 𝑒 1 𝑒 2 𝑒 1 𝑒 2 𝑒 3 27. October 2016 2

  3. What Is Authorship Obfuscation? β€œGiven two documents by the same author, modify one of them so that forensic tools cannot classify it as being written by the same author anymore.” βœ“ ✘ 𝑒 1 𝑒 2 27. October 2016 3

  4. Reasons for Obfuscating Authorship οƒ˜ General privacy concerns οƒ˜ Protection from prosecution οƒ˜ Anonymity of single / double blind reviews οƒ˜ Style imitation (writing contests) οƒ˜ Impersonation (malicious intents) οƒ˜ … 27. October 2016 4

  5. Corpus Setup Used corpus: PAN15 Corpus (English) οƒ˜ Training / test: 100 / 500 cases οƒ˜ Two classes with balanced number of cases οƒ˜ Each case consists of two documents either by the same or different author(s) οƒ˜ Test documents have 400-800 words on average Class: β€œsame author” Class: β€œdifferent authors” βœ“ ✘ 50% 50% 27. October 2016 5

  6. Reference Classifier Decision tree classifier with 8 features: οƒ˜ Kullback-Leibler divergence (KLD) οƒ˜ Skew divergence (smoothed KLD) οƒ˜ Jensen-Shannon divergence οƒ˜ Hellinger distance οƒ˜ Cosine similarity with TF weights οƒ˜ Cosine similarity with TF-IDF weights οƒ˜ Ratio between shared n-gram set and total text mass οƒ˜ Average sentence length difference in characters The first 7 features use character 3-grams 27. October 2016 6

  7. Classification Results Classification Accuracy (c@1) 78.00% 76.00% 74.00% 72.00% 76.8% 70.00% 75.7% 68.00% 69.4% 66.00% 64.00% Reference Classifier PAN15 Winner PAN15 Runner-Up 27. October 2016 7

  8. Obfuscation Idea (1) οƒ˜ Attack KLD as main feature οƒ˜ Assumes other features not to be independent 𝑄[𝑗] KLD(𝑄||𝑅) = 𝑄[𝑗] log 2 𝑅[𝑗] 𝑗 KLD Definition Variables: οƒ˜ 𝑗 : n-gram appearing in both texts 𝑒 1 and 𝑒 2 οƒ˜ 𝑄[𝑗] : relative frequency of n-gram 𝑗 in the portion of 𝑒 1 whose n-grams also appear in 𝑒 2 οƒ˜ 𝑅[𝑗] : analogous to 𝑄[𝑗] 27. October 2016 8

  9. KLD Properties οƒ˜ KLD range: [0, ∞) οƒ˜ KLD = 0 for identical texts οƒ˜ PAN15 corpus: 0.27 < KLD < 0.91 οƒ˜ KLD only defined for n-grams where 𝑅 𝑗 > 0 οƒ˜ PAN15 corpus: at least 25% text coverage by only using n-grams that appear in both texts 27. October 2016 9

  10. Obfuscation Idea (2) Idea: obfuscate by increasing the KLD οƒ˜ Assumption: not all n-grams are equally important for the KLD οƒ˜ Only touch those with highest impact οƒ˜ High-impact n-grams can be found by KLD summand derivative: πœ– π‘ž π‘ž πœ–π‘Ÿ π‘ž log 2 = βˆ’ π‘Ÿ π‘Ÿ ln 2 KLD Summand Derivative where π‘ž and π‘Ÿ denote probabilities 𝑄[𝑗] and 𝑅[𝑗] for any defined 𝑗 27. October 2016 10

  11. Obfuscator Implementation Only need to consider the (modifiable) n-gram 𝑗 that maximizes 𝑄[𝑗] 𝑅[𝑗] Three possible obfuscation strategies: N-gram 𝑗 in 𝑒 1 : … … N-gram 𝑗 in 𝑒 2 : … … - + I: Reduction II: Extension III: Hybrid 27. October 2016 11

  12. Obfuscation Results 27. October 2016 12

  13. Obfuscation Results 27. October 2016 13

  14. Obfuscation Results 27. October 2016 14

  15. Obfuscation Results 27. October 2016 15

  16. Obfuscation Results 27. October 2016 16

  17. Obfuscation Results 27. October 2016 17

  18. Obfuscation Results 27. October 2016 18

  19. Obfuscation Results Observation Hybrid: accuracy rises despite KLD increase Possible explanation: adding n- grams improves other features. Cross-validation with single features confirms explanation: Baseline Accuracy 20 Iterations KLD 67.2% 51.4% TF-IDF 74.4% 82.2% Solution: only use reductions 27. October 2016 19

  20. Results Analysis οƒ˜ Significant KLD increase possible with only few iterations οƒ˜ KLD histograms fully overlap after 10-20 iterations (~2% of text modified) οƒ˜ Overall classification accuracy down to ~66% οƒ˜ Extensions are problematic for TF-IDF 27. October 2016 20

  21. Corpus Flaws Results promising, but corpus appears to be flawed οƒ˜ Very short texts οƒ˜ Test corpus much larger than training corpus οƒ˜ Corpus-relative TF-IDF very strong feature (discrimination by topic) οƒ˜ Only chunks of 15 different stage plays by 5 unique authors οƒ˜ No proper text normalization 27. October 2016 21

  22. Development of New Corpus New corpus was developed with books from Project Gutenberg: οƒ˜ 274 cases from three genres and two time periods οƒ˜ Authors unique within genre / period οƒ˜ Avg. text length of 4000 words (few exceptions) οƒ˜ Proper text normalization οƒ˜ 70 / 30 split into training / test (192 / 82 cases) 27. October 2016 22

  23. Classifier Changes Cosine similarity (TF and TF-IDF) features were removed to avoid accidental classification by topic 27. October 2016 23

  24. Classification Results Classification Accuracy (c@1) 85.00% 80.00% 75.00% 70.00% 79.4% 72.0% 71.5% 65.00% 63.4% 60.00% Before Obfuscation After 160 Obfuscation Iterations Reference Classifier PAN15 Winner 27. October 2016 24

  25. Summary οƒ˜ Medium / high classification accuracy with only simple features οƒ˜ Obfuscation possible by attacking main feature οƒ˜ Results reproducible on more diverse corpus οƒ˜ Obfuscation also works against other verification systems 27. October 2016 25

  26. Future Work οƒ˜ Improve classifier by οƒ˜ …adding more features οƒ˜ …integrating β€œUnmasking” by Koppel and Schler [2004] οƒ˜ Attack more features οƒ˜ Use paraphrasing οƒ˜ Randomize obfuscation to harden against reversal 27. October 2016 26

  27. Thank you for your attention

Recommend


More recommend