tigerchen52/LOVE

typo rate problem

Closed this issue · 4 comments

Hi,
Recently, I used IMDB movie review data set to perform Bert_with_love classification task, but I found that the classification accuracy did not decrease with the increase of oov proportion, but would remain unchanged or even increase. For example:
微信图片编辑_20240726081437
The file typo_test_10-typo_test_90 was generated using get_random_attack() of attacks.py.
Could you tell me the reason?
Thank you.

Sorry for the late reply! I was busy preparing a rebuttal this week.

Regarding your question, I guess it is caused by the distinction of the IMDB dataset. I would suggest having a look at the corrupted samples. Also, it is useful to only attract keywords in the input (some stop words won't impact the prediction very much).

Thanks for your answer, my approach is to randomly select each word in the dataset to attack, however, the selected word is likely to have little impact on the classification. Could you guide me on how to choose keywords for each line of text?

Regarding the keywords 1) you can have a look at TF-IDF keyword extraction (link). It is possible to use other tools for keyword extraction 2) you can use gradients to find significant words, which is proposed by Adv-BERT.

Note that it's useful to repeat the experiment on another dataset to validate whether you meet the same issue.

Thank you very much for your guidance.!I know what to do now.