@article{jin2020bertrobust,
title={Is BERT Really Robust? A Strong Baseline for Natural Language Attack on Text Classification and Entailment},
author={Jin, Di and Jin, Zhijing and Zhou, Joey Tianyi and Szolovits, Peter},
journal={arXiv preprint arXiv:1907.11932},
year={2020}
}
@article{cer2018use,
title={Universal Sentence Encoder},
author={Cer, Daniel and Yang, Yinfei and Kong, Sheng-yi and Hua, Nan and Limtiaco, Nicole and St. John, Rhomni and Constant, Noah and Guajardo-Cespedes, Mario and Yuan, Steve and Tar, Chris and others},
journal={arXiv preprint arXiv:1803.11175},
year={2018}
}
@article{mrksic2016counterfitting,
title={Counter-fitting Word Vectors to Linguistic Constraints},
author={Mrkšić, Nikola and Séaghdha, Diarmuid Ó and Thomson, Blaise and Gašić, Milica and Rojas-Barahona, Lina M. and Su, Pei-Hao and Vandyke, David and Wen, Tsung-Hsien and Young, Steve},
journal={arXiv preprint arXiv:1603.00892},
year={2016}
}
@misc{textattack2020framework,
author = {Morris, John and Lifland, Eli and Yoo, Jin and Grigsby, Jake and Jin, Di and Qi, Yanjun},
title = {TextAttack: A Framework for Adversarial Attacks, Data Augmentation, and Adversarial Training in NLP},
year = {2020},
howpublished = {\url{https://arxiv.org/pdf/2005.05909.pdf}}
}
@article{jia2017adversarial,
title={Adversarial Examples for Evaluating Reading Comprehension Systems},
author={Jia, Robin and Liang, Percy},
journal={arXiv preprint arXiv:1707.07328},
year={2017}
}
@article{omar2022robust,
title={Robust Natural Language Processing: Recent Advances, Challenges, and Future Directions},
author={Omar, Marwan and Choi, Soohyeon and Nyang, DaeHun and Mohaisen, David},
journal={IEEE Access},
volume={10},
pages={86038--86056},
year={2022}
}
@article{huang2024semantic,
title={Defense against adversarial attacks via textual embeddings based on semantic associative field},
author={Huang, J. and Chen, L.},
journal={Neural Computing and Applications},
volume={36},
pages={289--301},
year={2024}
}
@article{chen2025worstcase,
title={Towards the Worst-case Robustness of Large Language Models},
author={Chen, H. and Dong, Y. and Wei, Z. and Su, H. and Zhu, J.},
journal={arXiv preprint arXiv:2501.19040},
year={2025}
}
@misc{wsj2025securityrisks,
author = {{The Wall Street Journal}},
title = {Large Language Models Pose Growing Security Risks},
year = {2025},
month = {February},
howpublished = {\url{https://www.wsj.com/articles/large-language-models-pose-growing-security-risks-f3c84ea9}},
note = {[Online]}
}
@misc{mistral2023,
author = {Mistral AI Team},
title = {Announcing Mistral 7B},
year = {2023},
howpublished = {\url{https://mistral.ai/news/announcing-mistral-7b}},
note = {[Online]}
}
@article{rajpurkar2016,
author = {P. Rajpurkar and J. Zhang and K. Lopyrev and P. Liang},
title = {SQuAD: 100,000+ Questions for Machine Comprehension of Text},
journal = {arXiv preprint arXiv:1606.05250},
year = {2016},
url = {https://arxiv.org/pdf/1606.05250.pdf}
}
@article{yang2024,
author = {Z. Yang and Z. Meng and X. Zheng and R. Wattenhofer},
title = {Assessing Adversarial Robustness of Large Language Models: An Empirical Study},
journal = {arXiv preprint arXiv:2405.02764},
year = {2024},
url = {https://arxiv.org/pdf/2405.02764.pdf}
}
@article{vitorino2024,
author = {J. Vitorino and E. Maia and I. Praça},
title = {Adversarial Evasion Attack Efficiency against Large Language Models},
journal = {arXiv preprint arXiv:2406.08050v1},
year = {2024},
url = {https://arxiv.org/pdf/2406.08050v1.pdf}
}