Privacy-preserving technologies for LLMs
Differential privacy is one of the privacy-preserving technologies that can be used for LLMs as well.
Text attacks on ML models and LLMs
TextAttack stands as a Python framework designed for conducting adversarial attacks, adversarial training, and data augmentation within the field of NLP. This versatile tool streamlines the process of exploring NLP model robustness, offering a seamless, rapid, and user-friendly experience. Furthermore, it proves invaluable for NLP model training, adversarial training, and data augmentation purposes. TextAttack offers various components tailored for typical NLP tasks, including sentence encoding, grammar checking, and word replacement, which can also be utilized independently.
Instructions on how to install the TextAttack package can be found at this GitHub URL: https://github.com/QData/TextAttack.
Install TextAttack framework using pip install
in the following way:
!pip install textattack...