Skip to content

isXinLiu/Awesome-MLLM-Safety

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

43 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Awesome-MLLM-Safety Awesome

A collection (won't be updated) of papers related to safety of Multimodal Large Language Models (MLLMs).

We follow the definition of safety from the paper Safety-Tuned LLaMAs: Lessons From Improving the Safety of Large Language Models that Follow Instructions:

Safety is defined as stopping models from following malicious instructions and generating toxic content.

The scope of our collection.
  • Robustness-related wrong prediction and downstream applications (e.g., robotic/medical/legal/financial domains, anomalies detection, fake news detection) are not involved.
  • We care about the safety of MLLMs, excluding other models like text-to-image models.
  • We mainly focus on images and text, and few about other modalities like audio and videos.

If you find some important work missed, it would be super helpful to let me know ([email protected]). Thanks!

If you find our survey useful for your research, please consider citing:

@article{liu:arxiv2024,
  title={Safety of Multimodal Large Language Models on Images and Text},
  author={Liu, Xin and Zhu, Yichen and Lan, Yunshi and Yang, Chao and Qiao, Yu},
  journal={arXiv preprint arXiv:2402.00357},
  year={2024}
}

Common terminologies related to safety: Taxonomy----safety of MLLMs on images and text:

Table of Contents


Evaluation

Attack

Defense

Other

About

Accepted by IJCAI-24 Survey Track

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages