MMDecodingTrust

Comprehensive Assessment of Trustworthiness in Multimodal Foundation Models

What is MMDecodingTrust?

MMDecodingTrust aims at providing a thorough assessment of trustworthiness in Multimodal foundation models (MMFMs).

This research endeavor is designed to help researchers and practitioners better understand the capabilities, limitations, and potential risks involved in deploying these state-of-the-art MMFMs.

This project is organized around the following six primary perspectives of trustworthiness, including:

  • Safety
  • Hallucination
  • Fairness
  • Privacy
  • Adversarial robustness
  • Out-of-Distribution Robustness

Trustworthiness Perspectives

⚠️ WARNING: our data contains model outputs that may be considered offensive.


MMDecodingTrust aims to provide a comprehensive trustworthiness evaluation on the recent MMFMs, from different perspectives, including safety, hallucination, fairness, privacy, adversarial robustness, and out-of-distribution robustness under different settings.


Examples of unreliable responses of MMFMs from different trustworthiness perspectives.