thu-ml/MMTrustEval
A toolbox for benchmarking trustworthiness of multimodal large language models (MultiTrust, NeurIPS 2024 Track Datasets and Benchmarks)
PythonCC-BY-SA-4.0
Issues
- 1
- 1
- 1
Regarding testing my own model
#4 opened by ZixianGao - 1
[BUG] LLava-RLHF run with BFloat16 failed
#3 opened by hxhcreate - 2
- 1
How to Replace Perceptive API?
#1 opened by hacker-jerry