|
--- |
|
license: apache-2.0 |
|
language: |
|
- en |
|
metrics: |
|
- accuracy |
|
base_model: |
|
- liuhaotian/llava-v1.6-mistral-7b |
|
pipeline_tag: image-classification |
|
library_name: transformers |
|
tags: |
|
- llm |
|
- mllm |
|
- deepfake |
|
--- |
|
# FFAA Model Card |
|
## Model details |
|
**Model type**: Face Forgery Analysis Assistant (FFAA) consists of a fine-tuned MLLM and Multi-answer Intelligent Decision System (MIDS). |
|
It is a Multi-modal Large Language Model dedicated to the face forgery analysis. |
|
Base MLLM: [liuhaotian/llava-v1.6-mistral-7b](https://huggingface.co/liuhaotian/llava-v1.6-mistral-7b) |
|
|
|
**Paper or resources for more information**: [https://ffaa-vl.github.io/](https://ffaa-vl.github.io/) |
|
|
|
**Where to send questions or comments about the model**: [https://github.com/thu-huangzc/FFAA/issues](https://github.com/thu-huangzc/FFAA/issues) |
|
|
|
## Intended use |
|
**Primary intended uses**: The primary use of FFAA is research on the applications of MLLMs in face forgery analysis, which is essential |
|
for understanding the model’s decision-making process and advancing real-world face forgery analysis. |
|
|
|
**Primary intended users**: The primary intended users of the model are researchers and hobbyists in computer vision, |
|
natural language processing, machine learning, and artificial intelligence. |
|
|
|
## Training dataset |
|
* 20K face forgery analysis VQA (FFA-VQA) dataset, captioned by GPT-4o. |
|
|
|
* 90K historical answer data generated by the MLLM fine-tuned on FFA-VQA. |
|
|
|
## Evaluation dataset |
|
Open-World Face Forgery Analysis Benchmark (OW-FFA-Bench), including 6 face forgery generalization test sets. |
|
The download link is [Google driver](https://drive.google.com/file/d/1867ZKwFCh_OLm-uUsiIiI9RjrUv0JMZX/view?usp=drive_link) |