Update README.md
Browse files
README.md
CHANGED
@@ -36,4 +36,41 @@ Prometheus-Vision is the first open-source VLM specialized for evaluation purpos
|
|
36 |
|
37 |
Prometheu-Vision is trained with two different sizes (7B and 13B).
|
38 |
You could check the 7B sized LM on [this page](https://huggingface.co/kaist-ai/prometheus-vision-7b-v1.0).
|
39 |
-
Also, check out our dataset as well on [this page](https://huggingface.co/datasets/kaist-ai/Perception-Collection).
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
36 |
|
37 |
Prometheu-Vision is trained with two different sizes (7B and 13B).
|
38 |
You could check the 7B sized LM on [this page](https://huggingface.co/kaist-ai/prometheus-vision-7b-v1.0).
|
39 |
+
Also, check out our dataset as well on [this page](https://huggingface.co/datasets/kaist-ai/Perception-Collection).
|
40 |
+
## Prompt Format
|
41 |
+
Prometheus-Vision requires 5 components in the input: An image, an instruction, a response to evaluate, a score rubric, and a reference answer. You could refer to the prompt format below.
|
42 |
+
You should fill in the instruction, response, reference answer, criteria description, and score description for score in range of 1 to 5.
|
43 |
+
```
|
44 |
+
###Task Description:
|
45 |
+
An instruction (might include an Input inside it), a response to evaluate, a reference answer that gets a score of 5, an image and a score rubric representing an evaluation criterion is given.
|
46 |
+
1. Write a detailed feedback that assess the quality of the response strictly based on the given score rubric, not evaluating in general.
|
47 |
+
2. After writing a feedback, write a score that is an integer between 1 and 5. You should refer to the score rubric.
|
48 |
+
3. The output format should look as follows: \"Feedback: (write a feedback for criteria) [RESULT] (an integer number between 1 and 5)\"
|
49 |
+
4. Please do not generate any other opening, closing, and explanations.
|
50 |
+
|
51 |
+
###The instruction to evaluate:
|
52 |
+
{instruction}
|
53 |
+
|
54 |
+
###Response to evaluate:
|
55 |
+
{response}
|
56 |
+
|
57 |
+
###Reference Answer (Score 5):
|
58 |
+
{reference_answer}
|
59 |
+
|
60 |
+
###Score Rubrics:
|
61 |
+
[{criteria_description}]
|
62 |
+
Score 1: {score1_description}
|
63 |
+
Score 2: {score2_description}
|
64 |
+
Score 3: {score3_description}
|
65 |
+
Score 4: {score4_description}
|
66 |
+
Score 5: {score5_description}
|
67 |
+
|
68 |
+
###Feedback:
|
69 |
+
```
|
70 |
+
Also, we use the following output format. During inference, you could parse the score by splitting the number that is generated next to the [RESULT] phrase.
|
71 |
+
```
|
72 |
+
{orig_feedback}
|
73 |
+
[RESULT] {orig_score}
|
74 |
+
```
|
75 |
+
## License
|
76 |
+
Perception Collection and Prometheus-Vision are subject to OpenAI's Terms of Use for the generated data. If you suspect any violations, please reach out to us.
|