Text Generation
Transformers
Safetensors
llama
conversational
text-generation-inference
MinjunZhu commited on
Commit
13db3bc
·
verified ·
1 Parent(s): 3fe714a

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +32 -27
README.md CHANGED
@@ -73,34 +73,8 @@ The main purposes of the WhizReviewer-ML series models are the following two:
73
 
74
  #### Open Source License
75
 
76
- The code in this repository is open-sourced under the Apache-2.0 license. The model weights are open-sourced under the WhizReviewer License, which introduces additional content based on the **Mistral Research License** to ensure the model is not misused.
77
 
78
- #### Intended Uses
79
-
80
- **Expected Use Cases** The WhizReviewer series models are suitable for research purposes in multiple languages. This includes but is not limited to the following objectives:
81
-
82
- 1. Paper Improvement: Assist in enhancing the quality and clarity of academic papers.
83
- 2. Writing Practice: Provide a platform for users to practice and refine their academic writing skills.
84
- 3. Self-assessment Tool: Enable researchers to evaluate their own work before submission.
85
- 4. Learning Aid: Support students and researchers in understanding the peer review process.
86
- 5. Feedback Simulation: Offer simulated peer review feedback to prepare authors for actual reviews.
87
- 6. Revision Guide: Provide structured guidance for revising academic papers.
88
- 7. Concept Validator: Help researchers validate their ideas and hypotheses.
89
- 8. Reward Model: Serve as a component in machine learning systems for academic writing improvement.
90
- 9. Educational Resource: Act as a teaching tool for academic writing and peer review processes.
91
- 10. Research Assistant: Aid in literature reviews and research methodology refinement.
92
- 11. Supplementary Tool: Complement human review in informal, non-official settings.
93
-
94
- **Out of Scope** We do not allow this model to be misused to influence the academic environment. In addition to what is not allowed under the Llama License and Mistral License, the following are also not permitted by us:
95
-
96
- 1. Official Reviews: The WhizReviewer-ML explicitly prohibits use for official peer reviews in any capacity.
97
- 2. Legal or Ethical Decisions: Not designed to make judgments on research ethics or legal compliance.
98
- 3. Factual Verification: While it can offer feedback, it should not be the sole source for fact-checking or verifying scientific claims.
99
- 4. Plagiarism Detection: Not equipped to serve as a plagiarism detection tool.
100
- 5. Publication Decisions: Cannot be used to make final decisions on whether a paper should be published.
101
- 6. Expert Consultation: Not a replacement for expert consultation in specialized fields.
102
-
103
- **If you are unsure whether you meet our License requirements, please send your relevant application to [email protected] for further inquiry**
104
 
105
 
106
 
@@ -121,6 +95,9 @@ We used 784 papers and their review comments from ICLR 2024 as test data, which
121
  | Score Min Acc | 36.96% | **42.70%** | 31.77% |
122
  | Score Max Acc | 24.73% | 23.69% | **49.09%** |
123
 
 
 
 
124
  #### How to use
125
 
126
  The models included in this repository can be used with the `transformers` or `vllm` code libraries.
@@ -356,6 +333,34 @@ We use Fast-Detect-GPT to avoid misuse of WhizReviewer. The table below shows th
356
 
357
  We mixed three hundred review comment samples from ICLR2024 and generated samples from WhizReviewer-ML as the evaluated dataset, with Llama-3.1-8B as the reference model. Detect Acc indicates the accuracy of being correctly detected by Fast-Detect-GPT.
358
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
359
  #### Case Study
360
 
361
  We take a (Poster Paper) from ICLR 2024 titled "Mastering Symbolic Operations: Augmenting Language Models with Compiled Neural Networks" as an example to showcase WhizReviewer-ML's review suggestions. Please refer to [Openreview](https://openreview.net/forum?id=9nsNyN0vox) for the original reviews. The human review scores were *[5,6,6,8]*. The following content is from the output of WhizReviewer-ML-Pro-123B:
 
73
 
74
  #### Open Source License
75
 
76
+ The code in this repository is open-sourced under the Apache-2.0 license. The model weights are open-sourced under the WhizReviewer License, which introduces additional content based on the **Llama 3.1 Community License** to ensure the model is not misused.
77
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
78
 
79
 
80
 
 
95
  | Score Min Acc | 36.96% | **42.70%** | 31.77% |
96
  | Score Max Acc | 24.73% | 23.69% | **49.09%** |
97
 
98
+ We instruct the WhizReviewer-ML model to simulate reviewers from low-scoring to high-scoring, generating review comments and final scores in sequence. After collecting all review comments, a Meta-Reviewer is generated, which can predict the final acceptance result. In the evaluation results, Decisions Acc represents the accuracy of predicting the correct outcome given a paper, while Score Avg Abs represents the absolute difference between the average predicted score and the original score.
99
+
100
+
101
  #### How to use
102
 
103
  The models included in this repository can be used with the `transformers` or `vllm` code libraries.
 
333
 
334
  We mixed three hundred review comment samples from ICLR2024 and generated samples from WhizReviewer-ML as the evaluated dataset, with Llama-3.1-8B as the reference model. Detect Acc indicates the accuracy of being correctly detected by Fast-Detect-GPT.
335
 
336
+ #### Intended Uses
337
+
338
+ **Expected Use Cases** The WhizReviewer series models are suitable for research purposes in multiple languages. This includes but is not limited to the following objectives:
339
+
340
+ 1. Paper Improvement: Assist in enhancing the quality and clarity of academic papers.
341
+ 2. Writing Practice: Provide a platform for users to practice and refine their academic writing skills.
342
+ 3. Self-assessment Tool: Enable researchers to evaluate their own work before submission.
343
+ 4. Learning Aid: Support students and researchers in understanding the peer review process.
344
+ 5. Feedback Simulation: Offer simulated peer review feedback to prepare authors for actual reviews.
345
+ 6. Revision Guide: Provide structured guidance for revising academic papers.
346
+ 7. Concept Validator: Help researchers validate their ideas and hypotheses.
347
+ 8. Reward Model: Serve as a component in machine learning systems for academic writing improvement.
348
+ 9. Educational Resource: Act as a teaching tool for academic writing and peer review processes.
349
+ 10. Research Assistant: Aid in literature reviews and research methodology refinement.
350
+ 11. Supplementary Tool: Complement human review in informal, non-official settings.
351
+
352
+ **Out of Scope** We do not allow this model to be misused to influence the academic environment. In addition to what is not allowed under the Llama License and Mistral License, the following are also not permitted by us:
353
+
354
+ 1. Official Reviews: The WhizReviewer-ML explicitly prohibits use for official peer reviews in any capacity.
355
+ 2. Legal or Ethical Decisions: Not designed to make judgments on research ethics or legal compliance.
356
+ 3. Factual Verification: While it can offer feedback, it should not be the sole source for fact-checking or verifying scientific claims.
357
+ 4. Plagiarism Detection: Not equipped to serve as a plagiarism detection tool.
358
+ 5. Publication Decisions: Cannot be used to make final decisions on whether a paper should be published.
359
+ 6. Expert Consultation: Not a replacement for expert consultation in specialized fields.
360
+
361
+ **If you are unsure whether you meet our License requirements, please send your relevant application to [email protected] for further inquiry**
362
+
363
+
364
  #### Case Study
365
 
366
  We take a (Poster Paper) from ICLR 2024 titled "Mastering Symbolic Operations: Augmenting Language Models with Compiled Neural Networks" as an example to showcase WhizReviewer-ML's review suggestions. Please refer to [Openreview](https://openreview.net/forum?id=9nsNyN0vox) for the original reviews. The human review scores were *[5,6,6,8]*. The following content is from the output of WhizReviewer-ML-Pro-123B: