Commit
Β·
84fadb9
1
Parent(s):
a5e37f8
Add description for the library-based code generation task
Browse files- src/tasks_content.py +16 -7
src/tasks_content.py
CHANGED
@@ -11,13 +11,22 @@ TASKS_PRETTY = {
|
|
11 |
TASKS_PRETTY_REVERSE = {value: key for key, value in TASKS_PRETTY.items()}
|
12 |
|
13 |
TASKS_DESCRIPTIONS = {
|
14 |
-
"library_based_code_generation": "
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
15 |
|
16 |
"ci_builds_repair": "cool description for Bug Localization on Build Logs task",
|
17 |
|
18 |
"project_code_completion": """# Project-Level Code Completion\n
|
19 |
|
20 |
-
Our Project-Level Code Completion π€ [JetBrains-Research/lca-code-completion](https://huggingface.co/datasets/JetBrains-Research/lca-code-completion) includes four datasets:
|
21 |
* `small-context`: 144 data points,
|
22 |
* `medium-context`: 224 data points,
|
23 |
* `large-context`: 270 data points,
|
@@ -34,7 +43,7 @@ TASKS_DESCRIPTIONS = {
|
|
34 |
|
35 |
For further details on the dataset and the baselines from ποΈ Long Code Arena Team, refer to `code_completion` folder in [our baselines repository](https://github.com/JetBrains-Research/lca-baselines) or to our preprint (TODO).
|
36 |
""",
|
37 |
-
|
38 |
"commit_message_generation": """# Commit Message Generation\n
|
39 |
|
40 |
Our Commit Message Generation benchmark π€ [JetBrains-Research/lca-commit-message-generation](https://huggingface.co/datasets/JetBrains-Research/lca-commit-message-generation) includes 163 manually curated commits from Python projects.
|
@@ -49,15 +58,15 @@ TASKS_DESCRIPTIONS = {
|
|
49 |
|
50 |
**Note.** The leaderboard is sorted by ROUGE-1 metric by default.
|
51 |
""",
|
52 |
-
|
53 |
"bug_localization": """# Bug Localization\n
|
54 |
|
55 |
Our Module-to-Text benchmark π€ [JetBrains-Research/lca-bug-localization](https://huggingface.co/datasets/JetBrains-Research/lca-bug-localization) includes 7,479 bug issue descriptions with information about pull request that fix them for Python, Java and Kotlin projects.
|
56 |
|
57 |
Moreover, 150 data points from the test split were manually verified and can be used for bug localization approaches evaluation.
|
58 |
-
We used information retrieval metrics such as R@k, P@k and F1-score for evaluation, taking k
|
59 |
""",
|
60 |
-
|
61 |
"module_summarization": """# Module Summarization\n
|
62 |
Our Module-to-Text benchmark π€ [JetBrains-Research/lca-module-summarization](https://huggingface.co/datasets/JetBrains-Research/lca-module-summarization) includes 216 manually curated text files describing different documentation of opensource permissive Python projects.
|
63 |
|
@@ -77,5 +86,5 @@ def get_submission_text_files_for_task(task_pretty: Optional[str]) -> str:
|
|
77 |
|
78 |
if task_id == "commit_message_generation":
|
79 |
return f"""**{task_pretty} Instructions:**\n\n* Please, attach files in [JSONLines format](https://jsonlines.org/). For an example, check the predictions provided by ποΈ Long Code Arena Team in π€ [JetBrains-Research/lca-results](https://huggingface.co/datasets/JetBrains-Research/lca-results/tree/main/commit_message_generation/predictions). Make sure to include `"prediction"` and `"reference"` fields for each example, the rest are optional."""
|
80 |
-
|
81 |
return f"**{task_pretty} Instructions:**\n\n* π§ There are no instructions for the current task yet."
|
|
|
11 |
TASKS_PRETTY_REVERSE = {value: key for key, value in TASKS_PRETTY.items()}
|
12 |
|
13 |
TASKS_DESCRIPTIONS = {
|
14 |
+
"library_based_code_generation": """# Library-Based Code Generation\n
|
15 |
+
|
16 |
+
Our Library-Based Code Generation benchmark π€ [JetBrains-Research/lca-library-based-code-generation](https://huggingface.co/datasets/JetBrains-Research/lca-library-based-code-generation) includes 150 manually curated instructions asking model to generate Python code using a particular library. Samples come from 62 Python repositories. All the samples in the dataset are based on reference example programs written by authors of the respective libraries.
|
17 |
+
|
18 |
+
For evaluation we use two metrics:
|
19 |
+
* `API Recall`: share of library-specific API calls used in the reference program that appear in the generated code,
|
20 |
+
* `ChrF`: textual similarity between the generated code and the reference program.
|
21 |
+
|
22 |
+
For further details on the dataset and the baselines from ποΈ Long Code Arena Team, refer to `library_based_code_generation` folder in [our baselines repository](https://github.com/JetBrains-Research/lca-baselines) or to our preprint (TODO).
|
23 |
+
""",
|
24 |
|
25 |
"ci_builds_repair": "cool description for Bug Localization on Build Logs task",
|
26 |
|
27 |
"project_code_completion": """# Project-Level Code Completion\n
|
28 |
|
29 |
+
Our Project-Level Code Completion benchmark π€ [JetBrains-Research/lca-code-completion](https://huggingface.co/datasets/JetBrains-Research/lca-code-completion) includes four datasets:
|
30 |
* `small-context`: 144 data points,
|
31 |
* `medium-context`: 224 data points,
|
32 |
* `large-context`: 270 data points,
|
|
|
43 |
|
44 |
For further details on the dataset and the baselines from ποΈ Long Code Arena Team, refer to `code_completion` folder in [our baselines repository](https://github.com/JetBrains-Research/lca-baselines) or to our preprint (TODO).
|
45 |
""",
|
46 |
+
|
47 |
"commit_message_generation": """# Commit Message Generation\n
|
48 |
|
49 |
Our Commit Message Generation benchmark π€ [JetBrains-Research/lca-commit-message-generation](https://huggingface.co/datasets/JetBrains-Research/lca-commit-message-generation) includes 163 manually curated commits from Python projects.
|
|
|
58 |
|
59 |
**Note.** The leaderboard is sorted by ROUGE-1 metric by default.
|
60 |
""",
|
61 |
+
|
62 |
"bug_localization": """# Bug Localization\n
|
63 |
|
64 |
Our Module-to-Text benchmark π€ [JetBrains-Research/lca-bug-localization](https://huggingface.co/datasets/JetBrains-Research/lca-bug-localization) includes 7,479 bug issue descriptions with information about pull request that fix them for Python, Java and Kotlin projects.
|
65 |
|
66 |
Moreover, 150 data points from the test split were manually verified and can be used for bug localization approaches evaluation.
|
67 |
+
We used information retrieval metrics such as R@k, P@k and F1-score for evaluation, taking k equal to 1 and 2.
|
68 |
""",
|
69 |
+
|
70 |
"module_summarization": """# Module Summarization\n
|
71 |
Our Module-to-Text benchmark π€ [JetBrains-Research/lca-module-summarization](https://huggingface.co/datasets/JetBrains-Research/lca-module-summarization) includes 216 manually curated text files describing different documentation of opensource permissive Python projects.
|
72 |
|
|
|
86 |
|
87 |
if task_id == "commit_message_generation":
|
88 |
return f"""**{task_pretty} Instructions:**\n\n* Please, attach files in [JSONLines format](https://jsonlines.org/). For an example, check the predictions provided by ποΈ Long Code Arena Team in π€ [JetBrains-Research/lca-results](https://huggingface.co/datasets/JetBrains-Research/lca-results/tree/main/commit_message_generation/predictions). Make sure to include `"prediction"` and `"reference"` fields for each example, the rest are optional."""
|
89 |
+
|
90 |
return f"**{task_pretty} Instructions:**\n\n* π§ There are no instructions for the current task yet."
|