egor-bogomolov commited on
Commit
84fadb9
Β·
1 Parent(s): a5e37f8

Add description for the library-based code generation task

Browse files
Files changed (1) hide show
  1. src/tasks_content.py +16 -7
src/tasks_content.py CHANGED
@@ -11,13 +11,22 @@ TASKS_PRETTY = {
11
  TASKS_PRETTY_REVERSE = {value: key for key, value in TASKS_PRETTY.items()}
12
 
13
  TASKS_DESCRIPTIONS = {
14
- "library_based_code_generation": "cool description for Library Usage Examples Generation task",
 
 
 
 
 
 
 
 
 
15
 
16
  "ci_builds_repair": "cool description for Bug Localization on Build Logs task",
17
 
18
  "project_code_completion": """# Project-Level Code Completion\n
19
 
20
- Our Project-Level Code Completion πŸ€— [JetBrains-Research/lca-code-completion](https://huggingface.co/datasets/JetBrains-Research/lca-code-completion) includes four datasets:
21
  * `small-context`: 144 data points,
22
  * `medium-context`: 224 data points,
23
  * `large-context`: 270 data points,
@@ -34,7 +43,7 @@ TASKS_DESCRIPTIONS = {
34
 
35
  For further details on the dataset and the baselines from 🏟️ Long Code Arena Team, refer to `code_completion` folder in [our baselines repository](https://github.com/JetBrains-Research/lca-baselines) or to our preprint (TODO).
36
  """,
37
-
38
  "commit_message_generation": """# Commit Message Generation\n
39
 
40
  Our Commit Message Generation benchmark πŸ€— [JetBrains-Research/lca-commit-message-generation](https://huggingface.co/datasets/JetBrains-Research/lca-commit-message-generation) includes 163 manually curated commits from Python projects.
@@ -49,15 +58,15 @@ TASKS_DESCRIPTIONS = {
49
 
50
  **Note.** The leaderboard is sorted by ROUGE-1 metric by default.
51
  """,
52
-
53
  "bug_localization": """# Bug Localization\n
54
 
55
  Our Module-to-Text benchmark πŸ€— [JetBrains-Research/lca-bug-localization](https://huggingface.co/datasets/JetBrains-Research/lca-bug-localization) includes 7,479 bug issue descriptions with information about pull request that fix them for Python, Java and Kotlin projects.
56
 
57
  Moreover, 150 data points from the test split were manually verified and can be used for bug localization approaches evaluation.
58
- We used information retrieval metrics such as R@k, P@k and F1-score for evaluation, taking k equals to 2.
59
  """,
60
-
61
  "module_summarization": """# Module Summarization\n
62
  Our Module-to-Text benchmark πŸ€— [JetBrains-Research/lca-module-summarization](https://huggingface.co/datasets/JetBrains-Research/lca-module-summarization) includes 216 manually curated text files describing different documentation of opensource permissive Python projects.
63
 
@@ -77,5 +86,5 @@ def get_submission_text_files_for_task(task_pretty: Optional[str]) -> str:
77
 
78
  if task_id == "commit_message_generation":
79
  return f"""**{task_pretty} Instructions:**\n\n* Please, attach files in [JSONLines format](https://jsonlines.org/). For an example, check the predictions provided by 🏟️ Long Code Arena Team in πŸ€— [JetBrains-Research/lca-results](https://huggingface.co/datasets/JetBrains-Research/lca-results/tree/main/commit_message_generation/predictions). Make sure to include `"prediction"` and `"reference"` fields for each example, the rest are optional."""
80
-
81
  return f"**{task_pretty} Instructions:**\n\n* 🚧 There are no instructions for the current task yet."
 
11
  TASKS_PRETTY_REVERSE = {value: key for key, value in TASKS_PRETTY.items()}
12
 
13
  TASKS_DESCRIPTIONS = {
14
+ "library_based_code_generation": """# Library-Based Code Generation\n
15
+
16
+ Our Library-Based Code Generation benchmark πŸ€— [JetBrains-Research/lca-library-based-code-generation](https://huggingface.co/datasets/JetBrains-Research/lca-library-based-code-generation) includes 150 manually curated instructions asking model to generate Python code using a particular library. Samples come from 62 Python repositories. All the samples in the dataset are based on reference example programs written by authors of the respective libraries.
17
+
18
+ For evaluation we use two metrics:
19
+ * `API Recall`: share of library-specific API calls used in the reference program that appear in the generated code,
20
+ * `ChrF`: textual similarity between the generated code and the reference program.
21
+
22
+ For further details on the dataset and the baselines from 🏟️ Long Code Arena Team, refer to `library_based_code_generation` folder in [our baselines repository](https://github.com/JetBrains-Research/lca-baselines) or to our preprint (TODO).
23
+ """,
24
 
25
  "ci_builds_repair": "cool description for Bug Localization on Build Logs task",
26
 
27
  "project_code_completion": """# Project-Level Code Completion\n
28
 
29
+ Our Project-Level Code Completion benchmark πŸ€— [JetBrains-Research/lca-code-completion](https://huggingface.co/datasets/JetBrains-Research/lca-code-completion) includes four datasets:
30
  * `small-context`: 144 data points,
31
  * `medium-context`: 224 data points,
32
  * `large-context`: 270 data points,
 
43
 
44
  For further details on the dataset and the baselines from 🏟️ Long Code Arena Team, refer to `code_completion` folder in [our baselines repository](https://github.com/JetBrains-Research/lca-baselines) or to our preprint (TODO).
45
  """,
46
+
47
  "commit_message_generation": """# Commit Message Generation\n
48
 
49
  Our Commit Message Generation benchmark πŸ€— [JetBrains-Research/lca-commit-message-generation](https://huggingface.co/datasets/JetBrains-Research/lca-commit-message-generation) includes 163 manually curated commits from Python projects.
 
58
 
59
  **Note.** The leaderboard is sorted by ROUGE-1 metric by default.
60
  """,
61
+
62
  "bug_localization": """# Bug Localization\n
63
 
64
  Our Module-to-Text benchmark πŸ€— [JetBrains-Research/lca-bug-localization](https://huggingface.co/datasets/JetBrains-Research/lca-bug-localization) includes 7,479 bug issue descriptions with information about pull request that fix them for Python, Java and Kotlin projects.
65
 
66
  Moreover, 150 data points from the test split were manually verified and can be used for bug localization approaches evaluation.
67
+ We used information retrieval metrics such as R@k, P@k and F1-score for evaluation, taking k equal to 1 and 2.
68
  """,
69
+
70
  "module_summarization": """# Module Summarization\n
71
  Our Module-to-Text benchmark πŸ€— [JetBrains-Research/lca-module-summarization](https://huggingface.co/datasets/JetBrains-Research/lca-module-summarization) includes 216 manually curated text files describing different documentation of opensource permissive Python projects.
72
 
 
86
 
87
  if task_id == "commit_message_generation":
88
  return f"""**{task_pretty} Instructions:**\n\n* Please, attach files in [JSONLines format](https://jsonlines.org/). For an example, check the predictions provided by 🏟️ Long Code Arena Team in πŸ€— [JetBrains-Research/lca-results](https://huggingface.co/datasets/JetBrains-Research/lca-results/tree/main/commit_message_generation/predictions). Make sure to include `"prediction"` and `"reference"` fields for each example, the rest are optional."""
89
+
90
  return f"**{task_pretty} Instructions:**\n\n* 🚧 There are no instructions for the current task yet."