jenyag commited on
Commit
9cb627e
Β·
verified Β·
1 Parent(s): 2fb6d29

Update src/tasks_content.py

Browse files
Files changed (1) hide show
  1. src/tasks_content.py +26 -24
src/tasks_content.py CHANGED
@@ -11,6 +11,30 @@ TASKS_PRETTY = {
11
  TASKS_PRETTY_REVERSE = {value: key for key, value in TASKS_PRETTY.items()}
12
 
13
  TASKS_DESCRIPTIONS = {
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
14
  "commit_message_generation": """# Commit Message Generation\n
15
 
16
  Our Commit Message Generation benchmark πŸ€— [JetBrains-Research/lca-commit-message-generation](https://huggingface.co/datasets/JetBrains-Research/lca-commit-message-generation) includes 163 manually curated commits from Python projects.
@@ -25,6 +49,7 @@ TASKS_DESCRIPTIONS = {
25
 
26
  **Note.** The leaderboard is sorted by ROUGE-1 metric by default.
27
  """,
 
28
  "bug_localization": """# Bug Localization\n
29
 
30
  Our Module-to-Text benchmark πŸ€— [JetBrains-Research/lca-bug-localization](https://huggingface.co/datasets/JetBrains-Research/lca-bug-localization) includes 7,479 bug issue descriptions with information about pull request that fix them for Python, Java and Kotlin projects.
@@ -32,6 +57,7 @@ TASKS_DESCRIPTIONS = {
32
  Moreover, 150 data points from the test split were manually verified and can be used for bug localization approaches evaluation.
33
  We used information retrieval metrics such as R@k, P@k and F1-score for evaluation, taking k equals to 2.
34
  """,
 
35
  "module_summarization": """# Module Summarization\n
36
  Our Module-to-Text benchmark πŸ€— [JetBrains-Research/lca-module-summarization](https://huggingface.co/datasets/JetBrains-Research/lca-module-summarization) includes 216 manually curated text files describing different documentation of opensource permissive Python projects.
37
 
@@ -40,30 +66,6 @@ TASKS_DESCRIPTIONS = {
40
 
41
  For further details on the dataset and the baselines from 🏟️ Long Code Arena Team, refer to `module2text` folder in [our baselines repository](https://github.com/JetBrains-Research/lca-baselines).
42
  """,
43
-
44
- "library_usage": "cool description for Library Usage Examples Generation task",
45
-
46
- "project_code_completion": """# Project-Level Code Completion\n
47
-
48
- Our Project-Level Code Completion πŸ€— [JetBrains-Research/lca-code-completion](https://huggingface.co/datasets/JetBrains-Research/lca-code-completion) includes four datasets:
49
- * `small-context`: 144 data points,
50
- * `medium-context`: 224 data points,
51
- * `large-context`: 270 data points,
52
- * `huge-context`: 296 data points.
53
-
54
- We use standard Exact Match (EM) metric for one-line code completion.
55
- We evaluate Exact Match for different line categories:
56
- * *infile* – functions and classes are from the completion file;
57
- * *inproject* – functions and files are from the repository snapshot;
58
- * *committed* – functions and classes are from the files that were added on the completion file commit;
59
- * *common* – functions and classes with common names, e.g., `main`, `get`, etc.;
60
- * *non-informative* – short/long lines, import/print lines, or comment lines;
61
- * *random* – lines that doesn't fit to any of previous categories.
62
-
63
- For further details on the dataset and the baselines from 🏟️ Long Code Arena Team, refer to `code_completion` folder in [our baselines repository](https://github.com/JetBrains-Research/lca-baselines) or to our preprint (TODO).
64
- """,
65
-
66
- "bug_localization_build_logs": "cool description for Bug Localization on Build Logs task",
67
  }
68
 
69
 
 
11
  TASKS_PRETTY_REVERSE = {value: key for key, value in TASKS_PRETTY.items()}
12
 
13
  TASKS_DESCRIPTIONS = {
14
+ "library_based_code_generation": "cool description for Library Usage Examples Generation task",
15
+
16
+ "ci_builds_repair": "cool description for Bug Localization on Build Logs task",
17
+
18
+ "project_code_completion": """# Project-Level Code Completion\n
19
+
20
+ Our Project-Level Code Completion πŸ€— [JetBrains-Research/lca-code-completion](https://huggingface.co/datasets/JetBrains-Research/lca-code-completion) includes four datasets:
21
+ * `small-context`: 144 data points,
22
+ * `medium-context`: 224 data points,
23
+ * `large-context`: 270 data points,
24
+ * `huge-context`: 296 data points.
25
+
26
+ We use standard Exact Match (EM) metric for one-line code completion.
27
+ We evaluate Exact Match for different line categories:
28
+ * *infile* – functions and classes are from the completion file;
29
+ * *inproject* – functions and files are from the repository snapshot;
30
+ * *committed* – functions and classes are from the files that were added on the completion file commit;
31
+ * *common* – functions and classes with common names, e.g., `main`, `get`, etc.;
32
+ * *non-informative* – short/long lines, import/print lines, or comment lines;
33
+ * *random* – lines that doesn't fit to any of previous categories.
34
+
35
+ For further details on the dataset and the baselines from 🏟️ Long Code Arena Team, refer to `code_completion` folder in [our baselines repository](https://github.com/JetBrains-Research/lca-baselines) or to our preprint (TODO).
36
+ """,
37
+
38
  "commit_message_generation": """# Commit Message Generation\n
39
 
40
  Our Commit Message Generation benchmark πŸ€— [JetBrains-Research/lca-commit-message-generation](https://huggingface.co/datasets/JetBrains-Research/lca-commit-message-generation) includes 163 manually curated commits from Python projects.
 
49
 
50
  **Note.** The leaderboard is sorted by ROUGE-1 metric by default.
51
  """,
52
+
53
  "bug_localization": """# Bug Localization\n
54
 
55
  Our Module-to-Text benchmark πŸ€— [JetBrains-Research/lca-bug-localization](https://huggingface.co/datasets/JetBrains-Research/lca-bug-localization) includes 7,479 bug issue descriptions with information about pull request that fix them for Python, Java and Kotlin projects.
 
57
  Moreover, 150 data points from the test split were manually verified and can be used for bug localization approaches evaluation.
58
  We used information retrieval metrics such as R@k, P@k and F1-score for evaluation, taking k equals to 2.
59
  """,
60
+
61
  "module_summarization": """# Module Summarization\n
62
  Our Module-to-Text benchmark πŸ€— [JetBrains-Research/lca-module-summarization](https://huggingface.co/datasets/JetBrains-Research/lca-module-summarization) includes 216 manually curated text files describing different documentation of opensource permissive Python projects.
63
 
 
66
 
67
  For further details on the dataset and the baselines from 🏟️ Long Code Arena Team, refer to `module2text` folder in [our baselines repository](https://github.com/JetBrains-Research/lca-baselines).
68
  """,
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
69
  }
70
 
71