Create README.md
Browse files
README.md
ADDED
@@ -0,0 +1,29 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: apache-2.0
|
3 |
+
arxiv: 2001.00059
|
4 |
+
pipeline_tag: fill-mask
|
5 |
+
tags:
|
6 |
+
- code
|
7 |
+
- cubert
|
8 |
+
---
|
9 |
+
|
10 |
+
# CuBERT: Learning and Evaluating Contextual Embedding of Source Code
|
11 |
+
|
12 |
+
## Overview
|
13 |
+
This model is the unofficial HuggingFace version of "[CuBERT](https://github.com/google-research/google-research/tree/master/cubert)". In particular, this version comes from [gs://cubert/20210711_Python/pre_trained_model_epochs_2__length_512](https://console.cloud.google.com/storage/browser/cubert/20210711_Python/pre_trained_model_epochs_2__length_512). It was trained 2021-07-11 for 2 epochs with a 512 token context window on the Python BigQuery dataset. I manually converted the Tensorflow checkpoint to PyTorch, the tokenizer to a HuggingFace tokenizer, and have uploaded them here. All credit goes to Aditya Kanade, Petros Maniatis, Gogul Balakrishnan, and Kensen Shi.
|
14 |
+
|
15 |
+
Citation:
|
16 |
+
```bibtex
|
17 |
+
@inproceedings{cubert,
|
18 |
+
author = {Aditya Kanade and
|
19 |
+
Petros Maniatis and
|
20 |
+
Gogul Balakrishnan and
|
21 |
+
Kensen Shi},
|
22 |
+
title = {Learning and evaluating contextual embedding of source code},
|
23 |
+
booktitle = {Proceedings of the 37th International Conference on Machine Learning,
|
24 |
+
{ICML} 2020, 12-18 July 2020},
|
25 |
+
series = {Proceedings of Machine Learning Research},
|
26 |
+
publisher = {{PMLR}},
|
27 |
+
year = {2020},
|
28 |
+
}
|
29 |
+
```
|