Update README.md
Browse files
README.md
CHANGED
@@ -21,12 +21,12 @@ tags:
|
|
21 |
<p align="center">
|
22 |
π€ <a href="https://modelscope.cn/organization/inclusionAI">ModelScope</a>
|
23 |
π€ <a href="https://huggingface.co/inclusionAI">Hugging Face</a>
|
24 |
-
π₯οΈ <a href="https://github.com/
|
25 |
<p>
|
26 |
|
27 |
## Introduction
|
28 |
|
29 |
-
Ling-Coder-Lite is a MoE LLM provided and open-sourced by InclusionAI, which has 16.
|
30 |
|
31 |
## Model Downloads
|
32 |
|
@@ -40,6 +40,18 @@ You can download the following table to see the various parameters for your use
|
|
40 |
| Ling-Coder-lite | 16.8B | 2.75B | 16K | [π€ HuggingFace](https://huggingface.co/inclusionAI/Ling-Coder-lite) |
|
41 |
</div>
|
42 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
43 |
## Evaluation
|
44 |
|
45 |
Detailed evaluation results are reported in our technical report [Ling-Coder-TR](https://huggingface.co/papers/2503.17793).
|
@@ -89,7 +101,7 @@ print(response)
|
|
89 |
```
|
90 |
|
91 |
## Deployment
|
92 |
-
Please refer to [Github](https://github.com/
|
93 |
|
94 |
## License
|
95 |
This code repository is licensed under [the MIT License](https://huggingface.co/inclusionAI/Ling-Coder-lite/blob/main/LICENCE).
|
@@ -99,7 +111,7 @@ This code repository is licensed under [the MIT License](https://huggingface.co/
|
|
99 |
```
|
100 |
@misc{codefuse2025samplemattersleveragingmixtureofexperts,
|
101 |
title={Every Sample Matters: Leveraging Mixture-of-Experts and High-Quality Data for Efficient and Accurate Code LLM},
|
102 |
-
author={Codefuse and Ling Team
|
103 |
year={2025},
|
104 |
eprint={2503.17793},
|
105 |
archivePrefix={arXiv},
|
|
|
21 |
<p align="center">
|
22 |
π€ <a href="https://modelscope.cn/organization/inclusionAI">ModelScope</a>
|
23 |
π€ <a href="https://huggingface.co/inclusionAI">Hugging Face</a>
|
24 |
+
π₯οΈ <a href="https://github.com/codefuse-ai/Ling-Coder-Lite">GitHub</a>
|
25 |
<p>
|
26 |
|
27 |
## Introduction
|
28 |
|
29 |
+
Ling-Coder-Lite is a MoE LLM provided and open-sourced by InclusionAI, which has 16.8B parameters with 2.75B activated parameters. This model demonstrates state-of-the-art performance on 12 coding benchmarks, while simultaneously offering competitive latency and throughput compared to code LLMs of similar size. In addition to open-sourcing the model itself, we also release a substantial amount of code-related data, including synthetic QA, SFT and DPO datasets. More details are described in the technique report [Ling-Coder-TR](https://huggingface.co/papers/2503.17793).
|
30 |
|
31 |
## Model Downloads
|
32 |
|
|
|
40 |
| Ling-Coder-lite | 16.8B | 2.75B | 16K | [π€ HuggingFace](https://huggingface.co/inclusionAI/Ling-Coder-lite) |
|
41 |
</div>
|
42 |
|
43 |
+
## Dataset Downloads
|
44 |
+
|
45 |
+
<div align="center">
|
46 |
+
|
47 |
+
| **Model** | **Samples** | **Download** |
|
48 |
+
| :------------: | :----------------: | :--------------------------------------------------------------------------------------------------------------------------------------------------: |
|
49 |
+
| Ling-Coder-SyntheticQA | 24M | [π€ HuggingFace](https://huggingface.co/datasets/inclusionAI/Ling-Coder-SyntheticQA) |
|
50 |
+
| Ling-Coder-SFT | 5M | [π€ HuggingFace](https://huggingface.co/datasets/inclusionAI/Ling-Coder-SFT) |
|
51 |
+
| Ling-Coder-DPO | 250K | [π€ HuggingFace](https://huggingface.co/datasets/inclusionAI/Ling-Coder-DPO) |
|
52 |
+
|
53 |
+
</div>
|
54 |
+
|
55 |
## Evaluation
|
56 |
|
57 |
Detailed evaluation results are reported in our technical report [Ling-Coder-TR](https://huggingface.co/papers/2503.17793).
|
|
|
101 |
```
|
102 |
|
103 |
## Deployment
|
104 |
+
Please refer to [Github](https://github.com/codefuse-ai/Ling-Coder-Lite/blob/master/README.md)
|
105 |
|
106 |
## License
|
107 |
This code repository is licensed under [the MIT License](https://huggingface.co/inclusionAI/Ling-Coder-lite/blob/main/LICENCE).
|
|
|
111 |
```
|
112 |
@misc{codefuse2025samplemattersleveragingmixtureofexperts,
|
113 |
title={Every Sample Matters: Leveraging Mixture-of-Experts and High-Quality Data for Efficient and Accurate Code LLM},
|
114 |
+
author={Codefuse and Ling Team},
|
115 |
year={2025},
|
116 |
eprint={2503.17793},
|
117 |
archivePrefix={arXiv},
|