Update README.md
Browse files
README.md
CHANGED
@@ -25,13 +25,13 @@ pipeline_tag: text-generation
|
|
25 |
- Original model: [NT-Java-1.1B](https://huggingface.co/infosys/NT-Java-1.1B)
|
26 |
|
27 |
<!-- description start -->
|
28 |
-
|
29 |
|
30 |
This repo contains GGUF format model files for [Infosys's NT-Java-1.1B](https://huggingface.co/infosys/NT-Java-1.1B).
|
31 |
|
32 |
<!-- description end -->
|
33 |
<!-- README_GGUF.md-about-gguf start -->
|
34 |
-
|
35 |
|
36 |
GGUF, introduced by the llama.cpp team on August 21st, 2023, is a new format designed to replace the outdated GGML, which is no longer maintained by llama.cpp. GGUF boasts several improvements over GGML, such as enhanced tokenization, support for special tokens, and metadata capabilities. It is also designed with extensibility in mind.
|
37 |
|
@@ -49,7 +49,7 @@ Below is a partial list of clients and libraries known to support GGUF:
|
|
49 |
<!-- README_GGUF.md-about-gguf end -->
|
50 |
|
51 |
<!-- prompt-template start -->
|
52 |
-
|
53 |
|
54 |
```
|
55 |
{prompt}
|
@@ -60,7 +60,7 @@ Below is a partial list of clients and libraries known to support GGUF:
|
|
60 |
|
61 |
|
62 |
<!-- compatibility_gguf start -->
|
63 |
-
|
64 |
|
65 |
The NT-Java-1.1B GGUFs are supported by llama.cpp and are compatible with a range of third-party user interfaces and libraries. For a detailed list, please refer to the beginning of this README.
|
66 |
# Explanation of quantisation methods
|
@@ -257,7 +257,7 @@ $env:CMAKE_ARGS = "-DLLAMA_OPENBLAS=on"
|
|
257 |
pip install llama-cpp-python
|
258 |
```
|
259 |
|
260 |
-
|
261 |
|
262 |
```python
|
263 |
from llama_cpp import Llama
|
@@ -290,7 +290,7 @@ Here are guides on using llama-cpp-python and ctransformers with LangChain:
|
|
290 |
<!-- footer start -->
|
291 |
<!-- 200823 -->
|
292 |
|
293 |
-
|
294 |
```
|
295 |
@article{li2023starcoder,
|
296 |
title={NARROW TRANSFORMER: STARCODER-BASED JAVA-LM FOR DESKTOP},
|
@@ -302,7 +302,7 @@ Here are guides on using llama-cpp-python and ctransformers with LangChain:
|
|
302 |
}
|
303 |
```
|
304 |
|
305 |
-
|
306 |
|
307 |
# **NT-Java-1.1B**
|
308 |
|
|
|
25 |
- Original model: [NT-Java-1.1B](https://huggingface.co/infosys/NT-Java-1.1B)
|
26 |
|
27 |
<!-- description start -->
|
28 |
+
## Description
|
29 |
|
30 |
This repo contains GGUF format model files for [Infosys's NT-Java-1.1B](https://huggingface.co/infosys/NT-Java-1.1B).
|
31 |
|
32 |
<!-- description end -->
|
33 |
<!-- README_GGUF.md-about-gguf start -->
|
34 |
+
### About GGUF
|
35 |
|
36 |
GGUF, introduced by the llama.cpp team on August 21st, 2023, is a new format designed to replace the outdated GGML, which is no longer maintained by llama.cpp. GGUF boasts several improvements over GGML, such as enhanced tokenization, support for special tokens, and metadata capabilities. It is also designed with extensibility in mind.
|
37 |
|
|
|
49 |
<!-- README_GGUF.md-about-gguf end -->
|
50 |
|
51 |
<!-- prompt-template start -->
|
52 |
+
## Prompt template: None
|
53 |
|
54 |
```
|
55 |
{prompt}
|
|
|
60 |
|
61 |
|
62 |
<!-- compatibility_gguf start -->
|
63 |
+
## Compatibility
|
64 |
|
65 |
The NT-Java-1.1B GGUFs are supported by llama.cpp and are compatible with a range of third-party user interfaces and libraries. For a detailed list, please refer to the beginning of this README.
|
66 |
# Explanation of quantisation methods
|
|
|
257 |
pip install llama-cpp-python
|
258 |
```
|
259 |
|
260 |
+
### Simple llama-cpp-python example code
|
261 |
|
262 |
```python
|
263 |
from llama_cpp import Llama
|
|
|
290 |
<!-- footer start -->
|
291 |
<!-- 200823 -->
|
292 |
|
293 |
+
## Citation
|
294 |
```
|
295 |
@article{li2023starcoder,
|
296 |
title={NARROW TRANSFORMER: STARCODER-BASED JAVA-LM FOR DESKTOP},
|
|
|
302 |
}
|
303 |
```
|
304 |
|
305 |
+
# Original model card: Infosys's [NT-Java-1.1B](https://huggingface.co/infosys/NT-Java-1.1B)
|
306 |
|
307 |
# **NT-Java-1.1B**
|
308 |
|