rajabmondal commited on
Commit
bd4d547
·
verified ·
1 Parent(s): d48fce8

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +7 -7
README.md CHANGED
@@ -25,13 +25,13 @@ pipeline_tag: text-generation
25
  - Original model: [NT-Java-1.1B](https://huggingface.co/infosys/NT-Java-1.1B)
26
 
27
  <!-- description start -->
28
- # Description
29
 
30
  This repo contains GGUF format model files for [Infosys's NT-Java-1.1B](https://huggingface.co/infosys/NT-Java-1.1B).
31
 
32
  <!-- description end -->
33
  <!-- README_GGUF.md-about-gguf start -->
34
- # About GGUF
35
 
36
  GGUF, introduced by the llama.cpp team on August 21st, 2023, is a new format designed to replace the outdated GGML, which is no longer maintained by llama.cpp. GGUF boasts several improvements over GGML, such as enhanced tokenization, support for special tokens, and metadata capabilities. It is also designed with extensibility in mind.
37
 
@@ -49,7 +49,7 @@ Below is a partial list of clients and libraries known to support GGUF:
49
  <!-- README_GGUF.md-about-gguf end -->
50
 
51
  <!-- prompt-template start -->
52
- # Prompt template: None
53
 
54
  ```
55
  {prompt}
@@ -60,7 +60,7 @@ Below is a partial list of clients and libraries known to support GGUF:
60
 
61
 
62
  <!-- compatibility_gguf start -->
63
- # Compatibility
64
 
65
  The NT-Java-1.1B GGUFs are supported by llama.cpp and are compatible with a range of third-party user interfaces and libraries. For a detailed list, please refer to the beginning of this README.
66
  # Explanation of quantisation methods
@@ -257,7 +257,7 @@ $env:CMAKE_ARGS = "-DLLAMA_OPENBLAS=on"
257
  pip install llama-cpp-python
258
  ```
259
 
260
- #### Simple llama-cpp-python example code
261
 
262
  ```python
263
  from llama_cpp import Llama
@@ -290,7 +290,7 @@ Here are guides on using llama-cpp-python and ctransformers with LangChain:
290
  <!-- footer start -->
291
  <!-- 200823 -->
292
 
293
- # Citation
294
  ```
295
  @article{li2023starcoder,
296
  title={NARROW TRANSFORMER: STARCODER-BASED JAVA-LM FOR DESKTOP},
@@ -302,7 +302,7 @@ Here are guides on using llama-cpp-python and ctransformers with LangChain:
302
  }
303
  ```
304
 
305
- #### Original model card: Infosys's [NT-Java-1.1B](https://huggingface.co/infosys/NT-Java-1.1B)
306
 
307
  # **NT-Java-1.1B**
308
 
 
25
  - Original model: [NT-Java-1.1B](https://huggingface.co/infosys/NT-Java-1.1B)
26
 
27
  <!-- description start -->
28
+ ## Description
29
 
30
  This repo contains GGUF format model files for [Infosys's NT-Java-1.1B](https://huggingface.co/infosys/NT-Java-1.1B).
31
 
32
  <!-- description end -->
33
  <!-- README_GGUF.md-about-gguf start -->
34
+ ### About GGUF
35
 
36
  GGUF, introduced by the llama.cpp team on August 21st, 2023, is a new format designed to replace the outdated GGML, which is no longer maintained by llama.cpp. GGUF boasts several improvements over GGML, such as enhanced tokenization, support for special tokens, and metadata capabilities. It is also designed with extensibility in mind.
37
 
 
49
  <!-- README_GGUF.md-about-gguf end -->
50
 
51
  <!-- prompt-template start -->
52
+ ## Prompt template: None
53
 
54
  ```
55
  {prompt}
 
60
 
61
 
62
  <!-- compatibility_gguf start -->
63
+ ## Compatibility
64
 
65
  The NT-Java-1.1B GGUFs are supported by llama.cpp and are compatible with a range of third-party user interfaces and libraries. For a detailed list, please refer to the beginning of this README.
66
  # Explanation of quantisation methods
 
257
  pip install llama-cpp-python
258
  ```
259
 
260
+ ### Simple llama-cpp-python example code
261
 
262
  ```python
263
  from llama_cpp import Llama
 
290
  <!-- footer start -->
291
  <!-- 200823 -->
292
 
293
+ ## Citation
294
  ```
295
  @article{li2023starcoder,
296
  title={NARROW TRANSFORMER: STARCODER-BASED JAVA-LM FOR DESKTOP},
 
302
  }
303
  ```
304
 
305
+ # Original model card: Infosys's [NT-Java-1.1B](https://huggingface.co/infosys/NT-Java-1.1B)
306
 
307
  # **NT-Java-1.1B**
308