Update README.md
Browse files
README.md
CHANGED
@@ -307,6 +307,6 @@ Here are guides on using llama-cpp-python and ctransformers with LangChain:
|
|
307 |
|
308 |
#### Original model card: Infosys's [NT-Java-1.1B](https://huggingface.co/infosys/NT-Java-1.1B)
|
309 |
|
310 |
-
# **NT-Java**
|
311 |
|
312 |
The Narrow Transformer (NT) model NT-Java-1.1B is an open-source specialized code model built by extending pre-training on StarCoderBase-1B, designed for coding tasks in Java programming. The model is a decoder-only transformer with Multi-Query Attention and with a context length of 8192 tokens. The model was trained with Java subset of the StarCoderData dataset, which is ~22B tokens.
|
|
|
307 |
|
308 |
#### Original model card: Infosys's [NT-Java-1.1B](https://huggingface.co/infosys/NT-Java-1.1B)
|
309 |
|
310 |
+
# **NT-Java-1.1B**
|
311 |
|
312 |
The Narrow Transformer (NT) model NT-Java-1.1B is an open-source specialized code model built by extending pre-training on StarCoderBase-1B, designed for coding tasks in Java programming. The model is a decoder-only transformer with Multi-Query Attention and with a context length of 8192 tokens. The model was trained with Java subset of the StarCoderData dataset, which is ~22B tokens.
|