Ashqar commited on
Commit
b8a6609
·
verified ·
1 Parent(s): f8b7865

Updating README

Browse files
Files changed (1) hide show
  1. README.md +43 -3
README.md CHANGED
@@ -9,14 +9,54 @@ tags:
9
  license: apache-2.0
10
  language:
11
  - en
 
12
  ---
13
 
14
  # Uploaded model
15
 
16
- - **Developed by:** Qutiba
17
  - **License:** apache-2.0
18
  - **Finetuned from model :** meta-llama/Llama-3.1-8B-Instruct
19
 
20
- This llama model was trained 2x faster with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library.
 
 
 
21
 
22
- [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth)
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
9
  license: apache-2.0
10
  language:
11
  - en
12
+ - tr
13
  ---
14
 
15
  # Uploaded model
16
 
17
+ - **Developed by:** Özgür Entegrasyon
18
  - **License:** apache-2.0
19
  - **Finetuned from model :** meta-llama/Llama-3.1-8B-Instruct
20
 
21
+ OZGURLUK-GPT-LinuxGeneral is a fine-tuned language model developed specifically to assist with Linux system administration,
22
+ network management, and database troubleshooting. By using a custom dataset focused on technical tasks, this model excels
23
+ at interpreting Linux commands and offering useful solutions. Whether you're troubleshooting a server, configuring a network,
24
+ or managing databases, this model offers step-by-step guidance.
25
 
26
+ The model is available in three formats: 4-bit, 8-bit, and 16-bit. The 8-bit version strikes a balance between performance
27
+ and memory usage, while the 16-bit version offers higher accuracy and precision for more demanding tasks. The 4-bit version
28
+ provides a lightweight alternative optimized for low-resource environments. Additionally, the HUGG format (GGUF) ensures
29
+ that the model has a smaller memory footprint with fast load times.
30
+
31
+
32
+ ---
33
+
34
+ ## Dataset
35
+
36
+ The dataset used for fine-tuning **OZGURLUK-GPT-LinuxGeneral** was developed internally, gathering knowledge from various sources and internet databases.
37
+
38
+ A total of **56,466 question-answer pairs** were collected, covering a broad range of technical domains including **Kubernetes**, **Linux**, **PostgreSQL**, **Docker**, and many others. The dataset is organized under different "ticker" headings, allowing the grouping of similar questions and answers within each domain. This enables the model to better understand the nuances of technical problem-solving and offer relevant responses tailored to users’ specific needs in system administration and other related fields.
39
+
40
+ The inclusion of diverse technical topics ensures that the model can assist with various tasks and adapt to a variety of Linux system-related challenges, providing highly accurate, domain-specific solutions.
41
+
42
+
43
+
44
+ --
45
+
46
+ ## Key Features
47
+
48
+ - **Optimized for Linux:** Trained with a dataset of Linux system commands, troubleshooting solutions, and network configurations.
49
+ - **Multi-format Support:** Choose between 4-bit, 8-bit, 16-bit, or HUGG for optimal performance based on your hardware.
50
+ - **Efficient Performance:** Trained using **Unsloth** and **TRL**, achieving faster training and efficient inference.
51
+ - **Custom Dataset:** Includes 56,466 question-answer pairs across multiple technical domains such as Kubernetes, Linux, PostgreSQL, Docker, etc.
52
+
53
+ ---
54
+
55
+ ## Training Details
56
+
57
+ - **Base Model:** meta-llama/Llama-3.1-8B-Instruct
58
+ - **Fine-tuning Method:** LoRA (Low-Rank Adaptation)
59
+ - **Training Hardware:** Google Colab with A100 GPUs
60
+ - **Dataset:** Custom-developed dataset of 56,466 question-answer pairs from various technical fields (Kubernetes, Linux, PostgreSQL, Docker, etc.).
61
+
62
+ ---