Ashqar
commited on
Updating README
Browse files
README.md
CHANGED
@@ -9,14 +9,54 @@ tags:
|
|
9 |
license: apache-2.0
|
10 |
language:
|
11 |
- en
|
|
|
12 |
---
|
13 |
|
14 |
# Uploaded model
|
15 |
|
16 |
-
- **Developed by:**
|
17 |
- **License:** apache-2.0
|
18 |
- **Finetuned from model :** meta-llama/Llama-3.1-8B-Instruct
|
19 |
|
20 |
-
|
|
|
|
|
|
|
21 |
|
22 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
9 |
license: apache-2.0
|
10 |
language:
|
11 |
- en
|
12 |
+
- tr
|
13 |
---
|
14 |
|
15 |
# Uploaded model
|
16 |
|
17 |
+
- **Developed by:** Özgür Entegrasyon
|
18 |
- **License:** apache-2.0
|
19 |
- **Finetuned from model :** meta-llama/Llama-3.1-8B-Instruct
|
20 |
|
21 |
+
OZGURLUK-GPT-LinuxGeneral is a fine-tuned language model developed specifically to assist with Linux system administration,
|
22 |
+
network management, and database troubleshooting. By using a custom dataset focused on technical tasks, this model excels
|
23 |
+
at interpreting Linux commands and offering useful solutions. Whether you're troubleshooting a server, configuring a network,
|
24 |
+
or managing databases, this model offers step-by-step guidance.
|
25 |
|
26 |
+
The model is available in three formats: 4-bit, 8-bit, and 16-bit. The 8-bit version strikes a balance between performance
|
27 |
+
and memory usage, while the 16-bit version offers higher accuracy and precision for more demanding tasks. The 4-bit version
|
28 |
+
provides a lightweight alternative optimized for low-resource environments. Additionally, the HUGG format (GGUF) ensures
|
29 |
+
that the model has a smaller memory footprint with fast load times.
|
30 |
+
|
31 |
+
|
32 |
+
---
|
33 |
+
|
34 |
+
## Dataset
|
35 |
+
|
36 |
+
The dataset used for fine-tuning **OZGURLUK-GPT-LinuxGeneral** was developed internally, gathering knowledge from various sources and internet databases.
|
37 |
+
|
38 |
+
A total of **56,466 question-answer pairs** were collected, covering a broad range of technical domains including **Kubernetes**, **Linux**, **PostgreSQL**, **Docker**, and many others. The dataset is organized under different "ticker" headings, allowing the grouping of similar questions and answers within each domain. This enables the model to better understand the nuances of technical problem-solving and offer relevant responses tailored to users’ specific needs in system administration and other related fields.
|
39 |
+
|
40 |
+
The inclusion of diverse technical topics ensures that the model can assist with various tasks and adapt to a variety of Linux system-related challenges, providing highly accurate, domain-specific solutions.
|
41 |
+
|
42 |
+
|
43 |
+
|
44 |
+
--
|
45 |
+
|
46 |
+
## Key Features
|
47 |
+
|
48 |
+
- **Optimized for Linux:** Trained with a dataset of Linux system commands, troubleshooting solutions, and network configurations.
|
49 |
+
- **Multi-format Support:** Choose between 4-bit, 8-bit, 16-bit, or HUGG for optimal performance based on your hardware.
|
50 |
+
- **Efficient Performance:** Trained using **Unsloth** and **TRL**, achieving faster training and efficient inference.
|
51 |
+
- **Custom Dataset:** Includes 56,466 question-answer pairs across multiple technical domains such as Kubernetes, Linux, PostgreSQL, Docker, etc.
|
52 |
+
|
53 |
+
---
|
54 |
+
|
55 |
+
## Training Details
|
56 |
+
|
57 |
+
- **Base Model:** meta-llama/Llama-3.1-8B-Instruct
|
58 |
+
- **Fine-tuning Method:** LoRA (Low-Rank Adaptation)
|
59 |
+
- **Training Hardware:** Google Colab with A100 GPUs
|
60 |
+
- **Dataset:** Custom-developed dataset of 56,466 question-answer pairs from various technical fields (Kubernetes, Linux, PostgreSQL, Docker, etc.).
|
61 |
+
|
62 |
+
---
|