Suparious commited on
Commit
75c04a4
·
verified ·
1 Parent(s): 5867fde

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +129 -1
README.md CHANGED
@@ -1,3 +1,131 @@
1
  ---
2
- license: apache-2.0
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ base_model:
3
+ - ResplendentAI/Datura_7B
4
+ - ChaoticNeutrals/Cookie_7B
5
+ datasets:
6
+ - ResplendentAI/Luna_NSFW_Text
7
+ - unalignment/toxic-dpo-v0.2
8
+ - ResplendentAI/Synthetic_Soul_1k
9
+ - grimulkan/theory-of-mind
10
+ - lemonilia/LimaRP
11
+ - PygmalionAI/PIPPA
12
+ library_name: transformers
13
+ tags:
14
+ - mistral
15
+ - 4-bit
16
+ - AWQ
17
+ - text-generation
18
+ - autotrain_compatible
19
+ - endpoints_compatible
20
+ - chatml
21
+ - not-for-all-audiences
22
+ license: other
23
+ language:
24
+ - en
25
+ pipeline_tag: text-generation
26
+ inference: false
27
+ prompt_template: '<|im_start|>system
28
+
29
+ {system_message}<|im_end|>
30
+
31
+ <|im_start|>user
32
+
33
+ {prompt}<|im_end|>
34
+
35
+ <|im_start|>assistant
36
+
37
+ '
38
+ quantized_by: Suparious
39
  ---
40
+ # ResplendentAI/DaturaCookie_7B AWQ
41
+
42
+ - Model creator: [ResplendentAI](https://huggingface.co/ResplendentAI)
43
+ - Original model: [DaturaCookie_7B](https://huggingface.co/ResplendentAI/DaturaCookie_7B)
44
+
45
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/626dfb8786671a29c715f8a9/5jG2dft51fgPcGUGc-4Ym.png)
46
+
47
+ ## Model Summary
48
+
49
+ Proficient at roleplaying and lightehearted conversation, this model is prone to NSFW outputs.
50
+
51
+ # Vision/multimodal capabilities:
52
+
53
+ If you want to use vision functionality:
54
+
55
+ You must use the latest versions of Koboldcpp. To use the multimodal capabilities of this model and use vision you need to load the specified mmproj file, this can be found inside this model repo.
56
+
57
+ You can load the mmproj by using the corresponding section in the interface:
58
+
59
+ ![image/png](https://cdn-uploads.huggingface.co/production/uploads/626dfb8786671a29c715f8a9/UxH8OteeRbD1av1re0yNZ.png)
60
+
61
+ ## How to use
62
+
63
+ ### Install the necessary packages
64
+
65
+ ```bash
66
+ pip install --upgrade autoawq autoawq-kernels
67
+ ```
68
+
69
+ ### Example Python code
70
+
71
+ ```python
72
+ from awq import AutoAWQForCausalLM
73
+ from transformers import AutoTokenizer, TextStreamer
74
+
75
+ model_path = "solidrust/DaturaCookie_7B-AWQ"
76
+ system_message = "You are DaturaCookie, incarnated as a powerful AI."
77
+
78
+ # Load model
79
+ model = AutoAWQForCausalLM.from_quantized(model_path,
80
+ fuse_layers=True)
81
+ tokenizer = AutoTokenizer.from_pretrained(model_path,
82
+ trust_remote_code=True)
83
+ streamer = TextStreamer(tokenizer,
84
+ skip_prompt=True,
85
+ skip_special_tokens=True)
86
+
87
+ # Convert prompt to tokens
88
+ prompt_template = """\
89
+ <|im_start|>system
90
+ {system_message}<|im_end|>
91
+ <|im_start|>user
92
+ {prompt}<|im_end|>
93
+ <|im_start|>assistant"""
94
+
95
+ prompt = "You're standing on the surface of the Earth. "\
96
+ "You walk one mile south, one mile west and one mile north. "\
97
+ "You end up exactly where you started. Where are you?"
98
+
99
+ tokens = tokenizer(prompt_template.format(system_message=system_message,prompt=prompt),
100
+ return_tensors='pt').input_ids.cuda()
101
+
102
+ # Generate output
103
+ generation_output = model.generate(tokens,
104
+ streamer=streamer,
105
+ max_new_tokens=512)
106
+
107
+ ```
108
+
109
+ ### About AWQ
110
+
111
+ AWQ is an efficient, accurate and blazing-fast low-bit weight quantization method, currently supporting 4-bit quantization. Compared to GPTQ, it offers faster Transformers-based inference with equivalent or better quality compared to the most commonly used GPTQ settings.
112
+
113
+ AWQ models are currently supported on Linux and Windows, with NVidia GPUs only. macOS users: please use GGUF models instead.
114
+
115
+ It is supported by:
116
+
117
+ - [Text Generation Webui](https://github.com/oobabooga/text-generation-webui) - using Loader: AutoAWQ
118
+ - [vLLM](https://github.com/vllm-project/vllm) - version 0.2.2 or later for support for all model types.
119
+ - [Hugging Face Text Generation Inference (TGI)](https://github.com/huggingface/text-generation-inference)
120
+ - [Transformers](https://huggingface.co/docs/transformers) version 4.35.0 and later, from any code or client that supports Transformers
121
+ - [AutoAWQ](https://github.com/casper-hansen/AutoAWQ) - for use from Python code
122
+
123
+ ## Prompt template: ChatML
124
+
125
+ ```plaintext
126
+ <|im_start|>system
127
+ {system_message}<|im_end|>
128
+ <|im_start|>user
129
+ {prompt}<|im_end|>
130
+ <|im_start|>assistant
131
+ ```