pankajmathur commited on
Commit
425ccb9
·
1 Parent(s): 68de732

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +14 -14
README.md CHANGED
@@ -56,7 +56,7 @@ Here are the results on metrics used by [HuggingFaceH4 Open LLM Leaderboard](htt
56
 
57
  ## Prompt Format
58
 
59
- Here is the prompt format
60
 
61
  ```
62
  ### System:
@@ -69,6 +69,17 @@ Tell me about Orcas.
69
 
70
  ```
71
 
 
 
 
 
 
 
 
 
 
 
 
72
  #### OobaBooga Instructions:
73
 
74
  This model required upto 45GB GPU VRAM in 4bit so it can be loaded directly on Single RTX 6000/L40/A40/A100/H100 GPU or Double RTX 4090/L4/A10/RTX 3090/RTX A5000
@@ -86,7 +97,7 @@ After that go to Default Tab/Page on OobaBooga Web UI and **copy paste above pro
86
 
87
  #### Code Instructions:
88
 
89
- Below shows a code example on how to use this model
90
 
91
  ```python
92
  import torch
@@ -113,18 +124,7 @@ print(tokenizer.decode(output[0], skip_special_tokens=True))
113
  ```
114
 
115
 
116
- Here is the Alpaca prompt format
117
-
118
- ```
119
-
120
- ### User:
121
- Tell me about Alpacas.
122
-
123
- ### Assistant:
124
-
125
- ```
126
-
127
- Below shows a code example on how to use this model
128
 
129
  ```python
130
  import torch
 
56
 
57
  ## Prompt Format
58
 
59
+ Here is the Orca prompt format
60
 
61
  ```
62
  ### System:
 
69
 
70
  ```
71
 
72
+ Here is the Alpaca prompt format
73
+
74
+ ```
75
+
76
+ ### User:
77
+ Tell me about Alpacas.
78
+
79
+ ### Assistant:
80
+
81
+ ```
82
+
83
  #### OobaBooga Instructions:
84
 
85
  This model required upto 45GB GPU VRAM in 4bit so it can be loaded directly on Single RTX 6000/L40/A40/A100/H100 GPU or Double RTX 4090/L4/A10/RTX 3090/RTX A5000
 
97
 
98
  #### Code Instructions:
99
 
100
+ Below shows a code example on how to use this model via Orca prompt
101
 
102
  ```python
103
  import torch
 
124
  ```
125
 
126
 
127
+ Below shows a code example on how to use this model via Alpaca prompt
 
 
 
 
 
 
 
 
 
 
 
128
 
129
  ```python
130
  import torch