File size: 8,027 Bytes
abbdd5c
 
03f1688
 
 
 
 
d17ae58
03f1688
 
 
d17ae58
abbdd5c
 
f221ed9
3aaa468
abbdd5c
bd5d89f
abbdd5c
c650ac0
 
 
 
 
 
 
 
 
 
 
 
c07bbe9
abbdd5c
c650ac0
4b8d003
abbdd5c
 
edb376b
 
abbdd5c
edb376b
c07bbe9
 
 
 
 
 
 
 
 
 
 
 
 
abbdd5c
69c61db
 
 
 
3b12e45
abbdd5c
25ecd6f
abbdd5c
edb376b
03f1688
25ecd6f
03f1688
 
 
 
14eec2e
 
69c61db
14eec2e
f997800
14eec2e
 
a9a7462
25ecd6f
14eec2e
 
25ecd6f
bd5d89f
 
69c61db
bd5d89f
 
 
69c61db
bd5d89f
 
 
69c61db
bd5d89f
 
 
 
69c61db
706ac6a
25ecd6f
 
69c61db
25ecd6f
 
 
 
 
 
 
c650ac0
 
 
25ecd6f
69c61db
 
 
 
 
25ecd6f
706ac6a
 
 
25ecd6f
 
 
77b337e
706ac6a
24156fb
706ac6a
61ef961
2c4942c
61ef961
706ac6a
 
25ecd6f
706ac6a
24156fb
61ef961
 
 
 
fa69537
61ef961
fa69537
 
 
f7327e7
 
61ef961
fa69537
61ef961
 
25ecd6f
61ef961
 
 
 
fa69537
61ef961
fa69537
24156fb
fa69537
61ef961
 
 
 
 
 
 
fa69537
61ef961
fa69537
24156fb
fa69537
61ef961
fa69537
61ef961
fa69537
61ef961
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
131
132
133
134
135
136
137
138
139
140
141
142
143
144
145
146
147
148
149
150
151
152
153
154
155
156
157
158
159
160
161
162
163
164
165
166
167
168
169
170
171
172
173
174
175
176
177
178
179
---
license: apache-2.0
datasets:
- Sweaterdog/Andy-3.5
language:
- en
base_model:
- deepseek-ai/DeepSeek-R1-Distill-Qwen-7B
- deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B
tags:
- Minecraft
- MindCraft
---

# 🚀 Welcome to a new generation of Minecraft with Andy 3.5 🚀

## Andy 3.5 is a collection of LOCAL LLM's designed for playing Minecraft
*Andy 3.5 is designed to be used with MindCraft, and is not designed nor intended to be used for any other applications*


>  # Please note! [!WARNING]
> 
> Andy-3.5 was trained on older data, and not the newest and latest versions of Mindcraft.
> 
> I **cannot** guarantee that Andy-3.5 will work on future versions as the model was tuned to play MindCraft with a specific version!
> 
> For the rest of the Andy-3.5 generation, this model will **ONLY** be supported on the version of Mindcraft in [this github repo!](https://github.com/Sweaterdog/Mindcraft-for-Andy-3.5)
>
> For more info, as well as the supported version of Mindcraft, please follow [this link to github](https://github.com/Sweaterdog/Mindcraft-for-Andy-3.5)


# How to Install / Setup

1. Select the model you would like to use *(The regular model, as well as the small model is recommended)*
2. Download the Modelfile
3. Once downloaded, open Modelfile in a text editor, and change the path to the download location of the gguf file
4. When changed, save the file, and open command terminal
5. *(Optional if CMD isn't opened via file explorer)* Navigate to the correct directory using "cd"
6. Run the command ```ollama create sweaterdog/Andy-3.5 -f Modelfile``` If you want multiple models, include a tag afterwards. Example: sweaterdog/Andy-3.5:mini-fp16 or sweaterdog/Andy-3.5:q2_k
7. Go to a profile in MindCraft
8. Change the model to be ```sweaterdog/Andy-3.5``` *Or whatever you named your model*
9. Ensure you have the emdedding tag set to Ollama, like below
```
{
    "name": "andy-3.5",

    "model": "Sweaterdog/Andy-3.5",

    "embedding": "ollama"

}
```

10. Enjoy playing with an AI that you are hosting!

> # Ollama Support [!NOTE]
> On Huggingface, there is an option to download GGUF models via Ollama
>
> However, this method **DOES NOT WORK** for models other than the base model of Andy-3.5!


# How was model trained?

The model was trained on the [MindCraft dataset](https://huggingface.co/datasets/Sweaterdog/Andy-3.5) for Andy-3.5, which includes ~12,000 prompts, featuring all things Minecraft.

# What are capabilities and Limitations?

Andy-3.5 was trained on EVERYTHING regarding Minecraft and MindCraft, it knows how to use commands natively without a system prompt.
Andy-3.5 also knows how to build / use !newAction to perform commands, it was trained on lots of building, as well as, using !newAction to do tasks like manually making something or strip mining.

# What models can I choose?

There are going to be 3 odel sizes avaliable, Regular, Small, and Mini
* Regular is a 7B parameter model, tuned from [Deepseek-R1 Distilled](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-7B)
* Small is a 3B parameter model, tuned from [Qwen2.5 3B](Qwen/Qwen2.5-3B-Instruct)
* Mini is a 1.5B parameter model, also tuned from [Deepseek-R1 Distilled](https://huggingface.co/deepseek-ai/DeepSeek-R1-Distill-Qwen-1.5B)

Small has a dedicated **"reasoning"** version released, Regular won't have a reasoning tune, Andy-3.6 will have built in case-by-case reasoning.

Out of all of the models, Teensy had the largest percent of parameters tuned, being 1/2 the models total size

# Safety and FAQ

Q: Is this model safe to use?

A. Yes, this model is non-volatile, and cannot generate malicous content

Q. Can this model be used on a server?

A. Yes, In theory and practice the model is only capable of building and performing manual tasks via newAction

Q. Who is responsible if this model does generate malicous content?

A. You are responsible, even though the model was never trained to be able to make malicous content, there is a ***very very slight chance*** it still generates malicous code.

Q. If I make media based on this model, like photos / videos, do I have to mention the Creator?

A. No, if you are making a post about MindCraft, and using this model, you only have to mention the creator if you mention the model being used.

# 🔥UPDATE🔥

**All models have their own folder, besides the main version of Andy-3.5**

To find models such as reasoning or mini, go into files and search inside the folder

**There is an Andy-3.5-reasoning-preview model, designed to demonstrate reasoning abilities in small language models to improve Minecraft skills**

Remember that this is a ***preview*** model and is **not** guaranteed to work, nor perform better or the same as Andy-3.5-*(Base)* 

When the full Andy-3.5-reasoning model is released, there will be the regular 7B model, as well as the small model, which is 3B parameters.

For future updates and generations there will **not** be a mini and a teensy version, of course the name may stay, but there wil **not** be a 1.5B **nor** a 360M model

> # I want to thank all supporters! [!NOTE]
> I would love to thank everyone who supported this project, there is a list of supporters in the files section.
> 
> You can find all of the supporters [here](https://huggingface.co/Sweaterdog/Andy-3.5/blob/main/Supporters.txt)

# Performance Metrics

These benchmarks are a-typical, since most standard benchmarks don't apply to Minecraft

The benchmarks below include models via API that are cheap, and other fine-tuned local models *(Excluding Andy-v2 and Andy-v3, since they are poor in quality)*

## Zero info Prompting
*How fast can a model collect 16 oak logs, and convert them all into sticks*

![image/png](https://cdn-uploads.huggingface.co/production/uploads/66960602f0ffd8e3a381106a/IEw1Gydg943qVSNGAL3RW.png)


Currently, Andy-3.5, Andy-3.5-small, and Andy-3.5-mini are the **ONLY** models that can play without command documentation, or any other instruction, and Andy-3.5-Mini *sometimes* fares better ***without*** the unnecessary data.
Test this for yourself using [this profile](https://huggingface.co/Sweaterdog/Andy-3.5/blob/main/local_demo.json)


## Time to get a stone pickaxe

![image/png](https://cdn-uploads.huggingface.co/production/uploads/66960602f0ffd8e3a381106a/frrT9IcJsNeUOLhszFrOq.png)

I am sure other models like Deepseek-R1 may be faster at getting a stone pickaxe, however the Demo was to show the performance of Andy-3.5

*For Andy-3.5-mini, I used the FP16 model, I had enough VRAM to do so*

*For Andy-3.5, I used the Q4_K_M quantization*

*For Andy-3.5-small, I used the Q8_0 quantization*

*Andy-3.5-reasoning-small was able to be the most efficient model producing the lowest amount of messages, but took a whopping 34.5 minutes to get a stone pickaxe.*

*For Andy-3.5-Teensy, I used the FP16 quantization*

*For Mineslayerv1 and Mineslayerv2, I used the default (and only) quantization, Q4_K_M*

## Notes about the benchmarks

**Zero Info Prompting**

Andy-3.5-Teensy was able to use one command successfully, but was not able to afterwards

Andy-3.5-Mini collected 32 oak_log instead of 16 oak_log

Andy-3.5-small *No notes*

Andy-3.5 attempted to continue playing, and make a wooden_pickaxe after the goal was done.

Both Mineslayerv1 and Mineslayerv2 hallucinated commands, like !chop or !grab

**Time to get a stone pickaxe**

Andy-3.5-teensy hallucinates too much for stable gameplay *(It is a 360M parameter model, what can be expected)*

Andy-3.5-Mini was unable to make itself a stone pickaxe, however it collected enough wood, but then got stuck on converting logs to planks, it kept trying "!craftRecipe("wooden_planks", 6) instead of oak_planks

Andy-3.5-small kept trying to make a stone_pickaxe first

Andy-3.5 Made a stone pickaxe the fastest out of all models, including GPT-4o-mini and Claude-3.5-Haiku

Mineslayerv1 Was unable to use !collectBlocks, instead kept trying !collectBlock

Mineslayerv2 Was unable to play, it kept hallucinating on the first command