File size: 1,730 Bytes
9652864
 
 
 
 
 
 
 
 
 
 
 
 
f0b9627
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
---
title: Basic Smol Talk Chatbot
emoji: 🐠
colorFrom: green
colorTo: blue
sdk: gradio
sdk_version: 5.15.0
app_file: app.py
pinned: false
license: apache-2.0
short_description: A replica of the basic smoltalk chatbot I run locally
---

# Basic SmolLM2 chatbot:

This is a very basic chatbot using HuggingFaceTB/SmolLM2-[x]-Instruct hosted on the same host as the app. It is basically a replica of the chatbot I have for running locally.

## Hardware Scale Up

- I recommend running this on a little better hardware than this is set up with.
- I'm using the free tier space, but it needs a few more CPU to write fast enough to be useful.
- It is rather slow on this setup, but when I run this on my laptop, it works very well on CPU without GPU.

## To run locally

- Download the files or clone the repo.
- Make sure you have a supported version of transformers and torch installed (or run `pip3 install -r requirements.txt` from the root folder of this repo).
- Run `python3 app.py` from the root folder of this repo.
- Set your browser to http://0.0.0.0:7860


## Configuration options

- In app.py, there are settings  of the screen.
    - If you run this locally on a laptop with at least 5CPU cores, I would recommend saving all your local work then set `MODEL` to HuggingFaceTB/SmolLM2-360M-Instruct.
    - If this works without signs of resource saturation, try setting `MODEL` it to HuggingFaceTB/SmolLM2-1.7B-Instruct. This will write well and works fine on my laptop that is about 2 years old.

```
MAX_NEW_TOKENS = 250
MODEL="HuggingFaceTB/SmolLM2-135M-Instruct"
# MODEL="HuggingFaceTB/SmolLM2-360M-Instruct"
# MODEL="HuggingFaceTB/SmolLM2-1.7B-Instruct"
TEMPERATURE = 0.6
TOP_P = 0.95
REPETITION_PENALTY = 1.2
```