|
FROM ubuntu:22.04 |
|
|
|
|
|
RUN apt-get update && \ |
|
apt-get install -y \ |
|
build-essential \ |
|
libssl-dev \ |
|
zlib1g-dev \ |
|
libboost-system-dev \ |
|
libboost-filesystem-dev \ |
|
cmake \ |
|
git \ |
|
python3-pip \ |
|
curl \ |
|
wget && \ |
|
rm -rf /var/lib/apt/lists/* |
|
|
|
|
|
RUN pip3 install huggingface-hub openai gradio |
|
|
|
|
|
RUN git clone https://github.com/ggerganov/llama.cpp && \ |
|
cd llama.cpp && \ |
|
mkdir build && \ |
|
cd build && \ |
|
cmake .. -DLLAMA_BUILD_SERVER=ON -DLLAMA_BUILD_EXAMPLES=ON -DCMAKE_BUILD_TYPE=Release && \ |
|
cmake --build . --config Release --target llama-server -j $(nproc) |
|
|
|
|
|
RUN mkdir -p /models && \ |
|
wget -O /models/model.q8_0.gguf https://huggingface.co/unsloth/DeepSeek-R1-Distill-Qwen-1.5B-GGUF/resolve/main/DeepSeek-R1-Distill-Qwen-1.5B-Q8_0.gguf |
|
|
|
|
|
COPY app.py /app.py |
|
COPY start.sh /start.sh |
|
RUN chmod +x /start.sh |
|
|
|
|
|
EXPOSE 7860 8080 |
|
|
|
|
|
CMD ["/start.sh"] |