File size: 1,384 Bytes
b73c955 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 |
# MiniGPT-4-Pipeline
A pipeline adding [MiniGPT-4](https://github.com/Vision-CAIR/MiniGPT-4) support to [text-generation-webui](https://github.com/oobabooga/text-generation-webui).
### Installation
Clone this repo into `extensions/multimodal/pipelines` directory in text-generation-webui, and install `requirements.txt`.
### Pipelines
This module provides 2 pipelines:
- `minigpt4-13b` - for use with Vicuna-v0-13B LLM
- `minigpt4-7b` - for use with Vicuna-v0-7B LLM
To use it in webui, select the appropriate LLM and run `server.py` with `--extensions multimodal --multimodal-pipeline minigpt4-13b` (or `minigpt4-7b`)
The supported parameter combinations for both the vision model, and the projector are: CUDA/32bit, CUDA/16bit, CPU/32bit
### Credits
Almost all the code in minigpt4 directory is taken from [the original MiniGPT4 repo](https://github.com/Vision-CAIR/MiniGPT-4), it was then cleaned up to leave mostly the parts, which are needed for inference. The only modifications are to `minigpt4/mini_gpt4.py`, but again, they are mostly removing not-needed parts of the code.
In short: I copied enough code from MiniGPT-4, so that inference works(but only for image embeds), then I added a pipeline descriptor.
### DISCLAIMER
This is not production-ready code, I take no liability whatsoever, and don't provide any warranty, nor support. Use it only for fun/research
|