|
--- |
|
tags: |
|
- ltx-video |
|
- text-to-video |
|
- image-to-video |
|
pinned: true |
|
language: |
|
- en |
|
license: other |
|
--- |
|
|
|
This is a special fork of LTX-Video designed to facilitate deployment to the Inference Endpoints, using diffusers + varnish. |
|
|
|
If you want to know how to use it once you've deployed it, please refer to this Python demo: |
|
|
|
https://huggingface.co/jbilcke-hf/LTX-Video-for-InferenceEndpoints/blob/main/example.py |
|
|
|
## Why? |
|
|
|
Using the Hugging Face Inference Endpoints gives you with reliable and controllable availability. |
|
|
|
The goal of this wrapper is to offer flexibility and low-level settings, to add things such as upscaling, interpolation, film grain, and compression controls. |
|
|
|
## Setup on Hugging Face Inference Endpoints |
|
|
|
### Pick a large machine |
|
|
|
It is recommended to use at least a NVIDIA L40S with 48 Gb of VRAM. |
|
|
|
### Downloads all assets |
|
|
|
Make sure to select "Download everything" when selecting the model. |
|
|
|
Otherwise some files ending in `.pth` won't be downloaded. |
|
|
|
### Select Text-to-Video or Image-to-Video |
|
|
|
By default the handler will do Text-to-Video. |
|
|
|
To do Image-Text-To-Video, you need to set the environment variable `SUPPORT_INPUT_IMAGE_PROMPT` to a trueish value (eg `1`, `True`). |
|
|
|
It is possible to support both pipelines at the same time if you modify the `handler.py`. |
|
|
|
But if you keep both pipelines active in parallel, this will consume more memory. |
|
|
|
### Using private LTX-Video LoRAs |
|
|
|
If you plan on using private LoRA models you will have to set the `HF_API_TOKEN` environment variable. |
|
|
|
## Credits |
|
|
|
For more information about this model, please see the [original HF repository here](https://huggingface.co/Lightricks/LTX-Video/). |
|
|