This directory contains usage examples of the Hugging Face Deep Learning Containers (DLCs) in Google Kubernetes Engine (GKE) for both training and inference, with a focus on Large Language Models (LLMs).
Example | Description |
---|---|
trl-full-fine-tuning | Full SFT fine-tuning of Gemma 2B in a multi-GPU instance with TRL on GKE. |
trl-lora-fine-tuning | LoRA SFT fine-tuning of Mistral 7B v0.3 in a single GPU instance with TRL on GKE. |
Example | Description |
---|---|
tgi-deployment | Deploying Llama3 8B with Text Generation Inference (TGI) on GKE. |
tgi-from-gcs-deployment | Deploying Qwen2 7B Instruct with Text Generation Inference (TGI) from a GCS Bucket on GKE. |
tei-deployment | Deploying Snowflake’s Arctic Embed (M) with Text Embeddings Inference (TEI) on GKE. |
tei-from-gcs-deployment | Deploying BGE Base v1.5 (English) with Text Embeddings Inference (TEI) from a GCS Bucket on GKE. |