Daniel Dahlmeier's picture
1 3

Daniel Dahlmeier

ddahlmeier

AI & ML interests

NLP

Recent Activity

updated a model 6 days ago
ddahlmeier/llama-3.1-1B-aws
published a model 6 days ago
ddahlmeier/llama-3.1-1B-aws
updated a model 6 days ago
ddahlmeier/llama-3.1-1B
View all activity

Organizations

None yet

ddahlmeier's activity

upvoted an article 19 days ago
view article
Article

Fine-Tuning 1B LLaMA 3.2: A Comprehensive Step-by-Step Guide with Code

56
New activity in McGill-NLP/feedbackQA 2 months ago

NonMatchingChecksumError

#3 opened 2 months ago by
ddahlmeier
reacted to philschmid's post with 👍 6 months ago
view post
Post
What's the best way to fine-tune open LLMs in 2024? Look no further! 👀 I am excited to share “How to Fine-Tune LLMs in 2024 with Hugging Face” using the latest research techniques, including Flash Attention, Q-LoRA, OpenAI dataset formats (messages), ChatML, Packing, all built with Hugging Face TRL. 🚀

It is created for consumer-size GPUs (24GB) covering the full end-to-end lifecycle with:
💡Define and understand use cases for fine-tuning
🧑🏻‍💻 Setup of the development environment
🧮 Create and prepare dataset (OpenAI format)
🏋️‍♀️ Fine-tune LLM using TRL and the SFTTrainer
🥇 Test and evaluate the LLM
🚀 Deploy for production with TGI

👉  https://www.philschmid.de/fine-tune-llms-in-2024-with-trl

Coming soon: Advanced Guides for multi-GPU/multi-Node full fine-tuning and alignment using DPO & KTO. 🔜
·