Fine-tuning large models on local hardware — Benjamin Bossan

preview_player
Показать описание
[EuroPython 2024 — Forum Hall on 2024-07-11]

Fine-tuning large models on local hardware by Benjamin Bossan

Fine-tuning big neural nets like Large Language Models (LLMs) has traditionally been prohibitive due to high hardware requirements. However, Parameter-Efficient Fine-Tuning (PEFT) and quantization enable the training of large models on modest hardware. Thanks to the PEFT library and the Hugging Face ecosystem, these techniques are now accessible to a broad audience.

Expect to learn:

- what the challenges are of fine-tuning large models
- what solutions have been proposed and how they work
- practical examples of applying the PEFT library

---
Рекомендации по теме