Skip to content

Fine-Tuning Llama3-8B LLM in a multi-GPU environment using DeepSpeed

Notifications You must be signed in to change notification settings

mallik3006/LLM_fine_tuning_llama3_8b

Repository files navigation

LLM_fine_tuning_llama3_8b

Introduction

The repo is a demonstration of fine tuning an open-source LLM (Llama-3-8B) utilizing different approaches and techniques. Fine-Tuning was done with ORPO technique that combines SFT and RLHF methods for preference alignment. The work explores fine tuning on multi-GPU environment utilizing distributed training methods like DeepSpeed, DDP and FSDP using the accelerate library provided by HuggingFace.

Stack

  • LLM - Meta-Llama-3-8B
  • Dataset (HF) - mlabonne/orpo-dpo-mix-40k
  • Fine-Tuning Method - ORPO
  • Accelerator Technique - DeepSpeed ZeRO-3
  • Trainer API - HuggingFace
  • Run-time environment - multi-GPU (2x Telsa T4 GPU - 15GB VRAM each)

Acknowledgments

Thanks for the work shared by Maxime Labonn in his blog here.

About

Fine-Tuning Llama3-8B LLM in a multi-GPU environment using DeepSpeed

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published