Skip to content

Releases: hiyouga/LLaMA-Factory

v0.2.1: Variant Models, NEFTune Trick

09 Nov 08:30
Compare
Choose a tag to compare

New features

  • Support NEFTune trick for supervised fine-tuning by @anvie in #1252
  • Support loading dataset in the sharegpt format - read data/readme for details
  • Support generating multiple responses in demo API via the n parameter
  • Support caching the pre-processed dataset files via the cache_path argument
  • Better LLaMA Board (pagination, controls, etc.)
  • Support push_to_hub argument #1088

New models

  • Base models
    • ChatGLM3-6B-Base
    • Yi (6B/34B)
    • Mistral-7B
    • BlueLM-7B-Base
    • Skywork-13B-Base
    • XVERSE-65B
    • Falcon-180B
    • Deepseek-Coder-Base (1.3B/6.7B/33B)
  • Instruct/Chat models
    • ChatGLM3-6B
    • Mistral-7B-Instruct
    • BlueLM-7B-Chat
    • Zephyr-7B
    • OpenChat-3.5
    • Yayi (7B/13B)
    • Deepseek-Coder-Instruct (1.3B/6.7B/33B)

New datasets

  • Pre-training datasets
    • RedPajama V2
    • Pile
  • Supervised fine-tuning datasets
    • OpenPlatypus
    • ShareGPT Hyperfiltered
    • ShareGPT4
    • UltraChat 200k
    • AgentInstruct
    • LMSYS Chat 1M
    • Evol Instruct V2

Bug fix

v0.2.0: Web UI Refactor, LongLoRA

15 Oct 13:06
Compare
Choose a tag to compare

New features

  • Support LongLoRA for the LLaMA models
  • Support training the Qwen-14B and InternLM-20B models
  • Support training state recovery for the all-in-one Web UI
  • Support Ascend NPU by @statelesshz in #975
  • Integrate MMLU, C-Eval and CMMLU benchmarks

Modifications

  • Rename repository to LLaMA Factory (former LLaMA Efficient Tuning)
  • Use the cutoff_len argument instead of max_source_length and max_target_length #944
  • Add a train_on_prompt option #1184

Bug fix

[1] huggingface/transformers#25598 (comment)

v0.1.8: FlashAttention-2 and Baichuan2

11 Sep 09:55
Compare
Choose a tag to compare

New features

  • Support FlashAttention-2 for LLaMA models. (RTX4090, A100, A800 or H100 GPU is required)
  • Support training the Baichuan2 models
  • Use right-padding to avoid overflow in fp16 training (also mentioned here)
  • Align the computation method of the reward score with DeepSpeed-Chat (better generation)
  • Support --lora_target all argument which automatically finds the applicable modules for LoRA training

Bug fix

v0.1.7: Script Preview and RoPE Scaling

18 Aug 09:39
Compare
Choose a tag to compare

New features

  • Preview training script in Web UI by @codemayq in #479 #511
  • Support resuming from checkpoints by @niuba in #434 (transformers>=4.31.0 required)
  • Two RoPE scaling methods: linear and NTK-aware scaling for LLaMA models (transformers>=4.31.0 required)
  • Support training the ChatGLM2-6B model
  • Support PPO training in bfloat16 data type #551

Bug fix

v0.1.6: DPO Training and Qwen-7B

11 Aug 15:43
Compare
Choose a tag to compare

v0.1.5: Patch release

02 Aug 08:13
Compare
Choose a tag to compare

v0.1.4: Dataset Streaming

01 Aug 04:20
Compare
Choose a tag to compare

v0.1.3: Patch release

21 Jul 08:49
Compare
Choose a tag to compare
release v0.1.3

v0.1.2: LLaMA-2 Models

20 Jul 14:42
Compare
Choose a tag to compare
  • Support LLaMA-2 (good issue #202 )
  • Advanced configurations in Web UI
  • Fix API (downgrade pydantic<2.0.0)
  • Fix baichuan lora hparam #194 #212
  • Fix padding #196
  • Fix ZeRO-3 #199
  • Allow pass args to app #213
  • Code simplification
  • Add ShareGPT dataset

v0.1.1

18 Jul 13:05
Compare
Choose a tag to compare
  • Web UI: source_prefix, max_length, dev set
  • Bug fix: reward token #179
  • Update template #171 #177
  • Bug fix: replace the Literal type with Enum for pydantic [1] #176
  • Add Web demo #180

[1] pydantic/pydantic#5821, fastapi/sqlmodel#67