Skip to content

Releases: hiyouga/LLaMA-Factory

v0.2.0: Web UI Refactor, LongLoRA

15 Oct 13:06
Compare
Choose a tag to compare

New features

  • Support LongLoRA for the LLaMA models
  • Support training the Qwen-14B and InternLM-20B models
  • Support training state recovery for the all-in-one Web UI
  • Support Ascend NPU by @statelesshz in #975
  • Integrate MMLU, C-Eval and CMMLU benchmarks

Modifications

  • Rename repository to LLaMA Factory (former LLaMA Efficient Tuning)
  • Use the cutoff_len argument instead of max_source_length and max_target_length #944
  • Add a train_on_prompt option #1184

Bug fix

[1] huggingface/transformers#25598 (comment)

v0.1.8: FlashAttention-2 and Baichuan2

11 Sep 09:55
Compare
Choose a tag to compare

New features

  • Support FlashAttention-2 for LLaMA models. (RTX4090, A100, A800 or H100 GPU is required)
  • Support training the Baichuan2 models
  • Use right-padding to avoid overflow in fp16 training (also mentioned here)
  • Align the computation method of the reward score with DeepSpeed-Chat (better generation)
  • Support --lora_target all argument which automatically finds the applicable modules for LoRA training

Bug fix

v0.1.7: Script Preview and RoPE Scaling

18 Aug 09:39
Compare
Choose a tag to compare

New features

  • Preview training script in Web UI by @codemayq in #479 #511
  • Support resuming from checkpoints by @niuba in #434 (transformers>=4.31.0 required)
  • Two RoPE scaling methods: linear and NTK-aware scaling for LLaMA models (transformers>=4.31.0 required)
  • Support training the ChatGLM2-6B model
  • Support PPO training in bfloat16 data type #551

Bug fix

v0.1.6: DPO Training and Qwen-7B

11 Aug 15:43
Compare
Choose a tag to compare

v0.1.5: Patch release

02 Aug 08:13
Compare
Choose a tag to compare

v0.1.4: Dataset Streaming

01 Aug 04:20
Compare
Choose a tag to compare

v0.1.3: Patch release

21 Jul 08:49
Compare
Choose a tag to compare
release v0.1.3

v0.1.2: LLaMA-2 Models

20 Jul 14:42
Compare
Choose a tag to compare
  • Support LLaMA-2 (good issue #202 )
  • Advanced configurations in Web UI
  • Fix API (downgrade pydantic<2.0.0)
  • Fix baichuan lora hparam #194 #212
  • Fix padding #196
  • Fix ZeRO-3 #199
  • Allow pass args to app #213
  • Code simplification
  • Add ShareGPT dataset

v0.1.1

18 Jul 13:05
Compare
Choose a tag to compare
  • Web UI: source_prefix, max_length, dev set
  • Bug fix: reward token #179
  • Update template #171 #177
  • Bug fix: replace the Literal type with Enum for pydantic [1] #176
  • Add Web demo #180

[1] pydantic/pydantic#5821, fastapi/sqlmodel#67

v0.1.0: All-in-one Web UI

17 Jul 16:35
Compare
Choose a tag to compare