-
Notifications
You must be signed in to change notification settings - Fork 834
Open
Description
This document includes the features in LMFlow's roadmap. We welcome any discuss or contribute to the specific features at related Issues/PRs. 🤗
Main Features
- Data
- DPO dataset format [Feature] reward model inferencer and dpov2 aligner #867
- Conversation template in DPO [Feature] Iterative DPO #883
- jinja template [usability] support qwen2.5 and deepseek #931
- Tools in conversation dataset function-call-finetune #884 Modify the formatter of function and observation #892 [usability] support qwen2.5 and deepseek #931
- Packing with block diagonal attention
- Add a tokenize-only script, allowing tokenization separately without gpu environment. (For those who has large datasets but limited gpu hrs)
- Model
- Backend
- 🏗️ Accelerate support [usability] Accelerate Support #936
- Tokenization
- Tokenization update, using hf method [usability] support qwen2.5 and deepseek #931
- Backend
- Pipeline
- Train/Finetune/Align
- DPO (multi-gpu) [Feature] reward model inferencer and dpov2 aligner #867
- Iterative DPO [Feature] reward model inferencer and dpov2 aligner #867 [Feature] Iterative DPO #883
- PPO
- LISA (multi-gpu, qwen2, chatglm)
- Batch size and learning rate recommendation (arxiv)
- No trainer version pipelines, allowing users to customize/modify based on their needs
- Sparse training for moe models [New Feature] Is Mixtral supported? #879
- Inference
- vllm inference [Feature] vllm inferencer and memory safe vllm inferencer #860 [Feature] Add vllm inference example #863
- Reward model scoring [Feature] reward model inferencer and dpov2 aligner #867
- Multiple instances inference (vllm, rm, others) [Feature] Iterative DPO #883
- Inference checkpointing and resume from checkpoints
- Inference accelerate EAGLE
- Inferencer for chat/instruction models, and
chatbot.pyupgrade Conversation_template #917
- Train/Finetune/Align
Usability
- Make some packages/functions (gradio, vllm, ray, etc.) optional, add conditional import. [usability] deps streamlining #905
- Inference method auto-downgrading (vllm>ds, etc.), and make
vllmpackage optional. [usability] deps streamlining #905 - Merging similar model methods into
hf_model_mixin - Set
torch_dtype='bfloat16'whenbf16is specified, etc. (bf16is inFinetunerArgumentsbuttorch_dtypeis inModelArguments, thus cannot handle in__post_init__(). )
Bug fixes
-
model.generate()with dsz3 [BUG] The text cannot be generated successfully during the Raft step #861 -
merge_loralora with abs path merging -
load_datasetlong data fix [Bug Fix] update load_dataset to support long data #878 - src/lmflow/utils/common.py
create_copied_dataclasscompatibility when python version >= 3.10 (kw_onlyissue) [BUG]TypeError: Field.__init__() missing 1 required positional argument: 'kw_only' #903 [usability] deps streamlining #905
Issues left over from history
-
use_accelerator->use_acceleratetypo fix (with Accelerate support PR) [usability] Accelerate Support #936 -
model_args.use_loraleads to truncation of the sequence, mentioned in [Feature] reward model inferencer and dpov2 aligner #867 - Make ports, addresses, and all other settings in distributed training tidy and clear (with Accelerate support PR)
Documentation
- Approx GPU memory requirement w.r.t model size & pipeline
- Dev handbook, indicating styles, test list, etc.
research4pan and shizhediao
Metadata
Metadata
Assignees
Labels
No labels