Skip to content
Permalink

Comparing changes

Choose two branches to see what’s changed or to start a new pull request. If you need to, you can also or learn more about diff comparisons.

Open a pull request

Create a new pull request by comparing changes across two branches. If you need to, you can also . Learn more about diff comparisons here.
base repository: InternLM/xtuner Loading
base: v0.1.20
Choose a base ref
...
head repository: InternLM/xtuner Loading
compare: v0.1.21
Choose a head ref
  • 5 commits
  • 41 files changed
  • 3 contributors

Commits on Jun 13, 2024

  1. [Feature] Support DPO, ORPO and Reward Model (#743)

    * Support reward model and dpo
    
    * support train reward model
    
    * fix config
    
    * fix lint
    
    * fix lint
    
    * support jsonl dataset
    
    * feat: support ORPO
    
    * reorg configs
    
    * rename collate function
    
    * rename collate function
    
    * use varlen attention in validation
    
    * fix lint
    
    * fix lint
    
    * rebase main
    
    * update
    
    * add reference and update dpo loss
    
    * inherit sft
    
    * fix broadcast
    
    * fix nan loss skip
    
    * support reward model sp
    
    * support dpo sp
    
    * support orpo sp
    
    * fix bugs
    
    * fix rebase
    
    * convert script
    
    * fix precommit
    
    * mv convert script to model
    
    * fix version check
    
    * fix import
    
    * add comments of reward token
    
    * fix orpo cfg
    
    * fix lint
    
    * fix lint
    
    * remove seed
    
    * remove seed
    
    * add sp config
    
    * add reward sp config
    
    * fix convert
    
    * fix lora reward model convert
    
    * fix qlora reward merge
    
    * update dpo loss
    
    * log reward acc and margin in dpo
    
    * update logits mask
    
    * unpack logits first
    
    * more loss setting in dpo cfgs
    
    * more loss setting in orpo cfgs
    
    ---------
    
    Co-authored-by: HIT-cwh <[email protected]>
    RangiLyu and HIT-cwh committed Jun 13, 2024
    Configuration menu
    Copy the full SHA
    a607fa3 View commit details
    Browse the repository at this point in the history

Commits on Jun 17, 2024

  1. [Bugs] fix dispatch bugs (#775)

    fix dispatch bugs
    HIT-cwh committed Jun 17, 2024
    Configuration menu
    Copy the full SHA
    c2328a0 View commit details
    Browse the repository at this point in the history
  2. [Bugs] Fix HFCheckpointHook bugs when training deepseekv2 and mixtral…

    … withou… (#774)
    
    fix HFCheckpointHook bugs when training deepseekv2 and mixtral without shard moe
    HIT-cwh committed Jun 17, 2024
    Configuration menu
    Copy the full SHA
    bddf85d View commit details
    Browse the repository at this point in the history
  3. [Feature] Support the scenario where sp size is not divisible by attn…

    … head num (#769)
    
    * Support the scenario where sp size is not divisible by attn head num
    
    * refactor attention.py
    
    * do not have to set sp_inner_size in config
    
    * rename
    
    * fix lint
    HIT-cwh committed Jun 17, 2024
    Configuration menu
    Copy the full SHA
    7646e7b View commit details
    Browse the repository at this point in the history
  4. bump version to 0.1.21 (#776)

    HIT-cwh committed Jun 17, 2024
    Configuration menu
    Copy the full SHA
    6bbc274 View commit details
    Browse the repository at this point in the history
Loading