Skip to content

Commit

Permalink
feat: remove need to add load_in* during merge (OpenAccess-AI-Collect…
Browse files Browse the repository at this point in the history
  • Loading branch information
NanoCode012 committed Dec 29, 2023
1 parent dec66d7 commit f6ecf14
Show file tree
Hide file tree
Showing 2 changed files with 11 additions and 3 deletions.
4 changes: 2 additions & 2 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -996,7 +996,7 @@ When you include these tokens in your axolotl config, axolotl adds these tokens
### Inference Playground

Axolotl allows you to load your model in an interactive terminal playground for quick experimentation.
The config file is the same config file used for training.
The config file is the same config file used for training.

Pass the appropriate flag to the inference command, depending upon what kind of model was trained:

Expand Down Expand Up @@ -1027,7 +1027,7 @@ Please use `--sample_packing False` if you have it on and receive the error simi
Add below flag to train command above

```bash
python3 -m axolotl.cli.merge_lora examples/your_config.yml --lora_model_dir="./completed-model" --load_in_8bit=False --load_in_4bit=False
python3 -m axolotl.cli.merge_lora examples/your_config.yml --lora_model_dir="./completed-model"
```

If you run out of CUDA memory, you can try to merge in system RAM with
Expand Down
10 changes: 9 additions & 1 deletion src/axolotl/cli/merge_lora.py
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,15 @@ def do_cli(config: Path = Path("examples/"), **kwargs):
return_remaining_strings=True
)
parsed_cli_args.merge_lora = True
parsed_cfg = load_cfg(config, merge_lora=True, **kwargs)

parsed_cfg = load_cfg(
config,
merge_lora=True,
load_in_8bit=False,
load_in_4bit=False,
flash_attention=False,
**kwargs
)

do_merge_lora(cfg=parsed_cfg, cli_args=parsed_cli_args)

Expand Down

0 comments on commit f6ecf14

Please sign in to comment.