Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

handle tokenization more correctly, clean up CI #62

Merged
merged 9 commits into from
May 30, 2024
Prev Previous commit
Next Next commit
25 epochs for all models except pythia
  • Loading branch information
charlesfrye committed May 30, 2024
commit 62e612d8abbf02a65a0cf5aecc8ff8bb6d236058
14 changes: 5 additions & 9 deletions ci/prep_for_ci.py
Original file line number Diff line number Diff line change
Expand Up @@ -18,18 +18,14 @@ def main(config: str, data: str):
cfg["weight_decay"] = 0

val_set_size = 0.5
num_epochs = 50
if (
"Meta-Llama-3-8B" in cfg["base_model"]
or "codellama/CodeLlama-7b-Instruct-hf" in cfg["base_model"]
):
num_epochs = num_epochs // 2
elif "pythia" in cfg["base_model"]:
num_epochs = num_epochs * 2

num_epochs = 25
if "pythia" in cfg["base_model"]:
num_epochs = num_epochs * 4

cfg["val_set_size"] = val_set_size
cfg["num_epochs"] = num_epochs
cfg["eval_steps"] = num_epochs // 10
cfg["eval_steps"] = num_epochs // 5
cfg.pop("evals_per_epoch", None) # incompatible with eval_steps
cfg.pop("sample_packing", False) # requires larger dataset

Expand Down
Loading