Skip to content

Commit

Permalink
Update NeoXArgs docs automatically
Browse files Browse the repository at this point in the history
  • Loading branch information
github-actions committed Feb 27, 2024
1 parent 7483246 commit 497485c
Showing 1 changed file with 15 additions and 25 deletions.
40 changes: 15 additions & 25 deletions configs/neox_arguments.md
Original file line number Diff line number Diff line change
Expand Up @@ -7,7 +7,7 @@ LR Scheduler Arguments



- **lr_decay_style**: Literal
- **lr_decay_style**: typing.Literal['constant', 'linear', 'cosine', 'exponential']

Default = linear

Expand Down Expand Up @@ -111,7 +111,7 @@ Logging Arguments

- **git_hash**: str

Default = 6bb4e62
Default = 7483246

current git hash of repository

Expand Down Expand Up @@ -253,7 +253,7 @@ Model Arguments



- **precision**: Literal
- **precision**: typing.Literal['fp16', 'fp32', 'bfloat16']

Default = None

Expand Down Expand Up @@ -288,16 +288,6 @@ Model Arguments



- **intermediate_size**: int

Default = None

Transformer intermediate size. Currently only used for "mlp_type": "llama".

If not passed, will be set to a reasonable default.



- **num_attention_heads**: int

Default = None
Expand Down Expand Up @@ -346,7 +336,7 @@ Model Arguments



- **norm**: Literal
- **norm**: typing.Literal['layernorm', 'rmsnorm', 'scalenorm']

Default = layernorm

Expand Down Expand Up @@ -394,7 +384,7 @@ Model Arguments



- **pos_emb**: Literal
- **pos_emb**: typing.Literal['learned', 'rotary', 'sinusoidal', 'rpe', 'alibi', 'none']

Default = learned

Expand Down Expand Up @@ -508,7 +498,7 @@ Model Arguments



- **activation**: Literal
- **activation**: typing.Literal['gelu', 'geglu', 'relu', 'softsign', 'swish', 'mish', 'silu']

Default = gelu

Expand Down Expand Up @@ -613,7 +603,7 @@ Model Arguments



- **init_method**: Literal
- **init_method**: typing.Literal['normal', 'scaled_normal', 'orthogonal', 'scaled_orthogonal', 'xavier_uniform', 'xavier_normal', 'wang_init', 'small_init']

Default = normal

Expand All @@ -622,7 +612,7 @@ Model Arguments



- **output_layer_init_method**: Literal
- **output_layer_init_method**: typing.Literal['normal', 'scaled_normal', 'orthogonal', 'scaled_orthogonal', 'xavier_uniform', 'xavier_normal', 'wang_init', 'small_init']

Default = scaled_normal

Expand Down Expand Up @@ -705,7 +695,7 @@ Model Arguments



- **output_layer_parallelism**: Literal
- **output_layer_parallelism**: typing.Literal['column']

Default = column

Expand All @@ -719,7 +709,7 @@ Optimizer Arguments



- **optimizer_type**: Literal
- **optimizer_type**: typing.Literal['adam', 'onebitadam', 'cpu_adam', 'cpu_torch_adam', 'sm3', 'madgrad_wd', 'sgd', 'lion']

Default = adam

Expand All @@ -736,7 +726,7 @@ Optimizer Arguments



- **zero_stage**: Union
- **zero_stage**: typing.Union[int, typing.List[int], typing.Literal['all']]

Default = None

Expand Down Expand Up @@ -1135,7 +1125,7 @@ Tokenizer Arguments



- **tokenizer_type**: Literal
- **tokenizer_type**: typing.Literal['GPT2BPETokenizer', 'HFTokenizer', 'HFGPT2Tokenizer', 'SPMTokenizer', 'CharLevelTokenizer', 'TiktokenTokenizer']

Default = GPT2BPETokenizer

Expand Down Expand Up @@ -1266,7 +1256,7 @@ Training Arguments



- **data_impl**: Literal
- **data_impl**: typing.Literal['infer', 'mmap', 'cached']

Default = infer

Expand Down Expand Up @@ -1330,7 +1320,7 @@ Training Arguments



- **checkpoint_scale**: Literal
- **checkpoint_scale**: typing.Literal['linear', 'log']

Default = linear

Expand Down Expand Up @@ -2088,7 +2078,7 @@ Args for deepspeed runner (deepspeed.launcher.runner).



- **launcher**: Literal
- **launcher**: typing.Literal['pdsh', 'openmpi', 'mvapich', 'slurm']

Default = pdsh

Expand Down

0 comments on commit 497485c

Please sign in to comment.