Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Update branch #38

Merged
merged 62 commits into from
Aug 4, 2023
Merged

Update branch #38

merged 62 commits into from
Aug 4, 2023

Conversation

kshitijkg
Copy link
Member

No description provided.

StellaAthena and others added 30 commits April 21, 2023 20:07
* llama

* spm tokenizer

* pipeline

* llama to neox conversion script

* llama checkin

* weights script update and pp reversion

* revert for PR

* configs

* 7B-specific tweak

* LLaMA updates

* PR feedback

* initialize multiple_of

---------

Co-authored-by: Quentin-Anthony <[email protected]>
)

* [bug-fix] enable finetuning option(set optimizer params correctly)

* change load_checkpoint

---------

Co-authored-by: logan.eo <[email protected]>
* fix list[tensor] typing in both scripts

* Update NeoXArgs docs automatically

* add bf16 saving to conversion scripts

* make precision check more complex for v1.0

* Update NeoXArgs docs automatically

* Update NeoXArgs docs automatically

---------

Co-authored-by: haileyschoelkopf <[email protected]>
Co-authored-by: github-actions <[email protected]>
Co-authored-by: Hailey Schoelkopf <[email protected]>
* add bf16 configuration

Signed-off-by: Dashiell Stander <[email protected]>

* Update NeoXArgs docs automatically

* pre commit

Signed-off-by: Dashiell Stander <[email protected]>

* Update NeoXArgs docs automatically

* Rework deriving precision

Signed-off-by: Dashiell Stander <[email protected]>

* Update NeoXArgs docs automatically

* Belt and suspenders

Signed-off-by: Dashiell Stander <[email protected]>

* Update NeoXArgs docs automatically

* Make the default setup (of only using fp16 dict) work

Signed-off-by: Dashiell Stander <[email protected]>

* Update NeoXArgs docs automatically

* Got rid of bf16 argument

Signed-off-by: Dashiell Stander <[email protected]>

* Update NeoXArgs docs automatically

* Update NeoXArgs docs automatically

* Update NeoXArgs docs automatically

* Re-add detailed bf16 message

* Update NeoXArgs docs automatically

* Remove unused import

* Update NeoXArgs docs automatically

* remove useless newline

* Update NeoXArgs docs automatically

* re-add detailed bf16 message to deepspeed_args

* Update NeoXArgs docs automatically

---------

Signed-off-by: Dashiell Stander <[email protected]>
Co-authored-by: github-actions <[email protected]>
Co-authored-by: Quentin Anthony <[email protected]>
* update torch and cuda

* Update NeoXArgs docs automatically

* Update NeoXArgs docs automatically

---------

Co-authored-by: github-actions <[email protected]>
Co-authored-by: Quentin Anthony <[email protected]>
Remove duplicate deepspeed config and allow forced multinode
* Pre-commit

Signed-off-by: Dashiell Stander <[email protected]>

* Update NeoXArgs docs automatically

* Do not check for overflow if not using fp16

Signed-off-by: Dashiell Stander <[email protected]>

* Update NeoXArgs docs automatically

---------

Signed-off-by: Dashiell Stander <[email protected]>
Co-authored-by: github-actions <[email protected]>
xu-song and others added 29 commits May 27, 2023 12:30
* added a simple script for multi-node data preparation.

* added a simple script for multi-node data preparation.

* fixed minor bugs regarding prefixing of the .bin and .idx files

* fixed minor bugs regarding prefixing of the .bin and .idx files

* fixed minor bugs regarding prefixing of the .bin and .idx files
…heck (#959)

* update conversion script instructions in readme

* rename v1.0 script (now default for 2.0) to module_to_hf

* Update NeoXArgs docs automatically

---------

Co-authored-by: github-actions <[email protected]>
* added HF to NeoX 2.0 conversion script with mp and pp sharding

* (1) added missing curly brace to pythial/1-4B config; (2) fixed a bug related to a hardcoded value withing the conversion script (3) fixed possible bugs in the conversion script wrt the mp sharding convention

* fill in minimal possible mask values

* initialize tensor on the target device

---------

Co-authored-by: Quentin Anthony <[email protected]>
* added HF to NeoX 2.0 conversion script with mp and pp sharding

* (1) added missing curly brace to pythial/1-4B config; (2) fixed a bug related to a hardcoded value withing the conversion script (3) fixed possible bugs in the conversion script wrt the mp sharding convention

* added GeLU fast for HF model, added barriers to enable conversion across multiple nodes, removed partially hardcoded pythia model name

* commented unecessary logging and timers

---------

Co-authored-by: Quentin Anthony <[email protected]>
* add an optional `label` field passed in parallel with training data.

* minor fix; Add doc

* fix

* fix data can be None

* prevent loading optimizer

* add script

* Remove some print() stmts, make mask documentation clearer

* Add documentation for preprocess_data_with_mask.py

---------

Co-authored-by: Hailey Schoelkopf <[email protected]>
* fix tensorboard version

* add setup wandb tensorboard
* Update gpt2_dataset.py

* Update NeoXArgs docs automatically

---------

Co-authored-by: github-actions <[email protected]>
For BPE tokenization, `tokenizer_type` argument should be a string instead of the list.
default tokenizer_type should be string
Replaced all torch.concat with torch.cat
@kshitijkg kshitijkg merged commit 6c1f7bc into CERC-AAI:main Aug 4, 2023
0 of 3 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

None yet