-
Notifications
You must be signed in to change notification settings - Fork 1k
Issues: EleutherAI/gpt-neox
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
loss stuck in overflow for RPE position embedding together with sparse attention
bug
Something isn't working
#292
opened May 4, 2021 by
sweinbach
updated May 4, 2021
Add shampoo optimizer
feature request
New feature or request
#314
opened May 10, 2021 by
sdtblck
updated Oct 18, 2021
ONNX Export / Inference Engine
feature request
New feature or request
#537
opened Feb 10, 2022 by
Mistobaan
updated Feb 10, 2022
2 tasks
Hosted Github Runners for CI
feature request
New feature or request
#531
opened Feb 9, 2022 by
Mistobaan
updated Feb 19, 2022
2 tasks
Fine-tuning GPT-NeoX doesn't work (for many scenarios) with the 16-bit stage-0 optimizer
bug
Something isn't working
#568
opened Feb 20, 2022 by
igor0
updated Feb 21, 2022
Increase Documentation Coverage
feature request
New feature or request
#458
opened Nov 7, 2021 by
sdtblck
updated Mar 23, 2022
OOM issues running inference with large contexts on 2x3090 system
bug
Something isn't working
#631
opened Jun 9, 2022 by
fpgaminer
updated Jun 9, 2022
getting error saying __init__ () got an unexpected keyword argument 'checkpointable_layers' when i started training
bug
Something isn't working
#632
opened Jun 13, 2022 by
whoislimshady
updated Sep 18, 2022
Training speed in bf16 mode is slow.
bug
Something isn't working
#660
opened Aug 29, 2022 by
frankang
updated Oct 3, 2022
Interactive inference fails
bug
Something isn't working
#699
opened Oct 1, 2022 by
tonsOfStu
updated Oct 7, 2022
Package Megatron
feature request
New feature or request
#708
opened Oct 26, 2022 by
cat-state
updated Oct 26, 2022
CUDA Out of Memory for 20B Model on 2 A100 40GB GPUs
#606
opened Apr 8, 2022 by
seeEssex
updated Nov 11, 2022
In interactive mode prompt length more than one word causes to crash
bug
Something isn't working
deprioritized
Issues that are not closed, but are low priority and unlikely to be solved soon
#758
opened Dec 27, 2022 by
ahmedavid
updated Jan 15, 2023
Implement Prefix-LM attention masking
feature request
New feature or request
#805
opened Mar 1, 2023 by
TokyoExpress
updated Mar 1, 2023
Unable to load model checkpoint with model parallelism
feature request
New feature or request
#773
opened Jan 20, 2023 by
RaoNikitha
updated Mar 14, 2023
Multi-node training without shared memory
deprioritized
Issues that are not closed, but are low priority and unlikely to be solved soon
feature request
New feature or request
#765
opened Jan 6, 2023 by
VHellendoorn
updated Apr 3, 2023
OOM error when training on a 220G Memory machine with 8 V100.
feature request
New feature or request
#867
opened Apr 2, 2023 by
SefaZeng
updated Apr 3, 2023
Cannot load the checkpoint
bug
Something isn't working
#782
opened Feb 6, 2023 by
jmlongriver12
updated Apr 6, 2023
Finetuning loss explode when not loading deepspeed zero optimal states
bug
Something isn't working
#843
opened Mar 19, 2023 by
sxthunder
updated Apr 12, 2023
Can't finetune 20B model from slim weights with zero optimizer enabled
bug
Something isn't working
#926
opened May 5, 2023 by
coreystatendet
updated May 11, 2023
Add support for pytorch 2.0 ?
deprioritized
Issues that are not closed, but are low priority and unlikely to be solved soon
feature request
New feature or request
#858
opened Mar 27, 2023 by
guozhiyao
updated May 11, 2023
Migrate tensor parallelism code to use OSLO
feature request
New feature or request
oslo
issues relating to refactoring NeoX to use OSLO
#578
opened Mar 1, 2022 by
sdtblck
updated May 18, 2023
3 tasks
Introduce improvements from OSLO
feature request
New feature or request
#571
opened Feb 23, 2022 by
hyunwoongko
updated May 20, 2023
Previous Next
ProTip!
Exclude everything labeled
bug
with -label:bug.