-
Notifications
You must be signed in to change notification settings - Fork 484
Issues: pytorch-labs/gpt-fast
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
getting different acceptance prob when using
torch.compile
after making a small change.
#184
opened Jun 22, 2024 by
kalradivyanshu
Question about the ENABLE_INTRA_NODE_COMM for speculative decoding
#183
opened Jun 22, 2024 by
jianc99
Hard-coded Llama-3 model name pattern matching breaks scripts/convert_hf_checkpoint.py
#177
opened May 31, 2024 by
ephremw
CUDA error if enabling compile_prefill for quantization model (int8)
#137
opened Mar 14, 2024 by
yanboliang
Reducing Latency in Application with Torch Compilation: Initialization and Inference Optimization
#127
opened Mar 8, 2024 by
daniyal214
Try Tensor Parallel on a server equipped with two V100 linked by NVLINK, but got a performance degradation
#111
opened Feb 27, 2024 by
duanzhaol
I try to speed up with llava,but this it slower then eager mode,why?
#92
opened Jan 31, 2024 by
bleedingfight
Previous Next
ProTip!
Mix and match filters to narrow down what you’re looking for.