You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
-[ ] add ability to gate in memorizing transformers knn attention layers
-[ ] figure out best way to cache the causal mask across all attention layers, but also reach out to pytorch team about some of the limitations for causal in q seq len != kv seq len scenario