Skip to content

PyTorch implementation of Infini-Transformer from "Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention" (https://arxiv.org/abs/2404.07143)

License

Notifications You must be signed in to change notification settings

xiechengmude/infini-transformer

 
 

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

28 Commits
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Infini-Transformer

Overview

Infini-Transformer (https://arxiv.org/abs/2404.07143) is a powerful and versatile transformer model designed for a wide range of natural language processing tasks. It leverages state-of-the-art techniques and architectures to achieve exceptional performance and scalability to infinite context lengths.

Features

  • Scalable architecture for handling long sequences
  • Large-scale pre-training on diverse datasets
  • Support for multiple downstream tasks, including text classification, question answering, and language generation
  • Efficient fine-tuning for task-specific adaptation
  • Includes a Mixture-of-Depths transformer layer that incorporates Infini-Attention (https://arxiv.org/abs/2404.02258)

Getting Started

To get started with Infini-Transformer:

  • Clone the repository:

        git clone https://github.com/dingo-actual/infini-transformer.git
  • Install it from source:

        pip install git+https://github.com/dingo-actual/infini-transformer.git

License

This project is licensed under the MIT License.

Acknowledgments

We would like to thank the researchers and developers whose work has inspired and contributed to the development of Infini-Transformer and Mixture-of-Depths Transformer.

Also, we'd like to give special thanks to all the contributors, collaborators and people who have given feedback. Your efforts have made what was a rough outline of an implementation into something actually usable.

If you have any questions or need further assistance, please feel free to reach out to me at [email protected].

About

PyTorch implementation of Infini-Transformer from "Leave No Context Behind: Efficient Infinite Context Transformers with Infini-attention" (https://arxiv.org/abs/2404.07143)

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%