Skip to content

ylwangy/slstm_pytorch

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

74 Commits
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Pytorch version of sentence-state LSTM and how to use it for masked language model pre-training.

Sentence-State LSTM for Text Representation

Pre-Training a Graph Recurrent Network for Language Representation

Requirements and Installation

  • fairseq
  • transformers
  • sentencepiece

Step

  1. we follow fairseq for pre-training language model

  2. put files in models to fairseq/fairseq/models

  3. pip install .

  4. preprocess dataset and run train.sh for lm pre-training

Checkpoints and Usage

See corresponding directories

Citation

  @article{wang2022pre,
    title={Pre-Training a Graph Recurrent Network for Language Representation},
    author={Wang, Yile and Yang, Linyi and Teng, Zhiyang and Zhou, Ming and Zhang, Yue},
    journal={arXiv preprint arXiv:2209.03834},
    year={2022}
  }

Contact

If you have any question, please create a issue or contact to the authors.

About

Pytorch version of sentence-state LSTM.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published