Skip to content
forked from lipiji/Guyu

pre-training and fine-tuning framework for text generation

License

Notifications You must be signed in to change notification settings

JerryX1110/Guyu

 
 

Repository files navigation

Guyu (谷雨)

pre-training and fine-tuning framework for text generation

backbone code for "An Empirical Investigation of Pre-Trained Transformer Language Models for Open-Domain Dialogue Generation": https://arxiv.org/abs/2003.04195

Pre-training:

./prepare_data.sh
./train.sh
./inference.sh

Fine-tuning

Example: chat-bot

cd chat_bot
./fine_tune.sh
./inference.sh

Web Api:

./deploy.sh

Pre-trained models

  • 12-layer, 768-hidden, 12-heads, Chinese (News + zhwiki, 200G) and English (Gigawords + Bookscorpus + enwiki, 60G)

  • 24-layer, 768-hidden, 12-heads, Chinese (News + zhwiki, 200G) and English (Gigawords + Bookscorpus + enwiki, 60G)

  • download them: https://github.com/lipiji/Guyu/tree/master/model

References:

About

pre-training and fine-tuning framework for text generation

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 96.3%
  • Shell 3.7%