From b0a70959f4b2276f248630adaa823c8bc2d27268 Mon Sep 17 00:00:00 2001 From: Pavel Grunt Date: Fri, 30 Sep 2022 17:20:05 +0200 Subject: [PATCH] Fix link to download 20B vocab Follow up to commit 0709327633973cfc656f5a12e3435cfb508e90e7 --- README.md | 2 +- 1 file changed, 1 insertion(+), 1 deletion(-) diff --git a/README.md b/README.md index 2039a9427..49374100c 100644 --- a/README.md +++ b/README.md @@ -207,7 +207,7 @@ Next make sure to download the GPT2 tokenizer vocab, and merge files from the fo Or use the 20B tokenizer (for which only a single Vocab file is needed): -- Vocab: https://mystic.the-eye.eu/public/AI/models/GPT-NeoX-20B/slim_weights/20B_tokenizer.json +- Vocab: https://the-eye.eu/public/AI/models/GPT-NeoX-20B/slim_weights/20B_tokenizer.json (alternatively, you can provide any tokenizer file that can be loaded by Huggingface's tokenizers library with the `Tokenizer.from_pretrained()` command)