Open
Description
Describe the bug
RuntimeError: Error(s) in loading state_dict for EmbeddingPipe: size mismatch for word_embeddings.weight: copying a param with shape torch.Size([25216, 6144]) from checkpoint, the shape in current model is torch.Size([50304, 6144]).
To Reproduce
- Download Slim weights
- Update the vocabulary and checkpoint path in
./configs/20B.yml
(HFTokenizer is used) - Run:
./deepy.py generate.py ./configs/20B.yml -i prompt.txt -o sample_outputs.txt
Environment (please complete the following information):
- GPUs: 2x RTX8000 (48G)