File size: 591 Bytes
8a737b0 db5608b f610dd6 db5608b f610dd6 db5608b |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 |
[PolyCoder](https://github.com/VHellendoorn/Code-LMs) uses GPT2 architecture, with BPE tokenizer trained on a random 5% subset of the data (all languages), and a context length of 2048. To study the effect of scaling of model size, the odel was trained in 3 different sizes. <div align="center"> |Model | # parameters | | - | - | | GPT2 | 160M | | GPT2 | 400M | | GPT2 | 2.7B | </div> PolyCoder is currently being integrated in `transformers`. Meanwhile it can be loaded following the instructions in the original Github [repo](https://github.com/vhellendoorn/code-lms#models). |