There are no reviews yet. Be the first to send feedback to the community and the maintainers!
Repository Details
GPT-α is a 124 million parameter decoder-only transformer based language model following the architecture of GPT-2 and training process of GPT-3. The model achieves state-of-the-art results for a model of this size.