Skip to content
#

gpt-2

Here are 736 public repositories matching this topic...

This project uses GPT-2 to generate realistic movie reviews from the IMDb dataset. By preprocessing data and fine-tuning the model, we achieved human-like text quality. The model's reviews were evaluated for coherence and diversity, showcasing GPT-2's potential in automated text generation.

  • Updated Jun 1, 2024
  • Jupyter Notebook

RWKV is an RNN with transformer-level LLM performance. It can be directly trained like a GPT (parallelizable). So it's combining the best of RNN and transformer - great performance, fast inference, saves VRAM, fast training, "infinite" ctx_len, and free sentence embedding.

  • Updated May 31, 2024
  • Python

Improve this page

Add a description, image, and links to the gpt-2 topic page so that developers can more easily learn about it.

Curate this topic

Add this topic to your repo

To associate your repository with the gpt-2 topic, visit your repo's landing page and select "manage topics."

Learn more