Register
Login
Resources
Docs Blog Datasets Glossary Case Studies Tutorials & Webinars
Product
Data Engine LLMs Platform Enterprise
Pricing Explore
Connect to our Discord channel
Type:  model Data Domain:  nlp
src
c2dae27c10
First commit
5 years ago
c2dae27c10
First commit
5 years ago
c2dae27c10
First commit
5 years ago
c2dae27c10
First commit
5 years ago
Storage Buckets

README.md

You have to be logged in to leave a comment. Sign In

gpt-2

Code and samples from the paper "Language Models are Unsupervised Multitask Learners"

Installation

Download the model data:

gsutil rsync -r gs://gpt-2/models/ models/

Install python packages:

pip install -r requirements.txt

Sample generation

WARNING: Samples are unfiltered and may contain offensive content.

To generate unconditional samples from the small model:

python3 src/main.py | tee samples

There are various flags for controlling the samples:

python3 src/main.py --top_k 40 --temperature 0.7 | tee samples
Tip!

Press p or to see the previous file or, n or to see the next file

About

This is the DAGsHub mirror of GPT-2 made by OpenAI.

Code for the paper "Language Models are Unsupervised Multitask Learners"

https://openai.com/blog/better-language-models/
Collaborators 1

Comments

Loading...