Register
Login
Resources
Docs Blog Datasets Glossary Case Studies Tutorials & Webinars
Product
Data Engine LLMs Platform Enterprise
Pricing Explore
Connect to our Discord channel
Type:  model Data Domain:  nlp
Jeff Wu 9d1e704fec
slight fix to batch size description
5 years ago
a3aa7de9e1
add conditional samples with default settings
5 years ago
src
9d1e704fec
slight fix to batch size description
5 years ago
68bf7a0036
add .gitattributes file to ensure files copied to docker container have LF line endings and all files stay unix on commit
5 years ago
c2dae27c10
First commit
5 years ago
99af6d7092
Add a Dockerfile and document usage in README
5 years ago
99af6d7092
Add a Dockerfile and document usage in README
5 years ago
aae26abd60
add license
5 years ago
ed49f03778
Add documentation for help flags (#81)
5 years ago
0aad2ab3f4
Fetch model using curl, add shebang to download_files.sh and mark it executable
5 years ago
825aa3d338
separate out tensorflow install
5 years ago
Storage Buckets

README.md

You have to be logged in to leave a comment. Sign In

gpt-2

Code and samples from the paper "Language Models are Unsupervised Multitask Learners".

For now, we have only released a smaller (117M parameter) version of GPT-2.

See more details in our blog post.

Installation

Git clone this repository, and cd into directory for remaining commands

git clone https://github.com/openai/gpt-2.git && cd gpt-2

Then, follow instructions for either native or Docker installation.

Native Installation

Download the model data

sh download_model.sh 117M

The remaining steps can optionally be done in a virtual environment using tools such as virtualenv or conda.

Install tensorflow 1.12 (with GPU support, if you have a GPU and want everything to run faster)

pip3 install tensorflow==1.12.0

or

pip3 install tensorflow-gpu==1.12.0

Install other python packages:

pip3 install -r requirements.txt

Docker Installation

Build the Dockerfile and tag the created image as gpt-2:

docker build --tag gpt-2 -f Dockerfile.gpu . # or Dockerfile.cpu

Start an interactive bash session from the gpt-2 docker image.

You can opt to use the --runtime=nvidia flag if you have access to a NVIDIA GPU and a valid install of nvidia-docker 2.0.

docker run --runtime=nvidia -it gpt-2 bash

Usage

WARNING: Samples are unfiltered and may contain offensive content.

Some of the examples below may include Unicode text characters. Set the environment variable:

export PYTHONIOENCODING=UTF-8

to override the standard stream settings in UTF-8 mode.

Unconditional sample generation

To generate unconditional samples from the small model:

python3 src/generate_unconditional_samples.py | tee /tmp/samples

There are various flags for controlling the samples:

python3 src/generate_unconditional_samples.py --top_k 40 --temperature 0.7 | tee /tmp/samples

To check flag descriptions, use:

python3 src/generate_unconditional_samples.py -- --help

Conditional sample generation

To give the model custom prompts, you can use:

python3 src/interactive_conditional_samples.py --top_k 40

To check flag descriptions, use:

python3 src/interactive_conditional_samples.py -- --help

GPT-2 samples

WARNING: Samples are unfiltered and may contain offensive content.

While we have not yet released GPT-2 itself, you can see some samples from it in the gpt-2-samples folder. We show unconditional samples with default settings (temperature 1 and no truncation), with temperature 0.7, and with truncation with top_k 40. We show conditional samples, with contexts drawn from WebText's test set, with default settings (temperature 1 and no truncation), with temperature 0.7, and with truncation with top_k 40.

Future work

We may release code for evaluating the models on various benchmarks.

We are still considering release of the larger models.

Tip!

Press p or to see the previous file or, n or to see the next file

About

This is the DAGsHub mirror of GPT-2 made by OpenAI.

Code for the paper "Language Models are Unsupervised Multitask Learners"

https://openai.com/blog/better-language-models/
Collaborators 1

Comments

Loading...