Skip to content

Commit fbae7db

Browse files
committed
update readmes
1 parent ac5d522 commit fbae7db

File tree

4 files changed

+6
-3
lines changed

4 files changed

+6
-3
lines changed

DEVELOPERS.md

+1
Original file line numberDiff line numberDiff line change
@@ -30,6 +30,7 @@ Download the model data
3030
python3 download_model.py 124M
3131
python3 download_model.py 355M
3232
python3 download_model.py 774M
33+
python3 download_model.py 1558M
3334
```
3435

3536
## Docker Installation

Dockerfile.cpu

+1
Original file line numberDiff line numberDiff line change
@@ -8,3 +8,4 @@ RUN pip3 install -r requirements.txt
88
RUN python3 download_model.py 124M
99
RUN python3 download_model.py 355M
1010
RUN python3 download_model.py 774M
11+
RUN python3 download_model.py 1558M

Dockerfile.gpu

+1
Original file line numberDiff line numberDiff line change
@@ -17,3 +17,4 @@ RUN pip3 install -r requirements.txt
1717
RUN python3 download_model.py 124M
1818
RUN python3 download_model.py 355M
1919
RUN python3 download_model.py 774M
20+
RUN python3 download_model.py 1558M

README.md

+3-3
Original file line numberDiff line numberDiff line change
@@ -2,11 +2,11 @@
22

33
# gpt-2
44

5-
Code from the paper ["Language Models are Unsupervised Multitask Learners"](https://d4mucfpksywv.cloudfront.net/better-language-models/language-models.pdf).
5+
Code and models from the paper ["Language Models are Unsupervised Multitask Learners"](https://d4mucfpksywv.cloudfront.net/better-language-models/language-models.pdf).
66

7-
We have currently released small (124M parameter), medium (355M parameter), and large (774M parameter) versions of GPT-2<sup>*</sup>, with only the full model as of yet unreleased. We have also [released a dataset](https://github.com/openai/gpt-2-output-dataset) for researchers to study their behaviors.
7+
You can read about GPT-2 and its staged release in our [original blog post](https://blog.openai.com/better-language-models/), [6 month follow-up post](https://openai.com/blog/gpt-2-6-month-follow-up/), and [final post](https://www.openai.com/blog/gpt-2-1-5b-release/).
88

9-
You can read about GPT-2 and release decisions in our [original blog post](https://blog.openai.com/better-language-models/) and [6 month follow-up post](https://openai.com/blog/gpt-2-6-month-follow-up/).
9+
We have also [released a dataset](https://github.com/openai/gpt-2-output-dataset) for researchers to study their behaviors.
1010

1111
<sup>*</sup> *Note that our original parameter counts were wrong due to an error (in our previous blog posts and paper). Thus you may have seen small referred to as 117M and medium referred to as 345M.*
1212

0 commit comments

Comments
 (0)