Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Update documentation #392

Merged
merged 32 commits into from
Aug 21, 2021
Merged
Changes from 1 commit
Commits
Show all changes
32 commits
Select commit Hold shift + click to select a range
962eacc
Update README.md
StellaAthena Jun 25, 2021
50ea5cc
Update README.md
StellaAthena Jun 25, 2021
5018594
Update README.md
StellaAthena Jun 25, 2021
7a75cd4
Update README.md
StellaAthena Jun 25, 2021
c221109
Update README.md
StellaAthena Jun 25, 2021
8f43bac
Update README.md
StellaAthena Jun 28, 2021
3783d7f
Update README.md
StellaAthena Jun 28, 2021
f22769e
Update README.md
StellaAthena Jun 28, 2021
0b7d2fe
Update README.md
StellaAthena Jul 7, 2021
5f978c8
add info about installing fused kernels
ShivanshuPurohit Jul 9, 2021
20cadc3
Update README.md
ShivanshuPurohit Jul 9, 2021
2686396
Update README.md
ShivanshuPurohit Jul 9, 2021
94980dd
sparsity + minor typos
ShivanshuPurohit Jul 9, 2021
7d44d97
change path to ssd-1
ShivanshuPurohit Jul 10, 2021
4333716
typo
ShivanshuPurohit Jul 10, 2021
05249ab
Update README.md
StellaAthena Jul 10, 2021
28a830e
Update README.md
StellaAthena Jul 10, 2021
ae00018
Update README.md
StellaAthena Jul 10, 2021
5245c6d
Update README.md
StellaAthena Jul 10, 2021
4c6469e
Update README.md
StellaAthena Jul 10, 2021
c695714
Update README.md
StellaAthena Jul 10, 2021
1cccfd2
Update README.md
StellaAthena Jul 10, 2021
486ed38
Update README.md
StellaAthena Jul 10, 2021
1e97f7d
Update README.md
ShivanshuPurohit Jul 16, 2021
a3d06bc
Update README.md
ShivanshuPurohit Jul 24, 2021
3e4f6d9
Merge pull request #380 from EleutherAI/main
StellaAthena Jul 24, 2021
173dfd4
Update README.md
StellaAthena Jul 27, 2021
91bb070
Update README.md
StellaAthena Jul 27, 2021
ff74c8a
Update README.md
StellaAthena Jul 27, 2021
74a6cdd
Update README.md
StellaAthena Jul 30, 2021
e84a344
Merge pull request #385 from EleutherAI/main
StellaAthena Jul 30, 2021
b6de20b
Update README.md
ShivanshuPurohit Aug 21, 2021
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Prev Previous commit
Next Next commit
Update README.md
  • Loading branch information
StellaAthena authored Jul 27, 2021
commit ff74c8aca62134690d8d752f6c7df376793922e8
6 changes: 3 additions & 3 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -37,13 +37,13 @@ GPT-NeoX is under active development.

## Why GPT-NeoX

**Straightforward configuration:*** Other libraries such as Megatron-LM require you configure them using command line arguments and global variables, which can often be difficult to work with and iterate upon. We offer straightforward configuration using .yaml files, which enables you to launch training runs across 100s of GPUs with a single line bash script. Additionally, we hope to make data preparation easier on the user by providing scripts to automatically download and pretokenize a number of large-scale datasets.
**Straightforward configuration:** Other libraries such as Megatron-LM require you configure them using command line arguments and global variables, which can often be difficult to work with and iterate upon. We offer straightforward configuration using .yaml files, which enables you to launch training runs across 100s of GPUs with a single line bash script. Additionally, we hope to make data preparation easier on the user by providing scripts to automatically download and pretokenize a number of large-scale datasets.

**Diverse Modeling Options:** We provide a wide collections of options for constructing your model.

**HuggingFace Integration:** Our code is designed to work with the HuggingFace `transformers` library. All models trained using this codebase can be uploaded to a custom HuggingFace class with ease.
**HuggingFace Integration:** Our code is designed to work with the HuggingFace `transformers` library. All models trained using this codebase can be uploaded to a custom HuggingFace class with ease, and all HuggingFace tokenizers and datasets can be used to train models.

**Large Pretrained Models:** We offer several large, pretrained models to iterate on
**Large Pretrained Models:** We offer several large, pretrained models to iterate on. For people who are unable to train billion parameter scale models themselves, this framework allows you to easily interact with models that we have released.

## Quick Start

Expand Down