justheuristic
commited on
Commit
·
f23d189
1
Parent(s):
636b67f
Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,8 @@
|
|
|
|
|
|
|
|
|
|
|
|
1 |
### Quantized EleutherAI/gpt-j-6b with 8-bit weights
|
2 |
|
3 |
This is a version of EleutherAI's GPT-J with 6 billion parameters that is modified so you can generate **and fine-tune the model in colab or equivalent desktop gpu (e.g. single 1080Ti)**.
|
|
|
1 |
+
Note: this model was superceded by the [`load_in_8bit=True` feature in transformers]https://github.com/huggingface/transformers/pull/17901)
|
2 |
+
by Younes Belkada and Tim Dettmers. Please see [this usage example](https://colab.research.google.com/drive/1qOjXfQIAULfKvZqwCen8-MoWKGdSatZ4#scrollTo=W8tQtyjp75O).
|
3 |
+
This legacy model was built for [transformers v4.15.0](https://github.com/huggingface/transformers/releases/tag/v4.15.0) and pytorch 1.11. Newer versions could work, but are not supported.
|
4 |
+
|
5 |
+
|
6 |
### Quantized EleutherAI/gpt-j-6b with 8-bit weights
|
7 |
|
8 |
This is a version of EleutherAI's GPT-J with 6 billion parameters that is modified so you can generate **and fine-tune the model in colab or equivalent desktop gpu (e.g. single 1080Ti)**.
|