File size: 757 Bytes
e41b458
 
bbb138b
 
 
 
 
 
 
 
4d402ef
 
 
 
e41b458
0463473
 
 
 
bbb138b
 
1dd285e
bbb138b
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
---
license: apache-2.0
tags:
- alpaca
- gpt4
- gpt-j
- instruction
- finetuning
- lora
- peft
datasets:
- vicgalle/alpaca-gpt4
pipeline_tag: conversational
base_model: EleutherAI/gpt-j-6b
---
GPT-J 6B model was finetuned on GPT-4 generations of the Alpaca prompts on [MonsterAPI](https://monsterapi.ai)'s no-code LLM finetuner, using LoRA for ~ 65,000 steps, auto-optmised to run on 1 A6000 GPU with no out of memory issues and without needing me to write any code or setup a GPU server with libraries to run this experiment. The finetuner does it all for us by itself.

Documentation on no-code LLM finetuner:
https://docs.monsterapi.ai/fine-tune-a-large-language-model-llm


![training loss](trainloss.png "Training loss")




---
license: apache-2.0
---