File size: 877 Bytes
d07e8fd 8d1dab4 21bc0ed 81c2812 2ebb1b1 81c2812 01abf9c 8d1dab4 3595599 8d1dab4 3f77f51 3595599 3f77f51 3595599 3f77f51 3595599 3f77f51 3595599 93f98a7 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 |
---
license: apache-2.0
datasets:
- Fishfishfishfishfish/Synthetic_text.txt
language:
- en
---
the only files needed for inference is inference.py, word2idk.pkl, and lstm_Hxxx.safetensors
input tokens must be space separated, as they aren't tokenized like the training data is.
>python inference.py --temp 0.5 --top_k 64 --model_file lstm_H256.safetensors --start_sequence "User : what is the capital of France ? Bot : " --max_length 32
usually results in something like
>The capital of the world of the world of the world of the world of the
its not very accurate yet, its trained on only 1.2mb of text
Each safetensors file represents a different hidden dim value.
Each trained for 1 epoch.
inference.py hidden dim value must be edited for each safetensors.
>sequence_length = 64
>
>batch_size = 16
>
>learning_rate = 0.0001
>
>embedding_dim = 256
>
>num_layers = 4
> |