mamba_0_5: | |
prompt_template: "zephyr-7b-alpha/prompt.txt" | |
fn_completions: "huggingface_local_completions" | |
completions_kwargs: | |
model_name: "/data/junxiong/sft/zephyr_0_5_dpo_open_not_openhermes_progressive_train_largest_dataset_ep3/" | |
model_kwargs: | |
torch_dtype: 'bfloat16' | |
max_new_tokens: 2048 | |
temperature: 0.7 | |
top_p: 1.0 | |
do_sample: True | |
pretty_name: "Mamba 0 5 From Zephyr 7B Beta" | |
link: "https://huggingface.co/HuggingFaceH4/zephyr-7b-beta" | |