id
stringlengths 4
117
| downloads
int64 0
18.7M
| likes
int64 0
4.28k
| created_at
stringlengths 25
25
|
---|---|---|---|
zepedrotrigo/llama-2-7b-fortnyce | 2 | 0 | 2023-10-20T01:07:53+00:00 |
zeping/codeparrot | 5 | 0 | 2022-03-02T23:29:05+00:00 |
zer0sh0t/programmer_ai_v2 | 1 | 0 | 2022-03-02T23:29:05+00:00 |
zerouchi/vchem-falcon-7b-pretrained | 0 | 0 | 2023-07-21T12:49:41+00:00 |
zetaqubit/llama-2-7b-miniguanaco | 2 | 0 | 2023-10-04T08:24:06+00:00 |
zetavg/pythia-6.9b | 2 | 0 | 2023-05-30T07:50:37+00:00 |
zetavg/pythia-70m | 1 | 0 | 2023-05-30T10:22:18+00:00 |
zetavg/zh-tw-llm-ta01-pythia-6.9b-ta8000-v1-a_1_embeddings-h100-t01-c5daa1-8bit-2 | 1 | 0 | 2023-05-19T21:57:56+00:00 |
zetavg/zh-tw-llm-ta01-pythia-6.9b-ta8000-v1-a_1_embeddings-h100-t01-c5daa1-f16 | 1 | 0 | 2023-05-19T23:46:41+00:00 |
zetavg/zh-tw-pythia-1b-230508-ckpt-20000 | 0 | 0 | 2023-05-08T20:07:45+00:00 |
zetavg/zh-tw-pythia-1b-230508-ckpt-21000 | 1 | 0 | 2023-05-08T20:12:34+00:00 |
zetavg/zh-tw-pythia-1b-a12k-f84566-embeddings-gcp-a100-trans-t3-d2ad | 1 | 0 | 2023-05-12T13:04:21+00:00 |
zetavg/zh_tw_pythia-1b-2023-05-01-05-12-16 | 1 | 0 | 2023-05-01T05:12:32+00:00 |
zetavg/zh_tw_pythia-2023-05-01-01-08-10 | 2 | 0 | 2023-05-01T01:08:18+00:00 |
zfchen/codeparrot | 6 | 0 | 2022-03-02T23:29:05+00:00 |
zgce/Emerhyst-20B-5bpw-hb6-exl2 | 6 | 1 | 2023-10-20T09:15:52+00:00 |
zgce/Mistral-RP-0.1-7B-3.5bpw-hb6-exl2 | 1 | 0 | 2023-10-21T11:54:27+00:00 |
zgce/Synatra-7B-v0.3-RP-3.5bpw-hb6-exl2 | 4 | 1 | 2023-11-26T00:42:02+00:00 |
zgce/Yi-34B-Chat-Spicyboros-limarpv3-4bpw-hb6-exl2 | 4 | 5 | 2023-11-24T01:34:56+00:00 |
zgce/Yi-34b-200K-alpaca-rpv3-4bpw-hb6-exl2 | 3 | 2 | 2023-11-11T17:04:31+00:00 |
zgce/Yi-34b-200K-alpaca-rpv3-scipy-4bpw-hb6-exl2 | 3 | 2 | 2023-11-12T14:51:43+00:00 |
zgce/Yi-34b-200K-alpaca-rpv3-scipy-6bpw-hb6-exl2 | 2 | 0 | 2023-11-13T03:32:24+00:00 |
zgce/acsr-v2-yi34b-4bpw-hb6-exl2 | 11 | 8 | 2023-11-25T08:00:06+00:00 |
zgce/acsr-y34b-4bpw-hb6-exl2 | 1 | 1 | 2023-11-16T05:12:13+00:00 |
zgce/mythalion-13b-chinese-alpaca2-8bpw-hb8-exl2 | 0 | 0 | 2023-11-04T14:43:03+00:00 |
zgotter/gpt2-test | 5 | 0 | 2022-03-02T23:29:05+00:00 |
zguo0525/vicuna-7b | 2 | 0 | 2023-06-05T22:15:54+00:00 |
zh-tw-llm-dv/sample-pythia-70m-dialogue | 3 | 0 | 2023-05-21T18:20:34+00:00 |
zh-tw-llm-dv/tw-pythia-6.9b-chat-v0_2-s2 | 3 | 0 | 2023-05-26T08:29:09+00:00 |
zh-tw-llm-dv/zh-tw-llm-ta01-pythia-1b-ta8000-v1-a_1_embeddings-a100-t02-3d435e | 4 | 1 | 2023-05-18T17:23:02+00:00 |
zh-tw-llm-dv/zh-tw-llm-ta01-pythia-1b-ta8000-v1-b_1_embeddings_and_attention-a100-t02-713b8e | 1 | 0 | 2023-05-18T18:08:15+00:00 |
zh-tw-llm-dv/zh-tw-llm-ta01-pythia-6.9b-ta8000-v1-a2_2_lora_instruction_tune-h100-t003-f19e35-merged-float16 | 1 | 0 | 2023-05-24T16:38:48+00:00 |
zh-tw-llm-dv/zh-tw-llm-ta01-pythia-6.9b-ta8000-v1-a_1_embeddings-h100-t01-c5daa1 | 1 | 0 | 2023-05-19T19:40:34+00:00 |
zh-tw-llm-dv/zh-tw-llm-ta01-pythia-6.9b-ta8000-v1-a_1_embeddings-h100-t01-c5daa1-8bit | 1 | 0 | 2023-05-19T21:05:39+00:00 |
zh-tw-llm-dv/zh-tw-llm-ta01-pythia-6.9b-ta8000-v1-a_1_embeddings-h100-t015-792f7c-float16 | 1 | 0 | 2023-05-21T03:14:36+00:00 |
zh-tw-llm-dv/zh-tw-llm-ta01-pythia-6.9b-ta8000-v1-a_2_lora_instruction_tune-h100-t002-3d42d8-merged-float16 | 1 | 0 | 2023-05-21T07:09:58+00:00 |
zh-tw-llm-dv/zh-tw-llm-ta01-pythia-70m-ta8000-v1-a_1_embeddings-a100-t4-ce784e-float16 | 1 | 0 | 2023-05-21T18:14:37+00:00 |
zh-tw-llm-dv/zh-tw-llm-ta01-pythia-70m-ta8000-v1-a_2_lora_instruction_tune-a100-t002-7a793a-merged | 1 | 0 | 2023-05-21T18:18:13+00:00 |
zh-tw-llm-dv/zh-tw-llm-ta01-pythia-70m-ta8000-v1-a_2_lora_instruction_tune-a100-t002-7a793a-merged-float16 | 2 | 0 | 2023-05-21T18:19:31+00:00 |
zh-tw-llm-dv/zh-tw-llm-ta01-pythia-70m-ta8000-v1-b_2_lora_instruction_tune-a100-t004-649aad-merged | 1 | 0 | 2023-05-19T10:17:04+00:00 |
zh-tw-llm-dv/zh-tw-pythia-6.9b-a12k-te01-embeddings-ea1 | 1 | 0 | 2023-05-13T15:41:02+00:00 |
zhangbo2008/gpt2-simulacra | 3 | 0 | 2023-07-03T09:55:41+00:00 |
zhangchuheng123/llama2-alpaca-sft-2epoch | 1 | 0 | 2023-10-19T09:28:53+00:00 |
zhangfaen/codeparrot-ds | 1 | 0 | 2023-05-23T09:59:17+00:00 |
zhangirazerbayev/llama-2-7b-roundtrip-private | 1 | 0 | 2023-08-10T04:13:50+00:00 |
zhangirazerbayev/llama_7b_code-no-matlab | 1 | 0 | 2023-07-18T15:19:04+00:00 |
zhangirazerbayev/llama_7b_code-v1 | 0 | 0 | 2023-07-31T00:30:36+00:00 |
zhangirazerbayev/llama_7b_code-v1-full-matlab | 2 | 0 | 2023-08-02T04:07:44+00:00 |
zhangirazerbayev/llama_7b_code-v1-with-tex | 1 | 0 | 2023-08-02T00:31:20+00:00 |
zhangirazerbayev/llama_7b_code-v2 | 2 | 0 | 2023-07-27T17:40:23+00:00 |
zhangirazerbayev/llama_7b_code-v2-full-matlab_rel-token-count | 2 | 0 | 2023-08-01T04:17:00+00:00 |
zhangirazerbayev/llama_7b_code-v2-with-tex_rel-token-count | 1 | 0 | 2023-08-01T00:05:31+00:00 |
zhangirazerbayev/llama_7b_code-v2_rel-token-count | 0 | 0 | 2023-07-31T23:20:43+00:00 |
zhangirazerbayev/llama_7b_code-v3 | 1 | 0 | 2023-07-26T23:39:49+00:00 |
zhangirazerbayev/llama_7b_mix_5e-2nl | 0 | 0 | 2023-07-18T15:08:08+00:00 |
zhangirazerbayev/llama_mix_2_7b_step10000 | 1 | 0 | 2023-07-29T00:36:18+00:00 |
zhangirazerbayev/mix_2_1b_step11632 | 1 | 0 | 2023-07-27T02:41:59+00:00 |
zhangirazerbayev/mix_3_1b_step11632 | 1 | 0 | 2023-08-07T00:06:01+00:00 |
zhangirazerbayev/mix_step11632 | 1 | 0 | 2023-07-13T04:18:35+00:00 |
zhangirazerbayev/open-web-math-52b_1b_step11632 | 2 | 0 | 2023-08-07T00:02:57+00:00 |
zhangirazerbayev/open-web-math-decontaminated_1b_step11632 | 2 | 0 | 2023-07-27T02:39:54+00:00 |
zhangirazerbayev/open-web-math-dev_step11632 | 2 | 0 | 2023-07-08T00:24:15+00:00 |
zhangirazerbayev/open-web-math-hq_step11632 | 1 | 0 | 2023-07-13T02:24:20+00:00 |
zhangirazerbayev/pile-sample_step11632 | 0 | 0 | 2023-07-11T00:57:15+00:00 |
zhangirazerbayev/proof-pile-v1_step11632 | 1 | 0 | 2023-07-08T06:56:55+00:00 |
zhangirazerbayev/proofgpt-v0.5-llama-7b-step20000 | 1 | 0 | 2023-06-15T22:38:41+00:00 |
zhangirazerbayev/proofgpt_v0.7_arxiv-pilev2_short | 1 | 0 | 2023-07-13T05:21:08+00:00 |
zhangirazerbayev/proofgpt_v0.7_arxiv-rp_short | 0 | 0 | 2023-07-13T05:05:10+00:00 |
zhangxy-2019/cu_dstc9_dialoGPT | 3 | 0 | 2022-03-02T23:29:05+00:00 |
zhangxy-2019/cunlp-gpt2-dialog | 3 | 0 | 2022-03-02T23:29:05+00:00 |
zhao-mm/mpt-30b-instruct-test | 1 | 0 | 2023-06-29T15:28:31+00:00 |
zhaolzhang/llama-2-7b-miniguanaco | 0 | 0 | 2023-09-05T05:42:49+00:00 |
zhaolzhang/llama-2-7b-resume | 1 | 0 | 2023-09-05T20:15:37+00:00 |
zhe0/aia-hw-lora | 0 | 0 | 2023-12-23T05:57:33+00:00 |
zhe0/peft-Taiwan-LLM-self-used | 12 | 0 | 2023-12-28T02:55:52+00:00 |
zhengchenphd/ICE-GRT | 1 | 3 | 2024-01-03T06:43:51+00:00 |
zhengkaitaken/Magical | 1 | 0 | 2023-08-02T18:30:54+00:00 |
zhengr/Mistral-7B-Instruct-v0.1-samsum | 0 | 0 | 2023-11-03T12:04:19+00:00 |
zhengr/Zhongjing-LLaMA-base | 0 | 0 | 2023-10-19T14:39:24+00:00 |
zhengr/llama-2-7b-miniguanaco | 1 | 0 | 2023-09-22T13:29:25+00:00 |
zhengudaoer/Wenzhong-GPT2-110M-finetuned-wikitext2 | 5 | 0 | 2023-03-14T03:39:52+00:00 |
zhengudaoer/Wenzhong-GPT2-110M-finetuned-wikitext2-2 | 5 | 0 | 2023-03-14T10:04:30+00:00 |
zhengudaoer/Wenzhong-GPT2-110M-finetuned-wikitext2-3 | 4 | 0 | 2023-03-14T10:25:49+00:00 |
zhengudaoer/distilgpt2-finetuned-wikitext2 | 5 | 0 | 2023-03-13T16:47:26+00:00 |
zhengxuanwen/Llama-2-7b-chat-finetune | 3 | 0 | 2023-11-28T05:31:20+00:00 |
zhengxuanzenwu/gpt2-5token-solver | 22 | 0 | 2023-11-29T09:53:30+00:00 |
zhenhai18/zhenhaitest | 0 | 0 | 2023-08-18T08:47:55+00:00 |
zhiqiulin/llava-v1.5-13b-flattened-multi-turn | 2 | 0 | 2023-12-13T08:10:06+00:00 |
zhiqiulin/llava-v1.5-13b-flattened-multi-turn-gpt4 | 2 | 0 | 2023-12-13T08:15:45+00:00 |
zhiqiulin/llava-v1.5-13b-flattened-multi-turn-gpt4_100k | 2 | 0 | 2023-12-13T08:21:28+00:00 |
zhiqiulin/llava-v1.5-13b-flattened-multi-turn-with-stage1-captions | 2 | 0 | 2023-12-13T08:26:56+00:00 |
zhiqiulin/llava-v1.5-13b-gpt4v_100k | 5 | 1 | 2023-12-13T08:37:56+00:00 |
zhiqiulin/llava-v1.5-13b-with-stage1-captions | 1 | 0 | 2023-12-13T08:43:26+00:00 |
zhk/G-SPEED | 1 | 0 | 2023-10-15T17:03:37+00:00 |
zhou12tao/pytorch_model | 3 | 0 | 2023-04-07T16:05:54+00:00 |
zhouning/lora-test | 1 | 0 | 2023-06-05T08:20:44+00:00 |
zhuimengshaonian/gpt2-ancient-base | 2 | 1 | 2022-11-25T14:46:33+00:00 |
zhuimengshaonian/gpt2-ancient-medium | 4 | 0 | 2022-12-09T05:58:56+00:00 |
zhuzilin/gpt2-summarize-sup4_ppo_rm4 | 5 | 0 | 2022-12-28T12:57:44+00:00 |
zhzhang93/test-model | 0 | 0 | 2023-12-12T02:11:43+00:00 |