rishiraj commited on
Commit
f2c58e5
·
1 Parent(s): 65d316e

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +54 -20
README.md CHANGED
@@ -24,30 +24,57 @@ model-index:
24
  results: []
25
  ---
26
 
27
- <!-- This model card has been generated automatically according to the information the Trainer had access to. You
28
- should probably proofread and complete it, then remove this comment. -->
29
-
30
  # 😼 CatPPT
31
 
32
- Introducing "CatPPT" - the purrfect alternative to that other big cat in town, known for keeping all the secrets to itself! Our feline friend here is a Large Language Model like no other, created through the magical process of merging openchat and neuralchat models using the enchanting Gradient SLERP method.
 
 
 
 
33
 
34
- This whiskered wonder boasts being the top-performing 7B model on the block, free from any whiff of evaluation data contamination. So go ahead, let your curiosity run wild, and engage with this independent, open-source kitty who's ready to pounce on all your language processing needs. Just remember, there's no need to feel left out in the cold when you have CatPPT warming up your cozy corner of the internet!
35
 
36
- This model is a fine-tuned version of [rishiraj/CatPPT-base](https://huggingface.co/rishiraj/CatPPT-base) on an unknown dataset.
37
- It achieves the following results on the evaluation set:
38
- - Loss: 2.0093
39
 
40
- ## Model description
41
 
42
- More information needed
 
 
 
 
 
 
 
 
 
 
43
 
44
- ## Intended uses & limitations
45
 
46
- More information needed
47
 
48
- ## Training and evaluation data
 
 
49
 
50
- More information needed
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
51
 
52
  ## Training procedure
53
 
@@ -78,10 +105,17 @@ The following hyperparameters were used during training:
78
  - Pytorch 2.1.2+cu121
79
  - Datasets 2.14.6
80
  - Tokenizers 0.15.0
81
- ## Training procedure
82
-
83
-
84
- ### Framework versions
85
-
86
-
87
  - PEFT 0.6.1
 
 
 
 
 
 
 
 
 
 
 
 
 
 
24
  results: []
25
  ---
26
 
 
 
 
27
  # 😼 CatPPT
28
 
29
+ Introducing "CatPPT" - the purrfect alternative to that other big cat in town, known for keeping all the secrets to itself! Our feline friend here is created through merging openchat and neuralchat models using Gradient SLERP method (resulting in [rishiraj/CatPPT-base](https://huggingface.co/rishiraj/CatPPT-base)) and then finetuned on no_robots dataset for chat.
30
+
31
+ This is the top-performing 7B model on the leaderboard, that's free from any whiff of evaluation data contamination.
32
+
33
+ ## Model date
34
 
35
+ rishiraj/CatPPT was trained between 15th and 17th December, 2023.
36
 
37
+ ## Evaluation
 
 
38
 
39
+ It achieves the following results on the [Open_LLM_Leaderboard](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard). At the time of release, CatPPT is the highest ranked 7B chat model on the leaderboard, that's free from evaluation data contamination.
40
 
41
+ |Model |Average|ARC |HellaSwag|MMLU |TruthfulQA|Winogrande|GSM8K|
42
+ |------------------------------------|-------|-----|---------|-----|----------|----------|-----|
43
+ |**rishiraj/CatPPT** |**72.32** |**68.09**|**86.69** |**65.16**|**61.55** |**81.61** |**70.81**|
44
+ |Intel/neural-chat-7b-v3-3 |69.83 |66.89|85.26 |63.07|63.01 |79.64 |61.11|
45
+ |openchat/openchat-3.5-1210 |68.89 |64.93|84.92 |64.62|52.15 |80.74 |65.96|
46
+ |meta-math/MetaMath-Mistral-7B |65.78 |60.67|82.58 |61.95|44.89 |75.77 |68.84|
47
+ |Deci/DeciLM-7B-instruct |63.19 |61.01|82.37 |60.24|49.75 |79.72 |46.02|
48
+ |mistralai/Mistral-7B-Instruct-v0.2 |65.71 |63.14|84.88 |60.78|68.26 |77.19 |40.03|
49
+ |mistralai/Mixtral-8x7B-Instruct-v0.1|72.62 |70.22|87.63 |71.16|64.58 |81.37 |60.73|
50
+ |meta-llama/Llama-2-70b-hf |67.87 |67.32|87.33 |69.83|44.92 |83.74 |54.06|
51
+ |tiiuae/falcon-180B |67.85 |69.45|88.86 |70.5 |45.47 |86.9 |45.94|
52
 
53
+ ## Inference procedure
54
 
55
+ Here's how you can run the model using the pipeline() function from 🤗 Transformers:
56
 
57
+ ```
58
+ import torch
59
+ from transformers import pipeline
60
 
61
+ pipe = pipeline("text-generation", model="rishiraj/CatPPT", torch_dtype=torch.bfloat16, device_map="auto")
62
+
63
+ # We use the tokenizer's chat template to format each message - see https://huggingface.co/docs/transformers/main/en/chat_templating
64
+ messages = [
65
+ {
66
+ "role": "system",
67
+ "content": "You are a friendly chatbot who always responds in the style of a pirate"
68
+ },
69
+ {
70
+ "role": "user",
71
+ "content": "How many helicopters can a human eat in one sitting?"
72
+ }
73
+ ]
74
+ prompt = pipe.tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
75
+ outputs = pipe(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
76
+ print(outputs[0]["generated_text"])
77
+ ```
78
 
79
  ## Training procedure
80
 
 
105
  - Pytorch 2.1.2+cu121
106
  - Datasets 2.14.6
107
  - Tokenizers 0.15.0
 
 
 
 
 
 
108
  - PEFT 0.6.1
109
+
110
+ ## Citation Information
111
+
112
+ ```
113
+ @misc{rishiraj2023catppt,
114
+ author = {Rishiraj Acharya},
115
+ title = {CatPPT},
116
+ year = {2023},
117
+ publisher = {Hugging Face},
118
+ journal = {Hugging Face repository},
119
+ howpublished = {\url{https://huggingface.co/rishiraj/CatPPT}}
120
+ }
121
+ ```