natolambert
commited on
Commit
•
2057cd9
1
Parent(s):
da8be66
Update src/md.py
Browse files
src/md.py
CHANGED
@@ -98,9 +98,8 @@ For more details, see the [dataset](https://huggingface.co/datasets/allenai/rewa
|
|
98 |
current_time = datetime.now().strftime("%H:%M, %d %b %Y")
|
99 |
|
100 |
TOP_TEXT = f"""# RewardBench: Evaluating Reward Models
|
101 |
-
Last restart: {current_time}
|
102 |
-
|
103 |
### Evaluating the capabilities, safety, and pitfalls of reward models
|
|
|
104 |
[Code](https://github.com/allenai/reward-bench) | [Eval. Dataset](https://huggingface.co/datasets/allenai/reward-bench) | [Prior Test Sets](https://huggingface.co/datasets/allenai/pref-test-sets) | [Results](https://huggingface.co/datasets/allenai/reward-bench-results) | [Paper](https://arxiv.org/abs/2403.13787) | Total models: {{}} | * Unverified models | ⚠️ Dataset Contamination
|
105 |
|
106 |
⚠️ Many of the top models were trained on unintentionally contaminated, AI-generated data, for more information, see this [gist](https://gist.github.com/natolambert/1aed306000c13e0e8c5bc17c1a5dd300)."""
|
|
|
98 |
current_time = datetime.now().strftime("%H:%M, %d %b %Y")
|
99 |
|
100 |
TOP_TEXT = f"""# RewardBench: Evaluating Reward Models
|
|
|
|
|
101 |
### Evaluating the capabilities, safety, and pitfalls of reward models
|
102 |
+
Last restart: {current_time}
|
103 |
[Code](https://github.com/allenai/reward-bench) | [Eval. Dataset](https://huggingface.co/datasets/allenai/reward-bench) | [Prior Test Sets](https://huggingface.co/datasets/allenai/pref-test-sets) | [Results](https://huggingface.co/datasets/allenai/reward-bench-results) | [Paper](https://arxiv.org/abs/2403.13787) | Total models: {{}} | * Unverified models | ⚠️ Dataset Contamination
|
104 |
|
105 |
⚠️ Many of the top models were trained on unintentionally contaminated, AI-generated data, for more information, see this [gist](https://gist.github.com/natolambert/1aed306000c13e0e8c5bc17c1a5dd300)."""
|