Add missing metadata
Browse filesThis PR adds a link to the paper [Control LLM: Controlled Evolution for Intelligence Retention in LLM](https://huggingface.co/papers/2501.10979), and
also adds the missing `library_name` and `pipeline_tag`.
README.md
CHANGED
@@ -74,12 +74,14 @@ model-index:
|
|
74 |
value: 0.4029255319148936
|
75 |
stderr: 0.004471732136513382
|
76 |
verified: false
|
|
|
|
|
77 |
---
|
|
|
78 |
# Control-LLM-Llama3.1-8B-OpenCoder8
|
79 |
-
This is a fine-tuned model of Llama-3.1-8B-Instruct for coding tasks on OpenCoder SFT dataset.
|
80 |
|
81 |
-
|
82 |
-
This model is associated with the paper: [Control-LLM](https://arxiv.org/abs/2501.10979).
|
83 |
|
84 |
## Evaluation Results
|
85 |
Here is an overview of the evaluation results and findings:
|
@@ -114,4 +116,4 @@ The table below summarizes evaluation results across coding tasks and original c
|
|
114 |
- **MLU**: MMLU (Massive Multitask Language Understanding)
|
115 |
- **MLUP**: MMLU Pro
|
116 |
- **O-Avg**: Original Capability - Size Weighted Average across ARC, GPQA, MMLU, and MMLU Pro
|
117 |
-
- **Overall**: Combined average across all tasks
|
|
|
74 |
value: 0.4029255319148936
|
75 |
stderr: 0.004471732136513382
|
76 |
verified: false
|
77 |
+
pipeline_tag: text-generation
|
78 |
+
library_name: transformers
|
79 |
---
|
80 |
+
|
81 |
# Control-LLM-Llama3.1-8B-OpenCoder8
|
82 |
+
This is a fine-tuned model of Llama-3.1-8B-Instruct for coding tasks on OpenCoder SFT dataset described in the paper: [Control LLM: Controlled Evolution for Intelligence Retention in LLM](https://huggingface.co/papers/2501.10979).
|
83 |
|
84 |
+
Code: https://github.com/linkedin/ControlLLM.
|
|
|
85 |
|
86 |
## Evaluation Results
|
87 |
Here is an overview of the evaluation results and findings:
|
|
|
116 |
- **MLU**: MMLU (Massive Multitask Language Understanding)
|
117 |
- **MLUP**: MMLU Pro
|
118 |
- **O-Avg**: Original Capability - Size Weighted Average across ARC, GPQA, MMLU, and MMLU Pro
|
119 |
+
- **Overall**: Combined average across all tasks
|