Text Generation
Transformers
Safetensors
English
Eval Results
Inference Endpoints
nielsr HF staff commited on
Commit
d218989
·
verified ·
1 Parent(s): 14662b1

Add missing metadata

Browse files

This PR adds a link to the paper [Control LLM: Controlled Evolution for Intelligence Retention in LLM](https://huggingface.co/papers/2501.10979), and
also adds the missing `library_name` and `pipeline_tag`.

Files changed (1) hide show
  1. README.md +6 -4
README.md CHANGED
@@ -74,12 +74,14 @@ model-index:
74
  value: 0.4029255319148936
75
  stderr: 0.004471732136513382
76
  verified: false
 
 
77
  ---
 
78
  # Control-LLM-Llama3.1-8B-OpenCoder8
79
- This is a fine-tuned model of Llama-3.1-8B-Instruct for coding tasks on OpenCoder SFT dataset.
80
 
81
- ## Linked Paper
82
- This model is associated with the paper: [Control-LLM](https://arxiv.org/abs/2501.10979).
83
 
84
  ## Evaluation Results
85
  Here is an overview of the evaluation results and findings:
@@ -114,4 +116,4 @@ The table below summarizes evaluation results across coding tasks and original c
114
  - **MLU**: MMLU (Massive Multitask Language Understanding)
115
  - **MLUP**: MMLU Pro
116
  - **O-Avg**: Original Capability - Size Weighted Average across ARC, GPQA, MMLU, and MMLU Pro
117
- - **Overall**: Combined average across all tasks
 
74
  value: 0.4029255319148936
75
  stderr: 0.004471732136513382
76
  verified: false
77
+ pipeline_tag: text-generation
78
+ library_name: transformers
79
  ---
80
+
81
  # Control-LLM-Llama3.1-8B-OpenCoder8
82
+ This is a fine-tuned model of Llama-3.1-8B-Instruct for coding tasks on OpenCoder SFT dataset described in the paper: [Control LLM: Controlled Evolution for Intelligence Retention in LLM](https://huggingface.co/papers/2501.10979).
83
 
84
+ Code: https://github.com/linkedin/ControlLLM.
 
85
 
86
  ## Evaluation Results
87
  Here is an overview of the evaluation results and findings:
 
116
  - **MLU**: MMLU (Massive Multitask Language Understanding)
117
  - **MLUP**: MMLU Pro
118
  - **O-Avg**: Original Capability - Size Weighted Average across ARC, GPQA, MMLU, and MMLU Pro
119
+ - **Overall**: Combined average across all tasks