|
roles_map = { |
|
'system': 'system', |
|
'user': 'user', |
|
'human': 'user', |
|
'assistant': 'assistant', |
|
'gpt': 'assistant', |
|
'AI': 'assistant', |
|
} |
|
|
|
|
|
contrain_datasets = [ |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
*[ |
|
{'path': 'mlabonne/open-perfectblend', 'split': f'train[{i}%:{i + 20}%]', 'field': 'conversations', 'transform': lambda msgs: [ |
|
{'role': roles_map[m['from']], 'content': m['value']} |
|
for m in msgs |
|
]} |
|
for i in range(0, 100, 20) |
|
], |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
*[ |
|
{'path': 'arcee-ai/The-Tome', 'split': f'train[{i}%:{i + 20}%]', 'field': 'conversations', 'transform': lambda msgs: [ |
|
{'role': roles_map[m['from']], 'content': m['value']} |
|
for m in msgs |
|
]} |
|
for i in range(0, 100, 20) |
|
], |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
*[ |
|
{'path': 'rombodawg/Everything_Instruct_Multilingual', 'split': f'train[{i}%:{i + 20}%]', 'transform': lambda r: [ |
|
{'role': 'system', 'content': r['instruction']}, |
|
{'role': 'user', 'content': r['input']}, |
|
{'role': 'assistant', 'content': r['output']}, |
|
]} |
|
for i in range(0, 100, 20) |
|
], |
|
|
|
|
|
|
|
|
|
|
|
{'path': 'NousResearch/hermes-function-calling-v1', 'field': 'conversations', 'transform': lambda msgs: [ |
|
{'role': roles_map[m['from']], 'content': m['value']} |
|
for m in msgs |
|
]}, |
|
|
|
|
|
|
|
|
|
|
|
*[ |
|
{'path': 'arcee-ai/agent-data', 'split': f'train[{i}%:{i + 20}%]', 'field': 'conversations', 'transform': lambda msgs: [ |
|
{'role': roles_map[m['from']], 'content': m['value']} |
|
for m in msgs |
|
]} |
|
for i in range(0, 100, 20) |
|
], |
|
|
|
|
|
|
|
|
|
*[ |
|
|
|
{'path': 'AtlasUnified/Atlas-Reasoning', 'data_files': 'reasoning.csv', 'transform': lambda r: [ |
|
{'role': 'user', 'content': r['Prompt']}, |
|
{'role': 'assistant', 'content': r['Step-by-step reasoning'] + '\n' + r['Solution']}, |
|
]}, |
|
], |
|
|
|
|
|
|
|
|
|
|
|
{'path': 'thesven/gsm8k-reasoning', 'transform': lambda r: [ |
|
{'role': 'user', 'content': r['question']}, |
|
{'role': 'assistant', 'content': (r['generation'] or '') + '\n' + r['answer'] + '\n' + r['short_answer']}, |
|
]}, |
|
|
|
|
|
{'path': 'AlgorithmicResearchGroup/math_reasoning_autoformalization_track', 'transform': lambda r: [ |
|
{'role': 'user', 'content': r['informal_statement']}, |
|
{'role': 'assistant', 'content': r['informal_proof'] + '\n' + r['formal_proof']}, |
|
]}, |
|
|
|
|
|
{'path': 'KingNish/reasoning-base-20k', 'transform': lambda r: [ |
|
{'role': 'user', 'content': r['user']}, |
|
{'role': 'assistant', 'content': r['reasoning'] + '\n' + r['assistant']}, |
|
]}, |
|
|
|
|
|
{'path': 'Aarushhh/math-reasoning-10k', 'transform': lambda r: [ |
|
{'role': 'user', 'content': r['problem']}, |
|
{'role': 'assistant', 'content': r['plan'] + '\n' + r['solution']}, |
|
]}, |
|
|
|
|
|
|
|
|
|
|
|
{'path': 'dvilasuero/reflection-v1-gpt-4o-judge', 'transform': lambda r: [ |
|
{'role': 'system', 'content': r['system']}, |
|
{'role': 'user', 'content': r['prompt']}, |
|
{'role': 'assistant', 'content': r['response']}, |
|
]}, |
|
|
|
{'path': 'dvilasuero/reflection-v1-openai-o-mini-judge', 'transform': lambda r: [ |
|
{'role': 'system', 'content': r['system']}, |
|
{'role': 'user', 'content': r['prompt']}, |
|
{'role': 'assistant', 'content': r['response']}, |
|
]}, |
|
|
|
{'path': 'dvilasuero/reflection-v1-final-dedup', 'transform': lambda r: [ |
|
{'role': 'system', 'content': r['system']}, |
|
{'role': 'user', 'content': r['prompt']}, |
|
{'role': 'assistant', 'content': r['response']}, |
|
]}, |
|
|
|
{'path': 'flozi00/reflection-qwen2.5-72b-260924', 'transform': lambda r: [ |
|
r['system'][0], |
|
{'role': 'user', 'content': r['input']}, |
|
{'role': 'assistant', 'content': r['reflection'] + '\n' + r['output']}, |
|
]}, |
|
|
|
|
|
|
|
|
|
|
|
{'path': 'HuggingFaceTB/smol-smoltalk', 'field': 'messages'}, |
|
] |
|
|