metadata
license: apache-2.0
datasets:
- teknium/openhermes
- argilla/ultrafeedback-binarized-preferences
- Intel/orca_dpo_pairs
language:
- en
library_name: transformers
DPOpenHermes 11B
This is a mergekit merge of DPOpenHermes-7B from seperate versions of it.
slices:
- sources:
- model: openaccess-ai-collective/DPOpenHermes-7B
revision: dpo-v0
layer_range: [0, 24]
- sources:
- model: openaccess-ai-collective/DPOpenHermes-7B
layer_range: [8, 32]
merge_method: passthrough
dtype: bfloat16