Commit History
ORPO Trainer replacement (#1551)
7d1d22f
unverified
winglian
commited on
Unsloth gradient checkpointing offload (#1528)
6319da1
unverified
winglian
commited on
DBRX Model Support (#1462)
132eb74
unverified
winglian
commited on
use locale agnostic seperator to make large nums easier to read (#1503)
da9b1a3
unverified
winglian
commited on
WIP: Support table logging for mlflow, too (#1506)
057fa44
unverified
Correctly handle splits for datasets.arrow_dataset.Dataset objects (#1504)
8fa0785
unverified
add field to sft dataset pydantic for completion support (#1497)
ff01c45
unverified
winglian
commited on
ignore issues with calculating # params when printing (#1493)
2fa65b9
unverified
winglian
commited on
Remove `validate_quantized_dora` (#1485)
9430b6e
unverified
xzuyn
commited on
fix: reduce sample_packing warning (#1484)
bda48f0
unverified
Nanobit
commited on
feat: validate sample packing requires flash_attention (#1465)
bf4cd67
unverified
Nanobit
commited on
add support for cohere chat template (#1478)
05b0b7e
unverified
winglian
commited on
refactor utils.data module for line count linter (#1476)
e0fcef4
unverified
winglian
commited on
Pretrain multipack v2 (#1470)
5aa5097
unverified
winglian
commited on
fix pretraining_ on odd datasets (#1463)
586bd8d
unverified
monsoon-nlp
commited on
reduce verbosity of the special tokens (#1472)
0b10377
unverified
winglian
commited on
qwen2_moe support w multipack (#1455)
6086be8
unverified
winglian
commited on
fix some of the edge cases for Jamba (#1452)
05b398a
unverified
winglian
commited on
Support loading datasets saved via save_to_disk (#1432)
e634118
unverified
fozziethebeat
commited on
Jamba (#1451)
02af082
unverified
winglian
commited on
fix layer_replication arg to peft (#1446)
4155e99
unverified
winglian
commited on
support layer replication for peft and fix rslora integration (#1445)
25afd35
unverified
winglian
commited on
fix for accelerate env var for auto bf16, add new base image and expand torch_cuda_arch_list support (#1413)
da265dd
unverified
winglian
commited on
Fix falcon tokenization step (#1441) [skip ci]
bcdc9b1
unverified
make sure to capture non-null defaults from config validation (#1415)
601b77b
unverified
winglian
commited on
fix(dataset): normalize tokenizer config and change hash from tokenizer class to tokenizer path (#1298)
ff939d8
unverified
Nanobit
commited on
strip out hacky qlora-fsdp workarounds now that qlora-fsdp fixes are upstreamed (#1428)
2a1589f
unverified
winglian
commited on
support galore once upstreamed into transformers (#1409)
dd449c5
unverified
winglian
commited on
Feat: Add sharegpt multirole (#1137)
40a88e8
unverified
Nanobit
commited on
fix(config): passing gradient_checkpoint_kwargs (#1412)
b1e3e1b
unverified
Nanobit
commited on
ORPO (#1419)
2ea70eb
unverified
winglian
commited on
beta support for multipack with gemmoe: (#1402)
8df7b88
unverified
winglian
commited on
Train parameters exclusively in specific ranges (#1390)
05bcc9e
unverified
seungduk
commited on
Update ChatTemplate enum to include alpaca and gemma (#1396)
0976781
unverified
chiragjn
commited on
chore: lint (#1389)
4326520
unverified
winglian
commited on
Add Glaive conversation format support (#1365)
b7d8a7d
unverified
support for rslora (#1387) [skip ci]
7659c00
unverified
winglian
commited on
validation for fsdp and deepspeed (#1388) [skip ci]
3fd8093
unverified
winglian
commited on
FDSP + QLoRA (#1378)
9b6ee83
unverified
winglian
commited on
support for DoRA w/ PEFT (#1363)
0cfdb2c
unverified
winglian
commited on
lora+ support (#1352)
decb66e
unverified
winglian
commited on
Fix validation for early stopping (#1358)
b5b4492
unverified
chiragjn
commited on
fix for protected model_ namespace w pydantic (#1345)
6b3b271
unverified
winglian
commited on
Fix `use_mlflow` to be bool instead of str (#1344)
3a5a2d2
unverified
chiragjn
commited on
more fixes 20240228 (#1342) [skip ci]
0f985e1
unverified
winglian
commited on