Commit History

add gptneox embeddings, fix phi2 inputs, also fix the casting (#1083)
78c5b19
unverified

winglian commited on

be more robust about checking embedding modules for lora finetunes (#1074) [skip ci]
0f10080
unverified

winglian commited on

fix: torch_dtype mistral default to fp32 (#1050)
c3e8165
unverified

Nanobit commited on

Phi2 rewrite (#1058)
732851f
unverified

winglian commited on

feature: better device mapping for large models (#918)
bdfefaf
unverified

kallewoof Karl-Johan Alm winglian commited on

RL/DPO (#935)
f243c21

winglian commited on

bump transformers and update attention class map name (#1023)
bcc78d8
unverified

winglian commited on

Adds chat templates (#1022)
f8ae59b
unverified

mhenrichsen commited on

feat: expose bnb kwargs (#1018)
41353d2
unverified

Nanobit hamel commited on

remove landmark attn and xpos rope implementations (#1010)
70b46ca
unverified

winglian commited on

Feat: Warns to add to modules_to_save when adding tokens or switching special_tokens (#787)
1ffa386
unverified

Nanobit commited on

Fix Deepspeed loading (#950)
5ea3aa3
unverified

winglian commited on

Flash attn hotfix (#951)
f1f60cb
unverified

winglian commited on

Mixtral official (#942)
7fabc4d
unverified

winglian commited on

Mixtral multipack (#928)
68b227a
unverified

winglian commited on

support for mamba (#915)
40a6362
unverified

winglian commited on

fix(tokenizer): handle fast tokenizer properly for bos/eos (#914)
fde091c
unverified

Nanobit commited on

feat: add check for quantized model (#913)
a581e9f
unverified

Nanobit winglian commited on

Support device_map=sequential & max_memory config parameters (#903)
992e742
unverified

Bryan Thornbury winglian commited on

fix for qwen w lora (#906)
3e3229e
unverified

winglian commited on

Feat: Add Qwen (#894)
1115c50
unverified

Nanobit commited on

Phi update 202311 (#876)
9bf854e
unverified

winglian commited on

allow overriding of model_config parameters from the YML (#853)
1bc1186
unverified

winglian commited on

fix model parallel (#816)
964d858
unverified

winglian commited on

fix(tokenizer): update log order after update (#806)
10388a8
unverified

Nanobit commited on

fix(config): Set eos/bos to tokenizer if different (#801)
637ed09
unverified

Nanobit commited on

refactor neft patch to be more re-usable similar to trl's impl (#796)
827ec3d
unverified

winglian commited on

chore: refactor truthy check and fix mypy (#780)
11d1d60
unverified

Nanobit commited on

Fix(model): Linear detected and added to target module with rope linear (#738)
440c3ab
unverified

Nanobit commited on

add noisy embedding (#721)
3bd9528
unverified

Maxime Maxime commited on

Fix: Higher vram usage for mistral and sample_packing (#691)
669f1d0
unverified

Nanobit commited on

flash_attention + sample packing for stablelm 3b (#671)
2d60ba3
unverified

winglian commited on

Fix: ValueError when FA + Mistral when padding_side=right (#681)
eb480df
unverified

Nanobit commited on

Fix(tokenizer): Set rstrip,lstrip,norm to False (#678)
e0b7eea
unverified

Nanobit commited on

chore: Clean up repetitive model kwargs (#670)
e62d590
unverified

Nanobit commited on

Feat: Allow usage of native Mistral FA when no sample_packing (#669)
697c50d
unverified

Nanobit commited on

remove patch fix for phi (#664)
f34648c
unverified

winglian commited on

Mistral flash attn packing (#646)
b6ab8aa
unverified

winglian commited on

skip some flash attn patches unless explicitly enabled (#643)
895f0a0
unverified

winglian commited on

Feat: Add support for upstream FA2 (#626)
19a600a
unverified

Nanobit commited on

misc fixes to add gptq tests (#621)
03e5907
unverified

winglian commited on

support to disable exllama for gptq (#604)
faecff9
unverified

winglian commited on

Delete duplicate lines (#606)
aa656e0
unverified

bofenghuang commited on

btlm and falcon monkey patches for flash attn (#566)
6b9b229
unverified

winglian commited on

make phi training work with Loras (#588)
62eaee7
unverified

winglian commited on

don't resize embeddings if it's already large enough (#577)
3607882
unverified

winglian commited on

Support Sample packing for phi arch (#586)
12a2dbb
unverified

winglian commited on

Add training callback to send predictions to WandB table (#521)
5b67ea9
unverified

Glavin001 commited on

fix for quant config from model (#540)
a94f9cb
unverified

winglian commited on