I am trying to finetune this mode, but it is being really slow. Therefore, I really hope that this model will support flash-attention-2 soon
flash-attention-2
· Sign up or log in to comment