Will you please clarify flash-attention 2 usage in this model

#2
by leocnj - opened

Thanks for posting this cool model !

Regarding using "flash-attention 2", do you mind elaborating more? Do you mean you are using FA-2 in your PEFT tuning?

Sign up or log in to comment