Flash Attention Support for XLM-R models?

#20
by ShaomuTan - opened

It would be great to support XLM-R models (base, large, xl, XXL) with flash attention 2.

Your need to confirm your account before you can post a new comment.

Sign up or log in to comment