Make flash attention configurable in user code
#26
by YenChunChen - opened
With this PR, users can specify whether to enable flash attention 2 in from_pretrain.
YenChunChen changed pull request status to open
@YenChunChen default should be flash_attention in readme, user can specify to use eager if they want
done
leoxiaobin changed pull request status to merged