FeepingCreature 7aceb9f91c
Add --use-flash-attention flag. (#7223)
* Add --use-flash-attention flag.
This is useful on AMD systems, as FA builds are still 10% faster than Pytorch cross-attention.
2025-03-14 03:22:41 -04:00
..
2024-12-20 16:24:55 -05:00
2025-01-16 17:45:37 -05:00
2024-12-12 17:59:16 -05:00
2024-12-16 19:35:40 -05:00
2024-12-27 18:02:30 -05:00
2025-02-27 07:22:42 -05:00