FeepingCreature 7aceb9f91c
Add --use-flash-attention flag. (#7223)
* Add --use-flash-attention flag.
This is useful on AMD systems, as FA builds are still 10% faster than Pytorch cross-attention.
2025-03-14 03:22:41 -04:00
..
2024-12-20 16:24:55 -05:00
2023-01-16 22:37:14 -05:00
2024-06-27 18:43:11 -04:00
2025-03-06 00:24:43 -05:00
2025-01-24 06:15:54 -05:00
2025-03-10 17:23:13 -04:00
2025-01-20 03:44:13 -05:00
2025-03-12 02:42:37 -04:00
2025-03-05 00:13:49 -05:00
2025-03-05 00:13:49 -05:00