Wing Lian
cae5cebb59
xformers attention with packing (#2619)
* xformers attention with packing
* wire up the patch
* fix xformers + packing validation
* fix warning
* reorder the packing check
* fix fp16 / bf16 reset when using fp16 with bf16 auto
* fix seq lens calc to drop hanging sequences
* handle xformers patch for inference too
* fix batch size setter
* fix xformers inference
* add colab callback to fix inference post train
* PR feedback
2025-05-07 16:10:16 -04:00
..
2025-05-07 16:10:15 -04:00
2025-01-29 00:10:19 -05:00
2025-03-21 12:26:47 -04:00
2025-02-25 16:09:37 +07:00
2024-04-04 13:43:40 +09:00
2025-04-24 01:03:43 -04:00
2025-05-07 16:10:16 -04:00
2025-03-21 10:18:01 -04:00
2025-04-07 12:41:13 -04:00
2025-03-21 12:26:47 -04:00
2025-02-25 16:09:37 +07:00
2025-04-25 17:14:48 -04:00
2025-04-28 10:11:20 -04:00
2024-07-11 09:19:29 -04:00
2025-03-17 08:39:04 -04:00
2025-03-17 08:39:04 -04:00
2025-02-25 16:09:37 +07:00
2025-04-24 01:03:43 -04:00
2025-04-02 09:33:46 -04:00
2025-01-24 12:56:28 -05:00
2025-02-25 16:09:37 +07:00
2025-04-11 09:52:43 -04:00
2025-02-25 16:09:37 +07:00
2025-05-07 16:10:10 -04:00
2024-03-21 22:28:36 -07:00
2025-02-25 16:09:37 +07:00
2025-02-25 16:09:37 +07:00
2025-03-17 08:38:19 -04:00
2025-04-28 10:07:45 -04:00
2025-04-16 13:50:48 -04:00
2025-02-25 16:09:37 +07:00
2025-02-25 16:09:37 +07:00