Commit Graph

21 Commits (0bdbf300ac347640e5b81dff2fa187bcfe05574c)

Author SHA1 Message Date
Disty0 d9e628574a SDNQ add 15, 13, 11 and 9 bit support 2026-03-11 03:39:02 +03:00
Disty0 78efbc7e85 update sdnq 2026-02-24 19:47:30 +03:00
Vladimir Mandic bfe014f5da modernize typing 2026-02-19 09:15:37 +01:00
Disty0 90cf83b90e update sdnq 2026-02-02 20:22:29 +03:00
Disty0 784cda80aa update sdnq 2026-01-14 16:23:26 +03:00
Disty0 db59d2b507 SDNQ handle packed floats in fp mm 2025-12-27 16:29:18 +03:00
Disty0 aaef4992c3 SDNQ fix svd + fp8 tw and fp16 mm 2025-11-28 22:31:09 +03:00
Disty0 b6e9332cfe SDNQ de-couple matmul dtype and add fp16 matmul 2025-11-22 02:16:20 +03:00
Disty0 1745ed53f8 Refactor SDNQDequantizer 2025-11-18 01:42:58 +03:00
Disty0 f12caf81f9 SDNQ skip bad layers on svd and fix svd with dequantize_fp32 2025-10-17 17:25:50 +03:00
Disty0 c7aba8589b SDNQ fix Qwen loading 2025-10-11 00:05:09 +03:00
Disty0 9e52d0c1fb SDNQ add SVDQuant quantization method 2025-10-05 22:50:30 +03:00
Disty0 54acf1760b Make SDNQ scales compatible with balanced offload 2025-10-03 18:13:55 +03:00
Disty0 6b67a9d0c4 SDNQ add check_mats to matmul 2025-09-30 01:58:13 +03:00
Disty0 e6715ba8d3 Cleanup SDNQ compile 2025-09-19 19:29:36 +03:00
Disty0 a12edc1e90 SDNQ use nan_to_num_ with fp8 quantization in case of zeros 2025-09-15 20:22:39 +03:00
Disty0 f324b7c0e5 SDNQ remove unnecessary .contiguous() 2025-08-21 02:21:05 +03:00
Disty0 8460be662c SDNQ use inplace transpose and use view instead of reshape 2025-08-17 05:07:55 +03:00
Disty0 dc7b25d387 Cleanup SDNQ and add SDNQ_USE_TENSORWISE_FP8_MATMUL env var 2025-08-11 14:50:17 +03:00
Disty0 22d86acda3 Make SDNQ MatMul listen to the dequantize fp32 setting 2025-08-09 01:10:07 +03:00
Disty0 c3d007b02c SDNQ split forward.py into layers and cleanup 2025-08-02 17:36:55 +03:00