Skip to content

Pull requests: intel/auto-round

Author
Filter by author
Loading
Label
Filter by label
Loading
Use alt + click/return to exclude labels
or + click/return for logical OR
Projects
Filter by project
Loading
Milestones
Filter by milestone
Loading
Reviews
Assignee
Filter by who’s assigned
Assigned to nobody Loading
Sort

Pull requests list

Fix FP8 MLA vllm-ext
#1343 opened Jan 26, 2026 by yiliu30 Loading…
9 tasks
add copilot-instructions
#1341 opened Jan 26, 2026 by xin3he Loading…
1 of 9 tasks
refactor init of compressor
#1339 opened Jan 26, 2026 by n1ck-guo Loading…
1 of 9 tasks
rm duplicate args of the quantization extra config
#1334 opened Jan 23, 2026 by WeiweiZhang1 Loading…
1 of 9 tasks
Fix cpu ut for transformers v5
#1333 opened Jan 23, 2026 by Kaihui-intel Loading…
1 of 9 tasks
0.10.0
add support for w4a16_mixed enhancement New feature or request ready only add when the PR is ready to merge
#1326 opened Jan 23, 2026 by n1ck-guo Loading…
6 of 17 tasks
Autoround in vLLM Office Hours documentation Improvements or additions to documentation
#1322 opened Jan 23, 2026 by yiliu30 Loading…
1 of 18 tasks
enable glm4_moe_lite quantization & generation
#1321 opened Jan 22, 2026 by WeiweiZhang1 Loading…
3 of 18 tasks
Add asym for XPU backend.
#1316 opened Jan 22, 2026 by luoyu-intel Draft
Update torch to 2.9.1 in CI
#1313 opened Jan 22, 2026 by XuehaoSun Loading…
align act_max of experts for qwen3-vl and qwen3-next
#1311 opened Jan 21, 2026 by xin3he Loading…
[WIP]Ds v32
#1291 opened Jan 16, 2026 by yiliu30 Draft
Fix ignore_layers not working for FP8 models
#1286 opened Jan 15, 2026 by Copilot AI Loading…
11 tasks done
fix disable_opt_rtn spelling error
#1250 opened Jan 9, 2026 by WeiweiZhang1 Loading…
add per-task lm_eval args for exprimental usage
#1017 opened Nov 11, 2025 by WeiweiZhang1 Loading…
[WIP] [STEP 2] split compressor into few quantizers
#841 opened Sep 23, 2025 by n1ck-guo Loading…
ProTip! What’s not been updated in a month: updated:<2025-12-26.