Go
Home
Pricing
FAQ
Install
Home
Pricing
FAQ
Install
Login
via GitHub
intel/auto-round
Pull Requests
Commits
try_to_fix_hadamard_regression
AutoAdamRound_bugfix
Chinesization
ZaneMark-patch-1
ZaneMark-patch-3
acp
actvation_quant
add_task_args_for_lmeval
agent-init
ar_agent
ark_zp
autoround_support_qbits_backend
awq_algorithm
bf16_scale
chore/claude-init
copilot/fix-corner-case-in-auto-round
copilot/fix-deprecated-fp-layers-handling
copilot/fix-docstrings-in-python-files
copilot/fix-issue-with-auto-rounding
copilot/fix-llm-type-70b-bits-setting
copilot/fix-typeerror-wrapped-fn
copilot/improve-pr-template-type-of-change
copilot/investigate-quantization-group-and-ffn
copilot/replace-getset-module-torch-api
copilot/sageattention
copilot/speedup-fp8-linear-convert
copilot/speedup-fp8-linear-convert-again
copilot/speedup-fp8-linear-convert-another-one
copilot/sub-pr-1237-again
copilot/sub-pr-1237
copilot/sub-pr-1324
copilot/sub-pr-1522-again
copilot/sub-pr-1532
copilot/update-user-settings-page
copilot/vscode-mo3shmf8-8qa6
ddp
debug_time_cost
debug/usable_rotation
debug-hang
debug-nvfp4
deepseekv3
ds-qwen
ds-v5
ds-v32
dsv4
enable_glm4_moe_lite_quantization
enable_llama4_int8_baseline
enable_llama4_quant
enable_mxfp_exporting
feat/activation-checkpointing
feat/autoround-quarot
fix_bug0627
fix_bug_0722
fix_bug_1105
fix_compile
fix_compile_1
fix_disable_act_dynamic_usage_in_mxfp.py
fix_dq
fix/fp-layers-deprecation-mapping
fix_gemma3_issue
fix_gguf_fp8
fix_gptqmodel
fix/issue-1751-fp8-ct-export
fix_low_cpu
fix_rotation
fix_save_quantized_func_nvfp_checker
fix_0107
fix_0109
fix_0113
fix-attn-mask-b60
fix-ds
fix-flashinfer
fix-gpt-oss
fix-hpu
fix-to-meta-assertion-error-1499
fixbug_0717
fp4_v2
fp4_v3
fp8-cache
fp8-cache-based-export
fp8-static-quant-patch
fp8_export_backup_stable
fp8_export_for_test
good-flux
hengguo/fix_cuda_ut
hengguo/fix_gguf_ds
hengguo/fix_qwen_bug
hengguo/quantizers
hengguo/refactor_calib
hengguo/refactor_init
hengguo/refactor_quant_step1
hengguo/smoothquant
hengguo/w4afp8_sim
henguo/update_so
hpu_only_kg
hpu_only_pkg
hpu/only/v1
hpu-limit-tran
kaihui/torch_dtype
lazy-model-replace
leq_opub
lib/pre-4.4.0
llama/new/9-610
llama/new/9
llm-main
llmc
llmc-backup
llmc-test
lm-head-quant
load-kv
load-w8a8-replace-mod
load-w8a8
lvl/autoscheme_ram_opt
lvl/cpu_ram_optimization
lvl/fix_mixed_acc_by_offload
lvl/fix_no_init_weights
lvl/fix_omni_long_audio
lvl/fix_vlm_large_ram_issue
lvl/fix_vlm_large_vram
lvl/general_moe_replacement
lvl/support_bagel_mot
lvl/support_fp8_with_ark
lvl/support_hunyuan_image
lvl/support_turbo_quant
lvl/support_wan2.2
lyt/numpy_fix
lyt/omni
main
marlin_modify
mengni/bug_fix
mengni/expert
mengni/mengni/block_wise
mengni/new_vllm
mengni/vllm
mengni/vlm
mengniwang95-patch-1
more-ar-ext
mxfp8
origin/block_wise
patch/for/ao/581/stable
patch-for-ao-2
pr1775-followup
pre-release/internal-inc/w4a8
quant-attn-hpu
quant-attn-hpu-o-scale
quant-attn-hpu-pr
quant-llama
quarot-llama
qwen3-vl
qwen3_vl_moe
qwen-split
qwen-v5
refine_device
refine-doc-table
replace-lm-head
revert_order
revert-318-fix/hpu/check
revert-1231-set_disable_opt_rtn_default_2_none
revert-1562-suyue/ut
save_memory
set_disable_opt_rtn_default_2_none
static_quant
support_gemma4
suyue/ark-ci
suyue/model
test-git
try_new_optimizer
try_to_fix_hadamard_regression
update_fp_compile
update_0522
update_0819
upstream-ao
use-ep
ut-time
v0.7.0rc
v0.7.1rc
v0.8.0rc
v0.8.0rc2
v0.9.1rc
v0.9.2-release
v0.9.2rc
v0.9.3rc
v0.9.4rc
v0.9.5rc
v0.9.6rc
v0.9.7rc
v0.10.0rc
v0.10.1rc
v0.10.2rc
v0.10.3rc
v0.12.0rc
v0.12.1rc
v0.12.2rc
v0.12.3rc
w4a4_int_quaro
w4int8dynamic
wangchang/fix_oom
wenhuach21-patch-1
wfp8-afp8-bk
xin3he-patch-1
xinhe/3-20c
xinhe/3-27c
xinhe/3-27d
xinhe/4-7
xinhe/4-15
xuehao/test_gptq
zhenzhong/arformat_fp8
zhenzhong/toolkit_release
update
wenhuach21
committed
13 days ago
096e5cdc
fix
wenhuach21
committed
13 days ago
84ecad2b
update
wenhuach21
committed
13 days ago
dba8c7bf
add int4 scheme
wenhuach21
committed
13 days ago
332266b1
fix rtn bug
wenhuach21
committed
21 days ago
0374bd20
much better result for INT4
wenhuach21
committed
21 days ago
abaea6e2
support unfused
wenhuach21
committed
25 days ago
8095ab12
update
wenhuach21
committed
25 days ago
2e1a3901
update
wenhuach21
committed
25 days ago
b860b406
try to support act clip
wenhuach21
committed
25 days ago
f1c42cca
tmp change
wenhuach21
committed
26 days ago
393950b3
[pre-commit.ci] auto fixes from pre-commit.com hooks
pre-commit-ci[bot]
committed
27 days ago
f8f1a349
split rtn and optrtn
wenhuach21
committed
27 days ago
73598522
[pre-commit.ci] auto fixes from pre-commit.com hooks
pre-commit-ci[bot]
committed
27 days ago
21ce3a77
Merge branch 'hadamard_change' of https://github.com/intel/auto-round into hadamard_change
wenhuach21
committed
27 days ago
7f9aab8d
update
wenhuach21
committed
27 days ago
d98bf85c
Apply suggestions from code review
wenhuach21
committed
27 days ago
Verified
39a61893
Merge branch 'hadamard_change' of https://github.com/intel/auto-round into hadamard_change
wenhuach21
committed
27 days ago
c937152d
update
wenhuach21
committed
27 days ago
d55d0107
[pre-commit.ci] auto fixes from pre-commit.com hooks
pre-commit-ci[bot]
committed
28 days ago
f49e7baf
fix
wenhuach21
committed
28 days ago
c52dc83e
fix
wenhuach21
committed
28 days ago
c78214d7
[pre-commit.ci] auto fixes from pre-commit.com hooks
pre-commit-ci[bot]
committed
28 days ago
1d941021
Merge branch 'hadamard_change' of https://github.com/intel/auto-round into hadamard_change
wenhuach21
committed
28 days ago
d9614255
remove 0.9 scale in act quantization
wenhuach21
committed
28 days ago
0678b0d2
upate
wenhuach21
committed
28 days ago
9ad51135
[pre-commit.ci] auto fixes from pre-commit.com hooks
pre-commit-ci[bot]
committed
28 days ago
ea4804d5
Merge branch 'hadamard_change' of https://github.com/intel/auto-round into hadamard_change
wenhuach21
committed
28 days ago
f1173e86
support group_size
wenhuach21
committed
28 days ago
bf79387c
[pre-commit.ci] auto fixes from pre-commit.com hooks
pre-commit-ci[bot]
committed
31 days ago
ef99f905
Older