Go
Home
Pricing
FAQ
Install
Home
Pricing
FAQ
Install
Login
via GitHub
intel/auto-round
Pull Requests
Commits
fp8_export_backup_stable
AutoAdamRound_bugfix
actvation_quant
add_task_args_for_lmeval
autoround_support_qbits_backend
bf16_scale
debug_time_cost
debug-nvfp4
deepseekv3
ds-qwen
enable_llama4_int8_baseline
enable_llama4_quant
enable_mxfp_exporting
fast_config
fix_bug0627
fix_bug_0722
fix_bug_1105
fix_dq
fix_gemma3_issue
fix_gguf_fp8
fix_save_quantized_func_nvfp_checker
fix-attn-mask-b60
fix-ds
fix-gpt-oss
fix-hpu
fixbug_0717
fp4_v2
fp8-cache
fp8-cache-based-export
fp8_export_backup_stable
fp8_export_for_test
hengguo/fix_cuda_ut
hengguo/fix_gguf_ds
hengguo/quantizers
hengguo/smoothquant
hengguo/w4afp8_sim
henguo/refactor_format_step2
henguo/update_so
hpu_only_kg
hpu_only_pkg
hpu/only/v1
kaihui/torch_dtype
leq_opub
lib/pre-4.4.0
llama/new/9-610
llama/new/9
llm-main
llmc
llmc-backup
llmc-test
lm-head-quant
load-kv
load-w8a8-replace-mod
load-w8a8
lyt/numpy_fix
lyt/omni
main
marlin_modify
mengni/arg_update
mengni/bug_fix
mengni/expert
mengni/vlm
mengniwang95-patch-1
mlperf-awq
more-ar-ext
mxfp8
new_teq
patch/for/ao/581/stable
patch-for-ao-2
pre-release/internal-inc/w4a8
quant-attn-hpu
quant-attn-hpu-o-scale
quant-attn-hpu-pr
quant-llama
qwen3-vl
qwen3_vl_moe
qwen-split
refine-doc-table
replace-lm-head
revert_order
revert-318-fix/hpu/check
save_memory
static_quant
suyue/ci
suyue/fix
suyue/version
test-git
tmp
try_new_optimizer
update_fp_compile
update_0522
update_0819
upstream-ao
use-ep
ut-time
v0.7.0rc
v0.7.1rc
v0.8.0rc
v0.8.0rc2
v0.9.1rc
v0.9.2-release
v0.9.2rc
v0.9.3rc
v0.9.4rc
w4a4_int_quaro
w4int8dynamic
wfp8-afp8-bk
xinhe/UT
xinhe/avg_bits
xinhe/device_bug
xinhe/eval
xinhe/exp
xinhe/fix_pp
xinhe/hp_level
xinhe/llama_tmp
xinhe/mix-precision
xinhe/mp
xinhe/new
xinhe/nvfp4
xinhe/release_bug
xinhe/target_loss_ratio
xinhe/tmp
xinhe/whisper
xuehao/cuda_ut
xuehao/fix_install
xuehao/v0.9.4_release
remove force fp16 dtype export (#192)
WeiweiZhang1
committed
1 year ago
Verified
d9f7ceee
Fix multimodal and moe issue (#191)
WeiweiZhang1
committed
1 year ago
Verified
ca59d36a
low_cpu_mem refinement (#186)
n1ck-guo
committed
1 year ago
Verified
7b9611ed
support autoround hpu format (#182)
yintong-lu
committed
1 year ago
Verified
72f5ce80
add initial support of mxfp4 (#187)
wenhuach21
committed
1 year ago
Verified
d48d0404
[pre-commit.ci] pre-commit autoupdate (#167)
pre-commit-ci[bot]
committed
1 year ago
Verified
9226a893
fix typos (#185)
wenhuach21
committed
1 year ago
Verified
24b2e740
Add layer wise mode to save memory (#136)
n1ck-guo
committed
1 year ago
Verified
e2e33f03
enable llava & Qwen-VL multimodal model quantization (#165)
WeiweiZhang1
committed
1 year ago
Verified
aeb9e408
Fix UT coverage report (#180)
XuehaoSun
committed
1 year ago
Verified
8d08400f
fix autoround format mixed precision issue and refine gptq format code (#183)
wenhuach21
committed
1 year ago
Verified
e2814996
Fix autoround format accuracy issue (#179)
wenhuach21
committed
1 year ago
Verified
0126180f
Add unit test (#173)
XuehaoSun
committed
1 year ago
Verified
81624095
add initial support for activation quantization (#176)
wenhuach21
committed
1 year ago
Verified
5f67048c
speedup the tuning a little (#175)
wenhuach21
committed
1 year ago
Verified
473f474d
add chat template in calib tokenization (#171)
yintong-lu
committed
1 year ago
Verified
735dfc9e
[Large impact]set the default nsamples to 128 and low_gpu_mem_usage to False (#174)
wenhuach21
committed
1 year ago
Verified
ab614824
support marlin in auto_round format (#172)
wenhuach21
committed
1 year ago
Verified
2b1448d4
revert the gptq format code to fix the regression (#168)
wenhuach21
committed
1 year ago
Verified
5947e9c0
fix typos, update overview img (#166)
WeiweiZhang1
committed
1 year ago
Verified
8d5765ac
1 fix a bug in autoround format with the latest transformers 2 rename n_samples n_blocks to nsamples nblocks (#163)
wenhuach21
committed
1 year ago
Verified
f9e7d79e
bugfix (#160)
WeiweiZhang1
committed
1 year ago
Verified
31c566cc
fix bug and limit numpy version (#159)
yintong-lu
committed
1 year ago
Verified
77320b0a
support calibration dataset concat (#147)
yintong-lu
committed
1 year ago
Verified
75e3fde0
remove gpt ppl eval from lm-0.4.2 (#158)
wenhuach21
committed
1 year ago
Verified
77d6a886
fix bug at whole block is excluded from quantization (#156)
wenhuach21
committed
1 year ago
Verified
edcec56e
auto round quantizer supports gptq kernel (#155)
wenhuach21
committed
1 year ago
Verified
9cae103d
fix qbits issue (#153)
wenhuach21
committed
1 year ago
Verified
c313fa33
Qbits related log (#151)
zhewang1-intc
committed
1 year ago
Verified
34274fb3
autoround_support_qbits_backend (#145)
zhewang1-intc
committed
1 year ago
Verified
dbdc4a39
Newer
Older