Go
Home
Pricing
FAQ
Install
Home
Pricing
FAQ
Install
Login
via GitHub
intel/auto-round
Pull Requests
Commits
hpu_only_kg
AutoAdamRound_bugfix
actvation_quant
add_task_args_for_lmeval
autoround_support_qbits_backend
bf16_scale
copilot/fix-corner-case-in-auto-round
copilot/fix-deprecated-fp-layers-handling
copilot/fix-issue-with-auto-rounding
copilot/fix-llm-type-70b-bits-setting
copilot/fix-typeerror-wrapped-fn
copilot/sub-pr-1237-again
copilot/sub-pr-1237
debug_time_cost
debug-nvfp4
deepseekv3
ds-qwen
enable_llama4_int8_baseline
enable_llama4_quant
enable_mxfp_exporting
fast_config
fix_bug0627
fix_bug_0722
fix_bug_1105
fix_dq
fix/fp-layers-deprecation-mapping
fix_gemma3_issue
fix_gguf_fp8
fix_low_cpu_new
fix_low_cpu
fix_save_quantized_func_nvfp_checker
fix_0107
fix_0109
fix_0113
fix-attn-mask-b60
fix-ds
fix-gpt-oss
fix-hpu
fixbug_0717
fp4_v2
fp8-cache
fp8-cache-based-export
fp8_export_backup_stable
fp8_export_for_test
hengguo/bug_fix_0115
hengguo/fix_cuda_ut
hengguo/fix_gguf_ds
hengguo/quantizers
hengguo/refactor_quant_step1
hengguo/smoothquant
hengguo/w4afp8_sim
henguo/update_so
hpu_only_kg
hpu_only_pkg
hpu/only/v1
kaihui/torch_dtype
lazy-model-replace
leq_opub
lib/pre-4.4.0
llama/new/9-610
llama/new/9
llm-main
llmc
llmc-backup
llmc-test
lm-head-quant
load-kv
load-w8a8-replace-mod
load-w8a8
lyt/numpy_fix
lyt/omni
main
marlin_modify
mengni/bug_fix
mengni/expert
mengni/fp8_sdpa
mengni/vllm
mengni/vlm
mengniwang95-patch-1
mlperf-awq
more-ar-ext
mxfp8
new_teq
patch/for/ao/581/stable
patch-for-ao-2
pre-release/internal-inc/w4a8
quant-attn-hpu
quant-attn-hpu-o-scale
quant-attn-hpu-pr
quant-llama
qwen3-vl
qwen3_vl_moe
qwen-split
refine_moe_modelling_2_reduce_peak_ram_usage
refine-doc-table
replace-lm-head
revert_order
revert-318-fix/hpu/check
revert-1231-set_disable_opt_rtn_default_2_none
save_memory
set_disable_opt_rtn_default_2_none
static_quant
suyue/ci
suyue/version
test-git
tmp
try_new_optimizer
update_fp_compile
update_0522
update_0819
upstream-ao
use-ep
ut-time
v0.7.0rc
v0.7.1rc
v0.8.0rc
v0.8.0rc2
v0.9.1rc
v0.9.2-release
v0.9.2rc
v0.9.3rc
v0.9.4rc
v0.9.5rc
w4a4_int_quaro
w4int8dynamic
wfp8-afp8-bk
xinhe/UT
xinhe/avg_bits
xinhe/device_bug
xinhe/eval
xinhe/exp
xinhe/export
xinhe/fix_pp
xinhe/fix_xpu_ci
xinhe/hp_level
xinhe/llama_tmp
xinhe/mix-precision
xinhe/mp
xinhe/new
xinhe/nvfp4
xinhe/release_bug
xinhe/target_loss_ratio
xinhe/tmp
xinhe/whisper
xuehao/cuda_ut
xuehao/fix_install
xuehao/v0.9.4_release
xuehao/version
add simple imports test
yiliu30
committed
1 year ago
98467a69
add requirements for hpu
yiliu30
committed
1 year ago
4db14d20
mllm eval bug fix (#297)
n1ck-guo
committed
1 year ago
Verified
0bb70a64
eval for MLLMs (#296)
n1ck-guo
committed
1 year ago
Verified
4384914e
refine forward hook (#290)
WeiweiZhang1
committed
1 year ago
Verified
25d977b3
use torch.compile by default for PyTorch versions 2.6 and above (#295)
wenhuach21
committed
1 year ago
Verified
c922f5b3
[Experimental Feature]support for common hf multimodel (#276)
n1ck-guo
committed
1 year ago
Verified
e6432125
fix bug of backend (#294)
wenhuach21
committed
1 year ago
Verified
4f228717
fix ipex tqdm mismatch issue (#293)
wenhuach21
committed
1 year ago
Verified
487abd6f
Add ipex support for intel cpu (#292)
wenhuach21
committed
1 year ago
Verified
168a1f69
Refine code (#291)
wenhuach21
committed
1 year ago
Verified
f41094a9
update readme (#287)
wenhuach21
committed
1 year ago
Verified
8efff6f4
fix mx_fp issues (#286)
wenhuach21
committed
1 year ago
Verified
99cff1fb
avoid deterministic algorithm warning in inference (#285)
wenhuach21
committed
1 year ago
Verified
ba5be40a
update readme for cpu inference
wenhuach21
committed
1 year ago
Verified
141c149f
update readme for v0.3.1 release (#283)
wenhuach21
committed
1 year ago
Verified
a3592220
refine AuoRound format and support marlin repacking (#280)
wenhuach21
committed
1 year ago
Verified
68138e82
qwen2_bugfix, add adamround vision UT (#281)
WeiweiZhang1
committed
1 year ago
Verified
7cfff967
refine eval (#282)
wenhuach21
committed
1 year ago
Verified
afa9e262
[Important update]set full range sym as the default (#278)
wenhuach21
committed
1 year ago
Verified
00122bc6
adamround bugfix, refine import (#275)
WeiweiZhang1
committed
1 year ago
Verified
a633aa70
change to even rounding for mantissa of mx_fp (#277)
wenhuach21
committed
1 year ago
Verified
98a9c755
fix mutable default value (#272)
wenhuach21
committed
1 year ago
Verified
8bf63f39
enable llama3.2-vision model quantization (#269)
WeiweiZhang1
committed
1 year ago
Verified
6b99d10a
keep the dtype after qdq (#268)
wenhuach21
committed
1 year ago
Verified
3a70be84
remove g_idx in gptq format (#267)
wenhuach21
committed
1 year ago
Verified
fdfd9711
Update readme for VLM support and integration (#266)
wenhuach21
committed
1 year ago
Verified
af3db170
Add a warning for improper export formats. (#265)
wenhuach21
committed
1 year ago
Verified
be32686b
Fix 3bit packing for auto-gptq format (#264)
wenhuach21
committed
1 year ago
Verified
6ee91a9f
better support quant_lm_head for larger models (#263)
wenhuach21
committed
1 year ago
Verified
82322ac9
Older