Go
Home
Pricing
FAQ
Install
Home
Pricing
FAQ
Install
Login
via GitHub
intel/auto-round
Pull Requests
Commits
AutoAdamRound_bugfix
AutoAdamRound_bugfix
Chinesization
actvation_quant
add_task_args_for_lmeval
ark_zp
autoround_support_qbits_backend
bf16_scale
copilot/fix-corner-case-in-auto-round
copilot/fix-deprecated-fp-layers-handling
copilot/fix-issue-with-auto-rounding
copilot/fix-llm-type-70b-bits-setting
copilot/fix-typeerror-wrapped-fn
copilot/replace-getset-module-torch-api
copilot/speedup-fp8-linear-convert
copilot/speedup-fp8-linear-convert-again
copilot/speedup-fp8-linear-convert-another-one
copilot/sub-pr-1237-again
copilot/sub-pr-1237
copilot/sub-pr-1324
copilot/update-user-settings-page
ddp
debug_time_cost
debug-nvfp4
deepseekv3
ds-qwen
ds-v5
ds-v32
enable_glm4_moe_lite_quantization
enable_llama4_int8_baseline
enable_llama4_quant
enable_mxfp_exporting
fast_config
fix_bug0627
fix_bug_0722
fix_bug_1105
fix_disable_act_dynamic_usage_in_mxfp.py
fix_dq
fix/fp-layers-deprecation-mapping
fix_gemma3_issue
fix_gguf_fp8
fix_gptqmodel
fix_low_cpu
fix_save_quantized_func_nvfp_checker
fix_0107
fix_0109
fix_0113
fix_0204
fix-attn-mask-b60
fix-ds
fix-flashinfer
fix-fp8-model
fix-gpt-oss
fix-hpu
fixbug_0717
fp4_v2
fp8-cache
fp8-cache-based-export
fp8_export_backup_stable
fp8_export_for_test
hengguo/fix_cuda_ut
hengguo/fix_gguf_ds
hengguo/quantizers
hengguo/refactor_init
hengguo/refactor_quant_step1
hengguo/smoothquant
hengguo/w4afp8_sim
henguo/update_so
hpu_only_kg
hpu_only_pkg
hpu/only/v1
kaihui/torch_dtype
lazy-model-replace
leq_opub
lib/pre-4.4.0
llama/new/9-610
llama/new/9
llm-main
llmc
llmc-backup
llmc-test
lm-head-quant
load-kv
load-w8a8-replace-mod
load-w8a8
lvl/cpu_ram_optimization
lvl/fix_no_init_weights
lvl/general_moe_replacement
lvl/ram_usage_optimization
lvl/support_omni
lyt/numpy_fix
lyt/omni
main
marlin_modify
mengni/bug_fix
mengni/expert
mengni/vllm
mengni/vlm
mengniwang95-patch-1
more-ar-ext
mxfp8
new_teq
patch/for/ao/581/stable
patch-for-ao-2
pre-release/internal-inc/w4a8
quant-attn-hpu
quant-attn-hpu-o-scale
quant-attn-hpu-pr
quant-llama
qwen3-vl
qwen3_vl_moe
qwen-split
qwen-v5
refine_global_scale_calculation_2_blockwise
refine-doc-table
replace-lm-head
revert_order
revert-318-fix/hpu/check
revert-1231-set_disable_opt_rtn_default_2_none
save_memory
set_disable_opt_rtn_default_2_none
static_quant
suyue/ci
suyue/test
suyue/version
test-git
tmp
try_new_optimizer
update_fp_compile
update_0522
update_0819
upstream-ao
use-ep
ut-time
v0.7.0rc
v0.7.1rc
v0.8.0rc
v0.8.0rc2
v0.9.1rc
v0.9.2-release
v0.9.2rc
v0.9.3rc
v0.9.4rc
v0.9.5rc
v0.9.6rc
v0.9.7rc
v0.10.0rc
w4a4_int_quaro
w4int8dynamic
wfp8-afp8-bk
xinhe/fix_ci
xinhe/gpt-oss
xinhe/qwen-nvfp4
xinhe/tmp
xinhe/2-10b
xuehao/cuda_ut
xuehao/fix_install
Update model link (#232)
WeiweiZhang1
committed
1 year ago
Verified
4455478e
add meta3.1-70B model, refine docs (#231)
WeiweiZhang1
committed
1 year ago
Verified
a9af34de
add quantized models by 3rd party (#230)
WeiweiZhang1
committed
1 year ago
Verified
985189f2
change the scale thresh generally (#229)
WeiweiZhang1
committed
1 year ago
Verified
b2f160c2
refine example (#224)
WeiweiZhang1
committed
1 year ago
Verified
ac77f1e5
Bump setuptools in /examples/multimodal-modeling/Phi-3-vision (#223)
dependabot[bot]
committed
1 year ago
Verified
0e2c4e1d
Enable phi3v tuning (#197)
WeiweiZhang1
committed
1 year ago
Verified
f2fef130
refine eval_042 to enable parallelize evaluation (#221)
WeiweiZhang1
committed
1 year ago
Verified
948f2282
update readme (#220)
wenhuach21
committed
1 year ago
Verified
1f1fe1a8
[Experimental Feature]fast tuning norm/bias at 2 bits (#208)
wenhuach21
committed
1 year ago
Verified
459c0c5c
add qwen int4 model, refine example (#217)
WeiweiZhang1
committed
1 year ago
Verified
fed34b76
avoid underflow and overflow for exllamav2 (#218)
wenhuach21
committed
1 year ago
Verified
04678e06
fix a bug in autoround format inference (#215)
wenhuach21
committed
1 year ago
Verified
a5f79c40
update xpu format exporting (#214)
WeiweiZhang1
committed
1 year ago
Verified
90d9fc07
remove local pile file (#213)
WeiweiZhang1
committed
1 year ago
Verified
c05c49bc
fix example typo (#212)
WeiweiZhang1
committed
1 year ago
Verified
77a7e601
limit the scale minimum value not to 0 (#211)
WeiweiZhang1
committed
1 year ago
Verified
1dd2708d
modify setup.py (#206)
n1ck-guo
committed
1 year ago
Verified
8610b888
add local file of pile-10k (#198)
WeiweiZhang1
committed
1 year ago
Verified
084f3db2
set autoround format as default to unify CPU/HPU/CUDA (#205)
wenhuach21
committed
1 year ago
Verified
1e75afd1
remove coverage check (#202)
XuehaoSun
committed
1 year ago
Verified
081556f6
support autoawq format (#115)
yintong-lu
committed
1 year ago
Verified
b08b967b
Add setseed in autoround (#201)
WeiweiZhang1
committed
1 year ago
Verified
9718d203
bugfix of groupsize dismatch with weight shape (#195)
WeiweiZhang1
committed
1 year ago
Verified
0daad3f2
fix memory issue (#193)
wenhuach21
committed
1 year ago
Verified
e24b9074
add check of mixed device (#190)
n1ck-guo
committed
1 year ago
Verified
61cf9eef
remove force fp16 dtype export (#192)
WeiweiZhang1
committed
1 year ago
Verified
d9f7ceee
Fix multimodal and moe issue (#191)
WeiweiZhang1
committed
1 year ago
Verified
ca59d36a
low_cpu_mem refinement (#186)
n1ck-guo
committed
1 year ago
Verified
7b9611ed
support autoround hpu format (#182)
yintong-lu
committed
1 year ago
Verified
72f5ce80
Newer
Older