Go
Home
Pricing
FAQ
Install
Home
Pricing
FAQ
Install
Login
via GitHub
intel/auto-round
Pull Requests
Commits
lyt/omni
AutoAdamRound_bugfix
actvation_quant
add_task_args_for_lmeval
autoround_support_qbits_backend
bf16_scale
debug_time_cost
debug-nvfp4
deepseekv3
ds-qwen
enable_llama4_int8_baseline
enable_llama4_quant
enable_mxfp_exporting
fast_config
fix_bug0627
fix_bug_0722
fix_bug_1105
fix_dq
fix_gemma3_issue
fix_gguf_fp8
fix_save_quantized_func_nvfp_checker
fix-attn-mask-b60
fix-ds
fix-gpt-oss
fix-hpu
fixbug_0717
fp4_v2
fp8-cache
fp8-cache-based-export
fp8_export_backup_stable
fp8_export_for_test
hengguo/fix_cuda_ut
hengguo/fix_gguf_ds
hengguo/quantizers
hengguo/smoothquant
hengguo/w4afp8_sim
henguo/refactor_format_step2
henguo/update_so
hpu_only_kg
hpu_only_pkg
hpu/only/v1
kaihui/torch_dtype
leq_opub
lib/pre-4.4.0
llama/new/9-610
llama/new/9
llm-main
llmc
llmc-backup
llmc-test
lm-head-quant
load-kv
load-w8a8-replace-mod
load-w8a8
lyt/numpy_fix
lyt/omni
main
marlin_modify
mengni/bug_fix
mengni/expert
mengni/vlm
mengniwang95-patch-1
mlperf-awq
more-ar-ext
mxfp8
new_teq
patch/for/ao/581/stable
patch-for-ao-2
pre-release/internal-inc/w4a8
quant-attn-hpu
quant-attn-hpu-o-scale
quant-attn-hpu-pr
quant-llama
qwen3-vl
qwen3_vl_moe
qwen-split
refine-doc-table
replace-lm-head
revert_order
revert-318-fix/hpu/check
save_memory
set_disable_opt_rtn_as_int8_default
static_quant
suyue/ci
suyue/fix
suyue/version
test-git
tmp
try_new_optimizer
update_fp_compile
update_0522
update_0819
upstream-ao
use-ep
ut-time
v0.7.0rc
v0.7.1rc
v0.8.0rc
v0.8.0rc2
v0.9.1rc
v0.9.2-release
v0.9.2rc
v0.9.3rc
w4a4_int_quaro
w4int8dynamic
wfp8-afp8-bk
xinhe/UT
xinhe/avg_bits
xinhe/device_bug
xinhe/eval
xinhe/exp
xinhe/fix_pp
xinhe/hp_level
xinhe/llama_tmp
xinhe/mix-precision
xinhe/mp
xinhe/new
xinhe/nvfp4
xinhe/release_bug
xinhe/target_loss_ratio
xinhe/tmp
xinhe/whisper
xuehao/cuda_ut
xuehao/fix_install
code update
yintong-lu
committed
1 year ago
508517f8
update
yintong-lu
committed
1 year ago
f0e7cadb
merge main into branch and fix conflicts
yintong-lu
committed
1 year ago
1bf6e627
fix typos (#140)
WeiweiZhang1
committed
1 year ago
Verified
4d2d2591
add files, WIP
yintong-lu
committed
1 year ago
b46dcad7
bump version into v0.2 (#139)
chensuyue
committed
1 year ago
Verified
aafb82ef
handling transformers version compatibility in lmhead export, bugfix (#130)
WeiweiZhang1
committed
1 year ago
Verified
4db22e1d
fix export issue with torch 2.0 (#129)
wenhuach21
committed
1 year ago
Verified
5bff86ee
Update falcon recipe (#128)
wenhuach21
committed
1 year ago
Verified
416ec7e9
fix falcon quant issue with disable_trust_remote_code (#126)
WeiweiZhang1
committed
1 year ago
Verified
e2985fdf
Update phi2 recipe (#124)
wenhuach21
committed
1 year ago
Verified
edca2980
remove fp32 conversion in exporting to autogptq (#123)
wenhuach21
committed
1 year ago
Verified
17024b16
update gemma recipe (#121)
wenhuach21
committed
1 year ago
Verified
ecc1dd65
Remove unused hook (#122)
XuehaoSun
committed
1 year ago
Verified
c7751c49
support `transformers.Conv1D` packing (#118)
Kaihui-intel
committed
1 year ago
Verified
02a6660e
Fix export format issue (#120)
wenhuach21
committed
1 year ago
Verified
ed29cf50
Fix doc (#119)
wenhuach21
committed
1 year ago
Verified
7bc9d8fb
fix lm-head quant issue at disable_quanted_input (#117)
wenhuach21
committed
1 year ago
Verified
b51cfa98
support real lm-head quantization and mixed precision inference (#114)
wenhuach21
committed
1 year ago
Verified
4d1caebb
fix lm-head gradient accumulation bug (#113)
wenhuach21
committed
1 year ago
Verified
ecca5349
update shells (#112)
WeiweiZhang1
committed
1 year ago
Verified
3c214db0
Adjust the default evaluation data type by selecting from the model path configuration (#107)
WeiweiZhang1
committed
1 year ago
Verified
c42eaa92
20% speedup by removing new zero tensor (#110)
wenhuach21
committed
1 year ago
Verified
c7434b63
1.8X speedup by disable_low_gpu_mem_usage and reduce memory usage by avoid using torch.cat (#106)
wenhuach21
committed
1 year ago
Verified
23b60c3a
remove costly operations
wenhuach21
committed
1 year ago
Verified
ad3a7bba
Consolidate dataloader&dataset_split to dataset (#105)
wenhuach21
committed
1 year ago
Verified
1f9cb4f8
disable quantizing lm-head with tied weights as a workaround (#102)
wenhuach21
committed
1 year ago
Verified
1fe6aae5
disable quantizing lm-head with tied weights as a workaround (#101)
wenhuach21
committed
1 year ago
Verified
c839e825
update readme of calibration dataset and lm-head usage (#98)
wenhuach21
committed
1 year ago
Verified
f95b8c7d
fix critic bug for gradient_accumulate_steps!=1 and reduce cpu memory of lm-head tuning (#97)
WeiweiZhang1
committed
1 year ago
Verified
7dd02eb4
Older