support real lm-head quantization and mixed precision inference #114
fix a bug in example
7d020db2
Merge branch 'main' of https://github.com/intel/auto-round
fbe69d5c
Merge branch 'main' of https://github.com/intel/auto-round
596a18f5
Merge branch 'main' of https://github.com/intel/auto-round
10add8ce
Merge branch 'main' of https://github.com/intel/auto-round
003b60a6
Merge branch 'main' of https://github.com/intel/auto-round
9d495140
Merge branch 'main' of https://github.com/intel/auto-round
3b7f386c
Merge branch 'main' of https://github.com/intel/auto-round
d3f14df6
Merge branch 'main' of https://github.com/intel/auto-round
76e4d909
Merge branch 'main' of https://github.com/intel/auto-round
08e46acb
Merge branch 'main' of https://github.com/intel/auto-round
15b756b0
Merge branch 'main' of https://github.com/intel/auto-round
40425bbd
Merge branch 'main' of https://github.com/intel/auto-round
f0b9ad00
Merge branch 'main' of https://github.com/intel/auto-round
04e70eca
Merge branch 'main' of https://github.com/intel/auto-round
43811bb9
Merge branch 'main' of https://github.com/intel/auto-round
c8817199
Merge branch 'main' of https://github.com/intel/auto-round
54920e56
Merge branch 'main' of https://github.com/intel/auto-round
e2c2f56c
Merge branch 'main' of https://github.com/intel/auto-round
4f718d45
Merge branch 'main' of https://github.com/intel/auto-round
022988aa
Merge branch 'main' of https://github.com/intel/auto-round
b4eb6790
Merge branch 'main' of https://github.com/intel/auto-round
0abbde8e
fix gradient_accmulate bug in lm-head
aed8be0d
init update
a129f011
[pre-commit.ci] auto fixes from pre-commit.com hooks
892cd8e7
fix bias zero and Non key mismatch, but autogptq way is not optimal
48fd07ee
Merge branch 'autoroundmodel' of https://github.com/intel/auto-round …
78b3fd10
[pre-commit.ci] auto fixes from pre-commit.com hooks
70960a8f
wip
1deda104
Merge branch 'autoroundmodel' of https://github.com/intel/auto-round …
fcb28573
[pre-commit.ci] auto fixes from pre-commit.com hooks
208ea37c
basically finished the exporting part, the polish of loading code is …
72dba51e
fix conflict
e791d863
fix typo
ec28b729
[pre-commit.ci] auto fixes from pre-commit.com hooks
f99b490e
basically finish the loading part
31e4bd56
fix conflict
6adf327d
[pre-commit.ci] auto fixes from pre-commit.com hooks
2ee273b3
fix some issues
03dd7f99
[pre-commit.ci] auto fixes from pre-commit.com hooks
ad2f3def
support lm eval for quantized lm-head model
8f2aa970
Merge branch 'autoroundmodel' of https://github.com/intel/auto-round …
5b26ccd7
force lm head to use gradient accumulate
b1229939
revert debug code
4e3e9716
tiny change
5f13bb0c
Merge branch 'main' into autoroundmodel
c8757264
Merge branch 'autoroundmodel' of https://github.com/intel/auto-round …
33d84537
[pre-commit.ci] auto fixes from pre-commit.com hooks
e5fc181b
fix line too long issue
209e1d04
Merge branch 'autoroundmodel' of https://github.com/intel/auto-round …
83424dec
[pre-commit.ci] auto fixes from pre-commit.com hooks
6a74a7df
wenhuach21
changed the title [WIP] support real lm-head quantization and mixed precision inference support real lm-head quantization and mixed precision inference 1 year ago
add mistral v0.1 lm-head acc and model
59ca3596
wenhuach21
deleted the autoroundmodel branch 1 year ago
Assignees
No one assigned
Login to write a write a comment.
Login via GitHub