DeepSpeed
FP [6,8,12] quantizer op
#5336
Merged

FP [6,8,12] quantizer op #5336

jeffra
jeffra test
010fa309
jeffra fp[6,8,12] quantizer op
99367962
jeffra op builder
cc3096f3
jeffra cleanup
483d8d93
jeffra jeffra requested a review from mrwyattii mrwyattii 1 year ago
jeffra jeffra requested a review from awan-10 awan-10 1 year ago
jeffra jeffra requested a review from arashb arashb 1 year ago
jeffra jeffra requested a review from tjruwase tjruwase 1 year ago
jeffra jeffra requested a review from loadams loadams 1 year ago
jeffra
jeffra
jeffra fp quantizer assumes ampere and above arch, also disable ninja for pr…
a53cd0b1
jeffra ifdef bf16 in reduction utils
4092e467
jeffra skip on cpu
90c13db7
jeffra cannot run fp quant on v100
bf390f4d
jeffra fix missing import
ee4aa3f9
jeffra move qtorch import to be after pytest skip
23c06bf5
jeffra Merge branch 'master' into fp-quantizer
fd564a72
loadams
jeffra
loadams
mrwyattii
mrwyattii approved these changes on 2024-04-03
JamesTheZ
JamesTheZ commented on 2024-04-03
sfc-gh-reyazda
JamesTheZ
JamesTheZ approved these changes on 2024-04-04
loadams loadams merged 3fbd01cc into master 1 year ago

Login to write a write a comment.

Login via GitHub

Assignees
No one assigned
Labels
Milestone