xla
Add w8a8 quantized matmul torchxla wrapper
#9290
Merged

Add w8a8 quantized matmul torchxla wrapper #9290

vanbasten23
vanbasten23 added the tests and the quantized matmul kernel. All test passed.
b6f5d4b7
vanbasten23 added tuned table and unit test
f1c96aa2
vanbasten23 fix the test
3609fa73
vanbasten23 fix a test
b0d3aa43
vanbasten23 convert scalar type to f32
6adfdd9b
vanbasten23 fix linter
ad184f42
vanbasten23 add vmem_limit_bytes to static_argnames
f699c0cf
vanbasten23 removed int4_weight
317c7282
vanbasten23 fix comments
6836baee
vanbasten23 vanbasten23 requested a review from yaochengji yaochengji 114 days ago
vanbasten23 vanbasten23 changed the base branch from master to xiowei/add_w8a8_quantized_matmul_kernel_v5e 114 days ago
vanbasten23 fix comments
a48e36d3
vanbasten23 add non-dynamo wrapper. Failing with vmem_limit is a tracer.
9c04c2bc
vanbasten23 non-dynamo case passed.
b4323593
vanbasten23 all tests pass.
2f9af5c2
vanbasten23 rebased
277d4b7d
vanbasten23 vanbasten23 force pushed from 2ce56ccc to 277d4b7d 113 days ago
yaochengji
yaochengji approved these changes on 2025-06-04
vanbasten23 vanbasten23 changed the base branch from xiowei/add_w8a8_quantized_matmul_kernel_v5e to master 113 days ago
vanbasten23 Merge branch 'master' into xiowei/add_w8a8_quantized_matmul_torchxla_…
8637b50a
vanbasten23
vanbasten23 vanbasten23 merged a6f2b27d into master 113 days ago

Login to write a write a comment.

Login via GitHub

Reviewers
Assignees
No one assigned
Labels
Milestone