refactored validation of test config, add skip for AMD w/CUDA Graph
2c88ab1a
cuda graph requires 1.10+
0872f9f7
ensure torch > 1.10 if cuda graph enabled, also fix typo
1d08254f
prune incompatible model+task pairs
816e746a
add pretty print names for test fixtures
42841c6e
use fp16 revision of gpt-j
2677d20c
convert pipe model to half for non-DS case
3ec3f9f8
Merge branch 'master' into jeffra/fill-mask-tests
d37f9d9f
jeffra
changed the title [inference] add fill-mask and q&a tests for bert/roberta [inference] test suite for ds-kernels (bert, roberta, gpt2, gpt-neo, gpt-j)3 years ago
remove dep on hf-accelerate
8b88eeff
manual convert gpt-j fp32->fp16, seeing issues with fp16 revision
50cbfced
force inference tests to be seq
f3e40aea
added correctness tests for text-generation gpt models
d3f6dfe9
fixed device string for lm tests
71a58405
during injection reshape data on cpu to avoid extra GPU memory
f99fc0d7
increased acceptable ppl diff val
1d21343a
added perf test and fix for gpt-j-6B OOM on V100 16GB
Login to write a write a comment.
Login via GitHub