transformers
eff263cd - Make Transformers more torch-exportable and dynamo-friendly (#42317)

Commit
49 days ago
Make Transformers more torch-exportable and dynamo-friendly (#42317) * make vlms export friendly * seq2seq lms * biogpt * more vlms * colqwen2 * vision models * more vlms * more vlms * more vlms * vectorized vision embedding * fixup * more vlms * more vlms * generate_masks_with_special_tokens_and_transfer_map * custom torch_check * use custom torch_check * revert grounding dino changes * fixup * remove file * undo * undo * testing * fixes * standard error message * use torch._check_with to raise value error instead of torch._check's runtime error * fix recurrent gemma * only itemize tensors * use spatial shapes list instead of tensor * fix udop use_cache default value * use tracable condition for seq2seq lms * make smolvlm exportable * fix fastvlm and t5gemma2 * fix qwen2_audio and idefics * remove script * tbc * skip mra model * helper * style and document * fix * set experts impl to batched * make xmod exportable and efficient * make more ssms exportable * fix * revert recurrent gemma * skip models that use chunked attention or rope_index * qwen3_next * assert async * tensorize (mm) grounding dino mask generation * style * fix repo * address comments * fix qwen2 audio and vits checks * skip two models using kernels by default * skip granite moe hybrid using custom kernels * disable mamba kernels * vits splinter and videomae
Parents
Loading