Make Transformers more torch-exportable and dynamo-friendly (#42317)
* make vlms export friendly
* seq2seq lms
* biogpt
* more vlms
* colqwen2
* vision models
* more vlms
* more vlms
* more vlms
* vectorized vision embedding
* fixup
* more vlms
* more vlms
* generate_masks_with_special_tokens_and_transfer_map
* custom torch_check
* use custom torch_check
* revert grounding dino changes
* fixup
* remove file
* undo
* undo
* testing
* fixes
* standard error message
* use torch._check_with to raise value error instead of torch._check's runtime error
* fix recurrent gemma
* only itemize tensors
* use spatial shapes list instead of tensor
* fix udop use_cache default value
* use tracable condition for seq2seq lms
* make smolvlm exportable
* fix fastvlm and t5gemma2
* fix qwen2_audio and idefics
* remove script
* tbc
* skip mra model
* helper
* style and document
* fix
* set experts impl to batched
* make xmod exportable and efficient
* make more ssms exportable
* fix
* revert recurrent gemma
* skip models that use chunked attention or rope_index
* qwen3_next
* assert async
* tensorize (mm) grounding dino mask generation
* style
* fix repo
* address comments
* fix qwen2 audio and vits checks
* skip two models using kernels by default
* skip granite moe hybrid using custom kernels
* disable mamba kernels
* vits splinter and videomae