optimum
a31e59ed - Add ORT inference (#113)

Commit
3 years ago
Add ORT inference (#113) * added gpu extras and added > transformers for token-classification pipeline issue * added numpy and huggingface hub to required packages * added modeling_* classes * adding tests and pipelines * remove vs code folder * added test model and adjusted gitignore * add readme for tests * working tests * added some documentation * will ci run? * added real model checkpoints * test ci * fix styling * fix some documentation * more doc fixes * added some feedback and wording from michael and lewis * renamed model class to ORTModelForXX * moved from_transformers to from_pretrained * applied ellas feedback * make style * first version of ORTModelForCausalLM without past-keys * added first draft of new .optimize method * added better quantize method * fix import * remove optimize and quantize * added lewis feedback * added style for test * added >>> to code snippets * style * added condition for staging tests * feedback morgan & michael * added action * forgot to install pytest * forgot sentence piece * made sure we won't have import conflicts * make style happy
Author
Parents
  • .github/workflows
    • File
      test_modeling_ort.yml
  • File
    .gitignore
  • docs/source
    • File
      _toctree.yml
    • onnxruntime
      • File
        modeling_ort.mdx
    • File
      pipelines.mdx
    • File
      quickstart.mdx
  • optimum
    • File
      modeling_base.py
    • onnxruntime
      • File
        __init__.py
      • File
        modeling_ort.py
      • File
        utils.py
    • File
      pipelines.py
    • utils
      • File
        __init__.py
      • File
        testing_utils.py
  • File
    setup.py
  • tests
    • File
      README.md
    • assets
      • hub
        • File
          config.json
      • onnx
        • File
          config.json
        • model.onnx
    • onnxruntime
      • File
        test_modeling_ort.py
    • File
      test_modeling_base.py