vllm
[Misc][OpenAI] deprecate max_tokens in favor of new max_completion_tokens field for chat completion endpoint
#9837
Merged

  • benchmarks
    • File
      backend_request_func.py
  • docs/source/serving
    • File
      run_on_sky.rst
  • examples
    • File
      offline_inference_openai.md
    • File
      openai_api_client_for_multimodal.py
    • File
      openai_example_batch.jsonl
  • File
    requirements-common.txt
  • tests
    • entrypoints/openai
      • File
        test_audio.py
      • File
        test_chat.py
      • File
        test_vision.py
    • tool_use
      • File
        test_chat_completions.py
      • File
        test_parallel_tool_calls.py
      • File
        test_tool_calls.py
  • vllm/entrypoints/openai
    • File
      protocol.py
    • File
      serving_engine.py

Loading comments...