Integrate huggingface_hub inference support (#651)
- Integrated Hugging Face Inference (hf-inference) as a new backend option for judge model evaluation.
- Added async inference logic using AsyncInferenceClient only for hf-inference, with support for retries, timeouts, and provider specification.
- Extended JudgeLM and JudgeLLM classes to support new configuration parameters like hf_provider, base_url, and max_tokens.