### Is your feature request related to a problem? Please describe Allow to run models with vllm ### Describe the solution you'd like Add a new inference server for vllm ### Describe alternatives you've considered _No response_ ### Additional context _No response_