Make inference requests on your model APIs.
run.blaxel.ai/your-workspace/models/your-model
(the base endpoint) will generate text based on a promptrun.blaxel.ai/your-workspace/models/your-model/v1/chat/completions
(the ChatCompletions API implementation) will generate response based on a list of messages--path
: