lmdeploy
lmdeploy copied to clipboard
[Feature] Support `response_format` for `TurboMind`
Motivation
I'm using TurboMind engine and I got an error while requesting response_format with json_schema. Code here:
https://github.com/InternLM/lmdeploy/blob/main/lmdeploy/serve/openai/api_server.py#L367
Is there any plan to support this for TurboMind?
Related resources
if request.response_format and request.response_format.type != 'text':
if VariableInterface.async_engine.backend != 'pytorch':
return create_error_response(
HTTPStatus.BAD_REQUEST,
'only pytorch backend can use response_format now')
response_format = request.response_format.model_dump()
Additional context
No response
Yes. We'll support it in December. Stay tuned.
Recently, we have been busy addressing the needs from our internal team. This feature won't be tackled until we finish them. Sorry for that.
Hey, any updates on this issue?
still waiting for updates for this issue, thanks.