You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
We should rather look at response.model for chat completion requests because this field is populated by the model (or adapter) itself.
The model/adapter must know better which model has been actually called.
Note also that model field doesn't exist in the chat completion request according to Azure OpenAI API:
Currently the
model
is derived fromrequest.model
field, which is not reliable, since the user may have provided anything in this field.ai-dial-analytics-realtime/aidial_analytics_realtime/app.py
Lines 105 to 107 in 68ab942
We should rather look at
response.model
for chat completion requests because this field is populated by the model (or adapter) itself.The model/adapter must know better which model has been actually called.
Note also that
model
field doesn't exist in the chat completion request according to Azure OpenAI API:https://github.com/Azure/azure-rest-api-specs/blob/main/specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2024-12-01-preview/inference.yaml#L2188
At the same time,
model
fields is a required field in the chat completion response:https://github.com/Azure/azure-rest-api-specs/blob/main/specification/cognitiveservices/data-plane/AzureOpenAI/inference/preview/2024-12-01-preview/inference.yaml#L4152
We should probably keep
request.model
as a fallback from a missingresponse.model
for the special case of theassistant
deployment:assistant
service receives the deployment id of the model it's going to call via therequest.model
field.Whether it makes any sense to report the
model
field for non-model deployments (applications and assisstant) in the first place, remains unclear.The text was updated successfully, but these errors were encountered: