15:44:41.862 [main] INFO c.a.a.i.i.C.completeSync - {"az.sdk.message":"HTTP response","statusCode":404,"url":"https://*.services.ai.azure.com/models/chat/completions?api-version=2024-08-01-preview","durationMs":946,"content-length":56,"Date":"Wed, 05 Mar 2025 23:44:41 GMT","redactedHeaders":"Content-Type,x-content-type-options,Strict-Transport-Security,apim-request-id","content-length":56,"body":"{\"error\":{\"code\":\"404\",\"message\": \"Resource not found\"}}"}
Exception in thread "main" com.azure.core.exception.ResourceNotFoundException: Status code 404, "{"error":{"code":"404","message": "Resource not found"}}"
azure-ai-inference playground sample does not work
It results in the 404 to
https://*.services.ai.azure.com/models/chat/completions?api-version=2024-08-01-preview.Workarounds
2024-05-01-previewwhen using inference endpointhttps://*.openai.azure.com/openai/deployments/gpt-4o-miniand not provide model tochatCompletionOptionsSuggestions
Based on the inference readme examples, it's not clear which type of the endpoint to use with/without the model:
.endpoint("https://{your resource}.openai.azure.com/openai/deployments/{model}")ChatCompletionOptions- it's not clear when to set/not set it