-
Notifications
You must be signed in to change notification settings - Fork 199
Description
I would like to discuss whether serving common AI features would be within this project's scope.
A good example would be the /v1/models endpoint. This is not implemented by every AI provider, but it is very commonly used by applications that allow users to choose their desired model. Right now, the ext-proc filter would fail for requests that go to any endpoint that is not the chat completions endpoint (#115 was created to address this), but probably the project could do more to facilitate the adoption to existing apps that rely on such APIs.
In the case of the /v1/models endpoint, for example, it would make a lot of sense that the ai-gateway could serve the response for such requests based on what has been configured in the ConfigMap, returning those models that have been configured and are allowed to be used.
What is the general feeling about ai-gateway directly implementing common AI features?