Skip to content

Support for VertexAI custom models (Endpoints) #654

Open
@rafal-dudek

Description

@rafal-dudek

Expected Behavior

I would like to be able to use my Custom Model (e.g. fine-tuned foundation model) deployed on GCP Endpoint with Spring AI library.

API for Endpoints:
https://cloud.google.com/vertex-ai/docs/reference/rest/v1/projects.locations.endpoints
Example for predict:
https://us-central1-aiplatform.googleapis.com/v1/projects/49...78/locations/us-central1/endpoints/433...736:predict
{"instances":[{"content":"Hello"}],"parameters":{"temperature":0.2,"maxOutputTokens":1024,"topP":0.8,"topK":40,"candidateCount":1}}

Current Behavior

Currently in VertexAI PaLM2 Chat and VertexAI Gemini Chat you can only use Foundation models: https://cloud.google.com/vertex-ai/generative-ai/docs/model-reference/text-chat

Context

There is a similar issue for langchain4j:
langchain4j/langchain4j#440

Metadata

Metadata

Assignees

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions