Install
openclaw skills install google-vertex-aiGoogle Vertex AI integration. Manage Projects. Use when the user wants to interact with Google Vertex AI data.
openclaw skills install google-vertex-aiGoogle Vertex AI is a machine learning platform that allows data scientists and ML engineers to build, deploy, and scale ML models. It provides a unified platform for the entire ML lifecycle, from data preparation to model deployment and monitoring. It's used by organizations looking to leverage Google's AI infrastructure and tools for their machine learning needs.
Official docs: https://cloud.google.com/vertex-ai/docs
This skill uses the Membrane CLI to interact with Google Vertex AI. Membrane handles authentication and credentials refresh automatically — so you can focus on the integration logic rather than auth plumbing.
Install the Membrane CLI so you can run membrane from the terminal:
npm install -g @membranehq/cli@latest
membrane login --tenant --clientName=<agentType>
This will either open a browser for authentication or print an authorization URL to the console, depending on whether interactive mode is available.
Headless environments: The command will print an authorization URL. Ask the user to open it in a browser. When they see a code after completing login, finish with:
membrane login complete <code>
Add --json to any command for machine-readable JSON output.
Agent Types : claude, openclaw, codex, warp, windsurf, etc. Those will be used to adjust tooling to be used best with your harness
Use membrane connection ensure to find or create a connection by app URL or domain:
membrane connection ensure "https://cloud.google.com/vertex-ai" --json
The user completes authentication in the browser. The output contains the new connection id.
This is the fastest way to get a connection. The URL is normalized to a domain and matched against known apps. If no app is found, one is created and a connector is built automatically.
If the returned connection has state: "READY", skip to Step 2.
If the connection is in BUILDING state, poll until it's ready:
npx @membranehq/cli connection get <id> --wait --json
The --wait flag long-polls (up to --timeout seconds, default 30) until the state changes. Keep polling until state is no longer BUILDING.
The resulting state tells you what to do next:
READY — connection is fully set up. Skip to Step 2.
CLIENT_ACTION_REQUIRED — the user or agent needs to do something. The clientAction object describes the required action:
clientAction.type — the kind of action needed:
"connect" — user needs to authenticate (OAuth, API key, etc.). This covers initial authentication and re-authentication for disconnected connections."provide-input" — more information is needed (e.g. which app to connect to).clientAction.description — human-readable explanation of what's needed.clientAction.uiUrl (optional) — URL to a pre-built UI where the user can complete the action. Show this to the user when present.clientAction.agentInstructions (optional) — instructions for the AI agent on how to proceed programmatically.After the user completes the action (e.g. authenticates in the browser), poll again with membrane connection get <id> --json to check if the state moved to READY.
CONFIGURATION_ERROR or SETUP_FAILED — something went wrong. Check the error field for details.
Search using a natural language description of what you want to do:
membrane action list --connectionId=CONNECTION_ID --intent "QUERY" --limit 10 --json
You should always search for actions in the context of a specific connection.
Each result includes id, name, description, inputSchema (what parameters the action accepts), and outputSchema (what it returns).
| Name | Key | Description |
|---|---|---|
| Cancel Tuning Job | cancel-tuning-job | Cancel a running tuning job in Vertex AI. |
| Create Tuning Job | create-tuning-job | Create a new tuning job to fine-tune a Gemini model with your custom data. |
| Get Tuning Job | get-tuning-job | Get details of a specific tuning job in Vertex AI. |
| List Tuning Jobs | list-tuning-jobs | List all tuning jobs in a Vertex AI project location. |
| Get Model | get-model | Get details of a specific model in Vertex AI. |
| List Models | list-models | List all models in a Vertex AI project location. |
| Count Tokens | count-tokens | Count the number of tokens in text content. |
| Embed Content | embed-content | Generate embeddings for text content using Vertex AI embedding models. |
| Generate Content | generate-content | Generate content with multimodal inputs using Gemini models. |
membrane action run <actionId> --connectionId=CONNECTION_ID --json
To pass JSON parameters:
membrane action run <actionId> --connectionId=CONNECTION_ID --input '{"key": "value"}' --json
The result is in the output field of the response.
When the available actions don't cover your use case, you can send requests directly to the Google Vertex AI API through Membrane's proxy. Membrane automatically appends the base URL to the path you provide and injects the correct authentication headers — including transparent credential refresh if they expire.
membrane request CONNECTION_ID /path/to/endpoint
Common options:
| Flag | Description |
|---|---|
-X, --method | HTTP method (GET, POST, PUT, PATCH, DELETE). Defaults to GET |
-H, --header | Add a request header (repeatable), e.g. -H "Accept: application/json" |
-d, --data | Request body (string) |
--json | Shorthand to send a JSON body and set Content-Type: application/json |
--rawData | Send the body as-is without any processing |
--query | Query-string parameter (repeatable), e.g. --query "limit=10" |
--pathParam | Path parameter (repeatable), e.g. --pathParam "id=123" |
membrane action list --intent=QUERY (replace QUERY with your intent) to find existing actions before writing custom API calls. Pre-built actions handle pagination, field mapping, and edge cases that raw API calls miss.