Documentation Index
Fetch the complete documentation index at: https://docs.getpatter.com/llms.txt
Use this file to discover all available pages before exploring further.
Google Gemini LLM
GoogleLLM plugs Google Gemini chat models into Patter’s pipeline mode via the google-genai SDK. It supports both the Gemini Developer API (with an API key) and Vertex AI (with GCP project + location). Streams normalise to Patter’s unified {type: "text" | "tool_call" | "done"} chunk protocol, and Gemini function_call parts map directly onto Patter tools.
This page covers Google Gemini in chat-completions mode for the pipeline (STT → LLM → TTS). For Gemini’s bidirectional speech-to-speech engine, see the separate
gemini-live adapter under Engines.Install
Usage
The namespaced import (
from getpatter.llm import google / import * as google from "getpatter/llm/google") auto-resolves the API key from GEMINI_API_KEY first, then GOOGLE_API_KEY for parity with other SDKs, and exposes a uniform LLM class.Supported models
Pricing in USD per 1M tokens.| Model | Input | Output | Notes |
|---|---|---|---|
gemini-2.5-flash (default) | $0.30 | $2.50 | Best price/perf for voice. |
gemini-2.5-pro | $1.25 | $10.00 | Highest quality. |
gemini-2.0-flash | n/a | n/a | Older fast model. |
gemini-2.0-flash-lite | n/a | n/a | Lightweight 2.0. |
gemini-1.5-flash | n/a | n/a | Legacy fast model. |
gemini-1.5-pro | n/a | n/a | Legacy pro model. |
gemini-live-2.5-flash-native-audio (input 2.50), see the Engines page — it is a separate Realtime adapter, not a chat-completions model.
Environment variables
| Variable | Required | Notes |
|---|---|---|
GEMINI_API_KEY | one of these | Preferred — Google’s CLI tooling uses this name. |
GOOGLE_API_KEY | one of these | Legacy/alt name accepted for parity. |
GOOGLE_GENAI_USE_VERTEXAI | optional | Set to 1 / true to default vertexai=True. |
GOOGLE_CLOUD_PROJECT | Vertex AI | GCP project ID when vertexai=True. |
GOOGLE_CLOUD_LOCATION | Vertex AI | GCP region (defaults to us-central1). |
Options
| Option | Default | Notes |
|---|---|---|
api_key / apiKey | None | Reads GEMINI_API_KEY, then GOOGLE_API_KEY. Ignored when vertexai=True. |
model | "gemini-2.5-flash" | Any Gemini chat model id. |
vertexai | False | Use Vertex AI instead of the Developer API. |
project | None | GCP project (Vertex AI). |
location | "us-central1" | GCP region (Vertex AI). |
temperature | unset | Optional sampling temperature. |
max_output_tokens / maxOutputTokens | unset | Output token cap. |
Vertex AI
Switch to Vertex AI when you need GCP-native auth (service accounts), VPC Service Controls, regional residency, or per-project billing isolation.google-genai SDK picks up Application Default Credentials automatically — set GOOGLE_APPLICATION_CREDENTIALS to a service-account key path or run gcloud auth application-default login for local dev.
Function calling
Gemini’sfunction_call parts map directly onto Patter tools — define a tool once and it works on every LLM provider. Patter assigns a monotonically increasing index per function_call part since Gemini does not provide a stable per-call index across stream chunks. Token usage is collected from usage_metadata (cumulative on each chunk; only the last value is yielded as a usage event to avoid double-counting).
