Skip to content
Open
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 1 addition & 1 deletion batchata/providers/gemini/gemini_provider.py
Original file line number Diff line number Diff line change
Expand Up @@ -48,7 +48,7 @@ def __init__(self, auto_register: bool = True):
if not api_key:
raise ValueError("GOOGLE_API_KEY environment variable is required")

self.client = genai_lib.Client(api_key=api_key)
self.client = genai_lib.Client(vertexai=False, api_key=api_key)
super().__init__()
self.models = GEMINI_MODELS
self._batches: Dict[str, Dict] = {}
Expand Down
85 changes: 28 additions & 57 deletions batchata/providers/gemini/models.py
Original file line number Diff line number Diff line change
Expand Up @@ -2,96 +2,67 @@

from ..model_config import ModelConfig

# updated based on current (Jan 28 2026) docs
# models - https://ai.google.dev/gemini-api/docs/models
# batch pricing - https://ai.google.dev/gemini-api/docs/pricing
# image file types - https://ai.google.dev/gemini-api/docs/image-understanding#supported-formats

# Google Gemini models with batch processing support
# Batch mode provides 50% discount on standard API pricing
GEMINI_MODELS = {
"gemini-3.0-pro-latest": ModelConfig(
name="gemini-3.0-pro-latest",
max_input_tokens=2097152, # 2M context
max_output_tokens=8192,
batch_discount=0.5, # 50% discount confirmed in docs
supports_images=True,
supports_files=True,
supports_citations=False,
supports_structured_output=True,
file_types=[".pdf", ".txt", ".jpg", ".png", ".gif", ".webp"]
),
"gemini-3.0-pro": ModelConfig(
name="gemini-3.0-pro",
max_input_tokens=2097152, # 2M context
max_output_tokens=8192,
batch_discount=0.5, # 50% discount confirmed in docs
supports_images=True,
supports_files=True,
supports_citations=False,
supports_structured_output=True,
file_types=[".pdf", ".txt", ".jpg", ".png", ".gif", ".webp"]
),
"gemini-3.0-flash-latest": ModelConfig(
name="gemini-3.0-flash-latest",
"gemini-3-pro-preview": ModelConfig(
name="gemini-3-pro-preview",
max_input_tokens=1048576, # 1M context
max_output_tokens=8192,
batch_discount=0.5,
supports_images=True,
supports_files=True,
supports_citations=False,
supports_structured_output=True,
file_types=[".pdf", ".txt", ".jpg", ".png", ".gif", ".webp"]
),
"gemini-3.0-flash": ModelConfig(
name="gemini-3.0-flash",
max_input_tokens=1048576, # 1M context
max_output_tokens=8192,
batch_discount=0.5,
max_output_tokens=65536,
batch_discount=0.5, # 50% discount confirmed in docs
supports_images=True,
supports_files=True,
supports_citations=False,
supports_structured_output=True,
file_types=[".pdf", ".txt", ".jpg", ".png", ".gif", ".webp"]
file_types=[".pdf", ".txt", ".jpg", ".png", ".webp"]
),
"gemini-3.0-flash-lite-latest": ModelConfig(
name="gemini-3.0-flash-lite-latest",
"gemini-3-flash-preview": ModelConfig(
name="gemini-3-flash-preview",
max_input_tokens=1048576, # 1M context
max_output_tokens=8192,
batch_discount=0.5,
max_output_tokens=65536,
batch_discount=0.5, # 50% discount confirmed in docs
supports_images=True,
supports_files=True,
supports_citations=False,
supports_structured_output=True,
file_types=[".pdf", ".txt", ".jpg", ".png", ".gif", ".webp"]
file_types=[".pdf", ".txt", ".jpg", ".png", ".webp"]
),
"gemini-3.0-flash-lite": ModelConfig(
name="gemini-3.0-flash-lite",
"gemini-2.5-pro": ModelConfig(
name="gemini-2.5-pro",
max_input_tokens=1048576, # 1M context
max_output_tokens=8192,
max_output_tokens=65536,
batch_discount=0.5,
supports_images=True,
supports_files=True,
supports_citations=False,
supports_structured_output=True,
file_types=[".pdf", ".txt", ".jpg", ".png", ".gif", ".webp"]
file_types=[".pdf", ".txt", ".jpg", ".png", ".webp"]
),
"gemini-2.0-flash": ModelConfig(
name="gemini-2.0-flash",
"gemini-2.5-flash": ModelConfig(
name="gemini-2.5-flash",
max_input_tokens=1048576, # 1M context
max_output_tokens=8192,
max_output_tokens=65536,
batch_discount=0.5,
supports_images=True,
supports_files=True,
supports_files=False,
supports_citations=False,
supports_structured_output=True,
file_types=[".pdf", ".txt", ".jpg", ".png", ".gif", ".webp"]
file_types=[".txt", ".jpg", ".png", ".webp"]
),
"gemini-2.0-flash-lite": ModelConfig(
name="gemini-2.0-flash-lite",
"gemini-2.5-flash-lite": ModelConfig(
name="gemini-2.5-flash-lite",
max_input_tokens=1048576, # 1M context
max_output_tokens=8192,
max_output_tokens=65536,
batch_discount=0.5,
supports_images=True,
supports_files=True,
supports_citations=False,
supports_structured_output=True,
file_types=[".pdf", ".txt", ".jpg", ".png", ".gif", ".webp"]
),
file_types=[".pdf", ".txt", ".jpg", ".png", ".webp"]
)
}