Gemini API
GCPAI & MLFree tier availableDirect API access to Google's most capable multimodal AI models with text, image, audio, and video understanding, long context windows, and function calling support
Attributes
- Multimodal
- Yes
- Context Window
- 1M+ tokens
- Streaming
- Yes
Sub-services (2)
Gemini Pro
General-purpose multimodal model
Gemini Flash
Fast, cost-efficient model for high-volume tasks
Compliance & Certifications
This service is attested for the following frameworks. Always verify with the provider before relying on a specific compliance posture.
Where this runs
Sovereign regions (2)
- T-Systems Sovereign Cloud · FrankfurtT-Systems Sovereign Cloud powered by Google Cloud
- S3NS Sovereign Cloud · ParisS3NS — Google Cloud + Thales joint venture
Commercial regions (42)
Europe (13)
- Belgium
- Finland
- Paris
- Berlin
- Frankfurt
- Milan
- Turin
- Netherlands
- Warsaw
- Madrid
- Stockholm
- Zurich
- London
North America (12)
- Montréal
- Toronto
- Querétaro
- Northern Virginia
- Columbus
- Iowa
- Dallas
- Las Vegas
- Los Angeles
- South Carolina
- Salt Lake City
- Oregon
South America (2)
- São Paulo
- Santiago
Asia (9)
- Hong Kong
- Delhi
- Mumbai
- Jakarta
- Osaka
- Tokyo
- Singapore
- Seoul
- Taiwan
Oceania (2)
- Melbourne
- Sydney
Middle East (3)
- Tel Aviv
- Doha
- Dammam
Africa (1)
- Johannesburg
Tags
Equivalent services on other platforms
Alibaba's flagship open-source foundation model family covering Qwen (text), Qwen-VL (vision-language), Qwen-Audio, and Qwen-Coder — accessible via the DashScope API with chat, completion, embeddings, and function-calling endpoints
Build generative AI applications with foundation models from Anthropic, Cohere, Meta, Mistral, Stability AI, and Amazon, accessed via a single API with fine-tuning and agents
Enterprise access to OpenAI models including GPT-4, GPT-3.5, and DALL-E with Azure security, private networking, regional deployments, and pay-as-you-go or provisioned throughput
Tencent's in-house family of large language models (Hunyuan-Pro, Standard, Lite, plus multimodal Hunyuan-Vision) accessible via the Hunyuan API, with enterprise-grade context windows up to 256K, function calling, embeddings, and tuning