Explore 24+ AI models from 49 providers. Filter by capability, tier, and pricing to find the right model.
24 results
Most powerful Gemini model with native multimodal understanding. Supports adjustable reasoning depth via thinking_level parameter.
Input
$3.50/M
Output
$10.50/M
Context
1.0M
Google's fast and cost-efficient thinking model with strong reasoning capabilities.
Input
$0.15/M
Output
$0.60/M
Context
1.0M
Google DeepMind's fourth-generation image synthesis model capable of producing images up to 2K resolution with exceptional photorealism and compositional accuracy. Imagen 4 includes SynthID watermarking by default for responsible AI deployment, supports advanced inpainting and outpainting, and demonstrates industry-leading performance on text rendering and spatial reasoning tasks.
Input
$4.00/M
Output
$20.00/M
Google's most capable thinking model with breakthrough performance on reasoning and coding.
Input
$1.25/M
Output
$10.00/M
Context
1.0M
Google DeepMind's flagship video generation model that natively produces joint audio-visual output in a single pass. Veo 3 leverages a Latent Diffusion Transformer to generate high-fidelity clips with synchronized dialogue, sound effects, and ambient audio without requiring a separate audio model. It demonstrates strong physical understanding and prompt adherence across diverse cinematic styles.
Input
$5.00/M
Output
$150.00/M
Google's frontier-class model at Flash-level latency and cost. 90.4% on GPQA Diamond, 78% on SWE-bench, 1M context window.
Input
$0.50/M
Output
$3.00/M
Context
1.0M
Google's most capable model. 94.3% on GPQA Diamond, 80.6% on SWE-bench, 77.1% on ARC-AGI-2. #1 on 12 of 18 tracked benchmarks.
Input
$2.00/M
Output
$12.00/M
Context
1.0M
Google's fastest multimodal model with native tool use and advanced agentic capabilities.
Input
$0.10/M
Output
$0.40/M
Context
1.0M
Google's ultra-efficient model offering better performance than Gemini 1.5 Flash at the same cost point.
Input
$0.07/M
Output
$0.30/M
Context
1.0M
An enhanced iteration of Google DeepMind's Veo series that produces 8-second clips that can be seamlessly extended up to 148 seconds through iterative generation. Veo 3.1 improves temporal consistency over long sequences, delivers higher resolution output, and refines audio synchronization for extended storytelling and commercial content production.
Input
$3.00/M
Output
$80.00/M
A multimodal extension of Google's Gemini 2.5 Flash model that adds native image generation and editing capabilities alongside text understanding. This model enables conversational image creation, iterative visual refinement, and combined text-image output within a single unified interface, making it particularly effective for design iteration and creative brainstorming workflows.
Input
$0.15/M
Output
$30.00/M
Smallest Gemma 2 model for efficient text processing on consumer hardware.
Input
$0.02/M
Output
$0.04/M
Context
8K
Specialized reasoning model designed for science, research, and complex engineering challenges.
Input
$5.00/M
Output
$15.00/M
Context
1.0M
Smallest Gemma 3 model for edge and mobile deployment. Text-only with 128K context.
Input
$0.02/M
Output
$0.02/M
Context
128K
Open vision-language model for image captioning, visual QA, and OCR tasks. Built on Gemma 2 backbone.
Input
$0.30/M
Output
$0.60/M
Context
8K
Mid-size PaliGemma for efficient vision-language tasks. Strong OCR and document understanding.
Input
$0.15/M
Output
$0.30/M
Context
8K
Experimental Gemini model with extended chain-of-thought reasoning. Transparent thinking process with strong performance on math and science.
Input
$0.15/M
Output
$0.60/M
Context
1.0M
Google's previous-gen flagship model with the longest context window in production.
Input
$1.25/M
Output
$5.00/M
Context
2.1M
Google's open-source multimodal model. Strong performance for its size with vision capabilities.
Input
$0.10/M
Output
$0.10/M
Context
128K
Efficient open-source model from Google with multimodal capabilities at 12B parameters.
Input
$0.05/M
Output
$0.05/M
Context
128K
Ultra-efficient open-source model from Google. Runs on mobile and edge devices.
Input
$0.02/M
Output
$0.02/M
Context
128K
Google's previous-gen open-source model with strong general capabilities.
Input
$0.07/M
Output
$0.07/M
Context
8K
Efficient open-source model from Google. Great performance-to-size ratio.
Input
$0.03/M
Output
$0.03/M
Context
8K
Google's open-source code-focused model based on the Gemma architecture.
Input
$0.03/M
Output
$0.03/M
Context
8K