Explore 114+ AI models from 49 providers. Filter by capability, tier, and pricing to find the right model.
114 results
DeepSeek
DeepSeek's open-source MoE model rivaling frontier models at a fraction of the cost.
Input
$0.27/M
Output
$1.10/M
Context
128K
DeepSeek
DeepSeek's reasoning model with transparent chain-of-thought. Open-source and highly competitive.
Input
$0.55/M
Output
$2.19/M
Context
128K
Meta
Meta's latest open-source MoE model with 17B active parameters and industry-leading 10M token context.
Input
$0.15/M
Output
$0.60/M
Context
10.5M
Anthropic
Anthropic's best balance of intelligence and speed. Excellent for production workloads.
Input
$3.00/M
Output
$15.00/M
Context
200K
Google's frontier-class model at Flash-level latency and cost. 90.4% on GPQA Diamond, 78% on SWE-bench, 1M context window.
Input
$0.50/M
Output
$3.00/M
Context
1.0M
Meta
Meta's open-source model matching GPT-4 class performance at 70B parameters.
Input
$0.18/M
Output
$0.18/M
Context
128K
Google's fastest multimodal model with native tool use and advanced agentic capabilities.
Input
$0.10/M
Output
$0.40/M
Context
1.0M
Google's fast and cost-efficient thinking model with strong reasoning capabilities.
Input
$0.15/M
Output
$0.60/M
Context
1.0M
OpenAI
OpenAI's efficient reasoning model, optimized for speed while maintaining strong analytical capabilities.
Input
$1.10/M
Output
$4.40/M
Context
200K
Meta
Meta's strong mid-range open-source model, predecessor to 3.3 with broad community support.
Input
$0.18/M
Output
$0.18/M
Context
128K
Meta
Meta's largest code-focused open-source model. Specialized for code generation and understanding.
Input
$0.18/M
Output
$0.18/M
Context
16K
Mistral AI
Mistral's first code-focused model with 32K context. Supports 80+ programming languages.
Input
$0.30/M
Output
$0.90/M
Context
32K
Mistral AI
Mistral's flagship multimodal model. Built on Mistral Large with vision capabilities.
Input
$2.00/M
Output
$6.00/M
Context
128K
Mistral AI
Mistral's large open-source MoE model with 176B total params. Strong coding and reasoning.
Input
$0.65/M
Output
$0.65/M
Context
66K
DeepSeek
DeepSeek's open-source code-focused MoE model. Competitive with GPT-4 Turbo on coding.
Input
$0.14/M
Output
$0.28/M
Context
128K
DeepSeek
Merged general and coder capabilities from V2 into a unified model.
Input
$0.14/M
Output
$0.28/M
Context
128K
DeepSeek
R1 reasoning capabilities distilled into a compact Qwen-based 32B model.
Input
$0.12/M
Output
$0.18/M
Context
128K
Cohere
Cohere's open-weight model optimized for RAG and tool use. Strong multilingual support.
Input
$0.15/M
Output
$0.60/M
Context
128K
Cohere
Cohere's open-source multilingual model covering 23 languages with strong performance.
Input
$0.50/M
Output
$1.50/M
Context
128K
Alibaba/Qwen
Alibaba's flagship open-source model. Competitive with Llama 3.1 405B at a fraction of the size.
Input
$0.30/M
Output
$0.30/M
Context
128K
Alibaba/Qwen
Strong mid-range open-source model from Alibaba with broad capabilities.
Input
$0.08/M
Output
$0.08/M
Context
128K
Alibaba/Qwen
Alibaba's open-source coding specialist. Matches GPT-4o on code benchmarks.
Input
$0.08/M
Output
$0.08/M
Context
128K
Alibaba/Qwen
Alibaba's open-source vision-language model with video understanding capabilities.
Input
$0.40/M
Output
$0.40/M
Context
32K
Alibaba/Qwen
Alibaba's open-source reasoning model with transparent chain-of-thought. Competitive with o1-mini.
Input
$0.10/M
Output
$0.30/M
Context
32K
Microsoft
Microsoft's open-source MoE model with 42B total params and only 6.6B active.
Input
$0.06/M
Output
$0.06/M
Context
128K
Microsoft
Microsoft's instruction-tuned MoE model based on Mixtral. Strong on complex reasoning tasks.
Input
$0.65/M
Output
$0.65/M
Context
66K
NVIDIA
NVIDIA's optimized Llama 3.1 variant with custom reward model training.
Input
$0.18/M
Output
$0.18/M
Context
128K
AI21 Labs
AI21's hybrid SSM-Transformer model with 256K context. Novel Mamba architecture.
Input
$2.00/M
Output
$8.00/M
Context
256K
TII
TII's largest open-source model. One of the first truly open 180B parameter models.
Input
$0.80/M
Output
$0.80/M
Context
2K
01.AI
01.AI's fast inference model with strong performance across benchmarks.
Input
$0.30/M
Output
$0.30/M
Context
16K
01.AI
01.AI's open-source 34B model with strong bilingual (English/Chinese) capabilities.
Input
$0.10/M
Output
$0.10/M
Context
4K
Shanghai AI Lab
Open-source model with 1M context from Shanghai AI Lab. Strong coding and math skills.
Input
$0.06/M
Output
$0.06/M
Context
1.0M
Shanghai AI Lab
Open-source vision-language model with strong image understanding capabilities.
Input
$0.08/M
Output
$0.08/M
Context
8K
Snowflake
Snowflake's open-source enterprise MoE model optimized for SQL and business tasks.
Input
$0.30/M
Output
$0.30/M
Context
4K
Databricks
Databricks' open-source MoE model with strong code and reasoning capabilities.
Input
$0.75/M
Output
$0.75/M
Context
32K
Zhipu AI
Zhipu AI's flagship model with strong Chinese and English bilingual capabilities.
Input
$1.00/M
Output
$3.00/M
Context
128K
OpenAI
OpenAI's cost-efficient reasoning model with multimodal input, strong math and coding performance at a fraction of o3 pricing.
Input
$1.10/M
Output
$4.40/M
Context
200K
Mistral AI
Mistral's mid-tier model offering 90% of Claude Sonnet quality at significantly lower cost.
Input
$0.40/M
Output
$2.00/M
Context
131K
Mistral AI
Mistral's specialized code model supporting 80+ languages with 256K context and fill-in-the-middle capability.
Input
$0.30/M
Output
$0.90/M
Context
256K
xAI
xAI's efficient reasoning model with fast inference and competitive performance at budget pricing.
Input
$0.30/M
Output
$0.50/M
Context
131K
Microsoft
Chain-of-thought reasoning variant of Phi-4, competitive with much larger models on math and logic tasks.
Input
$0.04/M
Output
$0.04/M
Context
32K
DeepSeek
R1's reasoning capability distilled into a Llama 3.1 70B architecture for efficient deployment.
Input
$0.18/M
Output
$0.18/M
Context
128K
Cohere
Cohere's 111B parameter model supporting 23 languages with enterprise tool use and 256K context.
Input
$2.50/M
Output
$10.00/M
Context
256K
Alibaba/Qwen
Alibaba's dense 32B model with dual thinking/non-thinking modes and strong reasoning performance.
Input
$0.08/M
Output
$0.20/M
Context
131K
Allen AI
Fully open model with all components public: data, code, weights, and checkpoints. Instruct, Think, and RL Zero variants.
Input
$0.25/M
Output
$0.75/M
Context
128K
Zhipu AI
Vision-language MoE model with superior performance at lower inference cost.
Input
$0.15/M
Output
$0.30/M
Context
128K
Zhipu AI
Open-source video generation model creating 6-second clips at 720x480. Supports LoRA fine-tuning.
Input
$0.05/M
Output
$0.05/M
BAAI
State-of-the-art multimodal embedding model for visual search applications.
Input
$0.02/M
Output
$0.02/M
Context
8K
Pika Labs
Pika Labs' video generation model focused on ultra-realistic output with enhanced physics simulation. Pika 2.5 handles complex material interactions such as fluid dynamics, cloth draping, and particle effects with high fidelity. Its intuitive prompt interface and style controls make it accessible for creators seeking photorealistic short-form video content.
Input
$2.00/M
Output
$50.00/M
Luma AI
Luma AI's video generation model delivering native 1080p output at 4x faster inference speeds than previous versions, with optional 4K upscaling. Ray3.14 specializes in photorealistic 3D-aware video synthesis with strong spatial understanding, making it particularly effective for product visualization, architectural walkthroughs, and immersive content creation.
Input
$2.00/M
Output
$45.00/M
MiniMax
MiniMax's Hailuo 2.3 video model combines photorealistic rendering with versatile style support including anime, watercolor, and cinematic looks. It features advanced motion control, accurate lip-sync for dialogue scenes, and sophisticated lighting effects that adapt dynamically to scene content and camera movement.
Input
$2.00/M
Output
$50.00/M
Lightricks
Lightricks' open-source video generation model capable of producing native 4K video at 50 frames per second with clips up to 20 seconds in length. LTX-2 includes native audio synthesis and offers full model weights under a permissive license, making it a leading choice for researchers and developers building custom video generation pipelines.
Input
Free/M
Output
Free/M
Alibaba/Qwen
Alibaba's open-source video generation model that achieved the number one ranking on the VBench video quality benchmark upon release. With 14 billion parameters, Wan 2.1 demonstrates exceptional prompt adherence, temporal consistency, and visual quality across diverse content types, establishing a new baseline for open-weight video synthesis models.
Input
Free/M
Output
Free/M
Alibaba/Qwen
The successor to Wan 2.1, this open-source model introduces a Mixture-of-Experts flow-matching architecture with approximately 27 billion total parameters and 14 billion active during inference. Wan 2.2 delivers significantly improved motion quality, fine-grained detail, and extended generation lengths while maintaining the accessibility of fully open weights.
Input
Free/M
Output
Free/M
Tencent
Tencent's open-source video generation model with 8.3 billion parameters, featuring a novel Spatial-Temporal Self-Attention (SSTA) mechanism for improved temporal coherence. HunyuanVideo 1.5 supports diverse aspect ratios, variable frame rates, and extended clip durations, making it a versatile foundation model for the open-source video generation community.
Input
Free/M
Output
Free/M
Stability AI
Stability AI's specialized model for 4D novel-view video synthesis, generating temporally consistent multi-angle video from a single input clip or image. Stable Video 4D 2.0 enables creators to produce orbiting camera paths, bullet-time effects, and 3D-aware video transformations that maintain geometric and photometric coherence throughout the sequence.
Input
$2.00/M
Output
$40.00/M
A multimodal extension of Google's Gemini 2.5 Flash model that adds native image generation and editing capabilities alongside text understanding. This model enables conversational image creation, iterative visual refinement, and combined text-image output within a single unified interface, making it particularly effective for design iteration and creative brainstorming workflows.
Input
$0.15/M
Output
$30.00/M
Black Forest Labs
The open-weights development version of FLUX.2 with the same 32 billion parameter architecture as the Pro variant, released for non-commercial research and experimentation. FLUX.2 Dev provides researchers full access to model weights for fine-tuning, distillation, and architectural exploration while delivering near-Pro-level quality for academic and personal projects.
Input
Free/M
Output
Free/M
Adobe
Adobe's fourth-generation Firefly image model offering improved quality, faster generation, and enhanced creative controls compared to its predecessors. Firefly Image 4 provides robust structure references, style transfer, and generative fill capabilities, all trained on Adobe's commercially licensed dataset to ensure IP safety for enterprise and professional use.
Input
$3.00/M
Output
$25.00/M
Stability AI
Stability AI's largest open-source image generation model built on the Multimodal Diffusion Transformer (MMDiT) architecture. SD 3.5 Large delivers high-quality results across photorealistic and artistic styles with strong prompt adherence, accurate text rendering, and diverse composition capabilities, available under an open license for both research and commercial use.
Input
$0.50/M
Output
$6.50/M
Ideogram
Ideogram's third-generation model combining exceptional photorealism with industry-leading text rendering accuracy within generated images. Ideogram 3.0 handles complex typography, logos, signs, and handwritten text with remarkable fidelity, making it the preferred choice for design professionals working on brand assets, marketing materials, and content requiring reliable in-image text.
Input
$2.00/M
Output
$20.00/M
NVIDIA
Input
Free/M
Output
Free/M
Recraft
Recraft's flagship image generation model that achieved the number one ranking on the HuggingFace text-to-image leaderboard, with native support for both raster and vector output formats. Recraft V3 excels at brand-consistent design, offering precise color palette control, style locking, and batch generation capabilities that make it uniquely suited for professional design systems.
Input
$2.00/M
Output
$20.00/M
Microsoft
Microsoft's first in-house image generation model developed by Microsoft AI, designed for integration across Microsoft's product ecosystem including Designer, Copilot, and Bing Image Creator. MAI-Image-1 focuses on safety, controllability, and consistent quality, with built-in content filtering and provenance metadata for responsible enterprise deployment.
Input
$2.00/M
Output
$15.00/M
01.AI
Mid-size Yi model with enhanced inference speed for extended prompts.
Input
$0.10/M
Output
$0.20/M
Context
128K
01.AI
Vision-language Yi model for image understanding and visual question answering.
Input
$0.30/M
Output
$0.60/M
Context
16K
BigCode
Mid-size code model matching CodeLlama 13B quality at half the parameters.
Input
$0.07/M
Output
$0.14/M
Context
16K
Stability AI
Mid-size Stable Diffusion optimized for consumer GPUs and edge devices.
Input
$0.02/M
Output
$0.02/M
Black Forest Labs
Fast open-source text-to-image model with 4-step generation. Apache 2.0 licensed.
Input
$0.02/M
Output
$0.02/M
Meta
Safety classification model for detecting unsafe content in LLM inputs and outputs.
Input
$0.05/M
Output
$0.05/M
Context
128K
LG AI Research
Korean sovereign AI model using MoE with hybrid attention for reduced computation.
Input
$0.25/M
Output
$0.75/M
Context
128K
Upstage
Agentic reasoning-focused model matching larger rivals. Strong multilingual capabilities.
Input
$0.20/M
Output
$0.60/M
Context
128K
BAAI
Most popular open embedding model. Multi-functionality, multi-linguality, multi-granularity in one model.
Input
$0.01/M
Output
$0.01/M
Context
8K
OpenAI
Gold standard speech recognition model supporting 99+ languages. 1.55B parameter encoder-decoder architecture.
Input
$0.0060/M
Output
$0.0060/M
Anthropic
High-intelligence Sonnet model with 1M token context window. Strong balance of performance and cost.
Input
$3.00/M
Output
$15.00/M
Context
1.0M
Open vision-language model for image captioning, visual QA, and OCR tasks. Built on Gemma 2 backbone.
Input
$0.30/M
Output
$0.60/M
Context
8K
Mid-size PaliGemma for efficient vision-language tasks. Strong OCR and document understanding.
Input
$0.15/M
Output
$0.30/M
Context
8K
Alibaba/Qwen
Dense model with hybrid thinking/non-thinking modes. Seamless switching between complex reasoning and general dialogue.
Input
$0.20/M
Output
$0.60/M
Context
128K
Alibaba/Qwen
Compact vision-language model excelling at video and image analysis. Top small multimodal model on Hugging Face.
Input
$0.10/M
Output
$0.30/M
Context
128K
Alibaba/Qwen
Audio-language model for speech recognition, audio understanding, and music analysis.
Input
$0.10/M
Output
$0.30/M
Context
128K
DeepSeek
Vision-language model for image understanding, OCR, and visual reasoning tasks.
Input
$0.14/M
Output
$0.28/M
Context
128K
Mistral AI
First Mistral reasoning model with 50% AIME-24 improvement via scalable RL. Reasoning in 8+ languages.
Input
$2.00/M
Output
$6.00/M
Context
128K
Mistral AI
Open-source reasoning model built on Small 3.1 with SFT and RL training. Efficient multilingual reasoning.
Input
$0.20/M
Output
$0.60/M
Context
128K
Mistral AI
Coding-specialized model outperforming Qwen 3 Coder Flash despite smaller size.
Input
$0.20/M
Output
$0.60/M
Context
128K
Mistral AI
Code model using Mamba SSM architecture for linear-time inference. Unlimited theoretical context.
Input
$0.10/M
Output
$0.30/M
Context
256K
Cohere
Multilingual model covering 23 languages for global enterprise deployment.
Input
$0.05/M
Output
$0.15/M
Context
128K
Cohere
State-of-the-art text embedding model for semantic search and RAG applications.
Input
$0.10/M
Output
$0.10/M
Context
8K
TII
Outperforms all models under 13B on HuggingFace leaderboard. Trained on 14T tokens with innovative 1.58-bit quantized variant.
Input
$0.10/M
Output
$0.30/M
Context
32K
TII
Versatile 7B model with 30 checkpoint variants including base, instruct, and quantized.
Input
$0.07/M
Output
$0.21/M
Context
32K
Nomic AI
First MoE embedding model. Trained on 1.6B pairs across ~100 languages with top-2 expert routing.
Input
$0.01/M
Output
$0.01/M
Context
8K
Jina AI
Universal multimodal embedding handling text, images, and documents in 30+ languages.
Input
$0.02/M
Output
$0.02/M
Context
8K
Amazon
Fast, cost-effective reasoning model with built-in code interpreter and web grounding.
Input
$0.80/M
Output
$2.40/M
Context
1.0M
Amazon
Speech-to-speech model for natural real-time conversations. Supports 7 languages.
Input
$0.50/M
Output
$0.50/M
Amazon
Image generation model with fine-grained control over composition, style, and content.
Input
$0.04/M
Output
$0.04/M
Reka AI
One of the few 21B models supporting full interleaved multimodal inputs. Videos up to 5 minutes.
Input
$0.80/M
Output
$2.40/M
Context
128K
Genmo
High-performance open text-to-video model excelling in text consistency.
Input
$0.05/M
Output
$0.05/M
IBM
Enterprise-grade model with strong instruction following for business applications.
Input
$0.10/M
Output
$0.20/M
Context
128K
IBM
Updated Granite with enhanced coding and tool-use capabilities for enterprise automation.
Input
$0.10/M
Output
$0.20/M
Context
128K
OpenBMB
Efficient vision-language model rivaling GPT-4V quality at a fraction of the size.
Input
$0.10/M
Output
$0.20/M
Context
128K
Naver
Korean sovereign AI with omnimodal capabilities. Specialized for Korean language and culture.
Input
$1.00/M
Output
$3.00/M
Context
128K
Mistral AI
Mid-size Mistral model bridging the gap between 8B edge models and large frontier offerings.
Input
$0.15/M
Output
$0.45/M
Context
128K
Alibaba/Qwen
Compact math-specialized model with chain-of-thought reasoning for mathematical problem solving.
Input
$0.07/M
Output
$0.14/M
Context
128K
Microsoft
Lightweight multimodal model with vision capabilities for on-device and edge visual understanding.
Input
$0.05/M
Output
$0.10/M
Context
128K
Amazon
Amazon's video generation model producing high-quality short clips for advertising and social media.
Input
$0.04/M
Output
$0.04/M
Experimental Gemini model with extended chain-of-thought reasoning. Transparent thinking process with strong performance on math and science.
Input
$0.15/M
Output
$0.60/M
Context
1.0M
Cohere
Cohere's open multimodal model for visual understanding across 23 languages. Strong image captioning and visual QA.
Input
$0.25/M
Output
$0.50/M
Context
128K
Google's previous-gen flagship model with the longest context window in production.
Input
$1.25/M
Output
$5.00/M
Context
2.1M
Mistral AI
Mistral's flagship model with strong multilingual and code generation capabilities.
Input
$2.00/M
Output
$6.00/M
Context
128K
Cohere
Cohere's enterprise-grade model optimized for RAG, tool use, and business workflows.
Input
$2.50/M
Output
$10.00/M
Context
128K
Alibaba/Qwen
Alibaba's efficient code-focused MoE model. 80B total params, 3B active, Apache 2.0 licensed.
Input
$0.12/M
Output
$0.75/M
Context
256K
Anthropic
Upgraded Claude 3.5 Sonnet with major coding and tool-use improvements, plus computer use capability.
Input
$3.00/M
Output
$15.00/M
Context
200K
Google's open-source multimodal model. Strong performance for its size with vision capabilities.
Input
$0.10/M
Output
$0.10/M
Context
128K
Google's previous-gen open-source model with strong general capabilities.
Input
$0.07/M
Output
$0.07/M
Context
8K
Meta
Meta's largest multimodal Llama model with image understanding capabilities.
Input
$0.35/M
Output
$0.40/M
Context
128K