Amazon Bedrock
Managed AWS service for building generative AI applications with multiple foundation models
Pricing: Pay-per-use
Best for: Enterprise AI applications on AWS with governance and compliance requirements
AI / ML Alternatives
TypeScript toolkit for building AI-powered streaming UIs with any LLM provider in Next.js and other frameworks
Managed AWS service for building generative AI applications with multiple foundation models
Pricing: Pay-per-use
Best for: Enterprise AI applications on AWS with governance and compliance requirements
Advanced AI assistant API known for safety, long context, and reasoning
Pricing: Pay-per-token (API pricing)
Best for: Long-form analysis, coding assistants, safe AI apps
Run inference on open models at Cloudflare edge with near-zero cold starts and no GPU provisioning
Pricing: Freemium — generous free neurons/day; pay-per-use beyond limit
Best for: Latency-sensitive AI inference, edge deployments, teams already on Cloudflare Workers
Google's multimodal model API for text, image, and reasoning workflows
Pricing: Pay-per-use
Best for: Multimodal AI features, Google ecosystem integration
AI inference platform using custom LPU hardware for ultra-fast LLM inference speeds
Pricing: Freemium - generous free tier, pay-per-token for production
Best for: Applications requiring the fastest possible LLM inference, real-time AI interactions
Platform and model hub for open-source AI models, datasets, inference APIs, and fine-tuning
Pricing: Freemium — model hub free; Inference Endpoints and Spaces paid
Best for: Accessing and deploying open-weight models, fine-tuning, ML research, and production inference
Framework for building LLM-powered applications with chains, agents, RAG pipelines, and tool integrations
Pricing: Free / Open Source
Best for: Complex LLM pipelines, RAG applications, AI agents with tool use
Data framework for building LLM applications with RAG pipelines, agents, and structured data ingestion
Pricing: Free / Open Source (LlamaCloud managed service paid)
Best for: RAG pipelines, document Q&A, AI agents that need to query private or structured data
European AI company providing high-performance open-weight and commercial LLM models via API
Pricing: Pay-per-token - from $0.1/1M tokens
Best for: Cost-efficient LLM inference, European data residency requirements, open-weight model access
Run large language models locally on your own hardware with a simple CLI and REST API
Pricing: Free / Open Source
Best for: Local LLM development, privacy-sensitive applications, offline AI workflows, cost-free inference
API platform for GPT, reasoning, and multimodal models for production applications
Pricing: Pay-per-use
Best for: AI assistants, workflow automation, text and image intelligence
Looking for more help deciding?