DeepSeek models now available via Lambda—starting at $0.34 per 1M input tokens. Lambda has officially launched support for two powerful Deep...
![]() |
DeepSeek models now available via Lambda—starting at $0.34 per 1M input tokens. |
Pricing is as follows:
- DeepSeek-V3: $0.34 per 1M input tokens, $0.88 per 1M output tokens
- DeepSeek-R1 671B: $0.54 per 1M input tokens, $2.18 per 1M output tokens
This pricing model puts Lambda’s offering among the most cost-efficient solutions in the generative AI space today. The DeepSeek models have been steadily gaining attention for their open licensing and performance, rivaling proprietary models like GPT-4 by OpenAI and Gemini by Google. With this launch, Lambda is positioning itself as a go-to infrastructure provider for those seeking transparency and scalability in AI deployment.
According to Lambda, users can integrate the models directly into apps using standard REST endpoints or through Python SDKs for fast experimentation and production scaling. The company’s infrastructure is powered by NVIDIA A100 and H100 GPUs, ensuring high performance for inference workloads.
The move highlights a broader trend toward open AI models becoming commercially viable alternatives to closed-source APIs. Organizations focused on data privacy and long-term cost control now have more competitive options for deploying large language models (LLMs) in real-world applications.
Developers can start experimenting immediately through the Lambda Inference Dashboard or automate deployments via their API documentation. As AI workloads continue to surge, offerings like this may help democratize access and reduce reliance on high-cost, centralized providers.