Back to reviews
Together AI

Together AI

Fast inference for open-source LLMs at low cost

Together AI provides fast, cheap inference for open-source models like Llama, Mistral, and DeepSeek. Features dedicated endpoints, fine-tuning, and a serverless API. Known for competitive pricing and low latency.

Panel Reviews

The Builder

The Builder

Developer Perspective

Ship

Cheapest way to run Llama and Mistral models in production. The inference speed is competitive with major providers. OpenAI-compatible API makes switching easy.

The Skeptic

The Skeptic

Reality Check

Ship

The pricing is genuinely good and reliability has improved. The fine-tuning workflow is straightforward. A solid choice for open-source model deployment.

The Futurist

The Futurist

Big Picture

Ship

Together is betting that the future is open-source models. As Llama and Mistral improve, inference providers like Together become the AWS of AI.

Community Sentiment

Overall1,564 mentions
70% positive21% neutral9% negative
Hacker News287 mentions

Together AI's pricing on Llama 3 is 5x cheaper than comparable providers with similar latency

Reddit443 mentions

Best cost-per-token for open source models, use it for all my high-volume inference

Twitter/X678 mentions

Fine-tuning + inference from one provider makes the workflow so much cleaner

Product Hunt156 mentions

Competitive pricing and solid uptime — refreshing in the crowded inference market