According to a recent LinkedIn post from FriendliAI, the company is highlighting beta support for Anthropic’s Messages API on its serverless and dedicated endpoints. The post suggests that applications built on this API can be redirected to open‑weight models such as GLM‑5.1, MiniMax‑M2.5, and DeepSeek‑V3.2 with minimal configuration changes.
Claim 55% Off TipRanks
- Unlock hedge fund-level data and powerful investing tools for smarter, sharper decisions
- Discover top-performing stock ideas and upgrade to a portfolio of market leaders with Smart Investor Picks
The company’s LinkedIn post emphasizes a large cost differential between Claude Sonnet 4.6 and the alternative models, citing output costs of $15 per 1M tokens versus as low as $1.20 per 1M tokens. This positioning appears aimed at high‑volume agent and workflow scenarios where token usage scales quickly and where prompt‑caching discounts could further widen effective cost gaps.
The post also notes an incentive offer of up to $50,000 in inference credits for teams migrating from Anthropic or other providers. For investors, this may indicate an aggressive customer‑acquisition strategy focused on price‑sensitive, infrastructure‑heavy AI users, which could support usage growth but may also pressure near‑term margins depending on unit economics.
Strategically, the post suggests FriendliAI is seeking to position itself as a cost‑efficient inference layer that is compatible with established APIs while steering demand toward open‑weight models. If successful, this approach could strengthen its competitive stance in the AI infrastructure market and deepen switching costs for enterprise customers, though long‑term impact will depend on model performance, customer retention, and pricing sustainability.

