According to a recent LinkedIn post from Sakana AI, the company’s researchers are exploring how large language models handle probabilistic tasks such as repeated coin flips. The post highlights that standard prompting can lead to biased distributions and limited diversity in outputs, affecting use cases like idea generation, recommendations, and other multi-option decisions.
Claim 55% Off TipRanks
- Unlock hedge fund-level data and powerful investing tools for smarter, sharper decisions
- Discover top-performing stock ideas and upgrade to a portfolio of market leaders with Smart Investor Picks
The LinkedIn post describes a new prompting technique called String Seed of Thought (SSoT), which instructs a model to generate a random string and then derive its answer from that string without external randomness. According to the post, SSoT reduces output bias across multiple open and closed models, generalizes to n-way choices and arbitrary distributions, and performs well on the NoveltyBench diversity benchmark.
As shared in the post, the work is slated for presentation at ICLR 2026, a major venue for machine learning research. For investors, the research suggests Sakana AI is targeting core limitations in current LLM behavior, which could strengthen its position in high-value applications where reliable sampling, diversity, and recommendation quality are critical.
If SSoT or related methods gain adoption, Sakana AI could enhance the competitiveness and perceived reliability of its own models or tooling in enterprise and consumer AI workflows. This type of methodological innovation may support future monetization via differentiated model offerings, licensing of techniques, or partnerships with platforms that rely on probabilistic generation and diverse content outputs.

