According to a recent LinkedIn post from Rad AI, the company is highlighting simulation work on photon-counting CT detector pixel sizes and their impact on charge sharing and spectral fidelity. The post describes 100,000-photon simulations through CdTe detectors at seven pixel sizes, showing that smaller pixels dramatically increase charge sharing and degrade energy resolution.
Claim 30% Off TipRanks
- Unlock hedge fund-level data and powerful investing tools for smarter, sharper decisions
- Discover top-performing stock ideas and upgrade to a portfolio of market leaders with Smart Investor Picks
The LinkedIn post suggests that medical CT applications operate at extremely high photon flux and therefore favor larger pixels, around 250–300 micrometers, to balance charge sharing with count-rate capability. In contrast, non-destructive testing runs at much lower flux, allowing smaller pixels and higher spatial resolution, underscoring that optimal detector design is highly application-specific.
For investors, the content implies that Rad AI is positioning itself as an expert provider of spectral CT simulation and detector optimization services rather than just a generic imaging technology player. By emphasizing physics-based design trade-offs and simulation-led workflows, the post points to potential demand from OEMs and industrial users seeking to de-risk detector development and improve performance.
This focus on simulation and optimization could translate into recurring project-based revenue and closer integration into customers’ R&D pipelines if Rad AI’s tools are adopted early in detector design cycles. It may also enhance the company’s strategic value in the emerging photon-counting CT segment, where detector architecture and spectral performance are key differentiators for both medical and industrial imaging systems.

