Tightly coordinated data movement and low-latency on-chip storage for real-time environments.
The race to build smarter and faster AI chips continues to surge. This is especially true in autonomous vehicles that interpret the world in milliseconds, edge accelerators that push trillions of operations per second, hyperscale data-center processors that drive massive workloads, and next-generation consumer devices that demand ever-higher intelligence. As modern system-on-chip (SoC) architectures become increasingly complex, they produce rapidly growing volumes of on-chip data. Managing this data requires increasingly efficient movement, storage, and access. Insufficient data delivery rates create bottlenecks that restrict overall system responsiveness. Cutting-edge designs require tremendous throughput, but they often sit idle because the data arrives too slowly. The result is data starvation. This creates evaporating performance and unpredictable latency.
With system interconnects and memory hierarchies now influencing performance as much as computation itself, SoC teams increasingly rely on FlexGen and FlexNoC-optimized data transport, paired with intelligent on-chip caching, to maintain responsiveness.

Fig. 1: The role of last-level cache in successful designs. (Source: Arteris, Inc.)
As these bottlenecks intensify, SoC designers are increasingly turning to on-chip memory resources to keep data close to compute. Reducing reliance on off-chip memory and minimizing wait cycles is essential to overcoming data starvation across diverse workloads. One of the most effective architectural tools for achieving this is a shared, on-chip last-level cache (LLC).
While an LLC plays a vital role in today’s SoCs, determining the correct configuration is a complex architectural challenge. For example, parameters such as the number of banks, parallel access capabilities, and partitioning strategies all influence the cache’s efficiency under real workloads.
Oversized LLCs add unnecessary silicon cost, while undersized LLCs fail to deliver meaningful benefits. Achieving the optimal balance requires detailed traffic analysis and workload simulation to ensure the cache is tailored precisely to the system’s data movement patterns.
Design teams must consider:
Arteris is uniquely qualified to address these limitations because data movement is the company’s core expertise. Leveraging this foundation, the company created CodaCache to mitigate the widening gap between rapidly advancing processors and comparatively slow main-memory access. By storing relevant in-flight data on-chip and coordinating efficiently with external DRAM, it provides a high-bandwidth, power-efficient reservoir that keeps critical data close to compute.
When paired with FlexGen, FlexNoC, CodaCache complements the interconnect’s ability to route traffic efficiently across heterogeneous systems:

Fig. 2: SoC performance is dependent upon data availability. (Source: Arteris, Inc.)
CodaCache is designed for seamless integration into heterogeneous network-on-chip architectures, delivering a well-balanced combination of capabilities.
The result is a subsystem that meaningfully boosts performance across advanced AI compute markets.
As SoC architectures become increasingly complex, with growing numbers of cores, accelerators, concurrent workloads, and chiplets, the need for intelligent, scalable, and automated data management infrastructure continues to grow.
It is no longer enough to build faster compute engines. Designers must feed them efficiently. The combination of FlexGen/FlexNoC and CodaCache represents a new class of system IP that acknowledges this shift, delivering tightly coordinated data movement and low-latency on-chip storage for demanding real-time environments.
Arteris CodaCache represents a new class of system IP that recognizes this shift in demanding real-time environments.
Leave a Reply