Stateful Sequential Dataflow AI Hardware

Stateful AI workloads such as autoregressive generation, sequence modeling, and streaming inference place unique pressure on conventional accelerator designs. At ARCH Lab, we study dataflow-driven architectures that better exploit temporal locality, predictable execution, and state reuse across long-running computations.


AI Generated Image

Our work investigates accelerator organizations that reduce data movement, improve memory efficiency, and provide scalable support for sequential and stateful computation. By combining architectural specialization with hardware-software co-design, we aim to build AI systems that are both efficient and structurally well suited for next-generation intelligent applications.

Research themes

  • Stateful and sequential AI execution models
  • Memory-centric accelerator design
  • Dataflow scheduling for long-context inference
  • Efficient support for structured and irregular workloads
  • Hardware-software co-design for emerging AI systems