Stateful Sequential Dataflow AI Hardware

Stateful AI workloads such as autoregressive generation, sequence modeling, and streaming inference place unique pressure on conventional accelerator designs. At ARCH Lab, we study dataflow-driven architectures that better exploit temporal locality, predictable execution, and state reuse across long-running computations.

Our work investigates accelerator organizations that reduce data movement, improve memory efficiency, and provide scalable support for sequential and stateful computation. By combining architectural specialization with hardware-software co-design, we aim to build AI systems that are both efficient and structurally well suited for next-generation intelligent applications.

Research themes

  • Stateful and sequential AI execution models
  • Memory-centric accelerator design
  • Dataflow scheduling for long-context inference
  • Efficient support for structured and irregular workloads
  • Hardware-software co-design for emerging AI systems