RPG Seminar – HIMSA: A Heterogeneous In-Memory Computing and Searching Architecture for Efficient Attention-Based Models
Zoom Link https://hku.zoom.us/j/99174148480?pwd=duVxaYZOJDT6MWxDh4OKOMmyo12A7A.1 Abstract The Transformer architecture, the foundation for modern large language models (LLMs), has revolutionized natural language processing and other AI domains. However, its significant computational and memory […]
