InMemQK: A Product Quantization Based MatMul Module for Compute-in-Memory Attention Macro
Large Language Models (LLMs), based on transformer architecture, have demonstrated remarkable capabilities in natural language processing tasks, enabling machines to generate human-like text and engage in meaningful dialogues. However, the exponential increase in model parameters has led to limitati...
Saved in:
| Main Authors: | Pengcheng Feng, Yihao Chen, Jinke Yu, Hao Yue, Zhelong Jiang, Yi Xiao, Wan’ang Xiao, Huaxiang Lu, Gang Chen |
|---|---|
| Format: | Article |
| Language: | English |
| Published: |
MDPI AG
2024-12-01
|
| Series: | Applied Sciences |
| Subjects: | |
| Online Access: | https://www.mdpi.com/2076-3417/14/23/11198 |
| Tags: |
Add Tag
No Tags, Be the first to tag this record!
|
Similar Items
-
On the preparation and characterization of thin NiTi shape memory alloy wires for MEMS
by: Riccardo Casati, et al.
Published: (2012-12-01) -
On the preparation and characterization of thin NiTi shape memory alloy wires for MEMS
by: Riccardo Casati, et al.
Published: (2012-12-01) -
Low-Power 8T SRAM Compute-in-Memory Macro for Edge AI Processors
by: Hye-Ju Shin, et al.
Published: (2024-11-01) -
Modeling and Characteristic Test for a Crank-Connecting Rod Mem-Inerter Device
by: Xiao-Liang Zhang, et al.
Published: (2024-12-01) -
On the preparation and characterization of thin NiTi shape memory alloy wires for MEMS
by: Ausonio Tuissi, et al.
Published: (2013-01-01)