InMemQK: A Product Quantization Based MatMul Module for Compute-in-Memory Attention Macro
Large Language Models (LLMs), based on transformer architecture, have demonstrated remarkable capabilities in natural language processing tasks, enabling machines to generate human-like text and engage in meaningful dialogues. However, the exponential increase in model parameters has led to limitati...
        Saved in:
      
    
          | Main Authors: | Pengcheng Feng, Yihao Chen, Jinke Yu, Hao Yue, Zhelong Jiang, Yi Xiao, Wan’ang Xiao, Huaxiang Lu, Gang Chen | 
|---|---|
| Format: | Article | 
| Language: | English | 
| Published: | MDPI AG
    
        2024-12-01 | 
| Series: | Applied Sciences | 
| Subjects: | |
| Online Access: | https://www.mdpi.com/2076-3417/14/23/11198 | 
| Tags: | Add Tag 
      No Tags, Be the first to tag this record!
   | 
Similar Items
- 
                
                    On the preparation and characterization of thin NiTi shape memory alloy wires for MEMS        
                          
 by: Riccardo Casati, et al.
 Published: (2012-12-01)
- 
                
                    On the preparation and characterization of thin NiTi shape memory alloy wires for MEMS        
                          
 by: Riccardo Casati, et al.
 Published: (2012-12-01)
- 
                
                    Low-Power 8T SRAM Compute-in-Memory Macro for Edge AI Processors        
                          
 by: Hye-Ju Shin, et al.
 Published: (2024-11-01)
- 
                
                    Modeling and Characteristic Test for a Crank-Connecting Rod Mem-Inerter Device        
                          
 by: Xiao-Liang Zhang, et al.
 Published: (2024-12-01)
- 
                
                    On the preparation and characterization of  thin NiTi shape memory alloy wires for MEMS        
                          
 by: Ausonio Tuissi, et al.
 Published: (2013-01-01)
 
       