心智圖資源庫 Transformer-based Model Compression: Techniques and Innovations

Transformer-based Model Compression: Techniques and Innovations

This mind map, created using EdrawMind, outlines key strategies for Transformer-based model compression. It covers selecting input data techniques like "Attention Probe," structural approaches such as weight sharing with MiniViT, and attention mechanisms including "Swin Transformer" and "Sheared LaMMA." The diagram highlights methods to reduce model size and enhance efficiency without compromising performance, showcasing innovations from recent papers.

編輯於2023-11-27 12:27:59
  • 推薦給您
  • 大綱