Skip to content

Commit

Permalink
Update S0-L26.md
Browse files Browse the repository at this point in the history
  • Loading branch information
qiyanjun authored Apr 21, 2024
1 parent a5320cb commit d212c29
Showing 1 changed file with 10 additions and 2 deletions.
12 changes: 10 additions & 2 deletions _contents/S0-L26.md
Original file line number Diff line number Diff line change
Expand Up @@ -40,6 +40,16 @@ In this session, our readings cover:
of LLMs (III/IV)](https://medium.com/mantisnlp/flashattention-techniques-for-efficient-inference-of-llms-iii-iv-0d619c9ca38c)


### State Space Model for New-Generation Network Alternative to Transformers: A Survey
+ [Submitted on 15 Apr 2024]
+ Xiao Wang, Shiao Wang, Yuhe Ding, Yuehang Li, Wentao Wu, Yao Rong, Weizhe Kong, Ju Huang, Shihao Li, Haoxiang Yang, Ziwen Wang, Bo Jiang, Chenglong Li, Yaowei Wang, Yonghong Tian, Jin Tang
+ In the post-deep learning era, the Transformer architecture has demonstrated its powerful performance across pre-trained big models and various downstream tasks. However, the enormous computational demands of this architecture have deterred many researchers. To further reduce the complexity of attention models, numerous efforts have been made to design more efficient methods. Among them, the State Space Model (SSM), as a possible replacement for the self-attention based Transformer model, has drawn more and more attention in recent years. In this paper, we give the first comprehensive review of these works and also provide experimental comparisons and analysis to better demonstrate the features and advantages of SSM. Specifically, we first give a detailed description of principles to help the readers quickly capture the key ideas of SSM. After that, we dive into the reviews of existing SSMs and their various applications, including natural language processing, computer vision, graph, multi-modal and multi-media, point cloud/event stream, time series data, and other domains. In addition, we give statistical comparisons and analysis of these models and hope it helps the readers to understand the effectiveness of different structures on various tasks. Then, we propose possible research points in this direction to better promote the development of the theoretical model and application of SSM. More related works will be continuously updated on the following GitHub: this https URL.



## More readings:


### JAMBA
+ Introducing Jamba: AI21's Groundbreaking SSM-Transformer Model
Debuting the first production-grade Mamba-based model delivering best-in-class quality and performance.
Expand All @@ -49,8 +59,6 @@ Debuting the first production-grade Mamba-based model delivering best-in-class q



## More readings:

### Mamba: Linear-Time Sequence Modeling with Selective State Spaces
+ Albert Gu, Tri Dao
+ https://arxiv.org/abs/2312.00752
Expand Down

0 comments on commit d212c29

Please sign in to comment.