463465810cz commited on
Commit
5c40565
·
1 Parent(s): 8cb8316

Former-commit-id: 9c52469c9ec250295ed02e18b0db068ad04bb4e9

Files changed (1) hide show
  1. README.md +2 -5
README.md CHANGED
@@ -6,7 +6,7 @@
6
 
7
  ---
8
 
9
- > **Abstract:** *Transformer-based methods have recently been widely used in low-level vision tasks, including image super-resolution (SR). These networks utilize self-attention along different dimensions, spatial or channel, and achieve impressive performance. This inspires us to combine the two dimensions in Transformer for a more powerful representation capability. Based on the above idea, we propose a novel Transformer model, Dual Aggregation Transformer (DAT), for image SR. Our DAT aggregates features across spatial and channel dimensions, in the inter-block and intra-block dual manner. Specifically, we alternately apply spatial and channel self-attention in consecutive Transformer blocks. The alternate strategy enables DAT to capture the global context and realize inter-block feature aggregation. Furthermore, we propose the adaptive interaction module (AIM) and the spatial-gate feed-forward network (SGFN) to achieve intra-block feature aggregation. AIM complements two self-attention mechanisms from corresponding dimensions. Meanwhile, SGFN introduces additional non-linear spatial information in the feed-forward network. Extensive experiments show that our DAT surpasses current state-of-the-art methods.*
10
  >
11
  > <p align="center">
12
  > <img width="800" src="figs/DAT.png">
@@ -150,10 +150,7 @@ We achieved state-of-the-art performance. Detailed results can be found in the p
150
  <img width="900" src="figs/Figure-4.png">
151
  <img width="900" src="figs/Figure-5.png">
152
  </p>
153
-
154
-
155
-
156
- - </details>
157
 
158
  ## Citation
159
 
 
6
 
7
  ---
8
 
9
+ > **Abstract:** Transformer-based methods have recently been widely used in low-level vision tasks, including image super-resolution (SR). These networks utilize self-attention along different dimensions, spatial or channel, and achieve impressive performance. This inspires us to combine the two dimensions in Transformer for a more powerful representation capability. Based on the above idea, we propose a novel Transformer model, Dual Aggregation Transformer (DAT), for image SR. Our DAT aggregates features across spatial and channel dimensions, in the inter-block and intra-block dual manner. Specifically, we alternately apply spatial and channel self-attention in consecutive Transformer blocks. The alternate strategy enables DAT to capture the global context and realize inter-block feature aggregation. Furthermore, we propose the adaptive interaction module (AIM) and the spatial-gate feed-forward network (SGFN) to achieve intra-block feature aggregation. AIM complements two self-attention mechanisms from corresponding dimensions. Meanwhile, SGFN introduces additional non-linear spatial information in the feed-forward network. Extensive experiments show that our DAT surpasses current methods.*
10
  >
11
  > <p align="center">
12
  > <img width="800" src="figs/DAT.png">
 
150
  <img width="900" src="figs/Figure-4.png">
151
  <img width="900" src="figs/Figure-5.png">
152
  </p>
153
+ </details>
 
 
 
154
 
155
  ## Citation
156