kth.sePublications KTH
Operational message
There are currently operational disruptions. Troubleshooting is in progress.
Change search
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf
DCTN: Dual-Branch Convolutional Transformer Network With Efficient Interactive Self-Attention for Hyperspectral Image Classification
East China Jiaotong University, School of Information Engineering, Nanchang, China, 330013.
East China Jiaotong University, School of Information Engineering, Nanchang, China, 330013.
Guangzhou University, School of Electronic and Communication Engineering, Guangzhou, China, 511370.
Hubei University, Hubei Key Laboratory of Applied Mathematics, Faculty of Mathematics and Statistics, Wuhan, China, 430062.
Show others and affiliations
2024 (English)In: IEEE Transactions on Geoscience and Remote Sensing, ISSN 0196-2892, E-ISSN 1558-0644, Vol. 62, p. 1-16, article id 5508616Article in journal (Refereed) Published
Abstract [en]

Hyperspectral image (HSI) classification is an essential task in remote sensing with substantial practical significance. However, most existing convolutional neural network (CNN)-based classification methods focus only on local spatial features while neglecting global spectral dependencies. Meanwhile, Transformer-based methods exhibit robust capabilities for global spectral feature modeling but struggle to extract local spatial features effectively. To fully exploit the local spatial feature extraction capabilities of CNN-based networks and the global spectral feature extraction capabilities of Transformer-based networks, this article proposes a dual-branch convolutional Transformer method with efficient interactive self-attention (EISA) for HSI classification, namely the dual-branch convolutional transformer network (DCTN), which can aggregate local and global spatial-spectral features fully. Specifically, DCTN includes two core modules: the spatial-spectral fusion projection module (SFPM) and the EISA module. The former utilizes 3-D convolution with adaptive pooling and 2-D group convolution with residual connection to parallel extract fused and grouped spatial-spectral features, respectively. The latter performs EISA across height, width, and spectral dimensions, enabling deep fusion of spatial-spectral features. Extensive experiments on three real HSI datasets demonstrate that the proposed DCTN method outperforms existing classification methods, yielding state-of-the-art classification performance. The code is available at https://github.com/AllFever/DeepHyperX-DCTN for reproducibility.

Place, publisher, year, edition, pages
Institute of Electrical and Electronics Engineers (IEEE) , 2024. Vol. 62, p. 1-16, article id 5508616
Keywords [en]
Convolution neural networks (CNNs), hyperspectral image (HIS) classification, self-attention mechanism, transformer
National Category
Computer graphics and computer vision
Identifiers
URN: urn:nbn:se:kth:diva-343990DOI: 10.1109/TGRS.2024.3364143ISI: 001164472500043Scopus ID: 2-s2.0-85185545267OAI: oai:DiVA.org:kth-343990DiVA, id: diva2:1841360
Note

QC 20240304

Available from: 2024-02-28 Created: 2024-02-28 Last updated: 2025-02-07Bibliographically approved

Open Access in DiVA

No full text in DiVA

Other links

Publisher's full textScopus

Authority records

Ban, Yifang

Search in DiVA

By author/editor
Ban, Yifang
By organisation
Geoinformatics
In the same journal
IEEE Transactions on Geoscience and Remote Sensing
Computer graphics and computer vision

Search outside of DiVA

GoogleGoogle Scholar

doi
urn-nbn

Altmetric score

doi
urn-nbn
Total: 82 hits
CiteExportLink to record
Permanent link

Direct link
Cite
Citation style
  • apa
  • ieee
  • modern-language-association-8th-edition
  • vancouver
  • Other style
More styles
Language
  • de-DE
  • en-GB
  • en-US
  • fi-FI
  • nn-NO
  • nn-NB
  • sv-SE
  • Other locale
More languages
Output format
  • html
  • text
  • asciidoc
  • rtf