Skip to content

jianghongcheng/SF-GPT

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

23 Commits
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

Spatial-Frequency Guided Pixel Transformer (SF-GPT) for NIR-to-RGB Translation

Official PyTorch implementation of the paper:

Spatial-Frequency Guided Pixel Transformer for NIR-to-RGB Translation
Infrared Physics & Technology, 2025
🔗 ScienceDirect
🔗 DOI: 10.1016/j.infrared.2025.105891

📌 Introduction

Near-Infrared (NIR) imaging provides enhanced contrast and sensitivity but lacks the spatial and textural richness of RGB images. NIR-to-RGB translation is a challenging task due to:

  • Spectral mapping ambiguity

  • Statistical weak correlation between NIR and RGB

We propose SF-GPT, a novel deep learning architecture that leverages both spatial and frequency domains through transformer-based mechanisms.

✨ Key Contributions

  1. SF-GPT: We propose a novel Spatial-Frequency Guided Pixel Transformer for NIR-to-RGB translation, combining spatial and frequency cues to capture both local textures and global context.
  2. Dual-domain Feature Extraction: We incorporate DCT or DWT to extract low- and high-frequency features, while pixel-wise cues are obtained via PixelUnshuffle for fine-grained reconstruction.
  3. SFG-MSA Module: We design a Spatial-Frequency Guided Multi-head Self-Attention mechanism that adaptively fuses pixel and frequency features, enhancing translation fidelity and feature discrimination.
  4. State-of-the-art Performance: Extensive experiments validate the effectiveness of SF-GPT, outperforming existing methods in both visual quality and quantitative metrics.

📈 PSNR Comparison on VCIP Test Dataset

🧠 Network Architecture

🔍 Visualization of DCT and DWT Decomposition

🖼️ Visual Comparison on VCIP Test Dataset

🖼️ Visual Comparison on SSMID Test Dataset

📊 Quantitative Comparison on VCIP Test Dataset

Method PSNR (↑) SSIM (↑) AE (↓) LPIPS (↓)
ATcycleGAN19.590.594.330.295
CoColor23.540.692.680.233
ColorMamba24.560.712.810.212
DCT-RCAN22.150.773.400.214
DRSformer20.180.564.220.254
HAT19.420.693.980.298
MCFNet20.340.613.790.208
MFF17.390.614.690.318
MPFNet22.140.633.680.253
NIR-GNN17.500.605.220.384
Restormer19.430.544.410.267
SPADE19.240.594.590.283
SST14.260.575.610.361
TTST18.570.674.460.320
SF-GPT (DCT)26.090.772.720.132
SF-GPT (DWT)25.820.792.570.114

📊 Quantitative Comparison on SSMID Test Dataset

Method PSNR (↑) SSIM (↑) AE (↓) LPIPS (↓)
DGCAN18.1330.601
Compressed DGCAN16.9730.565
DRSformer18.1760.6985.6980.238
Restormer17.9830.6935.5600.256
HAT17.6770.6925.8030.249
TTST17.7220.6965.7470.244
SF-GPT (DCT)19.0110.6995.5410.185
SF-GPT (DWT)19.9170.7105.4140.176

📚 Citation

If you find this work helpful in your research, please cite:

@article{jiang2025spatial,
  title={Spatial-Frequency Guided Pixel Transformer for NIR-to-RGB Translation},
  author={Jiang, Hongcheng and Chen, ZhiQiang},
  journal={Infrared Physics \& Technology},
  year={2025},
  pages={105891},
  publisher={Elsevier},
  doi={10.1016/j.infrared.2025.105891},
  note={doi: \href{https://doi.org/10.1016/j.infrared.2025.105891}{10.1016/j.infrared.2025.105891}}
}

📬 Contact

If you have any questions, feedback, or collaboration ideas, feel free to reach out:

About

No description, website, or topics provided.

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published