LSP-ST: Ladder Shape-Biased Side-Tuning for Robust Infrared Small Target Detection, link: LSP-ST
Fine-tuning the Segment Anything Model (SAM) for infrared small target detection poses significant challenges due to severe domain shifts. Existing adaptation methods often incorporate handcrafted priors to bridge this gap, yet such designs limit generalization and scalability. We identify a fundamental texture bias in foundation models, which overly depend on local texture cues for target localization. To address this, we propose Ladder Shape-Biased Side-Tuning (LSP-ST), a novel approach that introduces a shape-aware inductive bias to facilitate effective adaptation beyond texture cues. In contrast to prior work that injects explicit edge or contour features, LSP-ST models shape as a global structural prior, integrating both boundaries and internal layouts. We design a Shape-Enhanced Large-Kernel Attention Module to hierarchically and implicitly capture structural information in a fully differentiable manner, without task-specific handcrafted guidance. A theoretical analysis grounded in matched filtering and backpropagation reveals the mechanism by which the proposed attention improves structure-aware learning. With only 4.72M learnable parameters, LSP-ST achieves state-of-the-art performance on multiple infrared small target detection benchmarks. Furthermore, its strong generalization is validated across tasks such as mirror detection, shadow detection, and camouflaged object detection, while maintaining stable performance on texture-driven tasks like salient object detection, demonstrating that the introduced shape bias complements rather than competes with texture-based reasoning.
The link to the downstream prediction result figures mentioned in the supplementary materials is provided as follows:
If you find this paper useful, please cite it as:
@article{zhang2025vision,
title={Vision-centric representation-efficient fine-tuning for robust universal foreground segmentation},
author={Zhang, Guoyi and Chen, Siyang and Xu, Guangsheng and Wang, Han and Zhang, Xiaohu},
journal={arXiv preprint arXiv:2504.14481},
year={2025}
}