Incremental Few-shot Semantic Segmentation (iFSS) aims to learn novel classes with limited samples while preserving segmentation capability for base classes, addressing catastrophic forgetting. Existing methods, relying on knowledge distillation and background learning, still suffer from feature drift and poor generalization. To overcome these challenges, we propose a novel diffusion-based generative framework for iFSS. By mapping binary masks to three-channel representations and optimizing class-specific semantic embeddings, our method enhances foreground-background distinction and prevents feature interference. A lightweight post-processing module refines segmentation by converting generated images into binary masks. Leveraging the prior knowledge of diffusion models, we unify the learning of base and novel classes, eliminating complex training strategies and improving adaptability. Experiments on PASCAL-5i and COCO-20i datasets show our framework achieves state-of-the-art performance with minimal data. Additionally, our framework exhibits strong generalization in cross-domain few-shot segmentation (CD-FSS) benchmarks.
BibTex Code Here