Interactive Segment Anything NeRF with Feature Imitation

Xiaokang Chen*1, Jiaxiang Tang*1, Diwen Wan*1, Jingbo Wang*2, Gang Zeng1,

1 Peking University   2 Chinese University of Hong Kong  



This paper investigates the potential of enhancing Neural Radiance Fields (NeRF) with semantics to expand their applications. Although NeRF has been proven useful in real-world applications like VR and digital creation, the lack of semantics hinders interaction with objects in complex scenes. We propose to imitate the backbone feature of off-the-shelf perception models to achieve zero-shot semantic segmentation with NeRF. Our framework reformulates the segmentation process by directly rendering semantic features and only applying the decoder from perception models. This eliminates the need for expensive backbones and benefits 3D consistency. Furthermore, we can project the learned semantics onto extracted mesh surfaces for real-time interaction. With the state-of-the-art Segment Anything Model (SAM), our framework accelerates segmentation by 16 times with comparable mask quality. The experimental results demonstrate the efficacy and computational advantages of our approach.

Demo Video


Segmentation with different prompts

Click-based Text-based

User interaction

3D Mesh Segmentation


  title={Interactive Segment Anything NeRF with Feature Imitation},
  author={Chen, Xiaokang and Tang, jiaxiang and Wan, Diwen and Wang, Jingbo and Zeng, Gang},
  journal={arXiv preprint arXiv:2211.12368},