Abstract: Highlights•We propose a new disentangled perspective to treat motion and context information in the VCOD and VSOD tasks.•We further explore how to apply disentangled perspective into the self-supervised manner.•Extensive experiments demonstrate that our method outperforms the state-of-the-art both on VSOD and VCOD datasets
External IDs:dblp:journals/nn/XiaoTLLL26
Loading