Abstract: Highlights•We introduce a novel cross-wise transformer module (CTM), which employs cross-attention to capture dependencies.•We devise a feature supplement module (FSM) that employs an information complementary mechanism.•Our proposed network shows outstanding results across multiple datasets and effectively handle various visual tasks.
Loading