Abstract: Highlights•We investigate the role of multimodal CoT distilled knowledge from LLMs in KBVQA tasks.•We present a Multimodal Knowledge Reasoning via CoT (MuKCoT) model for reasoning and explanation.•Performance of MuKCoT fosters the higher-quality explanations for the KBVQA tasks.
Loading