Abstract: Highlights•A mixer attention is for the additivity of residual connections in the mixer layer.•Token sequence aggregation is to fuse prediction results of facial keypoints.•Our PointTransform can predict the depression level while protecting privacy.
Loading