Abstract: Hard parameter sharing in multi-domain learning (MDL) al-lows domains to share some of the model parameters to reduce storage cost while improving prediction accuracy. One com-mon sharing practice is to share the bottom layers of a deep neural network among domains while using separate top layers for each domain. In this work, we revisit this common practice via an empirical study on image classification tasks from a diverse set of visual domains and make two surprising obser-vations. (1) Using separate bottom-layer parameters could achieve significantly better performance than the common practice and this phenomenon holds with different experimen-tal settings. (2) A multi-domain model with a small proportion of domain-specific parameters from bottom layers can achieve competitive performance with independent models trained on each domain separately. Our observations suggest that people adopt the new strategy of using separate bottom-layer parame-ters as a stronger baseline for model design in MDL.
0 Replies
Loading