Task embeddings are task-specific vectors designed to construct a semantic space of tasks, which can be used to predict the most transferable source task for a given target task via the similarity between task embeddings. However, existing methods use optimized parameters and representations as task embeddings, resulting in substantial computational complexity and storage requirements. In this work, we draw inspiration from the operating mechanism of deep neural networks (DNNs) and biological brains, where neuronal activations are sparse and task-specific, and we use the connectivity patterns of neurons as a unique identifier associated with the task. The proposed method learns to assign importance masks for sub-structures of DNNs, and accordingly indicate the task-specific connectivity patterns. In addition to the storage advantages brought by the binary masking mechanism and structured sparsity, the early-bird nature of the sparse optimization process can deliver an efficient computation advantage. Experiments show that our method consistently outperforms other baselines in predicting inter-task transferability across data regimes and transfer settings, while keeping high efficiency in computation and storage.
CITATION STYLE
Xi, Z., Zheng, R., Zhang, Y., Huang, X., Wei, Z., Peng, M., … Gui, T. (2023). Connectivity Patterns are Task Embeddings. In Proceedings of the Annual Meeting of the Association for Computational Linguistics (pp. 11993–12013). Association for Computational Linguistics (ACL). https://doi.org/10.18653/v1/2023.findings-acl.759
Mendeley helps you to discover research relevant for your work.