Journal:Proceedings of the ACM on Web Conference 2024 (WWW), CCF-A
Abstract:Text classification is a fundamental task in web content mining. Although the existing supervised contrastive learning (SCL) approach combined with pre-trained language models (PLMs) has achieved leading performance in text classification, it lacks fundamental principles. Theoretically motivated by a derived lower bound of mutual information maximization, we propose a dual contrastive learning framework DualCL that satisfies three properties, i.e., parameter-free, augmentation-easy and label-aware. DualCL generates classifier parameters from the PLM and simultaneously uses them for classification and as augmented views of the input text for supervised contrastive learning. Extensive experiments conclusively demonstrate that DualCL excels in learning superior text representations and consistently outperforms baseline models.
Co-author:Junfan Chen,Richong Zhang, Yaowei Zheng, Qianben Chen,Chunming Hu, Yongyi Mao
Indexed by:国际学术会议
Page Number:4362-4371
Translation or Not:no
Date of Publication:2024-01-01
