Self-supervised Short Text Modeling through Auxiliary Context Generation
dc.contributor.author | Choudhary, Nurendra | en |
dc.contributor.author | Aggarwal, Charu | en |
dc.contributor.author | Subbian, Karthik | en |
dc.contributor.author | Reddy, Chandan K. | en |
dc.date.accessioned | 2022-09-23T20:34:19Z | en |
dc.date.available | 2022-09-23T20:34:19Z | en |
dc.date.issued | 2022-04-12 | en |
dc.date.updated | 2022-09-23T20:31:04Z | en |
dc.description.abstract | Short text is ambiguous and often relies predominantly on the domain and context at hand in order to attain semantic relevance. Existing classification models perform poorly on short text due to data sparsity and inadequate context. Auxiliary context, which can often provide sufficient background regarding the domain, is typically available in several application scenarios. While some of the existing works aim to leverage real-world knowledge to enhance short text representations, they fail to place appropriate emphasis on the auxiliary context. Such models do not harness the full potential of the available context in auxiliary sources. To address this challenge, we reformulate short text classification as a dual channel self-supervised learning problem (that leverages auxiliary context) with a generation network and a corresponding prediction model. We propose a self-supervised framework, Pseudo-Auxiliary Context generation network for Short text Modeling (PACS), to comprehensively leverage auxiliary context and is jointly learned with a prediction network in an end-to-end manner. Our PACS model consists of two sub-networks: a Context Generation Network (CGN) that models the auxiliary context?s distribution and a Prediction Network (PN) to map the short text features and auxiliary context distribution to the final class label. Our experimental results on diverse datasets demonstrate that PACS outperforms formidable state-of-the-art baselines. We also demonstrate the performance of our model on cold start scenarios (where contextual information is non-existent) during prediction. Furthermore, we perform interpretability and ablation studies to analyze various representational features captured by our model and the individual contribution of its modules to the overall performance of PACS, respectively. | en |
dc.description.version | Published version | en |
dc.format.mimetype | application/pdf | en |
dc.identifier.doi | https://doi.org/10.1145/3511712 | en |
dc.identifier.issue | 3 | en |
dc.identifier.other | 51 | en |
dc.identifier.uri | http://hdl.handle.net/10919/111992 | en |
dc.identifier.volume | 13 | en |
dc.language.iso | en | en |
dc.publisher | ACM | en |
dc.rights | Creative Commons Attribution 4.0 International | en |
dc.rights.holder | The author(s) | en |
dc.rights.uri | http://creativecommons.org/licenses/by/4.0/ | en |
dc.title | Self-supervised Short Text Modeling through Auxiliary Context Generation | en |
dc.title.serial | ACM Transactions on Intelligent Systems and Technology | en |
dc.type | Article - Refereed | en |
dc.type.dcmitype | Text | en |