Learning Improved Class Vector for Multi-Class Question Type Classification
- 10.2991/ahis.k.210913.015How to use a DOI?
- Class-specific vector, Deep Learning Models, Polysemy, Question Classification, Word2vec
Recent research in NLP has exploited word embedding to achieve outstanding results in various tasks such as; spam filtering, text classification and summarization and others. Present word embedding algorithms have power to capture semantic and syntactic knowledge about word, but not enough to portray the distinct meaning of polysemy word. Many work has utilized sense embeddings to integrate all possible meaning to word vector, which is computationally expensive. Context embedding is another way out to identify word’s actual meaning, but it is hard to enumerate every context with a small size dataset. This paper has proposed a methodology to generate improved class-specific word vector that enhance the distinctive property of word in a class to tackle light polysemy problem in question classification. The proposed approach is compared with baseline approaches, tested using deep learning models upon TREC, Kaggle and Yahoo questions datasets and respectively attain 93.6%, 91.8% and 89.2% accuracy.
- © 2021, the Authors. Published by Atlantis Press.
- Open Access
- This is an open access article distributed under the CC BY-NC license (http://creativecommons.org/licenses/by-nc/4.0/).
Cite this article
TY - CONF AU - Tanu Gupta AU - Ela Kumar PY - 2021 DA - 2021/09/13 TI - Learning Improved Class Vector for Multi-Class Question Type Classification BT - Proceedings of the 3rd International Conference on Integrated Intelligent Computing Communication & Security (ICIIC 2021) PB - Atlantis Press SP - 113 EP - 121 SN - 2589-4900 UR - https://doi.org/10.2991/ahis.k.210913.015 DO - 10.2991/ahis.k.210913.015 ID - Gupta2021 ER -