Humor Identification from Short Texts Using Roberta and Albert

  • Hemant P., Dr. Pramod Kumar, Dr. Nirmala C. R.

Abstract

In the recent times most of the chat based applications which are generative in nature suffer from not understanding the humor intent within the text written. In this paper, we showcase a novel approach for classifying text which contains humor using light weight version of BERT called the ALBERT.  Fine tuning the model takes a major step here with the current data as the existing pre trained model is created on Wikipedia corpus. The proposed system uses word embedding mechanism for text and ALBERT Tokenizer for generating the tokens. The approach reduces the training time to 2/3rd and also retains the accuracy at 98.7% where as other Language models like BERT takes more execution time to generate the approximate results. RoBERTa showcases an increase of 0.3% to 0.6% in accuracy as compared to other language models for Humor Classification.

Published
2020-10-03
How to Cite
Hemant P., Dr. Pramod Kumar, Dr. Nirmala C. R. (2020). Humor Identification from Short Texts Using Roberta and Albert. International Journal of Advanced Science and Technology, 29(04), 9592 - 9600. Retrieved from http://sersc.org/journals/index.php/IJAST/article/view/32988