Vis enkel innførsel

dc.contributor.authorYadav, Rohan Kumar
dc.date.accessioned2022-11-28T13:47:08Z
dc.date.available2022-11-28T13:47:08Z
dc.date.created2022-11-10T16:58:57Z
dc.date.issued2022
dc.identifier.citationYadav, R. K. (2022). Interpretable Architectures and Algorithms for Natural Language Processing [PhD. thesis]. University of Agder.en_US
dc.identifier.isbn978-82-8427-101-9
dc.identifier.issn1504-9272
dc.identifier.urihttps://hdl.handle.net/11250/3034519
dc.descriptionPaper V is excluded from the dissertation with respect to copyright.en_US
dc.description.abstractThis thesis has two parts: Firstly, we introduce the human level-interpretable models using Tsetlin Machine (TM) for NLP tasks. Secondly, we present an interpretable model using DNNs. The first part combines several architectures of various NLP tasks using TM along with its robustness. We use this model to propose logic-based text classification. We start with basic Word Sense Disambiguation (WSD), where we employ TM to design novel interpretation techniques using the frequency of words in the clause. We then tackle a new problem in NLP, i.e., aspect-based text classification using a novel feature engineering for TM. Since TM operates on Boolean features, it relies on Bag-of-Words (BOW), making it difficult to use pre-trained word embedding like Glove, word2vec, and fasttext. Hence, we designed a Glove embedded TM to significantly enhance the model’s performance. In addition to this, NLP models are sensitive to distribution bias because of spurious correlations. Hence we employ TM to design a robust text classification against spurious correlations. The second part of the thesis consists interpretable model using DNN where we design a simple solution for complex position dependent NLP task. Since TM’s interpretability comes with the cost of performance, we propose an DNN-based architecture using a masking scheme on LSTM/GRU based models that ease the interpretation for humans using the attention mechanism. At last, we take the advantages of both models and design an ensemble model by integrating TM’s interpretable information into DNN for better visualization of attention weights. Our proposed model can be efficiently integrated to have a fully explainable model for NLP that assists trustable AI. Overall, our model shows excellent results and interpretation in several open-sourced NLP datasets. Thus, we believe that by combining the novel interpretation of TM, the masking technique in the neural network, and the integrated ensemble model, we can build a simple yet effective platform for explainable NLP applications wherever necessary.en_US
dc.language.isoengen_US
dc.publisherUniversity of Agderen_US
dc.relation.ispartofseriesDoctoral Dissertations at the University of Agder; no. 388
dc.relation.haspartPaper I: Yadav, R. K., Jiao, L., Granmo, O. & Goodwin, M. (2021). Interpretability in Word Sense Disambiguation using Tsetlin Machine. In Proceedings of the 13th International Conference on Agents and Artificial Intelligence (2, p. 402-409). SciTePress. https://doi.org/10.5220/0010382104020409. Accepted version. Full-text is available in AURA as a separate file: .en_US
dc.relation.haspartPaper II: Yadav, R. K., Jiao, L., Granmo, O. & Goodwin, M. (2021). Human-Level Interpretable Learning for Aspect-Based Sentiment Analysis. Proceedings of the AAAI Conference on Artificial Intelligence, 35(16), 14203-14212. https://doi.org/10.1609/aaai.v35i16.1767. Accepted version. Full-text is available in AURA as a separate file: .en_US
dc.relation.haspartPaper III: Yadav, R. K., Jiao, L., Goodwin, M. & Granmo, O. (2021). Positionless aspect based sentiment analysis using attention mechanism. Knowledge-Based Systems, 226: 107136. https://doi.org/10.1016/j.knosys.2021.107136. Accepted version. Full-text is available in AURA as a separate file: https://hdl.handle.net/11250/2987280.en_US
dc.relation.haspartPaper IV: Yadav, R. K., Jiao, L., Granmo, O. & Goodwin, M. (2021). Enhancing Interpretable Clauses Semantically using Pretrained Word Representation. In Proceedings of the Fourth BlackboxNLP Workshop on Analyzing and Interpreting Neural Networks for NLP (p. 265–274). Association for Computational Linguistics. http://dx.doi.org/10.18653/v1/2021.blackboxnlp-1.19. Accepted version. Full-text is available in AURA as a separate file: .en_US
dc.relation.haspartPaper V: Yadav, R. K., Jiao, L., Granmo, O. & Goodwin, M. (2022). Robust Interpretable Text Classification against Spurious Correlations Using AND-rules with Negation. In Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence (p. 4439-4446). https://doi.org/10.24963/ijcai.2022/616. Accepted version. Full-text is not available in AURA as a separate file.en_US
dc.relation.haspartPaper VI: Yadav, R. K. & Nicolae, D. C. (2022). Enhancing Attention’s Explanation Using Interpretable Tsetlin Machine. Algorithms, 15(5): 143. https://doi.org/10.3390/a15050143. Accepted version. Full-text is available in AURA as a separate file: https://hdl.handle.net/11250/3030688.en_US
dc.rightsAttribution-NonCommercial-NoDerivatives 4.0 Internasjonal*
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/4.0/deed.no*
dc.titleInterpretable Architectures and Algorithms for Natural Language Processingen_US
dc.typeDoctoral thesisen_US
dc.description.versionpublishedVersionen_US
dc.rights.holder© 2022 Rohan Kumar Yadaven_US
dc.subject.nsiVDP::Teknologi: 500en_US
dc.source.pagenumber228en_US
dc.source.issue388en_US
dc.identifier.cristin2072077


Tilhørende fil(er)

Thumbnail
Thumbnail
Thumbnail
Thumbnail
Thumbnail
Thumbnail

Denne innførselen finnes i følgende samling(er)

Vis enkel innførsel

Attribution-NonCommercial-NoDerivatives 4.0 Internasjonal
Med mindre annet er angitt, så er denne innførselen lisensiert som Attribution-NonCommercial-NoDerivatives 4.0 Internasjonal