Vis enkel innførsel

dc.contributor.authorAbeyrathna, Kuruge Darshana
dc.contributor.authorAbouzeid, Ahmed Abdulrahem Othman
dc.contributor.authorBhattarai, Bimal
dc.contributor.authorGiri, Charul
dc.contributor.authorGlimsdal, Sondre
dc.contributor.authorGranmo, Ole-Christoffer
dc.contributor.authorLei, Jiao
dc.contributor.authorSaha, Rupsa
dc.contributor.authorSharma, Jivitesh
dc.contributor.authorTunheim, Svein Anders
dc.contributor.authorZhang, Xuan
dc.date.accessioned2024-04-11T12:47:58Z
dc.date.available2024-04-11T12:47:58Z
dc.date.created2023-11-15T17:18:19Z
dc.date.issued2023
dc.identifier.citationAbeyrathna, K. D., Abouzeid, A. A. O., Bhattarai, B., Giri, C., Glimsdal, S., Granmo, O.-C., Lei, J., Saha, R., Sharma, J., Tunheim, S. A. & Zhang, X. (2023). Building Concise Logical Patterns by Constraining Tsetlin Machine Clause Size. In E. Elkind (Ed.), Proceedings of the Thirty-Second International Joint Conference on Artificial Intelligence (pp. 3395-3403). International Joint Conferences on Artifical Intelligence.en_US
dc.identifier.isbn978-1-956792-03-4
dc.identifier.issn1045-0823
dc.identifier.urihttps://hdl.handle.net/11250/3126137
dc.descriptionAuthor's accepted manuscript.en_US
dc.description.abstractTsetlin machine (TM) is a logic-based machine learning approach with the crucial advantages of being transparent and hardware-friendly. While TMs match or surpass deep learning accuracy for an increasing number of applications, large clause pools tend to produce clauses with many literals (long clauses). As such, they become less interpretable. Further, longer clauses increase the switching activity of the clause logic in hardware, consuming more power. This paper introduces a novel variant of TM learning – Clause Size Constrained TMs (CSC-TMs) – where one can set a soft constraint on the clause size. As soon as a clause includes more literals than the constraint allows, it starts expelling literals. Accordingly, oversized clauses only appear transiently. To evaluate CSC-TM, we conduct classifcation, clustering, and regression experiments on tabular data, natural language text, images, and board games. Our results show that CSC-TM maintains accuracy with up to 80 times fewer literals. Indeed, the accuracy increases with shorter clauses for TREC, IMDb, and BBC Sports. After the accuracy peaks, it drops gracefully as the clause size approaches a single literal. We fnally analyze CSC-TM power consumption and derive new convergence properties.en_US
dc.language.isoengen_US
dc.publisherInternational Joint Conferences on Artifical Intelligenceen_US
dc.rightsAttribution-NonCommercial-NoDerivatives 4.0 Internasjonal*
dc.rights.urihttp://creativecommons.org/licenses/by-nc-nd/4.0/deed.no*
dc.titleBuilding Concise Logical Patterns by Constraining Tsetlin Machine Clause Sizeen_US
dc.typeAcademic articleen_US
dc.description.versionacceptedVersionen_US
dc.rights.holder© 2023 International Joint Conferences on Artificial Intelligenceen_US
dc.subject.nsiVDP::Teknologi: 500::Informasjons- og kommunikasjonsteknologi: 550en_US
dc.source.pagenumber3395-3403en_US
dc.source.journalIJCAI International Joint Conference on Artificial Intelligence
dc.identifier.doihttps://doi.org/10.24963/ijcai.2023/378
dc.identifier.cristin2197244
dc.relation.projectUniversity of Agder: CAIRen_US
cristin.qualitycode1


Tilhørende fil(er)

Thumbnail

Denne innførselen finnes i følgende samling(er)

Vis enkel innførsel

Attribution-NonCommercial-NoDerivatives 4.0 Internasjonal
Med mindre annet er angitt, så er denne innførselen lisensiert som Attribution-NonCommercial-NoDerivatives 4.0 Internasjonal