Detailed Information

Cited 0 time in webofscience Cited 0 time in scopus
Metadata Downloads

How Does a Transformer Learn Compression? An Attention Study on Huffman and LZ4

Full metadata record
DC Field Value Language
dc.contributor.authorSeo, Beomseok-
dc.contributor.authorNo, Albert-
dc.date.accessioned2024-01-03T06:00:13Z-
dc.date.available2024-01-03T06:00:13Z-
dc.date.issued2023-
dc.identifier.issn2169-3536-
dc.identifier.urihttps://scholarworks.bwise.kr/hongik/handle/2020.sw.hongik/32404-
dc.description.abstractTransformers have excelled in natural language processing and vision domains. This leads to the intriguing proposition: can Transformers be adapted to a more generalized framework, such as understanding general finite state machines? To explore this, we trained a Transformer network on compression algorithms such as Huffman and LZ4, viewing them as stepping stones towards mastering general finite state machines. Our analysis indicates that Transformers can adeptly internalize these methods and replicate essential states, echoing human-like interpretation via their attention mechanisms. This provides evidence of their capability to decipher practical finite state machines. Examining the attention maps offers insights into the Transformer's methodology when engaging with these compression techniques. In Huffman encodings, the encoder predominantly focuses on input statistics to define the present state, which the decoder subsequently leverages to produce output bits. Remarkably, with a 2nd-order Markov input, the encoder's attention is prominently directed at the previous two symbols, underscoring its proficiency in summarizing input statistics. The cross-attention maps further elucidate the exact association between input symbols and output bits. For the more complex LZ4 compression, the attention maps vividly portray the Transformer's processing of input sequences and the close linkage between the input and resulting output bit stream. This study underscores the Transformer's proficiency in comprehending compression algorithms and its keen ability to grasp input statistics, implying that its mechanisms, as illustrated by attention maps, provide profound interpretability.-
dc.format.extent10-
dc.language영어-
dc.language.isoENG-
dc.publisherIEEE-INST ELECTRICAL ELECTRONICS ENGINEERS INC-
dc.titleHow Does a Transformer Learn Compression? An Attention Study on Huffman and LZ4-
dc.typeArticle-
dc.publisher.location미국-
dc.identifier.doi10.1109/ACCESS.2023.3341512-
dc.identifier.scopusid2-s2.0-85179829744-
dc.identifier.wosid001129536100001-
dc.identifier.bibliographicCitationIEEE ACCESS, v.11, pp 140559 - 140568-
dc.citation.titleIEEE ACCESS-
dc.citation.volume11-
dc.citation.startPage140559-
dc.citation.endPage140568-
dc.type.docTypeArticle-
dc.description.isOpenAccessY-
dc.description.journalRegisteredClassscie-
dc.description.journalRegisteredClassscopus-
dc.relation.journalResearchAreaComputer Science-
dc.relation.journalResearchAreaEngineering-
dc.relation.journalResearchAreaTelecommunications-
dc.relation.journalWebOfScienceCategoryComputer Science, Information Systems-
dc.relation.journalWebOfScienceCategoryEngineering, Electrical & Electronic-
dc.relation.journalWebOfScienceCategoryTelecommunications-
dc.subject.keywordAuthorAttention-
dc.subject.keywordAuthorcompression-
dc.subject.keywordAuthorfinite state machine-
dc.subject.keywordAuthorHuffman coding-
dc.subject.keywordAuthorLZ4-
dc.subject.keywordAuthortransformer-
Files in This Item
There are no files associated with this item.
Appears in
Collections
College of Engineering > School of Electronic & Electrical Engineering > 1. Journal Articles

qrcode

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.

Altmetrics

Total Views & Downloads

BROWSE