{"title":"Egalitarian Language Representation in Language Models: It All Begins with Tokenizers","authors":"Menan Velayuthan, Kengatharaiyer Sarveswaran","doi":"arxiv-2409.11501","DOIUrl":null,"url":null,"abstract":"Tokenizers act as a bridge between human language and the latent space of\nlanguage models, influencing how language is represented in these models. Due\nto the immense popularity of English-Centric Large Language Models (LLMs),\nefforts are being made to adapt them for other languages. However, we\ndemonstrate that, from a tokenization standpoint, not all tokenizers offer fair\nrepresentation for complex script languages such as Tamil, Sinhala, and Hindi,\nprimarily due to the choice of pre-tokenization methods. We go further to show\nthat pre-tokenization plays a more critical role than the tokenization\nalgorithm itself in achieving an egalitarian representation of these complex\nscript languages. To address this, we introduce an improvement to the Byte Pair\nEncoding (BPE) algorithm by incorporating graphemes, which we term Grapheme\nPair Encoding (GPE). Our experiments show that grapheme-based character\nextraction outperforms byte-level tokenizers for complex scripts. We validate\nthis approach through experiments on Tamil, Sinhala, and Hindi.","PeriodicalId":501030,"journal":{"name":"arXiv - CS - Computation and Language","volume":"50 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Computation and Language","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.11501","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Tokenizers act as a bridge between human language and the latent space of
language models, influencing how language is represented in these models. Due
to the immense popularity of English-Centric Large Language Models (LLMs),
efforts are being made to adapt them for other languages. However, we
demonstrate that, from a tokenization standpoint, not all tokenizers offer fair
representation for complex script languages such as Tamil, Sinhala, and Hindi,
primarily due to the choice of pre-tokenization methods. We go further to show
that pre-tokenization plays a more critical role than the tokenization
algorithm itself in achieving an egalitarian representation of these complex
script languages. To address this, we introduce an improvement to the Byte Pair
Encoding (BPE) algorithm by incorporating graphemes, which we term Grapheme
Pair Encoding (GPE). Our experiments show that grapheme-based character
extraction outperforms byte-level tokenizers for complex scripts. We validate
this approach through experiments on Tamil, Sinhala, and Hindi.