Tokenization of Multilingual Texts using Language-Specific Tokenizers
Multi-Tokenizer is a Python package that provides tokenization of multilingual texts using language-specific tokenizers. The package is designed to be used in a variety of applications, including natural language processing, machine learning, and data analysis. Behind the scenes, the package uses lingua
library to detect the language of the text segments, the tokenizers
library to create language-specific tokenizers, and then tokenizes the text segments using the appropriate tokenizer. Multi-tokenizer introduces additional special tokens to handle the language-specific tokenization, which can be used to reconstruct the original text segments after tokenization and allows the models to differentiate between the languages in the text segments.
pip install multi-tokenizer
git clone https://github.com/chandralegend/multi-tokenizer.git
cd multi-tokenizer
pip install .
from multi_tokenizer import MultiTokenizer, PretrainedTokenizers
# specify the language tokenizers to be used
lang_tokenizers = [
PretrainedTokenizers.ENGLISH,
PretrainedTokenizers.CHINESE,
PretrainedTokenizers.HINDI,
]
# create a multi-tokenizer object (split_text=True to split the text into segments, for better language detection)
tokenizer = MultiTokenizer(lang_tokenizers, split_text=True)
sentence = "Translate this hindi sentence to english - बिल्ली बहुत प्यारी है."
# Pretokenize the text
pretokenized_text = tokenizer.pre_tokenize(sentence) # [('<EN>', (0, 1)), ('Translate', (1, 10)), ('Ġthis', (10, 15)), ('Ġhindi', (15, 21)), ...]
# Encode the text
ids, tokens = tokenizer.encode(sentence) # [3, 7235, 6614, 86, 755, 775, 10763, 83, 19412, 276, ...], ['<EN>', 'Tr', 'ans', 'l', 'ate', 'Ġthis', 'Ġhind', ...]
# Decode the tokens
decoded_text = tokenizer.decode(ids) # Translate this hindi sentence to english - बिल्ली बहुत प्यारी है.
- Use the VSCode Dev Containers for easy setup (Recommended)
- Install dev dependencies
pip install poetry poetry install
- Add the directory to safe.directory
git config --global --add safe.directory /workspaces/multi-tokenizer
- Run the following command to lint and format the code
pre-commit run --all-files
- To install pre-commit hooks, run the following command (Recommended)
pre-commit install
Run the tests using the following command
pytest -n "auto"
- Approach 1: Individual tokenizers for each language
- Approach 2: Unified tokenization approach across languages using utf-8 encondings