# Copyright 2018 T5 Authors and HuggingFace Inc. team. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. """Tokenization class for model T5.""" import re from tokenizers import Tokenizer, decoders, pre_tokenizers, processors from tokenizers.models import Unigram from ...tokenization_utils_tokenizers import TokenizersBackend from ...utils import logging logger = logging.get_logger(__name__) VOCAB_FILES_NAMES = {"vocab_file": "spiece.model", "tokenizer_file": "tokenizer.json"} class T5Tokenizer(TokenizersBackend): """ Construct a T5 tokenizer (backed by HuggingFace's *tokenizers* library). Based on [Unigram](https://huggingface.co/docs/tokenizers/python/latest/components.html?highlight=unigram#models). This tokenizer inherits from [`TokenizersBackend`] which contains most of the main methods. Users should refer to this superclass for more information regarding those methods. Args: vocab_file (`str`, *optional*): [SentencePiece](https://github.com/google/sentencepiece) file (generally has a *.spm* extension) that contains the vocabulary necessary to instantiate a tokenizer. eos_token (`str`, *optional*, defaults to `""`): The end of sequence token. When building a sequence using special tokens, this is not the token that is used for the end of sequence. The token used is the `sep_token`. unk_token (`str`, *optional*, defaults to `""`): The unknown token. A token that is not in the vocabulary cannot be converted to an ID and is set to be this token instead. pad_token (`str`, *optional*, defaults to `""`): The token used for padding, for example when batching sequences of different lengths. extra_ids (`int`, *optional*, defaults to 100): Add a number of extra ids added to the vocabulary for use as sentinels. These tokens are accessible as "" where "{%d}" is a number between 0 and extra_ids-1. These tokens can be retrieved by calling get_sentinel_tokens method and token ids can be by calling get_sentinel_token_ids method additional_special_tokens (`list[str]`, *optional*): Additional special tokens used by the tokenizer. vocab (`str`, `dict` or `list`, *optional*): Custom vocabulary dict. If not provided, a minimal vocabulary is created using the special tokens. """ vocab_files_names = VOCAB_FILES_NAMES model_input_names = ["input_ids", "attention_mask"] model = Unigram def __init__( self, vocab: str | list[tuple[str, float]] | None = None, eos_token="", unk_token="", pad_token="", extra_ids=100, additional_special_tokens=None, **kwargs, ): self._extra_ids = extra_ids # Handle extra_ids and additional_special_tokens if additional_special_tokens is not None: extra_tokens = [x for x in additional_special_tokens if "" for i in range(extra_ids)] elif extra_ids > 0 and extra_ids != len(extra_tokens): raise ValueError( f"Both extra_ids ({extra_ids}) and additional_special_tokens ({additional_special_tokens}) are" " provided to T5Tokenizer. In this case the additional_special_tokens must include the extra_ids" " tokens" ) else: extra_tokens = [f"" for i in range(extra_ids)] additional_special_tokens = extra_tokens # T5 vocab structure: =0, =1, =2, then regular vocab, then extra_ids in reverse if vocab is not None: self._vocab_scores = vocab else: self._vocab_scores = [ (str(pad_token), 0.0), (str(eos_token), 0.0), (str(unk_token), 0.0), ("▁", -2.0), # Space token ] for i in range(extra_ids - 1, -1, -1): self._vocab_scores.append((f"", 0.0)) self._tokenizer = Tokenizer( Unigram( self._vocab_scores, unk_id=2, byte_fallback=False, ) ) self._tokenizer.normalizer = None self._tokenizer.pre_tokenizer = pre_tokenizers.Sequence( [ pre_tokenizers.WhitespaceSplit(), pre_tokenizers.Metaspace(replacement="▁", prepend_scheme="always", split=True), ] ) self._tokenizer.decoder = decoders.Metaspace(replacement="▁", prepend_scheme="always", split=True) super().__init__( eos_token=eos_token, unk_token=unk_token, pad_token=pad_token, extra_ids=extra_ids, additional_special_tokens=additional_special_tokens, **kwargs, ) self._tokenizer.post_processor = processors.TemplateProcessing( single=["$A", ""], pair=["$A", "", "$B", ""], special_tokens=[ ("", self.eos_token_id), ], ) def get_sentinel_tokens(self): """Get the list of sentinel tokens (extra_id tokens) from additional_special_tokens.""" return list( set(filter(lambda x: bool(re.search(r"", x)) is not None, self.additional_special_tokens)) ) def get_sentinel_token_ids(self): """Get the token IDs for sentinel tokens.""" return [self.convert_tokens_to_ids(token) for token in self.get_sentinel_tokens()] __all__ = ["T5Tokenizer"]