vllm.tokenizers ¶
Modules:
| Name | Description |
|---|---|
deepseek_v32 | |
deepseek_v4 | |
deepseek_v4_encoding | DeepSeek-V4 Encoding |
detokenizer_utils | |
fastokens |
|
grok2 | Tokenizer for Grok-2 .tok.json format. |
hf | |
kimi_audio | Tokenizer for Kimi-Audio using TikToken. |
mistral | |
qwen_vl | |
registry | |
get_tokenizer ¶
get_tokenizer(
tokenizer_name: str | Path,
*args,
tokenizer_cls: type[_T] = TokenizerLike,
trust_remote_code: bool = False,
revision: str | None = None,
download_dir: str | None = None,
**kwargs,
) -> _T
Gets a tokenizer for the given model name via HuggingFace or ModelScope.
Source code in vllm/tokenizers/registry.py
maybe_make_thread_pool ¶
maybe_make_thread_pool(tokenizer: _T, copies: int = 1)
If tokenizer is a PreTrainedTokenizerFast, modify the tokenizer in-place to make the public interface thread-safe by routing calls through a deep-copied tokenizer pool.
Note that: - Only TokenizerLike's public interface is thread-safe. This doesn't include _tokenizer property nor any mutation methods like add_special_tokens or add_tokens. - Adjacent method calls could happen on different deep copies.