tokenizers

View on PyPIReverse Dependencies (652)

0.21.1 tokenizers-0.21.1-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl
tokenizers-0.21.1-cp39-abi3-manylinux_2_17_i686.manylinux2014_i686.whl
tokenizers-0.21.1-cp39-abi3-manylinux_2_17_armv7l.manylinux2014_armv7l.whl
tokenizers-0.21.1-cp39-abi3-manylinux_2_17_s390x.manylinux2014_s390x.whl
tokenizers-0.21.1-cp39-abi3-manylinux_2_17_ppc64le.manylinux2014_ppc64le.whl
tokenizers-0.21.1-cp39-abi3-manylinux_2_17_aarch64.manylinux2014_aarch64.whl
tokenizers-0.21.1-cp39-abi3-win_amd64.whl
tokenizers-0.21.1-cp39-abi3-win32.whl
tokenizers-0.21.1-cp39-abi3-macosx_10_12_x86_64.whl
tokenizers-0.21.1-cp39-abi3-musllinux_1_2_x86_64.whl
tokenizers-0.21.1-cp39-abi3-musllinux_1_2_i686.whl
tokenizers-0.21.1-cp39-abi3-musllinux_1_2_armv7l.whl
tokenizers-0.21.1-cp39-abi3-musllinux_1_2_aarch64.whl
tokenizers-0.21.1-cp39-abi3-macosx_11_0_arm64.whl

Wheel Details

Project: tokenizers
Version: 0.21.1
Filename: tokenizers-0.21.1-cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64.whl
Download: [link]
Size: 3038874
MD5: 0364da4a197b81b040b9704d1018441a
SHA256: 2dd9a0061e403546f7377df940e866c3e678d7d4e9643d0461ea442b4f89e61a
Uploaded: 2025-03-13 10:51:06 +0000

dist-info

METADATA

Metadata-Version: 2.4
Name: tokenizers
Version: 0.21.1
Author: Anthony MOI <m.anthony.moi@gmail.com>
Author-Email: Nicolas Patry <patry.nicolas[at]protonmail.com>, Anthony Moi <anthony[at]huggingface.co>
Project-Url: Homepage, https://github.com/huggingface/tokenizers
Project-Url: Source, https://github.com/huggingface/tokenizers
Keywords: NLP,tokenizer,BPE,transformer,deep learning
Classifier: Development Status :: 5 - Production/Stable
Classifier: Intended Audience :: Developers
Classifier: Intended Audience :: Education
Classifier: Intended Audience :: Science/Research
Classifier: License :: OSI Approved :: Apache Software License
Classifier: Operating System :: OS Independent
Classifier: Programming Language :: Python :: 3
Classifier: Programming Language :: Python :: 3.9
Classifier: Programming Language :: Python :: 3.10
Classifier: Programming Language :: Python :: 3.11
Classifier: Programming Language :: Python :: 3.12
Classifier: Programming Language :: Python :: 3.13
Classifier: Programming Language :: Python :: 3 :: Only
Classifier: Topic :: Scientific/Engineering :: Artificial Intelligence
Requires-Python: >=3.9
Requires-Dist: huggingface-hub (<1.0,>=0.16.4)
Requires-Dist: pytest; extra == "testing"
Requires-Dist: requests; extra == "testing"
Requires-Dist: numpy; extra == "testing"
Requires-Dist: datasets; extra == "testing"
Requires-Dist: black (==22.3); extra == "testing"
Requires-Dist: ruff; extra == "testing"
Requires-Dist: sphinx; extra == "docs"
Requires-Dist: sphinx-rtd-theme; extra == "docs"
Requires-Dist: setuptools-rust; extra == "docs"
Requires-Dist: tokenizers[testing]; extra == "dev"
Provides-Extra: testing
Provides-Extra: docs
Provides-Extra: dev
Description-Content-Type: text/markdown; charset=UTF-8; variant=GFM
[Description omitted; length: 5006 characters]

WHEEL

Wheel-Version: 1.0
Generator: maturin (1.8.2)
Root-Is-Purelib: false
Tag: cp39-abi3-manylinux_2_17_x86_64.manylinux2014_x86_64

RECORD

Path Digest Size
tokenizers-0.21.1.dist-info/METADATA sha256=GL78TpkKOH8KBbB1Bba01wXgp9ImYCYPAtmQl3J4dZk 6774
tokenizers-0.21.1.dist-info/WHEEL sha256=f5fytCivPqoitFei8k2YKm7f1DR3oNmqLTfZvuAlYZc 127
tokenizers/implementations/bert_wordpiece.py sha256=sKCum0FKPYdSgJFJN8LDerVBoTDRSqyqSdrcm-lvQqI 5520
tokenizers/implementations/sentencepiece_bpe.py sha256=LwrofoohnUfME2lK2lQYoyQIhP84RP0CIlHRaj0hyNs 3738
tokenizers/implementations/sentencepiece_unigram.py sha256=SYiVXL8ZtqLXKpuqwnwmrfxgGotu8yAkOu7dLztEXIo 7580
tokenizers/implementations/char_level_bpe.py sha256=Q2ZEAW0xMQHF7YCUtmplwaxbU-J0P2NK4PJGMxUb-_c 5466
tokenizers/implementations/__init__.py sha256=VzAsplaIo7rl4AFO8Miu7ig7MfZjvonwVblZw01zR6M 310
tokenizers/implementations/base_tokenizer.py sha256=2TFZhLupaJiMDYGJuUNmxYJv-cnR8bDHmbMzaYpFROs 14206
tokenizers/implementations/byte_level_bpe.py sha256=OA_jyy3EQmYTa6hnf-EKwLOFuyroqFYOJz25ysM2BUk 4289
tokenizers/tools/__init__.py sha256=xG8caB9OHC8cbB01S5vYV14HZxhO6eWbLehsb70ppio 55
tokenizers/tools/visualizer.py sha256=0W90s4Qm8Nd6P-npqQX-bCMLQCfAEPk2qgj-K8r7OMc 14624
tokenizers/tools/visualizer-styles.css sha256=zAydq1oGWD8QEll4-eyL8Llw0B1sty_hpIE3tYxL02k 4850
tokenizers/processors/__init__.py sha256=xM2DEKwKtHIumHsszM8AMkq-AlaqvBZFXWgLU8SNhOY 307
tokenizers/processors/__init__.pyi sha256=hx767ZY8SHhxb_hiXPRxm-f_KcoR4XDx7vfK2c0lR-Q 11357
tokenizers/decoders/__init__.py sha256=hfwM6CFUDvlMGGL4-xsaaYz81K9P5rQI5ZL5UHWK8Y4 372
tokenizers/decoders/__init__.pyi sha256=qylMinPKn2_NAMC6BzcXqFyndoicLws34Q6OLH1gcp4 7378
tokenizers/normalizers/__init__.py sha256=_06w4cqRItveEgIddYaLMScgkSOkIAMIzYCesb5AA4U 841
tokenizers/normalizers/__init__.pyi sha256=lSFqDb_lPZBfRxEG99EcFEaU1HlnIhIQUu7zZIyP4AY 20898
tokenizers/trainers/__init__.py sha256=UTu22AGcp76IvpW45xLRbJWET04NxPW6NfCb2YYz0EM 248
tokenizers/trainers/__init__.pyi sha256=3TwFKts4me7zQfVRcSTmtXYiP4XwcRjfAYtwqoZVtoQ 5382
tokenizers/models/__init__.py sha256=eJZ4HTAQZpxnKILNylWaTFqxXy-Ba6OKswWN47feeV8 176
tokenizers/models/__init__.pyi sha256=clPTwiyjz7FlVdEuwo_3Wa_TmQrbZhW0SGmnNylepnY 16929
tokenizers/__init__.py sha256=ZE5ZagUvobBScrHBQdEobhx4wqM0bsq9F9aLYkBNjYQ 2615
tokenizers/__init__.pyi sha256=jw34WZXaYu8NBBJ2_cypfOqJYxI7CXKPzlveisXw4XQ 40182
tokenizers/pre_tokenizers/__init__.py sha256=wd6KYQA_RsGSQK-HeG9opTRhv4ttSRkyno2dk6az-PM 557
tokenizers/pre_tokenizers/__init__.pyi sha256=k9Jnez-t4ww_Pyj8OkVeo78v9TntcNn6ri9kwB590KE 23606
tokenizers/tokenizers.abi3.so sha256=kA2fIqNga2xHoJAhXfwqsJZQhWnWNipHoi0kZwt5yP8 9086856
tokenizers-0.21.1.dist-info/RECORD