File: control

package info (click to toggle)
tokenizers 0.20.3%2Bdfsg-1
  • links: PTS, VCS
  • area: main
  • in suites: experimental
  • size: 5,480 kB
  • sloc: python: 4,499; javascript: 419; makefile: 124
file content (69 lines) | stat: -rw-r--r-- 2,780 bytes parent folder | download
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
Source: tokenizers
Section: python
Priority: optional
Maintainer: Debian Deep Learning Team <debian-ai@lists.debian.org>
Uploaders: Kohei Sendai <kouhei.sendai@gmail.com>
Rules-Requires-Root: no
#Testsuite: autopkgtest-pkg-python
Standards-Version: 4.7.2
Homepage: https://github.com/huggingface/tokenizers
Vcs-Browser: https://salsa.debian.org/deeplearning-team/tokenizers
Vcs-Git: https://salsa.debian.org/deeplearning-team/tokenizers.git
Build-Depends: debhelper-compat (= 13),
               dh-sequence-python3,
               cargo,
               librust-derive-builder-dev,
               librust-env-logger-dev,
               librust-esaxx-rs-dev,
               librust-indicatif-dev,
               librust-itertools-dev,
               librust-libc-dev, librust-macro-rules-attribute-dev,
               librust-monostate-dev,
               librust-ndarray-dev,
               librust-numpy-dev,
               librust-onig-dev,
               librust-pyo3-dev,
               librust-rand-dev,
               librust-rayon-cond-dev, librust-rayon-dev,
               librust-regex-dev,
               librust-regex-syntax-dev,
               librust-serde-derive-dev,
               librust-serde-dev,
               librust-serde-json-dev,
               librust-spm-precompiled-dev,
               librust-thiserror-1-dev,
               librust-unicode-categories-dev,
               librust-unicode-normalization-alignments-dev,
               pybuild-plugin-pyproject,
               python3-all,
               python3-filelock,
               python3-fsspec,
               python3-huggingface-hub,
               python3-maturin,
               python3-packaging,
               python3-requests,
               python3-setuptools,
               python3-tqdm,
               python3-typing-extensions,
               python3-yaml,
               rustc

Package: python3-tokenizers
Section: python
Architecture: any
Depends: ${misc:Depends}, ${python3:Depends}, ${shlibs:Depends}
Description: Fast State-of-the-Art Tokenizers for Research and Production
 Provides an implementation of today's most used tokenizers, with a focus on
 performance and versatility.
 .
 Main features:
  * Train new vocabularies and tokenize, using today's most used tokenizers.
  * Extremely fast (both training and tokenization), thanks to the Rust
    implementation. Takes less than 20 seconds to tokenize a GB of text on
    a server's CPU.
  * Easy to use, but also extremely versatile.
  * Designed for research and production.
  * Normalization comes with alignments tracking. It's always possible to get
    the part of the original sentence that corresponds to a given token.
  * Does all the pre-processing: Truncate, Pad, add the special tokens your
    model needs.