[package]
edition = "2024"
rust-version = "1.85"
name = "miktik"
version = "0.2.0"
build = false
exclude = [
".local/**",
"scripts/**",
"tests/**",
"Cargo.lock",
]
autolib = false
autobins = false
autoexamples = false
autotests = false
autobenches = false
description = "A unified, multi-backend tokenizer library for LLMs"
homepage = "https://github.com/Darkatse/MikTik"
documentation = "https://docs.rs/miktik"
readme = "README.md"
keywords = [
"llm",
"tokenizer",
"tiktoken",
"huggingface",
"rust",
]
categories = ["text-processing"]
license = "MIT"
repository = "https://github.com/Darkatse/MikTik"
[features]
default = ["openai"]
full = [
"openai",
"huggingface",
"sentencepiece",
]
huggingface = ["dep:tokenizers"]
openai = ["dep:tiktoken-rs"]
sentencepiece = [
"huggingface",
"dep:sentencepiece-model",
]
[lib]
name = "miktik"
path = "src/lib.rs"
[dependencies.sentencepiece-model]
version = "0.1.4"
optional = true
[dependencies.thiserror]
version = "2"
[dependencies.tiktoken-rs]
version = "0.9.1"
optional = true
[dependencies.tokenizers]
version = "0.22.2"
features = ["onig"]
optional = true
default-features = false
[profile.dev.package.sentencepiece-model]
opt-level = 3
[profile.dev.package.tiktoken-rs]
opt-level = 3
[profile.dev.package.tokenizers]
opt-level = 3
[profile.test.package.sentencepiece-model]
opt-level = 3
[profile.test.package.tiktoken-rs]
opt-level = 3
[profile.test.package.tokenizers]
opt-level = 3