code2prompt_core 4.2.0

A command-line (CLI) tool to generate an LLM prompt from codebases of any size, fast.
Documentation
//! This module encapsulates the logic for counting the tokens in the rendered text.
use log::debug;
use serde::{Deserialize, Serialize};
use std::fmt;
use std::sync::OnceLock;
use tiktoken_rs::{CoreBPE, cl100k_base, o200k_base, p50k_base, p50k_edit, r50k_base};

#[derive(Default, Debug, Clone, Copy, PartialEq, Deserialize, Serialize)]
#[serde(rename_all = "lowercase")]
pub enum TokenFormat {
    #[default]
    Raw,
    Format,
}

impl fmt::Display for TokenFormat {
    fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
        match self {
            TokenFormat::Raw => write!(f, "Raw"),
            TokenFormat::Format => write!(f, "Formatted"),
        }
    }
}

/// Tokenizer types supported by tiktoken.
#[derive(Debug, Clone, Copy, PartialEq, Eq, Default, Serialize, Deserialize)]
pub enum TokenizerType {
    #[serde(alias = "o200k")]
    O200kBase,
    #[default]
    #[serde(alias = "cl100k")]
    Cl100kBase,
    #[serde(alias = "p50k")]
    P50kBase,
    #[serde(alias = "p50k_edit")]
    P50kEdit,
    #[serde(alias = "r50k")]
    R50kBase,
}

impl fmt::Display for TokenizerType {
    fn fmt(&self, f: &mut fmt::Formatter<'_>) -> fmt::Result {
        match self {
            TokenizerType::O200kBase => write!(f, "o200k (GPT-4o)"),
            TokenizerType::Cl100kBase => write!(f, "cl100k (ChatGPT)"),
            TokenizerType::P50kBase => write!(f, "p50k (Code models)"),
            TokenizerType::P50kEdit => write!(f, "p50k_edit (Edit models)"),
            TokenizerType::R50kBase => write!(f, "r50k (GPT-3)"),
        }
    }
}

/// Returns a description of the tokenizer type.
impl TokenizerType {
    pub fn description(&self) -> &'static str {
        match self {
            TokenizerType::O200kBase => "OpenAI models, ChatGPT-4o",
            TokenizerType::Cl100kBase => "ChatGPT models, text-embedding-ada-002",
            TokenizerType::P50kBase => "Code models, text-davinci-002, text-davinci-003",
            TokenizerType::P50kEdit => {
                "Edit models like text-davinci-edit-001, code-davinci-edit-001"
            }
            TokenizerType::R50kBase => "GPT-3 models like davinci",
        }
    }
}

// Cache tokenizers to avoid expensive re-initialization
static O200K_BASE: OnceLock<CoreBPE> = OnceLock::new();
static CL100K_BASE: OnceLock<CoreBPE> = OnceLock::new();
static P50K_BASE: OnceLock<CoreBPE> = OnceLock::new();
static P50K_EDIT: OnceLock<CoreBPE> = OnceLock::new();
static R50K_BASE: OnceLock<CoreBPE> = OnceLock::new();

/// Counts the tokens in the provided text using the specified tokenizer type.
///
/// # Arguments
///
/// * `rendered` - The text to count tokens in
/// * `tokenizer_type` - The tokenizer encoding to use
///
/// # Returns
///
/// * `usize` - The number of tokens in the text
pub fn count_tokens(rendered: &str, tokenizer_type: &TokenizerType) -> usize {
    use std::time::Instant;
    let start = Instant::now();

    let bpe = match tokenizer_type {
        TokenizerType::O200kBase => O200K_BASE.get_or_init(|| o200k_base().unwrap()),
        TokenizerType::Cl100kBase => CL100K_BASE.get_or_init(|| cl100k_base().unwrap()),
        TokenizerType::P50kBase => P50K_BASE.get_or_init(|| p50k_base().unwrap()),
        TokenizerType::P50kEdit => P50K_EDIT.get_or_init(|| p50k_edit().unwrap()),
        TokenizerType::R50kBase => R50K_BASE.get_or_init(|| r50k_base().unwrap()),
    };

    let token_count = bpe.encode_with_special_tokens(rendered).len();

    if std::env::var("DEBUG_TOKENIZER").is_ok() {
        debug!(
            "Tokenized {} chars in {:?}",
            rendered.len(),
            start.elapsed()
        );
    }

    token_count
}