LLM-specific code like tokenization and kv-cache logic which (currently) has to live outside the model graph.