[input]
tokenizer = { bert = { vocab = "bert-base-german-cased-vocab.txt" } }
[labeler]
labels = "sticker.labels"
encoders = [
{ name = "dep", encoder = { dependency = { encoder = { relativepos = "xpos" }, root_relation = "root" } } },
{ name = "lemma", encoder = { lemma = "form" } },
{ name = "pos", encoder = { sequence = "xpos" } },
]
[model]
parameters = "epoch-99"
position_embeddings = "model"
pretrain_config = "bert_config.json"
pretrain_type = "bert"