i am using ner.teach recipe for long documents, after saving few annotations to db and then to a .jsonl file (via db-out recipe) i have noticed that the "text" of the annotations is very short. Why does't it save the real text? ( i mean the real text on the source)
How this will impact the training after the teaching ?
I mean...training the model with these new annotations will ""corrupt""" the current NER model that i have trained with long documents?
P.S. My current model does not have a parser (or sentencizer), so, for the moment, i do not split sentences.