What is meaning of "fully correct" in NER active learning instructions?

hi @AbeHandler,

Thanks for your question and welcome to the Prodigy community :wave:

For ner.teach, decisions are only for one span being correct at a time. This is why it's also called a "binary" annotation in the docs:

The binary ner.teach workflow implements uncertainty sampling with beam search: for each example, the annotation model gets a number of analyses and asks you to accept or reject the entity analyses it’s most uncertain about. Based on your decisions, the model is updated in the loop and guided towards better predictions . Even if you don’t update with the full gold-standard annotation, confirming correct analyses and eliminating “dead ends” can very quickly move the model towards the right answer, since all token-based tagging decisions influence each other. Collecting binary feedback is also very fast and allows the annotator to focus on one concept and decision at a time, reducing the potential for human error and inconsistent data.

Yes, it would be correct because it's only evaluating whether "Germany" is correct. Very likely, the next span it'll predict would be "Spain".

If you want to "fully correct", that's what the ner.correct recipe is for. Suggestions for the ner.correct should be accepted when all entities are identified.

There are some old posts that mention the "information loss" when using binary annotations as opposed to "gold" (or manual) annotations:

Hope this helps!