hi @alvaro.marlo!
Check out this previous post:
The "E" is epochs and it seems like your dataset is so large (240,000+ training), it can't make a full pass (epoch) on your entire dataset.
I found this spaCy discussion post on advice on training large NER datasets. They recommended a few suggestions like creating a labels file (more to speed up the process), modifying the learning rate (will need some experimentation), and reducing the evaluation set. If you have further questions related to training, I would recommend posting on the spaCy discussion as the spaCy core team answers those posts (and prodigy train
is really just a wrapper for spacy train
, so your questions are really spaCy questions).