Training table doubts

hi @alvaro.marlo!

Check out this previous post:

The "E" is epochs and it seems like your dataset is so large (240,000+ training), it can't make a full pass (epoch) on your entire dataset.

I found this spaCy discussion post on advice on training large NER datasets. They recommended a few suggestions like creating a labels file (more to speed up the process), modifying the learning rate (will need some experimentation), and reducing the evaluation set. If you have further questions related to training, I would recommend posting on the spaCy discussion as the spaCy core team answers those posts (and prodigy train is really just a wrapper for spacy train, so your questions are really spaCy questions).