I've been developing a document parser using Prodigy, and the tool has made it very easy to fly through documents. It seems to have almost made it too easy. While training I take periodic breaks to retrain and check the scores of the model through the train recipe. Through almost all of the labeling process the training scores have only increased. However, now they are decreasing. This could be for a lot of reasons, accidental mislabels, underfit, overfit, etc. but whatever the reason I'm trying to find a way to loop through training sessions using different amounts of annotated data. I know how I can easily loop through the training sessions, but i cant seem to find any way to store the best score from each session. The only way i can see as of now is to manually record each session. Is there a way i can store the training scores of each model for review later?