I have a dataset with multiple labels which I have annotated.
When I run textcat.batch train - would I expect the performance of my labels to be the same as training on a data set with just a single label?
dataset A 2 labels: HOTDOG & NOTHOTDOG
dataset B: 1 label HOTDOG
dataset C 1 label NOTHOTDOG
textcat.batchtrain model datasetA have the same performance as training 2 separate models on datasets B & C and combining their outputs?
For separate textcat.batch labels is each label trained separately? or is there any ‘leak’
Currently working on comparing these empirically - but some insight and tips would be great