I started by creating a dataset using prodigy dataset ... and then using prodigy ner.manual ... with a bunch of my own labels I annotated a bunch of examples.
I was initially planning on using the BERT model: en_trf_bertbaseuncased_lg, but trying to run batch-train with:
Hi! We currently do not have an NER model implementation using the transformer weights. See here for details:
So running a transformer model with ner.batch-train doesn't really make sense – you'd always be training a regular spaCy NER model (so you might as well use a blank en model). To use the transformer models with Prodigy likely also require slightly modified training recipes, since the updating works slightly differently in those cases (and has additional configuration options).