Hey everyone! We've been making good progress on some really cool new features for Prodigy v1.10. One of them is going to be a fully new and improved manual UI for annotating images So if that's an area you're working in, andy you're a Prodigy user, maybe you want to help us test it?
Features include
- allow moving and resizing existing shapes
- allow moving polygon points
- freehand mode (lasso) for drawing fully custom shapes
- enhanced data format: all bounding boxes now also expose their
width
,height
,x
,y
andcenter
, and all imagespans
include atype
("rect"
,"polygon"
,"freehand"
) - new config settings:
-
image_manual_stroke_width
: stroke width of boxes, also used to calculate transform points -
image_manual_font_size
: font size of the box label -
image_manual_show_labels
: (not pictured) default setting for toggle to show/hide box labels -
image_manual_modes
: annotation modes to allow and display, in order. Defaults to["rect", "polygon", "freehand"]
. -
image_manual_from_center
: enable drawing and resizing boxes starting with the center, which can be much more efficient
-
Beta testing and requirements
If you want to help beta test the new feature and try them on your data, feel free to send me an email at ines@explosion.ai or a DM on Twitter. Requirements are:
- Current Prodigy user on v1.9 (just include your order ID starting with #EX... when you get in touch)
- Should have an immediate project you can test it on, ideally involving some of the features added in the new interface
- Big plus: you already have a downstream pipeline in place that uses the data to train a model, so you can test the end-to-end workflow
Obvious disclaimer: This is a beta, so things may be broken (Hopefully not too much, though!) Definitely test it with a fresh and separate dataset, though.
If you have any questions, I'm also happy to answer them in this thread! P.S. If you're working with dependency/relation annotation, including complex NER relations, coref etc. tasks, keep an eye out for another call for beta testers soon