hi @nlp-guy !
Unfortunately, most of the functionality you're looking for isn't available out-of-the-box with Prodigy. As you've probably seen, typically we would recommend converting into images for every n-th frame or use some combination/similar functionality from the audio interface . I've included several of related posts where other users have asked for similar features.
I have made an internal issue about your request. Let us know if you have any additional details of what the functionality would look like. This could help us out if we look more out developing these features in the future.
Hi! You could probably build an interface like this using custom HTML templates and JavaScript, but Prodigy doesn’t have a built-in interface for playing and labelling videos. If you need to actually play videos, jump around frames etc., there are probably much more specialised tools that will do a much better job at this.
That said, in many cases, what you’re annotating in videos isn’t actually the live video, but rather every n-th frame as an image. That’s a workflow you can probably build qu…
Hi! By keypoint annotation, do you mean annotating specific points on an image? There's no separate annotation mode for just points but depending on the task, one solution could be to use the regular bounding box and enable visualising the center point with show_bounding_box_center: https://prodi.gy/docs/api-interfaces#image_manual-settings You can also set image_manual_from_center to enable drawing from the center if that feels more natural. And you can customise the box to make it less promine…
Hi! The audio and video UI support annotating segments in the audio track – there's currently no interface for object tracking annotation in video files.
However, if your goal is to just mark the speaker (and not track the speaker's movement across frames etc.), you could probably use a combination of the audio or video interface, and the image_manual interface with a still image from the video. See here for the docs on custom interfaces with blocks. The solution here re…
Yay, I like that spirit
I haven't worked on object tracking use cases myself, but just to make sure I understand the requirements correctly: your goal is essentially to annotate a given object / bounding box across the frames of your video, right? Since you'd be working on the individual frames, one option could be to use a workflow like image.manual and stream in the selected frames as images, in order. You can probably do this all programmatically in Python as part of …