Alright, that's perfect - I'm looking forward to the new update! 
In the meantime, using the looped setup described in the last paragraph of Issue in multi-session mode: duplicated annotation tasks and different order?, two annotators repeatedly receive internal server error. Looking at the logs, I find:
Exception when serving /give_answers
Traceback (most recent call last):
File "/usr/local/lib/python3.6/site-packages/waitress/channel.py", line 336, in service
task.service()
File "/usr/local/lib/python3.6/site-packages/waitress/task.py", line 175, in service
self.execute()
File "/usr/local/lib/python3.6/site-packages/waitress/task.py", line 452, in execute
app_iter = self.channel.server.application(env, start_response)
File "/usr/local/lib/python3.6/site-packages/hug/api.py", line 451, in api_auto_instantiate
return module.__hug_wsgi__(*args, **kwargs)
File "/usr/local/lib/python3.6/site-packages/falcon/api.py", line 244, in __call__
responder(req, resp, **params)
File "/usr/local/lib/python3.6/site-packages/hug/interface.py", line 789, in __call__
raise exception
File "/usr/local/lib/python3.6/site-packages/hug/interface.py", line 762, in __call__
self.render_content(self.call_function(input_parameters), context, request, response, **kwargs)
File "/usr/local/lib/python3.6/site-packages/hug/interface.py", line 698, in call_function
return self.interface(**parameters)
File "/usr/local/lib/python3.6/site-packages/hug/interface.py", line 100, in __call__
return __hug_internal_self._function(*args, **kwargs)
File "/usr/local/lib/python3.6/site-packages/prodigy/_api/hug_app.py", line 282, in give_answers
controller.receive_answers(answers, session_id=session_id)
File "cython_src/prodigy/core.pyx", line 148, in prodigy.core.Controller.receive_answers
File "/usr/local/lib/python3.6/site-packages/prodigy/components/db.py", line 378, in add_examples
input_hash=eg[INPUT_HASH_ATTR],
KeyError: '_input_hash'
What could be the issue here? Restarting the server helps, but unfortunately only temporarily. The full recipe is:
import prodigy
from prodigy.components.db import connect
from prodigy.components.loaders import JSONL
@prodigy.recipe('newstsa',
dataset=prodigy.recipe_args['dataset'],
file_path=("Path to texts", "positional", None, str))
def sentiment(dataset, file_path):
"""Annotate the sentiment of texts using different mood options."""
stream = get_stream_loop(file_path, dataset)
return {
'dataset': dataset, # save annotations in this dataset
'view_id': 'choice', # use the choice interface
"config": {
"choice_auto_accept": True, # auto-accept example, once the users selects an option
"instructions": "/prodigy/manual.html"
},
'on_exit': on_exit,
'stream': stream,
}
def get_stream_loop(file_path, dataset):
# to prevent that no tasks are shown even though there are still unlabeled tasks left
# https://support.prodi.gy/t/struggling-to-create-a-multiple-choice-image-classification/1345/2
db = connect()
while True:
stream = get_stream(file_path)
hashes_in_dataset = db.get_task_hashes(dataset)
yielded = False
for eg in stream:
# Only send out task if its hash isn't in the dataset yet, which should mean that we will not have duplicates
if eg["_task_hash"] not in hashes_in_dataset:
yield eg
yielded = True
if not yielded:
break
def get_stream(file_path):
stream = JSONL(file_path) # load in the JSONL file
stream = add_options(stream) # add options to each task
for eg in stream:
eg = prodigy.set_hashes(eg)
yield eg
def add_options(stream):
"""Helper function to add options to every task in a stream."""
options = [{'id': 'positive', 'text': '😊 positive'},
{'id': 'neutral', 'text': '😶 neutral'},
{'id': 'negative', 'text': '🙁 negative'},
{'id': 'posneg', 'text': '😊+🙁 pos. and neg.'}]
for task in stream:
task['options'] = options
yield task
def on_exit(controller):
"""Get all annotations in the dataset, filter out the accepted tasks,
count them by the selected options and print the counts."""
# taken from https://prodi.gy/docs/workflow-custom-recipes#example-choice
examples = controller.db.get_dataset(controller.dataset)
examples = [eg for eg in examples if eg['answer'] == 'accept']
for option in ('positive', 'neutral', 'negative', 'posneg'):
count = get_count_by_option(examples, option)
print('Annotated {} {} examples'.format(count, option))
def get_count_by_option(examples, option):
filtered = [eg for eg in examples if option in eg['accept']]
return len(filtered)