When I run:
prodigy ner.batch-train my_dataset en_core_web_sm --output ./ --n-iter 25 --eval-split 0.2 --dropout 0.2
It returns:
Traceback (most recent call last): File "/usr/lib/python3.5/runpy.py", line 184, in _run_module_as_main "__main__", mod_spec) File "/usr/lib/python3.5/runpy.py", line 85, in _run_code exec(code, run_globals) File "/home/rohan/.local/lib/python3.5/site-packages/prodigy/__main__.py", line 380, in <module> controller = recipe(*args, use_plac=True) File "cython_src/prodigy/core.pyx", line 212, in prodigy.core.recipe.recipe_decorator.recipe_proxy File "/usr/local/lib/python3.5/dist-packages/plac_core.py", line 328, in call cmd, result = parser.consume(arglist) File "/usr/local/lib/python3.5/dist-packages/plac_core.py", line 207, in consume return cmd, self.func(*(args + varargs + extraopts), **kwargs) File "/home/rohan/.local/lib/python3.5/site-packages/prodigy/recipes/ner.py", line 602, in batch_train examples = list(split_sentences(model.orig_nlp, examples)) File "cython_src/prodigy/components/preprocess.pyx", line 39, in split_sentences File "/home/rohan/.local/lib/python3.5/site-packages/spacy/language.py", line 708, in pipe for doc, context in izip(docs, contexts): File "/home/rohan/.local/lib/python3.5/site-packages/spacy/language.py", line 736, in pipe for doc in docs: File "nn_parser.pyx", line 221, in pipe File "/home/rohan/.local/lib/python3.5/site-packages/spacy/util.py", line 460, in minibatch batch = list(itertools.islice(items, int(batch_size))) File "nn_parser.pyx", line 226, in pipe File "nn_parser.pyx", line 244, in spacy.syntax.nn_parser.Parser.predict File "nn_parser.pyx", line 257, in spacy.syntax.nn_parser.Parser.greedy_parse File "/home/rohan/.local/lib/python3.5/site-packages/thinc/neural/_classes/model.py", line 169, in __call__ return self.predict(x) File "/home/rohan/.local/lib/python3.5/site-packages/thinc/neural/_classes/model.py", line 133, in predict y, _ = self.begin_update(X, drop=None) File "_parser_model.pyx", line 214, in spacy.syntax._parser_model.ParserModel.begin_update File "_parser_model.pyx", line 262, in spacy.syntax._parser_model.ParserStepModel.__init__ File "/home/rohan/.local/lib/python3.5/site-packages/thinc/neural/_classes/feed_forward.py", line 46, in begin_update X, inc_layer_grad = layer.begin_update(X, drop=drop) File "/home/rohan/.local/lib/python3.5/site-packages/thinc/api.py", line 295, in begin_update X, bp_layer = layer.begin_update(layer.ops.flatten(seqs_in, pad=pad), drop=drop) File "/home/rohan/.local/lib/python3.5/site-packages/thinc/neural/_classes/feed_forward.py", line 46, in begin_update X, inc_layer_grad = layer.begin_update(X, drop=drop) File "/home/rohan/.local/lib/python3.5/site-packages/thinc/neural/_classes/resnet.py", line 25, in begin_update y, bp_y = self._layers[0].begin_update(X, drop=drop) File "/home/rohan/.local/lib/python3.5/site-packages/thinc/neural/_classes/feed_forward.py", line 46, in begin_update X, inc_layer_grad = layer.begin_update(X, drop=drop) File "/home/rohan/.local/lib/python3.5/site-packages/thinc/neural/_classes/convolution.py", line 33, in begin_update X__bo = self.ops.seq2col(X__bi, self.nW) File "ops.pyx", line 557, in thinc.neural.ops.NumpyOps.seq2col File "ops.pyx", line 401, in thinc.neural.ops.NumpyOps.allocate MemoryError
I was reading This Topic and This Another Topic and this doesn't help. I checked numpy, and everything seems fine. Still not sure, why am I not able to train it.