Running NER on tokenized data only: KeyError: "Parameter 'E' for model 'hashembed' has not been allocated yet." #10507
-
Hello, I am trying to only run Named Entity Recognition and my dataset already contains tokenized text which I need to use since I need this exact tokenization for the indices. I do not need to train the pipeline on my dataset, I only need to apply it for the results. I was trying to apply only the NER-pipeline on a custom
But then I recieved the following error:
After researching this "Parameter 'E' for model 'hashembed' has not been allocated yet.", I realized that I have to run
But I recieve this error as well already when trying to apply the
What am I missing here? I can find no other infos about this error. Thank you in advance!
|
Beta Was this translation helpful? Give feedback.
Replies: 2 comments 4 replies
-
This is kind of the same issue as #10508, just with slightly different (admittedly confusing) error messages. I'm not sure exactly what you're trying to do, but you probably want to keep the The |
Beta Was this translation helpful? Give feedback.
-
Hello This error is popping up when I am trying to train a 'ner' component, while keeping the tok2vec component frozen. Can I use spacy-alignment? If yes, how to go about it? Here is my base_config.cfg:
Here is the error:
|
Beta Was this translation helpful? Give feedback.
This is kind of the same issue as #10508, just with slightly different (admittedly confusing) error messages.
I'm not sure exactly what you're trying to do, but you probably want to keep the
ner
component fromen_ner_bc5cdr_md
rather than excluding it and trying to replace it with new uninitialized/untrained one.The
Doc
construction is fine, but it's unlikely thaten_ner_bc5cdr_md
has been trained on BPE/wordpiece-y tokens like##na
so you might not see good results. Is there a particular reason that you're using this tokenization?