We read every piece of feedback, and take your input very seriously.
To see all available qualifiers, see our documentation.
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
small observation on pos-tags:
count("GERMAPARL", query = '"muss"', cqp = T, breakdown = T, p_attribute = "pos")
results in ~30k pos = "NN", which is about 20% of all hits.
As I'm not familiar with the pos-tagger, i've no idea, whether it is possible or feasible to optimize the results...
The text was updated successfully, but these errors were encountered:
This issue is old ... but still relevant. I just inspected the the NN-tagged "muss" matches using this snippet:
k <- kwic("GERMAPARL", query = '[word = "muss" & pos = "NN"]', cqp = T)
I would hope that our pos tagger performed better than that. We have started to use StanfordNLP - and need to inspect these POS tags.
Sorry, something went wrong.
No branches or pull requests
small observation on pos-tags:
count("GERMAPARL", query = '"muss"', cqp = T, breakdown = T, p_attribute = "pos")
results in ~30k pos = "NN", which is about 20% of all hits.
As I'm not familiar with the pos-tagger, i've no idea, whether it is possible or feasible to optimize the results...
The text was updated successfully, but these errors were encountered: