New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
refactor: clean up checks #1155
Comments
Good ideas, and I share many of your concerns. However, there are several distinct issues that you've bundled here in #1155 and they should be broken down into smaller issues that can be discussed and completed independently. Here are some possible standalone issues:
|
Also, note that proselint/proselint/.proselintrc Line 13 in 372ebf0
As a more general point, we've been wary of any checks that attempt to categorically ban words. The only time that's seemed like a good idea so far is for needless variants, where the determination has already been made for us that the word has no need. |
Strong suggestion, actually. Ultimately I just wanted to put this down as an RFC to get people's thoughts prior to doing any real work.
I believe so, as that's what we saw with the split between
That would definitely be the right thing to do going forward I think. I foresee it making maintenance quite a lot easier, and will overall help people to understand the general scope of these checks better.
That likely falls under the same problem we discussed relating to flag-based parsing honestly.
I would suggest improving them rather than deleting altogether. Principally speaking, many of these things are genuinely words that should be avoided in most contexts, and if we can tighten the error margin much more and make them more definitive, they may very well be suitable for our usage.
I would be more than happy to do this. Ultimately it would be good to concretely define and lay out our process for making these decisions and the criteria required for linguistic constructs. For part of this, we could use something similar to my language suitability evaluation framework
That would be quite an easy fix, too. Perhaps one best placed in the same restructure as a categorization evaluation.
I wasn't aware of that, actually - thanks for the tip. I'll be sure to consider
That's for the best, things like that can get authoritative or out of hand quite quickly. It's nice to see these things taken as seriously as they should be. It'll be easier for us to make those decisions once a framework is in place. |
@Nytelife26 Thanks for the response, we're on the same page on every point :) |
This relates to #1362. |
It has come to my attention that a lot of checks within proselint are dubious at best, or misguided. For instance:
cursing.nfl
check - some of them are just numbers, and others have many variations included, almost like a poorly-designed censoring system, in contrast with using regex.Et cetera. I feel it may be necessary to do a refactor of these checks and categorizations with a formal review to make maintainability easier in future and also to maintain a better linguistic ecosystem.
The text was updated successfully, but these errors were encountered: