Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Optionally delegate classifiers to XGBoost for finetuning and inference #114

Open
JackHopkins opened this issue Dec 4, 2023 · 0 comments
Labels
enhancement New feature or request

Comments

@JackHopkins
Copy link
Contributor

Is your feature request related to a problem? Please describe.
LLMs are extremely inefficient at classification. XGBoost is better if the data is available. We could use the aligned data from the LLM to train an XGBoost model, which would be much faster to run.

Describe the solution you'd like
When the output types denote a classification task (i.e where the goal is to sample one type in a union of literal types, or an enum), we optionally distil the teacher model into a decision forest using the XGBoost library.

Additional context
We could represent student models as optional packages, sort of like drivers, that the user could install through PIP.

E.g pip3 install tanuki.py[xgboost]

@JackHopkins JackHopkins added the enhancement New feature or request label Dec 4, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request
Projects
None yet
Development

No branches or pull requests

1 participant