Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Interesting paper #1

Open
blazespinnaker opened this issue Jan 30, 2023 · 0 comments
Open

Interesting paper #1

blazespinnaker opened this issue Jan 30, 2023 · 0 comments

Comments

@blazespinnaker
Copy link

I've been playing around with the ideas proposed in this paper as they are similar to some results that I am seeing.

In particular, I'm using a formation like this:

preds = .. #Assume preds have been generated by some ensemble/model to evaluate / fix up

def test_is_better(preds, disagreement_index, new_prob):
  new_preds = preds.copy()
  new_preds [disagreement_index] = new_prob
  model_evaluator = lightgbm.LGBMRegressor()
  model_evaluator.fit(X_Unlabeled, new_preds)
  for i in range(5):
      model_evaluator_1 =  lightgbm.LGBMRegressor()
      model_evaluator_1.fit(X_Unlabeled, model_evaluator_1.predict(X_Unlabeled))
      model_evaluator =  lightgbm.LGBMRegressor()
      model_evaluator.fit(X_Unlabeled, model_evaluator_1.predict(X_Unlabeled))
  return roc_auc_score(y_holdout, model_evaluator.predict(X_holdout)) > BASE_ROC_AUC_SCORE

I'm finding it works for not just out of distribution data, but also somewhat for in distribution data as well.

What's interesting is that iterating on the model_evaluator is what is important and improves the accuracy of this metric.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant