Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Investigate performance of ExperimentData regarding adding more data to it #1394

Open
wshanks opened this issue Feb 6, 2024 · 0 comments
Open

Comments

@wshanks
Copy link
Contributor

wshanks commented Feb 6, 2024

In #1360, there was some discussion that we give up performance in order to allow experiments to be extended with additional data. We should look at how much ExperimentData could benefit from only allowing a fixed set of data and if it would be better to write a function that could run analysis on multiple ExperimentData objects. Or perhaps calling .run() a second time should generate a new ExperimentData object and new experiment ID that combines the old data and the new data? Perhaps there would need to be a new method .extend_run() instead of changing run.

          Another option would be to stop over-engineering. We currently allow users to run experiment on existing experiment data to get more statistically confident result. But I don't think this is heavy used feature and we can remove. In principle we run very expensive operation for the sake of something useless; if we have 100 xvals, we need to iterate over them and create new table with size=1 to average... Even if we switch to polars there still be overhead of object creation.

Originally posted by @nkanazawa1989 in #1360 (comment)

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant