Issues: princeton-nlp/SWE-bench
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Author
Label
Projects
Milestones
Assignee
Sort
Issues list
model_name_or_path
is None when running models without adapters, causing an error in run_evaluation.py
bug
#126
opened May 25, 2024 by
rucnyz
what's the difference between environment_setup_commit and base_commit?
documentation
Improvements or additions to documentation
#125
opened May 21, 2024 by
ramsey-coding
how to download one task instance from SWE-bench dataset?
documentation
Improvements or additions to documentation
#124
opened May 20, 2024 by
ramsey-coding
What's the best way to browse the SWE-bench dataset?
documentation
Improvements or additions to documentation
#123
opened May 20, 2024 by
ramsey-coding
Using New feature or request
uv pip
instead of pip
for significant speedup
enhancement
#120
opened May 15, 2024 by
klieret
Has anyone successfully ran an eval on patches against early versions of astropy, sympy, scipy etc? I'm really struggling to run things from earlier python versions
documentation
Improvements or additions to documentation
#119
opened May 14, 2024 by
PandelisZ
swe-bench eval stops running after a point
bug
Something isn't working
#118
opened May 13, 2024 by
ssh-randy
run_live.py: clone_repo() takes 3 positional arguments but 5 were given
bug
Something isn't working
#117
opened May 11, 2024 by
stevensu1977
Is it possible to evaluate the train set?
documentation
Improvements or additions to documentation
#116
opened May 10, 2024 by
chriscremer
Why AutoCodeRover not mentioned?
documentation
Improvements or additions to documentation
#115
opened May 9, 2024 by
kripper
environment is lost when running pip install
bug
Something isn't working
#111
opened Apr 30, 2024 by
waterson
improve eval performance by caching per-repo/version conda environments
#104
opened Apr 26, 2024 by
waterson
Dataset field & set up reliable environment
documentation
Improvements or additions to documentation
#99
opened Apr 21, 2024 by
Hodge931
Repository not found while running python3 create_text_dataset.py
inference
This issue is related to running inference
#98
opened Apr 18, 2024 by
hsm1997
/bin/sh: pytest: command not found
when running evaluations
bug
#96
opened Apr 17, 2024 by
psykhi
Remove This issue is related to running evaluation
pre_install
from install specs
evaluation
#95
opened Apr 11, 2024 by
carlosejimenez
Adding LLM tokens to the generated inference for cost calculation
inference
This issue is related to running inference
#91
opened Apr 10, 2024 by
moresearch
Running create_text_dataset.py gets Killed and takes too long
inference
This issue is related to running inference
#88
opened Apr 7, 2024 by
vishwa27yvs
inference issue run live
inference
This issue is related to running inference
#87
opened Apr 7, 2024 by
kneeraj-AIDE
Issue with Django gold patch results not evaluating to "Resolved" when logs say "All tests passed"
evaluation
This issue is related to running evaluation
#80
opened Apr 5, 2024 by
brombaut
Error when running evaluation - ModuleNotFoundError: No module named 'conda'
evaluation
This issue is related to running evaluation
#77
opened Apr 4, 2024 by
brombaut
Unable to replicate basic results
evaluation
This issue is related to running evaluation
#74
opened Apr 3, 2024 by
bytesuji
Previous Next
ProTip!
no:milestone will show everything without a milestone.