Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Help students run MLPerf inference at the Student Cluster Competition'24 #26

Open
12 tasks
gfursin opened this issue Apr 29, 2024 · 0 comments
Open
12 tasks
Assignees

Comments

@gfursin
Copy link
Contributor

gfursin commented Apr 29, 2024

We were asked to help students run MLPerf inference benchmark at the Student Cluster Competition'24 and automate their submission and grading via the MLCommons CM automation framework.

Current plan is to use MLPerf inference Stable Diffusion benchmark with Stability AI’s Stable Diffusion XL model (2.6 billion parameters) and COCO data set. This popular model is used to create compelling images through a text-based prompt.

We must check the following:

  • Check current CM workflows to run reference MLPerf SD benchmark
  • Check CM workflows to run optimized MLPerf SD benchmark v4.0
    • Intel
    • Nvidia
  • Check if support for AMD GPUs can be provided
  • Check how to support multi-node inference
  • Prepare tutorial about MLPerf, loadgen, this benchmark and CM
  • Check MLCommons Croissant format for the dataset?
  • Automate submission and grading
    • Need to agree how to report accuracy
    • We may train a smaller model to analyze produced images
    • Create live scoreboard (W&B?)
@gfursin gfursin self-assigned this Apr 29, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant