Skip to content

Commit 0b24af9

Browse files
Release (#209)
1 parent 14fc8ac commit 0b24af9

File tree

6 files changed

+25
-22
lines changed

6 files changed

+25
-22
lines changed

.github/workflows/test_cli_misc.yaml

Lines changed: 1 addition & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -8,7 +8,6 @@ on:
88
paths:
99
- .github/workflows/test_cli_misc.yaml
1010
- "optimum_benchmark/**"
11-
- "docker/**"
1211
- "tests/**"
1312
- "setup.py"
1413
pull_request:
@@ -17,7 +16,6 @@ on:
1716
paths:
1817
- .github/workflows/test_cli_misc.yaml
1918
- "optimum_benchmark/**"
20-
- "docker/**"
2119
- "tests/**"
2220
- "setup.py"
2321

@@ -31,7 +29,7 @@ jobs:
3129
fail-fast: false
3230
matrix:
3331
os: [ubuntu-latest]
34-
python: ["3.8", "3.10"]
32+
python: ["3.8", "3.9", "3.10"]
3533

3634
runs-on: ${{ matrix.os }}
3735

.github/workflows/update_llm_perf_cuda_pytorch.yaml

Lines changed: 1 addition & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -29,7 +29,6 @@ jobs:
2929
- name: Run benchmarks
3030
uses: addnab/docker-run-action@v3
3131
env:
32-
IMAGE: ${{ env.IMAGE }}
3332
SUBSET: ${{ matrix.subset }}
3433
MACHINE: ${{ matrix.machine.name }}
3534
HF_TOKEN: ${{ secrets.HF_TOKEN }}
@@ -49,5 +48,5 @@ jobs:
4948
run: |
5049
pip install packaging && pip install flash-attn einops scipy auto-gptq optimum bitsandbytes autoawq codecarbon
5150
pip install -U transformers huggingface_hub[hf_transfer]
52-
pip install -e .
51+
pip install optimum-benchmark
5352
python llm_perf/update_llm_perf_cuda_pytorch.py

README.md

Lines changed: 10 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,15 +1,22 @@
1-
<p align="center"><img src="logo.png" alt="Optimum-Benchmark Logo" width="350" style="max-width: 100%;" /></p>
1+
<p align="center"><img src="https://raw.githubusercontent.com/huggingface/optimum-benchmark/main/logo.png" alt="Optimum-Benchmark Logo" width="350" style="max-width: 100%;" /></p>
22
<p align="center"><q>All benchmarks are wrong, some will cost you less than others.</q></p>
33
<h1 align="center">Optimum-Benchmark 🏋️</h1>
44

5+
[![PyPI - Python Version](https://img.shields.io/pypi/pyversions/optimum-benchmark)](https://pypi.org/project/optimum-benchmark/)
6+
[![PyPI - Version](https://img.shields.io/pypi/v/optimum-benchmark)](https://pypi.org/project/optimum-benchmark/)
7+
[![PyPI - Downloads](https://img.shields.io/pypi/dm/optimum-benchmark)](https://pypi.org/project/optimum-benchmark/)
8+
[![PyPI - Implementation](https://img.shields.io/pypi/implementation/optimum-benchmark)](https://pypi.org/project/optimum-benchmark/)
9+
[![PyPI - Format](https://img.shields.io/pypi/format/optimum-benchmark)](https://pypi.org/project/optimum-benchmark/)
10+
[![PyPI - License](https://img.shields.io/pypi/l/optimum-benchmark)](https://pypi.org/project/optimum-benchmark/)
11+
512
Optimum-Benchmark is a unified [multi-backend & multi-device](#backends--devices-) utility for benchmarking [Transformers](https://github.com/huggingface/transformers), [Diffusers](https://github.com/huggingface/diffusers), [PEFT](https://github.com/huggingface/peft), [TIMM](https://github.com/huggingface/pytorch-image-models) and [Optimum](https://github.com/huggingface/optimum) libraries, along with all their supported [optimizations & quantization schemes](#backends--devices-), for [inference & training](#scenarios-), in [distributed & non-distributed settings](#launchers-), in the most correct, efficient and scalable way possible.
613

714
*News* 📰
815

9-
- PyPI package is now available for installation: `pip install optimum-benchmark` 🎉 check it out !
16+
- PyPI package is now available for installation: `pip install optimum-benchmark` 🎉 [check it out](https://pypi.org/project/optimum-benchmark/) !
1017
- Hosted 4 minimal docker images (`cpu`, `cuda`, `rocm`, `cuda-ort`) in [packages](https://github.com/huggingface/optimum-benchmark/pkgs/container/optimum-benchmark) for testing, benchmarking and reproducibility 🐳
1118
- Added vLLM backend for benchmarking [vLLM](https://github.com/vllm-project/vllm)'s inference engine 🚀
12-
- Hosted the codebase of the LLM-Perf Leaderboard [LLM-Perf](https://huggingface.co/spaces/optimum/llm-perf-leaderboard) 🥇
19+
- Hosted the codebase of the [LLM-Perf Leaderboard](https://huggingface.co/spaces/optimum/llm-perf-leaderboard) 🥇
1320
- Added Py-TXI backend for benchmarking [Py-TXI](https://github.com/IlyasMoutawwakil/py-txi/tree/main) 🚀
1421
- Introduced a Python API for running isolated benchmarks from the comfort of your Python scripts 🐍
1522
- Simplified the CLI interface for running benchmarks using the Hydra CLI 🧪

llm_perf/utils.py

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -4,13 +4,11 @@
44

55
from optimum_benchmark.report import BenchmarkReport
66

7-
OPEN_LLM_LEADERBOARD = pd.read_csv("hf://datasets/optimum-benchmark/open-llm-leaderboard/open-llm-leaderboard.csv")
8-
9-
107
INPUT_SHAPES = {"batch_size": 1, "sequence_length": 256}
118
GENERATE_KWARGS = {"max_new_tokens": 64, "min_new_tokens": 64}
129

1310

11+
OPEN_LLM_LEADERBOARD = pd.read_csv("hf://datasets/optimum-benchmark/llm-perf-leaderboard/llm-df.csv")
1412
OPEN_LLM_LIST = OPEN_LLM_LEADERBOARD.drop_duplicates(subset=["Model"])["Model"].tolist()
1513
PRETRAINED_OPEN_LLM_LIST = (
1614
OPEN_LLM_LEADERBOARD[OPEN_LLM_LEADERBOARD["Type"] == "pretrained"]
@@ -44,7 +42,9 @@
4442
# "Qwen",
4543
# ],
4644
# ]
47-
# CANONICAL_PRETRAINED_OPEN_LLM_LIST = [model for model in PRETRAINED_OPEN_LLM_LIST if model.split("/")[0] in CANONICAL_ORGANIZATIONS]
45+
# CANONICAL_PRETRAINED_OPEN_LLM_LIST = [
46+
# model for model in PRETRAINED_OPEN_LLM_LIST if model.split("/")[0] in CANONICAL_ORGANIZATIONS
47+
# ]
4848
CANONICAL_PRETRAINED_OPEN_LLM_LIST = [
4949
"01-ai/Yi-34B",
5050
"01-ai/Yi-6B",

optimum_benchmark/version.py

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,4 +12,4 @@
1212
# See the License for the specific language governing permissions and
1313
# limitations under the License.
1414

15-
__version__ = "0.2.0"
15+
__version__ = "0.2.1"

setup.py

Lines changed: 8 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -98,26 +98,25 @@
9898
extras_require=EXTRAS_REQUIRE,
9999
entry_points={"console_scripts": ["optimum-benchmark=optimum_benchmark.cli:main"]},
100100
description="Optimum-Benchmark is a unified multi-backend utility for benchmarking "
101-
"Transformers, Timm, Diffusers and Sentence-Transformers with full support of Optimum's "
102-
"hardware optimizations & quantization schemes.",
103-
long_description=open("README.md", "r", encoding="utf-8").read(),
104-
long_description_content_type="text/markdown",
101+
"Transformers, Timm, Diffusers and Sentence-Transformers with full support of "
102+
"Optimum's hardware optimizations & quantization schemes.",
103+
url="https://github.com/huggingface/optimum-benchmark",
105104
classifiers=[
106-
"License :: OSI Approved :: Apache Software License",
107-
"Intended Audience :: Developers",
108105
"Intended Audience :: Education",
106+
"Intended Audience :: Developers",
107+
"Operating System :: POSIX :: Linux",
109108
"Intended Audience :: Science/Research",
110-
"Operating System :: OS Independent",
111-
"Programming Language :: Python :: 3.7",
112109
"Programming Language :: Python :: 3.8",
113110
"Programming Language :: Python :: 3.9",
114111
"Programming Language :: Python :: 3.10",
112+
"License :: OSI Approved :: Apache Software License",
115113
"Topic :: Scientific/Engineering :: Artificial Intelligence",
116114
],
117115
keywords="benchmaek, transformers, quantization, pruning, optimization, training, inference, onnx, onnx runtime, intel, "
118116
"habana, graphcore, neural compressor, ipex, ipu, hpu, llm-swarm, py-txi, vllm, auto-gptq, autoawq, "
119117
"sentence-transformers, bitsandbytes, codecarbon, flash-attn, deepspeed, diffusers, timm, peft",
120-
url="https://github.com/huggingface/optimum-benchmark",
118+
long_description=open("README.md", "r", encoding="utf-8").read(),
119+
long_description_content_type="text/markdown",
121120
author="HuggingFace Inc. Special Ops Team",
122121
include_package_data=True,
123122
name="optimum-benchmark",

0 commit comments

Comments
 (0)