Skip to content

Latest commit

 

History

History

tests

Folders and files

NameName
Last commit message
Last commit date

parent directory

..
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Tests

Currently, following tests are supported:

  1. Converter level tests
  2. Module level tests
  3. Accuracy tests

The goal of Converter tests are to tests individual converters againsts specific subgraphs. The current tests in core/conveters are good examples on how to write these tests. In general every converter should have at least 1 test. More may be required if the operation has switches that change the behavior of the op.

Module tests are designed to test the compiler against common network architectures and verify the integration of converters together into a single engine.

In addition to the above, we have lowering tests (//core/lowering) which test the functionality of lowering passes and partitioning tests (//core/partitioning ) which test different cases of torch fallback on test networks.

You can run the whole test suite with bazel. But be aware you may exhaust GPU memory (this may be seen as a cuDNN initialization error) running them naively, you therefore may need to limit the number of concurrent tests. Also because the inputs to tests are random it may make sense to run tests a few times.

Here are some settings that we usually test with:

bazel test //tests --compilation_mode=dbg --test_output=errors --jobs=4 --runs_per_test=5

--runs_per_test is optional and can be performed to check if numerical issues in outputs persist across multiple runs.

--jobs=4 is useful and is sometimes required to prevent too many processes to use GPU memory and cause CUDA out of memory issues.

Additionally, accuracy tests are supported for Python backend using Nox. Please refer setup_nox.sh for reference.

# To run complete Python accuracy + API tests
nox

# To list supported sessions
nox -l

Note: Supported Python tests

* download_datasets-3
* download_models-3
* train_model-3
* finetune_model-3
* ptq_test-3
* qat_test-3
* cleanup

Testing using pre-built Torch-TensorRT library

Currently, the default strategy when we run all the tests (bazel test //tests) is to build the testing scripts along with the full Torch-TensorRT library (libtorchtrt.so) from scratch. This can lead to increased testing time and might not be needed incase you already have a pre-built Torch-TensorRT library that you want to link against.

In order to not build the entire Torch-TensorRT library and only build the test scripts, please use the following command.

bazel test //tests  --compilation_mode=dbg --test_output=summary --define torchtrt_src=prebuilt --jobs 2

The flag --define torchtrt_src=prebuilt signals bazel to use pre-compiled library as an external dependency for tests. The pre-compiled library path is defined as a local_repository rule in root WORKSPACE file (https://github.com/pytorch/TensorRT/blob/master/WORKSPACE).

# External dependency for torch_tensorrt if you already have precompiled binaries.
# This is currently used in pytorch NGC container CI testing.
local_repository(
    name = "torch_tensorrt",
    path = "/opt/pytorch/torch_tensorrt"
)