Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

RuntimeError: result type ComplexFloat can't be cast to the desired output type Float #2600

Open
yunchu opened this issue Nov 3, 2023 · 3 comments
Assignees

Comments

@yunchu
Copy link
Contributor

yunchu commented Nov 3, 2023

Describe the bug

tested on branch releases/1.5.0

failures on regression test for semantic segmentation

  • train type: class incr
  • template
    • SegNext-B, SegNext-t, SegNext-s

image

Steps to Reproduce

  1. copy ci datasets to local
  2. run regression test with following command
CI_DATA_ROOT=/mnt/hdd1/data/ci_datasets/ tox -vvv -e tests-seg-py310-pt1 -- tests/regression/semantic_segmentation/test_segmentation.py::TestRegressionSegmentation::test_otx_train[Custom_Semantic_Segmentation_SegNext_B] tests/regression/semantic_segmentation/test_segmentation.py::TestRegressionSegmentation::test_otx_train_cls_incr[Custom_Semantic_Segmentation_SegNext_B]

Environment:

  • OS:
  • Framework version:
  • Python version:
  • OpenVINO version:
  • CUDA/cuDNN version:
  • GPU model and memory:
@yunchu yunchu modified the milestone: 1.5.0 Nov 7, 2023
@kprokofi
Copy link
Contributor

kprokofi commented Nov 7, 2023

@yunchu , I investigated this issue and found some weird behavior in a test environment. This issue happens because of in train_cls_incr we change datasets, but the number of iterations is much lower than expected, so, runner.current_iter < runner,max_iters, and there are some issues in the scheduler emerge.
Can it be some possible shift in two testing environments (otx_train and train_cls_incr) leading to broken information about proper dataset size/number of epochs?

@kprokofi
Copy link
Contributor

kprokofi commented Nov 7, 2023

BTW,
This problem is not reproducible outside of tests. I tried the class incremental approach with the same data and everything is fine. The number of iterations is what it should be without any errors.

@yunchu
Copy link
Contributor Author

yunchu commented Nov 16, 2023

@kprokofi I guess you can check the regression test on the repo. I've just checked that was happened on 'develop' branch as well.
AFAIK, the current regression test run only 10 epoch for each training. do you think it could make any issues?

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants