New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Reduce the used memory for big map #1217
Comments
Also waiting for this critical feature. Cannot load full vocabulary due to huge memory needs to allocate and dont manage to get same accuracy while activating WM/LTM. |
@alexk1976 You can try to use |
Here some options:
|
Thank you! |
i dont think it's a real solution..a bit bigger area like we have and its impossible to load dictionary even when we set MaxFeatures=500. If we dont use FlanTree - have performance issues. Fixed dictionary - gives worse accuracy. We need a way to load full graph and only part of the dictionary |
@alexk1976 Agreed, it is kinda included in that other issue #1201 . |
Hello @matlabbe,
We created a pretty good map of a large area (multi-session mapping: 5). It ended up with >4 million words in the vocabulary, which led to rtabmap crashing every time It ran in localization mode for minutes or when I called the backup service explicitly. I am aware that you already opened issue #1201, however, I am wondering if any intermediate step to reduce the size of words?
Some parameters that may related for you to check:
Let me know if I can provide more information to help you with support.
Thank you for your contribution and hope that will receive your reply soon.
The text was updated successfully, but these errors were encountered: