Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Bloating memory with write_tables in beam mode #5907

Open
jsu6 opened this issue Oct 20, 2023 · 0 comments
Open

Bloating memory with write_tables in beam mode #5907

jsu6 opened this issue Oct 20, 2023 · 0 comments

Comments

@jsu6
Copy link

jsu6 commented Oct 20, 2023

Hello,

I have a 36GB entries file and I tried to run ~/kythe-v0.0.63/tools/write_tables --entries entries --out serving/ -experimental_beam_pipeline -experimental_beam_columnar_data -num_workers 20 to get its serving table in beam mode. However, ~200GB RAM seems not enough. I did an experiment showing ~200GB RAM barely survived from ~13GB entries file. Is this expected? ~200GB RAM works with write_tables without beam but some links in serving table are missing in this case. Is there any way to limit the RAM usage for write_tables in beam mode? I saw statsK/beamK is set to 128 so the max RAM usage should also be 128GB? Am I missing or misunderstanding something? Any feedback is appreciated.

@schroederc schroederc removed their assignment May 1, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants