Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Almond on windows is not use spark config in spark-default and hadoop config in hadoop config dir #1092

Open
AnhQuanTran opened this issue Apr 20, 2023 · 0 comments

Comments

@AnhQuanTran
Copy link

AnhQuanTran commented Apr 20, 2023

Hi,

I setup almond kernel on windows. It's work well with scala and spark local. But i want to use remote hdfs and remote yarn to submit job. So i need to override some configs like this

hadoop config 

fs.defaultFS
hadoop.security.authentication 
hadoop.http.authentication.type

spark config 
spark.yarn.stagingDir
spark.history.fs.logDirectory

All of these configs existed in spark-default file and hadoop config dir. I already set system env var and also set env var in kernel json like this. But almond kernel still not use these configs

"HADOOP_HOME": "D:\\DATA\\Environment\\hadoop-2.7.2",
"HADOOP_CONF_DIR": "D:\\DATA\\Environment\\hadoop-2.7.2\\etc\\hadoop",
"SPARK_HOME": "D:\\DATA\\Environment\\spark-3.2.2-bin-hadoop2.7",
"SPARK_CONF_DIR": "D:\\DATA\\Environment\\spark-3.2.2-bin-hadoop2.7\\conf",

When i open spark ui, view environment. It still use default config of almond instead my override config
How to override config above and use it from spark-default and hadoop config dir. Thank you

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

1 participant