You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
I am trying to filter table data from Databricks connection using dplyr. I am encountering an issue if I skip the 'select' statement - it seems that if you skip the selection of the columns, when translating the statement to SQL, instead of SELECT * it creates SELECT <table_name>.*, which leads to throwing Spark SQL error:
Error in py_call_impl(callable, call_args$unnamed, call_args$named) :
pyspark.errors.exceptions.connect.AnalysisException: [UNRESOLVED_COLUMN.WITH_SUGGESTION] A column, variable, or function parameter with name `table`.`*` cannot be resolved. Did you mean one of the following? [...]. SQLSTATE: 42703; line 1 pos 7;
'Project ['table.*]
What works - when I specify column names to select
Hi, this should have been resolved in the latest version of sparkly, which was published 3/25 (#3430). Would you mind confirming that you have sparklyr 1.8.5? And if you don't, can you confirm that the issue goes away after you upgrade to 1.8.5? Thank you
Hello,
I am trying to filter table data from Databricks connection using dplyr. I am encountering an issue if I skip the 'select' statement - it seems that if you skip the selection of the columns, when translating the statement to SQL, instead of
SELECT *
it createsSELECT <table_name>.*
, which leads to throwing Spark SQL error:What works - when I specify column names to select
What doesn't work - when I want to select all and do not specify column names to select
This code leads to above shared error.
Thanks in advance!
S.
The text was updated successfully, but these errors were encountered: