[BUG] Using subscriptionKeyCol
argument results in SparkException [FAILED_EXECUTE_UDF]
#2064
Open
3 of 19 tasks
SynapseML version
0.11.2
System information
Describe the problem
We are using the
AnalyseHealthText
Cognitive Services transformer, and have been up to now passing a single subscription key via thesubscriptionKey
construction argument with no problems.After some throttling issues, we've needed to fan out across multiple accounts so have implemented the following code (shortened for brevity):
Checking the DataFrame before sending through the transformer, I can see that a new column has been added successfully with the randomised subscription keys; however as soon as it's passed through transform we get the following exception:
As far as I can tell when preparing the query to cog services and extracting the cog services key in
HasCognitiveServicesInput.addHeaders
, its finding a WrappedArray and not a string and thus hitting the above error.Code to reproduce issue
Derived from https://microsoft.github.io/SynapseML/docs/Explore%20Algorithms/AI%20Services/Advanced%20Usage%20-%20Async,%20Batching,%20and%20Multi-Key/#step-5-multi-key
Other info / logs
What component(s) does this bug affect?
area/cognitive
: Cognitive projectarea/core
: Core projectarea/deep-learning
: DeepLearning projectarea/lightgbm
: Lightgbm projectarea/opencv
: Opencv projectarea/vw
: VW projectarea/website
: Websitearea/build
: Project build systemarea/notebooks
: Samples under notebooks folderarea/docker
: Docker usagearea/models
: models related issueWhat language(s) does this bug affect?
language/scala
: Scala source codelanguage/python
: Pyspark APIslanguage/r
: R APIslanguage/csharp
: .NET APIslanguage/new
: Proposals for new client languagesWhat integration(s) does this bug affect?
integrations/synapse
: Azure Synapse integrationsintegrations/azureml
: Azure ML integrationsintegrations/databricks
: Databricks integrationsThe text was updated successfully, but these errors were encountered: