Hello, have any of you encountered this error?
Error: Data Stream Out (11): Error SQLExecute: [Simba][Hardy] (35) Error from server: error code: '0' error message: 'org.apache.hive.service.cli.HiveSQLException: Error running query: org.apache.spark.SparkException: Job aborted due to stage failure: Total size of serialized results of 8 tasks (4.1 GiB) is bigger than spark.driver.maxResultSize 4.0 GiB.
at org.apache.spark.sql.hive.thriftserver.HiveThriftServerErrors$.runningQueryError(HiveThriftServerErrors.scala:48)
at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.$anonfun$execute$1(SparkExecuteStatementOperation.scala:611)
at scala.runtime.java8.JFunction0$mcV$sp.apply(JFunction0$mcV$sp.java:23)
at com.databricks.unity.EmptyHandle$.runWith(UCSHandle.scala:124)
at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation.org$apache$spark$sql$hive$thriftserver$SparkExecuteStatementOperation$$execute(SparkExecuteStatementOperation.scala:501)
at org.apache.spark.sql.hive.thriftserver.SparkExecuteStatementOperation$$anon$2$$anon$3.$anonfun$run$2(SparkExecuteState
I am trying to run a delta table in databricks using the Simba ODBC connector in Alteryx but it seems the size of the results are too high for the spark driver. I wasnt sure if the spark driver needs to be set at a higher size capacity to allow the total number of tasks.
https://kb.databricks.com/jobs/job-fails-maxresultsize-exception.html
talk to your DB admin - this is a spark cluster setting vs an Odbc/Alteryx setting.
Can you confirm if this worked?