Free Trial

Alteryx Designer Desktop Discussions

Find answers, ask questions, and share expertise about Alteryx Designer Desktop and Intelligence Suite.
SOLVED

Error SQLExecute: [Simba][BigQuery] (115) Operation timeout. Timed out pol

Heriberto_Prieto
5 - Atom

Hello everyone,

 

Trying to pull data from Google big query using the Simba ODBC driver but every time a query takes more then five minutes it timeouts. Running the query directly on google console the timeout doesn't happen. Does anyone know if there is a way to increase that timeout limit?

 
 

8-8-2024 10-45-28 AM.png

7 REPLIES 7
Heriberto_Prieto
5 - Atom

Thanks,

going through the documentation found out that adding a timeout to the connection string solved the issue

JMB001
8 - Asteroid

@Heriberto_Prieto Could you please share details on how you did this with the ODBC - I don't see an option in Advanced Options in the DSN Setup. 

 

@STP - this timeout message is happening on Gallery runs.

apathetichell
19 - Altair

Are you connecting via JWT on Gallery/Server? page 100 of the guide I posted shows you the setting you can control for timeout.

Heriberto_Prieto
5 - Atom

Hi JMB001,

 

If you are using windows the timeout has to be added to the connection string, the driver won't have that setting, here is a screenshot on how to add the timeout to the connection

 
 
 

 

 
 

 

JMB001
8 - Asteroid

Thanks guys.

 

I wasn't clear enough: I get this issue when OUTPUTTING to BigQuery using a InDB Connection File to open the connection in the Connect In-DB tool in workflows that are entirely In-DB tooling.

 

I also get the error using the InDB Connection File and output using the DCM - Google Bulk Loader.

 

I can run them locally but not on our Gallery Server. But locally, I have no special line/code/setting for timeout.

 

@apathetichell - I've looped in our Server admin as I dont know how to answer that.

@Heriberto_Prieto - thanks for the screen shot. Where is that connection string location?

apathetichell
19 - Altair

If I was your architect - I would recommend doing this all via scripted GCP CLI where your Server has preauthenticated to a specific GCP project/account - or connects in a wofklow. I would recommend moving your data to a storage bucket and having BQ query the storage bucket  vs uploading directly to BQ. 

Labels
Top Solution Authors