This site uses different types of cookies, including analytics and functional cookies (its own and from other sites). To change your cookie settings or find out more, click here. If you continue browsing our website, you accept these cookies.
The current version of the Publish to Tableau macro retrieves an API key at the start of the workflow run. Often times the workflow may take several hours to run before it's ready to write to Tableau by which time the API may have expired. (I think the default tableau server setting times out in 2 hrs) It's one of those soul crushing "I should've forked the output!"
Sample Log Error -
Tool #46: TableauServer.UploadChunks (238): Iteration #1: Tool #19: Tool #4: Tableau Server API Request (Upload file) Error Code 401002: Unauthorized Access -- Invalid authentication credentials were provided.
Tool #46: Tool #252: Tool #4: Tableau Server API Request (Publish file) Error Code 401002: Unauthorized Access -- Invalid authentication credentials were provided.
The idea would be to change when the macro obtains the API from when the workflow is initiated to just before the workflow is ready to write to the Tableau avoiding these timeouts.
(If you're having this issue in the meantime you can have your Tableau server admin up the timeout)
When building custom tools for Alteryx using the Python SDK, there is no current way to test these outside of the Alteryx Designer.
This means that your development process is:
- write some code (no code-sense; intellisense; auto-complete because Jupyter; VSCode; Visual Studio; etc cannot access AlteryxEngine or any of the other imports)
- copy that .py module into your C:\Users\<username>\AppData\roaming\Alteryx\Tools\<toolname>
- fire up Alteryx
- drop this new custom tool on a canvas
- run it to see if you get any errors
- then copy these errors out of Alteryx result window into Notepad to be able to read them
- then go back into your development environment to make changes
This is very painful, and this will directly scare most people away from learning how to create custom tools since it's not only inefficient - but also scary and frustrating for beginners.
Could we instead create mock python libraries; and a development harness (like Google does with Android development in Eclipse) in this SDK where:
- you have full code intelligence (intellisense, autocomplete)
- you can simulate engine events in a test harness (for example in the Android SDK; you can simulate the user rotating their phone, turning off GPS, hitting a volume button, etc).
- you can also write test cases which can run automatically
- then once you know that your tool will work - only then you drop it into the Alteryx Designer environment.
NOTE: This IDE way of thinking also allows you to bring the configuration pieces (like number of inputs; etc) out of raw code and into configuration options.
Although you may be able to do remote debugging by using platforms like PyCharm - that really does not give you the full ability to check in the code of your tool; along with all the test cases; in a harness that allows you to automatically check different events; or to make sure that your tool works in the test harness before deploying.
While Alteryx allows for a proxy username and password in the settings, these are not passed properly to an NTLM proxy. Support for NTLM authentication would be incredibly useful for a number of corporations who utilize this firewall setup.
We currently have to either download via Python or cURL through batch commands called by Alteryx. Since Alteryx uses a cURL back-end, this should be a fairly simple addition to the existing download tool by allowing a selection of proxy server, port, and authentication method in addition to the proxy username and password. This could be done either in the tool itself or in User Settings.
Adobe analytics api current version is 2.0 but the connector still uses 1.4. This works for log in access but does not for automated access since the end points are changed and 2.0 uses a new key structure
API Security requirements are constantly evolving and strengthening. As API architectures migrate from traditional authentication models (Basic, OAuth, etc.) to more secure, certificate-based models, like MTLS/MSSL, leveraging Alteryx Designer will become increasingly difficult, especially for larger organizations trying to scale the use of Alteryx across a large user base, with vastly diverse skillsets.
I realize issuing API calls with certificates is possible via the Run Command tool. We consider this a temporary workaround, and not a permanent, strategic solution. The Run Command tool can be clunky to use when passing in variables and passing the output back into the workflow for downstream processing.
Therefore, I would like to request a more scalable approach to issuing MTLS/MSSL API calls. Can an option be added to the Download Tool to allow for certificates to be passed on API calls?
Most organizations are moving to the cloud and Microsoft is the biggest one in the game.
I pleased that you have updated the List tool to connect to SP 365 but soon all of the shared drives will be online and if its vital for Alteryx success to be able to communicate, pull and export data to those locations.
With the Desktop, there is a way to make it work by authenticating and for the previous SP version by logging on to SharePoint first on the server then the scheduler would run successfully. Now the scheduler isn't able to take advantage of the authentication already established.
If Alteryx isn't able to keep up it will soon find itself behind in innovation and tool of choice.
My company has recently purchased some Alteryx licences with the hope of advancing their Data Science capability. The business is currently moving all their POS data from in-premise to cloud environment and have identified Azure Cosmos DB as a perfect enviornment to house the streaming data. Having purchased the Alteryx licences, we have now a challenge of not being able to connect to the Azure Cosmos DB environment and we would like Alteryx to consider speeding up the development of this process.
In the current version of alteryx, we are only able to pull the first 2000 records from an existing salesforce report, a lot of times it becomes difficult to automate my monthly/weekly requests because it is depended on an existing salesforce report and because of that I need to recreate the report on alteryx using the salesforce objects and fields.
If we had the ability to bring in the entire salesforce report to alteryx environment it would save a lot of time and effort for analysts like me because in that way, we won't have to re-invent the wheel.
Using the download tool is great and easy to use. However, if there was a connection problem with a request the workflow errors out. Having an option to not error out, ability to skip failed records, and retrying records that failed would be A LIFE CHANGER. Currently I have been using a Python tool to create multi-threaded requests and is proven to be time consuming.
My idea is to have the AlteryxEngineCMD.exe to run a workflow as part of the standard Alteryx license.
Use case - be able to run Alteryx from the command line without the need to buy the entire Scheduler package (at $6,500/seat).
I understand why certain features are add on, but the ability to run AlteryxEngineCMD.exe (I feel) should be part of the standard license which is already $5K+. For those who only need to be able to run a command line execution of a workflow $6.5K is a lot of money!
Recently we had a situation where installing the data packages was expected to take over 20 hours! We do not have the ability to run a machine undisturbed for this length of time at the office, and VPN automatically times out after 12 hours. Okay, so these items are company-specific, but how nice would it be to be able to download/install the data in smaller portions so you don't have to worry about setting up shop for 8-10 hours?!
I was able to copy the DataInstall.ini and name only the portions of the data I wanted to install in each session. But then I had to separate the installs into different network folders otherwise you end up overwriting the DataInstall.exe file, which users need for each install in order to register from a network location (for each install).
Needless to say, it took SEVERAL weekends and quite a few mistakes before I was able to do it this way successfully!
Please vote for data installs in smaller portions!
For those with large web and streaming-media server logs, the ability to geocode IP addresses be an excellent feature, similar to Alteryx's ability go geocode street addresses. Several IP geocoding services exist, with different levels of accuracy and cost. Ideally, the user should be able to choose their own service if they have one, in addition to a default service built-in to Alteryx.
Right now as far as I know you need to add each DB connection manually. This works... but is quite time consuming when trying to run tasks against a cluster of prod databases. It would be awesome to pass a JSON config file,example below, to the Alteryx Engine and have Alteryx create those connections upon parsing the file. This would save tons of time, and allow teams to share a central config file with consistent aliases across their clusters to ensure their app connections point to the same DBs across workflows. It would also make on boarding a breeze for new developers on team.
I work with a large group of individuals, close to 30,000, and a lot of our files are ran as .dif/.kat files used to import to certain applications and softwares that pertain to our work. We were wondering if this has been brought up before and what the possibility might be.