Alteryx Designer

Share your Designer product ideas - we're listening!

1 Review

Our submission guidelines & status definitions before getting started

2 Search

The community for a solution or existing idea before posting

3 Vote

By clicking the like in the top left corner of an idea you support

4 Submit

A new idea to suggest a product enhancement or new feature


Suggest an idea

Please upgrade the "curl.exe" that are packaged with Designer from 7.15 to 7.55 or greater to allow for -k flags. Also please allow the -k functionality for the Atleryx Download tool.  

 

-k, --insecure

(TLS) By default, every SSL connection curl makes is verified to be secure. This option allows curl to proceed and operate even for server connections otherwise considered insecure.

The server connection is verified by making sure the server's certificate contains the right name and verifies successfully using the cert store.

 

Regards,

John Colgan

The current version of the Publish to Tableau macro retrieves an API key at the start of the workflow run.  Often times the workflow may take several hours to run before it's ready to write to Tableau by which time the API may have expired.  (I think the default tableau server setting times out in 2 hrs)  It's one of those soul crushing "I should've forked the output!"

Sample Log Error - 

  •         Tool #46: TableauServer.UploadChunks (238): Iteration #1: Tool #19: Tool #4: Tableau Server API Request (Upload file) Error Code 401002: Unauthorized Access -- Invalid authentication credentials were provided. 
  •          Tool #46: Tool #252: Tool #4: Tableau Server API Request (Publish file) Error Code 401002: Unauthorized Access -- Invalid authentication credentials were provided.

The idea would be to change when the macro obtains the API from when the workflow is initiated to just before the workflow is ready to write to the Tableau avoiding these timeouts.

 

(If you're having this issue in the meantime you can have your Tableau server admin up the timeout)

7-9-2020 2-53-42 PM.png

When building custom tools for Alteryx using the Python SDK, there is no current way to test these outside of the Alteryx Designer.

This means that your development process is:

- write some code (no code-sense; intellisense; auto-complete because Jupyter; VSCode; Visual Studio; etc cannot access AlteryxEngine or any of the other imports)

- hope

- copy that .py module into your C:\Users\<username>\AppData\roaming\Alteryx\Tools\<toolname>

- fire up Alteryx

- drop this new custom tool on a canvas

- run it to see if you get any errors

- then copy these errors out of Alteryx result window into Notepad to be able to read them

- then go back into your development environment to make changes

- repeat.

 

This is very painful, and this will directly scare most people away from learning how to create custom tools since it's not only inefficient - but also scary and frustrating for beginners.

 

Proposal:

Could we instead create mock python libraries; and a development harness (like Google does with Android development in Eclipse) in this SDK where:

- you have full code intelligence (intellisense, autocomplete)

- you can simulate engine events in a test harness (for example in the Android SDK; you can simulate the user rotating their phone, turning off GPS, hitting a volume button, etc).

- you can also write test cases which can run automatically

- then once you know that your tool will work - only then you drop it into the Alteryx Designer environment.

NOTE: This IDE way of thinking also allows you to bring the configuration pieces (like number of inputs; etc) out of raw code and into configuration options.

 

Although you may be able to do remote debugging by using platforms like PyCharm - that really does not give you the full ability to check in the code of your tool; along with all the test cases; in a harness that allows you to automatically check different events; or to make sure that your tool works in the test harness before deploying.

 

Thank you

 

 

 

 

 

cc: @BlytheE @SteveA @Ozzie @tlarsen7572 @c2willis @jdunkerley79 

While Alteryx allows for a proxy username and password in the settings, these are not passed properly to an NTLM proxy. Support for NTLM authentication would be incredibly useful for a number of corporations who utilize this firewall setup.

 

We currently have to either download via Python or cURL through batch commands called by Alteryx. Since Alteryx uses a cURL back-end, this should be a fairly simple addition to the existing download tool by allowing a selection of proxy server, port, and authentication method in addition to the proxy username and password. This could be done either in the tool itself or in User Settings.

When developing in Python using custom objects - you often use print( object) or str(object) to see what's in this object quickly.

For example 

myDictionary = {
'CarType': 'Ford',
'Cost': 20000}

this defines a dictionary.   If I want to quickly look into these to see what's there I can use:

 

print(myDictionary)

# gives {'Cost': 20000, 'CarType': 'Ford'}

str(myDictionary) 

"{'Cost': 20000, 'CarType': 'Ford'}"

 

This is incredibly useful for debugging and to understand how these custom objects / classes work.

 

Please could you add an overload to the str() method to allow this kind of simple debugging and understanding for all the classes used in the Alteryx Python SDK (https://help.alteryx.com/20193/developer-help/sdks/build-custom-tools/python-engine-sdk/classes)

 

For example:

str(record_info_in) which is of type <class 'AlteryxPythonSDK.RecordInfo'>) gives you <AlteryxPythonSDK.RecordInfo object at 0x000001A2C48C3190> which is not very helpful.

Much more useful would be to flatten this into a string format or dictionary so that users can see what's in the RecordInfo object that they're working with to make delivery and debugging easier.

 

 

cc: @Ozzie @BlytheE 

 

 

 

 

Adobe analytics api current version is 2.0 but the connector still uses 1.4. This works for log in access but does not for automated access since the end points are changed and 2.0 uses a new key structure

API Security requirements are constantly evolving and strengthening.  As API architectures migrate from traditional authentication models (Basic, OAuth, etc.) to more secure, certificate-based models, like MTLS/MSSL, leveraging Alteryx Designer will become increasingly difficult, especially for larger organizations trying to scale the use of Alteryx across a large user base, with vastly diverse skillsets.

 

I realize issuing API calls with certificates is possible via the Run Command tool.  We consider this a temporary workaround, and not a permanent, strategic solution.  The Run Command tool can be clunky to use when passing in variables and passing the output back into the workflow for downstream processing.

 

Therefore, I would like to request a more scalable approach to issuing MTLS/MSSL API calls.  Can an option be added to the Download Tool to allow for certificates to be passed on API calls?

In response to my question here: https://community.alteryx.com/t5/Alteryx-Designer-Discussions/Singin-Error-to-Tableau-server-using-P...

 

The Publish to Tableau Server does not support SAML/SSO. I would like this feature to be added to this tool as it will make our business process more efficient.

 

Thank you.

  • API

Most organizations are moving to the cloud and Microsoft is the biggest one in the game.

I pleased that you have updated the List tool to connect to SP 365 but soon all of the shared drives will be online and if its vital for Alteryx success to be able to communicate, pull and export data to those locations. 

 

With the Desktop, there is a way to make it work by authenticating and for the previous SP version by logging on to SharePoint first on the server then the scheduler would run successfully. Now the scheduler isn't able to take advantage of the authentication already established. 

 

 

If Alteryx isn't able to keep up it will soon find itself behind in innovation and tool of choice.

My company has recently purchased some Alteryx licences with the hope of advancing their Data Science  capability. The business is currently moving all their POS data from in-premise to cloud environment and have identified Azure Cosmos DB as a perfect enviornment to house the streaming data. Having purchased the Alteryx licences, we have now a challenge of not being able to connect to the Azure Cosmos DB environment and we would like Alteryx to consider speeding up the development of this process.

In the current version of alteryx, we are only able to pull the first 2000 records from an existing salesforce report, a lot of times it becomes difficult to automate my monthly/weekly requests because it is depended on an existing salesforce report and because of that I need to recreate the report on alteryx using the salesforce objects and fields.

 

If we had the ability to bring in the entire salesforce report to alteryx environment it would save a lot of time and effort for analysts like me because in that way, we won't have to re-invent the wheel.

  • API

Using the download tool is great and easy to use. However, if there was a connection problem with a request the workflow errors out. Having an option to not error out, ability to skip failed records, and retrying records that failed would be A LIFE CHANGER. Currently I have been using a Python tool to create multi-threaded requests and is proven to be time consuming.

My idea is to have the AlteryxEngineCMD.exe to run a workflow as part of the standard Alteryx license.

 

Use case - be able to run Alteryx from the command line without the need to buy the entire Scheduler package (at $6,500/seat).

 

I understand why certain features are add on, but the ability to run AlteryxEngineCMD.exe (I feel) should be part of  the standard license which is already $5K+. For those who only need to be able to run a command line execution of a workflow $6.5K is a lot of money!

Recently we had a situation where installing the data packages was expected to take over 20 hours!  We do not have the ability to run a machine undisturbed for this length of time at the office, and VPN automatically times out after 12 hours.  Okay, so these items are company-specific, but how nice would it be to be able to download/install the data in smaller portions so you don't have to worry about setting up shop for 8-10 hours?!

 

I was able to copy the DataInstall.ini and name only the portions of the data I wanted to install in each session.  But then I had to separate the installs into different network folders otherwise you end up overwriting the DataInstall.exe file, which users need for each install in order to register from a network location (for each install).

 

Needless to say, it took SEVERAL weekends and quite a few mistakes before I was able to do it this way successfully!

 

Please vote for data installs in smaller portions!

  • API

For those with large web and streaming-media server logs, the ability to geocode IP addresses be an excellent feature, similar to Alteryx's ability go geocode street addresses. Several IP geocoding services exist, with different levels of accuracy and cost. Ideally, the user should be able to choose their own service if they have one, in addition to a default service built-in to Alteryx.

 

A proof of concept is documented in this post

 

The IP geocoding tools should include:

  • IP address to IP number conversion, and the reverse
  • IPv4 and IPv6
  • Output for postal code, country, city, state, latitude/longitude centeroids

 

0 Likes

Hello Alteryx, Please fix Salesforce Input and Output tools. 

 

The input tool has a lot of issues with Login , , Custom SQL, Json issues, Machine Encryption, and saving to the gallery , . 

Unable to use Crew macros with Salesforce input Workflows etc.... there is a lot of post's with all the issues with Salesforce V.1 Versions. 

0 Likes

Right now as far as I know you need to add each DB connection manually. This works... but is quite time consuming when trying to run tasks against a cluster of prod databases. It would be awesome to pass a JSON config file,example below, to the Alteryx Engine and have Alteryx create those connections upon parsing the file. This would save tons of time, and allow teams to share a central config file with consistent aliases across their clusters to ensure their app connections point to the same DBs across workflows. It would also make on boarding a breeze for new developers on team.

 

Alteryx-DB-Config.png

0 Likes

Hi,

 

The Adobe Analytics API token is currently set to expire 30 days after the call has been configured. When the token expires, I have to re-authenticate AND reconfigure the API call.

 

The API call shuoldn't expire when the token expires. Upon re-authenticating, the call should persist as it was originally configuired.

 

Thanks,

Mihail

 

 

  • API
0 Likes

Good afternoon,

 

I work with a large group of individuals, close to 30,000, and a lot of our files are ran as .dif/.kat files used to import to certain applications and softwares that pertain to our work. We were wondering if this has been brought up before and what the possibility might be.

  • API
Top Liked Authors