Inspire EMEA 2022 On-Demand is live! Watch now, and be sure to save the date for Inspire 2023 in Las Vegas next May.

Alteryx Designer Ideas

Share your Designer product ideas - we're listening!
Submitting an Idea?

Be sure to review our Idea Submission Guidelines for more information!

Submission Guidelines

Featured Ideas

Connecting to Smartsheets using Alteryx Desktop (and by extension, Alteryx Server) is extremely cumbersome. If a user wants to read data from Smartsheet, they are required to get an API token (preferred) or use a username/password


Then do one of the following to read data from Smartsheets:

1. a. Install a ODBC driver

    b. Configure a DSN connection for ODBC
    c. Use the input data using a generic ODBC connection
2. Use python


To write data to Smartsheets, a user can use Python or upload the data using an API call - both very hard for end users to use especially if they're not Python developers.


Regardless, all of these are problematic. On the server I manage, I have over 15 ODBC connections to Smartsheets and it's getting very hard to upgrade the server hardware because of them. Creating a native connector for input/output of data to Smartsheets will eliminate a headache of managing ODBC connections, and make it simple for Alteryx Desktop users to read and write data.



When email body gets imported using latest version of the Outlook 365 tool, this tool removes the new line separators from the message body, which makes it difficult to parse relevant information out of the message body. New line separators are there prior to message being imported into Alteryx as can be verified when importing same message using different tools (for example, Python or Power Automate). Without new line separators it is not possible to accurately parse message body using Alteryx. Please add the enhancement to the Outlook 365 tool so that it doesn't remove new line separators from the message body.

This limitation of the Outlook  365 tool has been discussed in the community 

Outlook 365 Input - Separate Body lines - Alteryx Community

Hi all,


Currently, only the Sharepoint list tool (deprecated) is working with DCM, it would be amazing to add the Sharepoint files input/output to also work with DCM.



Thank you,

Fernando Vizcaino

The Sharepoint file tools are certainly a step in the right direction, but it would be great to enhance the files types that it is possible to write to sharepoint from Alteryx.


The format missing that I think is probably most in demand is pdf. If we're using the Alteryx reporting suite to create PDF reports, it would be awesome to have an easy way to output these to Sharepoint.


A few suggestions that I think can improve the Sharepoint Files Output Tool:


  1. Maybe I'm missing it, but I cannot see how you can delete a file from the output list once you've added it:
    1. Joe_Lipski_1-1643795491697.png



  2. Have the write headers output checkbox ticked by default as I expect this is the more common expectation:
    1. Joe_Lipski_2-1643795558506.png


  3. Take the file extension by default based on the users selection in the Options tab as I shouldn't have to write .xlsx for the extension:
    1. Joe_Lipski_3-1643795713015.png


    2. Joe_Lipski_4-1643795723518.png



Currently Anaplan Input and Output tool only supports basic authentication and 
does not have option to connect using CA Certificate Authentication.
Drawback of basic authentication is that critical information (email and password) is easily accessible.
If unauthorized person retrieves the credentials, he can access all of Anaplan interface and functionality.
Using CA Certification authentication adds a layer of security as no critical information is stored.
If certificate information is retrieved, the unauthorized person can only trigger APIs and has no access to other parts of Anaplan interface and functionality
We hope that the Anaplan tools can be enhanced to have CA Certificate Authentication so
as to provide added security to transfer data between Alteryx and Anaplan.

In order to make the connections between Alteryx and Snowflake even more secure we would like to have the possibility to connect to snowflake with OAuth in an easier way. 


The connections to snowflake via OAuth are very similar to the connections Alteryx already does with O365 applications. It requires: 

  • Tenant URL 
  • Client ID 
  • Client Secret 


  1. Get Authorization token provided by the snowflake authorization endpoint.  
  2. Give access consent (a browser popup will appear) 
  3. With the Authorization Code, the client ID and the Client Secret make a call to retrieve the Refresh Token and TTL information for the tokens 
  4. Get the Access Token every time it expires 


With this an automated workflow using OAuth between Alteryx and Snowflake will be possible.


You can find a more detailed explanation in the attached document.

Hello Alteryx,


It seems that the Endpoint parameter for the Amazon S3 Upload tool only support "Path Like" URL. It would be great if the Endpoint parameter could also take into account "Virual Hosted" URL.


When we enter a "Virtual Hosted" URL, the "Bucket Name" and "Object Name" parameters don't respond correctly.

The three dots option for the "Bucket Name" parameter returns the bucket name and the object name at the same time. And the three dots option for the "Object Name" parameter doesn't suggest any object name.

We can enter those manually but we lose some of the Alteryx functionnality. 


It would be a great improvement that the Endpoint parameter takes into account "Virtual Hosted" URL so we keep "Bucket Name" and "Object Name" suggestions once the Endpoint is registered.


Is it in the roadmap?




Hi all,


The SalesForce Input tool is great.. but has some really bad limitations when it comes to report. 

I think there are 2 main limitations :


A - It can only consume 2000 rows due to the rest api limitation. There plenty of articles about it in the community.

B - Long string such as text comment are cutout after a certain number of characters. 


Thanks to this great article : , I had the idea of going through a csv file export to then import the data into Alteryx. 

I've done it using two consequent download tool. The first download is used to get the session id and the second to export a report into a csv in the temp folder. This temp file can then be read using a dynamic input workflow. 


Long story short, I think Alteryx should upgrade the Salesforce connector to make it more robust and usable. Using the export to csv feature, this should enable Alteryx to be fully compatible with Salesforce report.



Currently Alteryx does not support writing to SharePoint document libraries.

However there are success sometimes but not at other times.

Please see attachment where we ran into an issue.

See this link for additional information.


We need official support for reading and writing to SharePoint document libraries.

It's an important Output target, and will becoming more so, as Alteryx enhances its reporting capabilities.


It would be wonderful for Alteryx to be able to connect to and query OData feeds natively, rather than using a 3rd-party driver or custom macro.   


OData querying is supported by quite a few familiar products, including Excel and PowerBISSIS/SSRS, FME SafeTableau, and many others. And the protocol is used to publish feeds from Microsoft Dynamics and Sharepoint, as well as many of the 10,000 publically available government datasets with API's (esp. those hosted by Socrata)   


I didn't see it as in the Idea section, but questions and workarounds have been discussed in the community a few times (11/15, 3/18, 4/18), and suggestions seem to be just to buy the $400-600 ODBC driver from CDATA (or ZappySys), or I could use a VBA script in Excel trigger a refresh, or create my own Alteryx connector macro (great series btw, though most was beyond my understanding!) 


While not opposed paying, kludging, or learning to program, they're just one more thing to build/buy, install, maintain, and break at the most inconvenient time 🙂




OData Overview:

OData (Open Data Protocol) is an ISO/IEC approvedOASIS standard that defines a set of best practices for building and consuming RESTful APIs. OData helps you focus on your business logic while building RESTful APIs without having to worry about the various approaches to define request and response headers, status codes, HTTP methods, URL conventions, media types, payload formats, query options, etc. OData also provides guidance for tracking changes, defining functions/actions for reusable procedures, and sending asynchronous/batch requests.  OData RESTful APIs are easy to consume. The OData metadata, a machine-readable description of the data model of the APIs, enables the creation of powerful generic client proxies and tools.

More info at at

Hi all,


The Publish to Tableau Server tool is great.. but requires username and password. If you are using AD, there is a chance that your users don't have a password. In that case, you probably have a technical user that you share across the team. This is not an ideal situation and you loose the governance around the data. 


Fortunately, there is an easy workaround. You can leverage personal token authentication : 


The advantage of this method is that it logs in with your user and your data source is uploaded under your name. This is still using the Tableau REST API so the changes to do in the current macro is MINOR. 


Changes to do in the current macro : 


1- Add a parameter authentication method with choices : Username/Password ; Personal Token 

2- If Personal Token is selected, add two parameters : Token_Name and Token_Value

3 - In the TableauServer.Login supporting macro, improve the formula(13) to change the payload based on user selection. If Username/Password, keep it as is. Else use the syntax here : 



This is quite a straight forward change but could help a lot of companies using Alteryx.

Can you please implement that changes to strengthen this tool ? 


Thanks a lot,

After talking with support we found out that Oracle Financial Cloud ERP is not listed among supported Data Sources as stated in the url below:


We would like this added as our company will begin working heavily with Oracle Financial Cloud ERP to bring data from that into our SQL servers. Is there a reason why that connection is not currently being investigated and set up?




I recently began using the SharePoint Files v2.0.1 tools to read and write data. The SharePoint Files Output tool allows you to take a sheet or filename from a column but that column is still included in the output. The standard Output Data tool has a "Keep Field in Output" checkbox that allows you to control if the column stays in the XLSX of CSV file. It would be great if this same functionality could be included in the SharePoint Files Output tool.

Ouput Tool Checkbox.PNG

We aren't getting a huge amount of help from support on this, so I'm posting this idea to raise awareness for the product teams responsible for the Salesforce connectors and the embedded Python environment.


This post from user Dubya describes the issue in detail:


I have a workflow with several salesforce tools in it, which works fine on my machine. But we need another alteryx user in our office to be able to access, run and maintain the workflow too, via their machine and copy of alteryx designer.

However we're finding that the salesforce inputs and outputs can only be authenticated on one machine at a time.

When the other new user opens the original workflow from the shared network location, the salesforce tools display an error "Salesforce Input (1): {'error': 'invalid_grant', 'error_description': 'authentication failure'}" and the tools fail to load any data. But we can see the full query in the tool and we can even set the custom query option and validate the query successfully, which suggests the source is being correctly connected to and queried, but we just cant run the tool.

The only way to run the tool successfully is to change the credentials and re-authenticate the tool. However this then de-authenticates the original machine, and when we open up the workflow on there and try to run ying the workflow brings back the same error.

We've both tried this authentication back and forth on our own machines and each time one of us re-authenticates, it de-authenticates the other, leading to it triggering the error.

Can someone help explain what's going on and how to fix it, as this doesn't bode well for our collaboration.

We're both running:
The latest build of version of designer 2021.2 (original machine also running desktop automation)
Salesforce Input Tool v4.1.0
Salesforce Output Tool v1.3.0

My response here identifies that this is a problem for our organization as well:


We're experiencing the same issue. It appears to be related to how the tool handles password and security token decryption. I've found that when you modify the related registry entry from "true" to "false", you can see in the tool's xml that the encrypted password and security token are still in there. I'm not sure what else is going on behind the scenes beyond that, but that ought to be addressable by the product teams handling the Salesforce connectors and the Python installation embedded in Designer.

The only differences in our environment compared to u/Dubya's are that we're running on 2020.4 and attempting to use Salesforce Input Tool v4.2.4.


This is a must have for anyone who needs the ability to share workflows among multiple users. This is part of a series of problems that these updated connectors have been plagued with since introducing them years ago, and no one at Alteryx seems to care enough to truly fix the problems. Salesforce is a core system for our organization, so having tools that utilize the latest version of Salesforce's APIs is very important to us. The additional features that the Input tool provides are welcome, but these bugs have to be sorted out in order for us to extract any kind of value out of them. If the "deprecated" Salesforce tools were ever to be removed from Designer while there are issues with the "new" connectors, we would have no choice other than to never upgrade Designer/Server again and be forced to look for another product to serve as our ETL platform.


Please, please, please address this.

Hey all,


The join tool currently does not allow case-insensitive joins, but the find/replace tool does.    Additionally- even if both sides are identical, the join tool will not join "Sean's house" to "Sean's house" because of the non-letter character in the middle.    Finally - if one side is a string(2), and the other is a vString(200) - even if you have a single identical character on both sides you get uncertain outcomes unless you force the type


Please could you consider amending the join tool to include 3 new options or capabilities:

- Case insensitive join

- Allow full Unicode character set in join

- Full match across text types (irrespective of string size) - this would allow a string(2) value to match to a string(100) value as long as the string(100) value only has the same 2 characters in it as the string(2) value


That would remove a load of work from every text-join that's being done on every canvas we do.


Thank you 




The ‘Existing File Action’ configuration setting needs to address the situation where columns change.

Currently the following options exist, with instruction as follows at





There are two expectations when using this tool:

1. Reloads are built to completely replace the contents of a dataset, i.e. an Append is not being performed

2. Columns will change over time with continued development


There is therefore a need for an ‘Overwrite (update columns)’ option. However, when this Existing File Action is used it updates column names, but it does not delete the contents prior to upload. An append onto the existing data, but with new column names is therefore performed.

I acknowledge that the instructions do not say that existing rows are deleted.


This leaves the need to perform a workaround:

  1. Publish with ‘Overwrite (update columns)’
  2. Publish (immediately after) with ‘Overwrite (keep existing columns)’

If step 2 is not done data will be appended which would lead to duplication issues.


Either these Existing File Actions need to be renamed to be clearer as to their operation, or preferably an option that updates the columns and sets up new (non-appended) data is required.


A lot of companies now are deploying on both AWS and Microsoft Azure.

Alteryx supports AWS S3 object storage out of the box, it would be important to support Microsoft Azure blob as part of the native Alteryx product as well. 



  • Category Connectors
The Azure Machine Learning Training and Scoring Tools seems great to improve Azure ML process.

Introducing: The Azure Machine Learning Training and Scoring Tools 

We tried to use this tool but can't log in to Azure ML correctly. We have several Tenant ID then log in to another tenant for office 365 not Azure ML.

====================== <Error Message> ==========================================================
Error: Azure ML Training (367): UserErrorException:
    Message: You are currently logged-in to 55f0a...-.............................................. tenant. You don't have access to d846a...-............................................. subscription, please check if it is in this tenant. All the subscriptions that you have access to in this tenant are =
 [SubscriptionInfo(subscription_name='Microsoft Azure Enterprise', subscription_id='754c5...-...........................')].
 Please refer to for different authentication mechanisms in azureml-sdk.
    InnerException None

Microsoft states that tenant needs to be specified if we have access to multiple tenants.

Set up authentication for Azure Machine Learning resources and workflows 

Could you add Tenant ID into Azure credentials so that we can use this tool? 


At our organization we are required to change our passwords every few months forcing a change to my Tableau Server password.  How does this relate to Alteryx?  Well, every 90 days I have to change my password in the "Publish to Tableau Server Tool" for all of my workflows.  This is quite a cumbersome process that could be eliminated with AD.


If you dislike manually changing your for each workflow that uses this tool then "star" this post!




  • Category Connectors
Top Liked Authors