Bring your best ideas to the AI Use Case Contest! Enter to win 40 hours of expert engineering support and bring your vision to life using the powerful combination of Alteryx + AI. Learn more now, or go straight to the submission form.
Start Free Trial

Alteryx Designer Desktop Ideas

Share your Designer Desktop product ideas - we're listening!
Submitting an Idea?

Be sure to review our Idea Submission Guidelines for more information!

Submission Guidelines

Featured Ideas

Please add support for Databricks' Unity Catalog

 

Currently, when selecting a Databricks-connection in the “Connect In-DB”-tool, and opening the “Query Builder”, only tables in the catalog named “hive_metastore” are listed. That is, Alteryx submits the following SQL query to Databricks:

Listing tables 'catalog : hive\_metastore, schemaPattern : %, tableTypes : null, tableName : %'

 

However, with Unity Catalog in Databricks the namespace is three-tier and there may be multiple catalogs (and not just the "hive_metastore" catalog), see https://docs.microsoft.com/en-gb/azure/databricks/lakehouse/data-objects#--what-is-a-catalog

 

I reached out to Alteryx support, which replied that you currently have a feature request for implementing this change (ID TDCB-4056) and they furthermore suggested that I post here.

 

Thanks in advance.

The Output tool currently creates a zero-row file if you filter all the rows out (an empty file)

 

However - there's a very common need to say "don't generate a file at all if there are zero rows" - for example if you are generating a "Daily Details" file and there are no daily details today, you want to suppress this file.   This frequently trips up newer users since the only way to suppress a zero-row output file today is to it with macros.

 

Could you add a flag to the output tool to "Suppress empty files" so that if there are no rows to be written, Alteryx does NOT create an empty file?   This way the user can be in control of what to do with empty record sets?

 

Hello,

According to wikipedia :

 

A partition is a division of a logical database or its constituent elements into distinct independent parts. Database partitioning is normally done for manageability, performance or availability reasons, or for load balancing. It is popular in distributed database management systems, where each partition may be spread over multiple nodes, with users at the node performing local transactions on the partition. This increases performance for sites that have regular transactions involving certain views of data, whilst maintaining availability and security. 

 

 

Well, basically, you split your table in several parts, according to a field. it's very useful in term of performance when your workflows are in delta or when all your queries are based on a date. (e.g. : my table helps me to follow my sales month by month, I partition my table by month).

So the idea is to support that in Alteryx, it will add a good value, especially in In-DB workflows.

Best regards,

Simon

Data Cleansing Tool:   There should be a sub-category on the "Punctuation" cleansing.  Ideally to have an option to "Include Only" or conversely "Exclude these characters" which would allow you for example to remove all characters except   "." from a dollar formatted field .  There are times when you need to clean almost everything except a certain punctuation or not. 

I would like to suggest creating a fix to allow In-DB Connect tool's custom SQL to read Common Table Expressions. As of 2018.2, the SQL fails due to the fact that In-DB tools wrap everything in a select * statement. Since CTE's need to start with With, this causes the SQL to error out. This would be a huge help instead of having to write nested sub selects in a long, complex SQL code!

The Append Fields tool will issue a Warning if/when the Source data stream has no records that reads something like this:

Append Fields (823) There are no records present in the source.

 

I can imagine many situations when this issue should be flagged as a Warning. However, I have use cases when both the Source and Target data streams are expected to be empty. Because it is a common, expected scenario, I do not want it flagged as a Warning for the user. 

 

My Idea: provide another option to suppress warnings for this situation.

 

Perhaps it could be a standalone checkbox, for example:

                 [x] Suppress Warning when both source and target streams are empty

 

Alternatively, the tool currently has 3 options to manage warnings or errors related to "too many" records. Perhaps this could be added as a 4th option to the dropdown list, although that would necessitate changing the label slightly.

 

It would be ever so helpful and save a couple extra steps if a count distinct option could be added to the crosstab tool.  Seems like a slam dunk since plain ole 'count' is already a choice.

Capture.JPG

 

We all love seeing this.  And, it's fairly easy to fix, just go find the macro and insert a new copy.  But, then you have to remember the configuration and hope that it was simple. 

With the tool that's there, the XML still contains the configuration, all that's missing is the tool path.    It would be great to be able to right click and repair the path from the context of the missing macro.

Hi

 

The action of the 'tab' key in configuration window recently appears to have changed from indenting to a navigation function. 

 

The user should be able to select which action the tab key performs. 

 

Alternatively, tab should indent and shift-tab (or alternative) navigate. I'm not the only one who would appreciate the choice.

 

PuffinPanic

 

Hello all,

I suggest a new string function Repeat()

 

Repeat() forms a string consisting of the input string repeated the number of times defined by the second argument.

Repeat(text[, repeat_count])

 

Repeat('to',3) gives tototo

It's also a standard SQL function
https://www.w3schools.com/sql/func_mysql_repeat.asp

Best regards,

Simon

I would love to be able to see the actual curl statement that is executed as part of the download tool. Maybe something like a debug switch can be added which would produce 1 extra output field which is the curl statement itself? This would greatly enhance the ability to debug when things aren't working as expected from the download tool.

Hello there!

Currently the email tool has the following configuration:

TheOC_0-1653663519956.png



It is a fairly easy tool to use - however one part that I would like to be improved is testing the SMTP settings - similar to how it is done on the Alteryx Server. It would be awesome to have a button as part of this page, that would send a test email, and return true/display to the user that the email sent correctly. This would stop the need to setup dummy data and a dummy output to test a connection before rolling the email tool out into a live environment/use-case.

 

I imagine something along these lines for this functionality:

TheOC_2-1653663925251.png


Clicking test (and passing):

TheOC_3-1653663937900.png



Clicking test (and failing):

TheOC_4-1653663975804.png




Thanks,
TheOC

 



This year, Microsoft updated improve their API (GraphAPI) to access Office365 enviroment.

 

Alteryx have launched on Microsoft District in Public Gallery the Dataverse , Onedrive & Sharepoint Connectors.

 

Alteryx must develop as soon as possible an connector/email with same authenticator options as connectors Above and improve the emails settings.

 

 

It´s important to release and documentation to show wich permissions on Azure it's necssary to send the emails. 

 

References:

 

Microsoft Announce:

https://docs.microsoft.com/pt-br/lifecycle/announcements/exchange-online-basic-auth-deprecated 

 

Azure Permissions:

https://docs.microsoft.com/en-us/graph/api/user-sendmail?view=graph-rest-1.0&tabs=csharp 

Parquet is a very fast, efficient and widely used data format, currently only below Parquet compression algorithms are supported and we cannot use Alteryx to read the parquet file that generated by other processes. This limits our usage in Alteryx.  

  • Read support: Snappy and Gzip compression algorithms.

  • Write support: Snappy only

It would be great for Alteryx to support all types of Parquet format so we can maximize the use of Alteryx in data analysis.

Who needs a 1073741823 sized string anyways?  No one, or close enough to no one.  But, if you are creating some fancy new properties in the formula tool and just cranking along and then you see that your **bleep** data stream is 9G for nine rows of data you find yourself wondering what the hell is going on.  And then, you walk your way way down the workflow for a while finding slots where the default 1073741823 value got set, changing them to non-insane sized strings, and the your data flow is more like 64kb and your workflow runs in 3 seconds instead of 30 seconds.  

 

Please set the default value for formula tools to a non-insane value that won't be changed by default by 99.99999% of use cases.  Thank you.

 

 

Hi 

 

The wording of the tool tip displayed in results window cells with long strings is misleading.  The current wording is "This cell has truncated characters".   

danilang_0-1616587137476.png

 

 

New users tend to infer that this means that the data value has been truncated somewhere upstream.  See here, here and here.   Changing this message to something like "Only a portion of long strings is displayed" will help reduce the confusion immensely.

 

Dan  

Hello all,

Change Data Capture ( https://en.wikipedia.org/wiki/Change_data_capture ) is an effective way to deal with changes in a database, allowing streaming or delta functionning. Several technos, more or less intrusive, can be applied (and combined). Ex : logs reading.

Qlik  : https://www.qlik.com/us/streaming-data/data-streaming-cdc

Talend : https://www.talend.com/resources/change-data-capture/

 

Best regards,

Simon

for iterative macro, generally it had 2 anchors, one if it is for iterative, and it normally no output (whether got error or not)


it good to have option to remove this anchor when using it in workflow. 

so other user no need to identify which one is the True output and which one is just iteration.

 

additional, if this can apply to input anchor.
(i just built one macro where i don't need the start input, but the input need to be iterate input)

 

Ability to ‘name’ the point created in the “Create Points” tool.

 

Instead of sticking a select tool after it to rename it from ‘centroid’ to Starting Location or Store location or whatever.

A common problem with the R tool is that it outputs "False Errors" like the following: "The R.exe exit code (4294967295) indicted an error"

I call this a false error because data passes out of the R script the same as if there were no error. As such, this error can generally be ignored. In my use case, however, my R tool is embedded within an iterative macro, and the error causes the iterator to stop running.

 

I was able to create a workaround by moving the R tool to a separate workflow and calling it from the CReW runner macro within my iterator, effectively suppressing the error message, but this solution is a bit clumsy, requires unnecessary read/writes, and uses nonstandard macros.

 

I propose the solution suggested by @mbarone (https://community.alteryx.com/t5/Alteryx-Designer-Discussions/Boosted-Model-Error/td-p/5509) to only generate an error when the R return code is 1, indicating a true error, and to either ignore these false errors or pass them as warnings. This will allow R scripts and R-based tools to be embedded within iterative macros without breaking.

 

 

Autoren, für die die meisten Kudos vergeben wurden