Join the Alteryx Community’s Maveryx Summer Cup event! Compete, network with others, and earn your gold through a series of challenges from July 24th to August 11th. Learn more about the event here.
The Product Idea boards have gotten an update to better integrate them within our Product team's idea cycle! However this update does have a few unique behaviors, if you have any questions about them check out our FAQ.

Alteryx Designer Desktop Ideas

Share your Designer Desktop product ideas - we're listening!
Submitting an Idea?

Be sure to review our Idea Submission Guidelines for more information!

Submission Guidelines

Featured Ideas

There is no tool that exists that outputs all records that are duplicates (those sharing the selected values with at least one other record) and also outputs the records that are not duplicates (those not sharing the selected values with at least one other record).

 

The Unique Tool is not sufficient.  It only provides the first record of a unique duplicate group along with any non-duplicates and then provides a secondary output that only contains the additional records of a duplicate group.  Sometimes you only care about the duplicates and want to quickly see what differs between the unique groups.

 

For example, if there are 4 records with the City of Austin and I am looking for duplicates on City I want to see all 4 records with Austin in the output so I can quickly compare additional fields to see what might differ, or if they are all indeed truly duplicates.

0 Likes

Many of today's APIs, like MS Graph, won't or can't return more than a few hundred rows of JSON data. Usually, the metadata returned will include a complete URL for the NEXT set of data.

 

Example:  https://graph.microsoft.com/v1.0/devices?$count=true&$top=999&$filter=(startswith(operatingSystem,'W...') or startswith(operatingSystem,'Mac')) and (approximateLastSignInDateTime ge 2022-09-25T12:00:00Z)

This will require that the "Encode URL" checkbox in the download tool be checked, and the metadata "nextLevel" output will have the same URL plus a $skiptoken=xxxxx value.  That "nextLevel" url is what you need to get the next set of rows.

 

The only way to do this effectively is an Iterative Macro . 

 

Now, your download tool is "encode URL" checked, BUT the next url in the metadata is already URL Encoded . . . so it will break, badly, when using the nextLevel metadata value as the iterative item.

 

So, long story short, we need to DECODE the url in the nextLevel metadata before it reaches the Iterative Output point . . . but no such tool exists.

I've made a little macro to decode a url, but I am no expert. Running the url through a Find Replace tool against a table of ASCII replacements pulled from w3school.com probably isn't a good answer.

 

We need a proper tool from Alteryx!

 

Someone suggested I use the Formula UrlEncode ability . . .

 

Unfortunately, the Formula UrlEncode does NOT work. It encodes things based upon a straight ASCII conversion table, and therefore it encodes things like ? and $ when it should not.  Whoever is responsible for that code in the formula tool needs to re-visit it.

 

Base URL: https://graph.microsoft.com/v1.0/devices?$count=true&$top=999&$filter=(startswith(operatingSystem,'W...') or startswith(operatingSystem,'Mac')) and (approximateLastSignInDateTime ge 2022-09-25T12:00:00Z)

 

Correct Encoding: 

 
Formula UrlEncode:
 
As you can see, the initial "?" was altered, as were all the "$", and the result is rejected by the http server. Also, the spaces around "and" and "or" should convert to + in this case, but I THINK the standard %20 would also work.
 
So, I don't know what the Download Tool is doing when it encodes, but it does a good job at it. Formula UrlEncode, while one would expect it to do the same encoding as the Download Tool, is not and therefore fails.
 
So, I have a couple of asks:
1. Copy the code from inside the Download Tool for Encoding and make it into a separate tool. (UrlEncode Tool, perhaps in Parse Tab?)
2. Using that copied code but in reverse, create a Url Decode Tool. (UrlDecode Tool, again - perhaps in Parse Tab?)
3. Fix the formula UrlEncode

Hello all,

Like many softwares in the market,  Alteryx uses third-party components developed by other teams/providers/entities. This is a good thing since it means standard features for a very low price. However, these components are very regurarly upgraded (usually several times a year) while Alteryx doesn't upgrade it... this leads to lack of features, performance issues, bugs let uncorrected or worse, safety failures.

Among these third-party components :

- CURL (behind Download tool for API) : on Alteryx 7.15 (2006) while the current release is 8.0 (2023)
- Active Query Builder (behind Visual Query Builder) : several years behind

- R : on Alteryx 4.1.3 (march 2022)  while the next is 4.3 (april 2023)
- Python : on Alteryx 3.8.5 (2020) whil the current is 3.10 (april 2023)
-etc, etc....
-

of course, you can't upgrade each time but once a year seems a minimum...

Best regards,

Simon

Add ability to name the columns for the text to column fields tool. 

It looks like we can choose which fields to include in a workflow with the Listbox interface + Select tool., but we cannot ORDER or REORDER fields.

 

I did stumble across this post, it looks like it can be done but it isnt very elegant.

https://community.alteryx.com/t5/Alteryx-Designer-Discussions/User-Dynamic-Reorder-Columns/td-p/7437...

 

 

0 Likes

HI, 

Not sure if this Idea was already posted (I was not able to find an answer), but let me try to explain.

 

When I am using Mongo DB Input tool to query AlteryxService Mongo DB (in order to identify issues on the Gallery) I have to extract all data from Collection AS_Result.

The problem is that here we have huge amount of data and extracting and then parsing _ServiceData_ (blob) consume time and system resources.

 

This solution I am proposing is to add Sorting option to Mongo input tool. Simple choice ASC or DESC order. 

Sebastian_Chaieb2_0-1679473890913.png

 

Thanks to that I can extract in example last 200 records and do my investigation instead of extracting everything 

 

In addition it will be much easier to estimate daily workload and extract (via scheduler) only this amount of data we need to analyze every day ad load results to external BD.

 

Thanks,

Sebastian 

I hope have a flexibility to adjust the print area workflow.

in the canvas.

 

it similar to excel print area, so we can adjust all the area easily.

for security. it may add function to check whether all tools is covered in print areas.

Pang_Hee_Choy_0-1679451852663.png

it is hard to cut the workflow to pieces when it is huge. 

 

 

I've used the Table tool with large data sets to make tables with conditional formatting etc. There's a couple of suggestions I'd like to see.

 

1. I noticed an issue where if you disconnect from the tool prior to the Table tool before it forgets your settings quite easily and you may need to redo them. This is quite frustrating if you have lots of columns

2. The controls for sorting and interacting with columns aren't very good, if they were more like the select tool controls that would be fantastic. Perhaps this could be resolved with a select tool beforehand but I still think it is worth putting on the table tool itself.

3. Render output. when making excel outputs with multiple sheets of varying sizes, its very difficult to control. The sheets all stretch to the largest size. I've found I've had to put in white space in Report Text tools on one side of a table tool in order to make up the space and prevent stretching. (I found that solution on the forums)

 

Thanks.

 

Frank

Hi!

 

Currently, the only visual the user has on which column they have selected in the Results window, is the 3 dots for column options (Data Cleanse/Filter/Sort). It would be incredible if Alteryx would add a border around the entire column that is selected (like you currently do when selecting a whole row) & maybe possibly even bold the header of the selected column.

1.JPG

 

I think you might be changing the background color from light grey to white, but it's so subtle it's hard to tell. Make it pop!

 

 

Connect to PowerPivot data model or an Atomsvc file 

 

It would be nice to have a tool that automatically normalized data, or calculated percentages.  This could be overall or in groups. 

 

For example, maybe I have a dataset with 2 columns: US State and Number of amusement parks.  So I know the count of amusement parks by state.  But maybe I want a distribution so I can see which percentage are in what state.  What I want is (# of parks in a state) / (total across all states).  Currently you need at least 3 tools to do this calculation (summarize, then join or append fields, then formula).  This is a very common operation, and often I want counts and percentages next to each other in a table.

 

Such a tool could be called "normalize" or "rescale" or "scale".  It could be more general - maybe not just normalizing so values add to 1 (or 100%), but to other magnitudes, recentering the data, or doing a "standard normal" (z score) transormation as suggested here:
https://community.alteryx.com/t5/Alteryx-Designer-Discussions/How-to-do-Feature-Normalization-in-Alt...

Can we have an option to disable all tool containers at once? Similar to disable all browse tools or tools that write output.

Hello,

A lot of time, I would like to manipulate data between my interface tool (let's say a dropdown or a control parameter) and the action tool.

This can be things like parsing (text to column..), regex, multifield, etc...

As of today, we can just write formula and this is really poor.

Best regards,

Simon

0 Likes

Alteryx currently shows 100% in the profiling of spatial fields in the results window, regardless of if there are rows with missing spatial features. I opened a ticket about this & was told it is expected behavior.

 

Therefore, I submit the idea that the profiling for spatial fields should give an accurate profile of the field, & if there are nulls in the field, it should identify that column isn't 100% OK and show the % of records that have null values, like the profiling does for every other column in workflows.

1.JPG

 

 

 

 

 

 

 

 

 

 

Thank you!

With data sharing, you can share live data with relative security and ease across Amazon Redshift clusters, AWS accounts, or AWS Regions for read purposes.

Data sharing can improve the agility of your organization. It does this by giving you instant, granular, and high-performance access to data across Amazon Redshift clusters without the need to copy or move it manually.

 

aws Datasharing feature in Alteryx. It's not working in Alteryx v 2020.4

We would like to know whether Alteryx is support redshift data share or not ?

it uses same redhshift ODBC or Simba ODBC drivers only but the functionality would be sharing the data across the clusters. Earlier we have seen a limitation from Alteryx end (not able to read data share objects) so we wanted to check if it’s resolved in newer versions of Alteryx.  Reference link

https://docs.aws.amazon.com/redshift/latest/dg/datashare-overview.html 

Improve error messages to include possible tools or solutions to help solve the error, in some cases that may be too many possibilities and it won't work but in other cases there really is only a few tools to solve some of these error messages. I'm a new user so I'm probably seeing a lot more error messages than everyone else but what would help is a list of tools or suggestions on how a user could solve some of these error messages.

From what I can tell using ProcMon, presently when using the Directory tool to list files (including subdirectories) the Alteryx Engine runs a single threaded process.

When you're trying to find files by checking recursively in large network paths, this can take hours to run.

It would be great if the tools would split up lists of directories (maybe by getting two or three levels down first) and then run each of those recursive paths in parallel. 

While it is possible to do this using a custom Python or cmd->PS command, it would be great if this could just be a native part of the application.

My organization use the SharePoint Files Input and SharePoint Files Output (v2.1.0) and connect with the Client ID, Client Secret, and Tenant ID. After a workflow is saved and scheduled on the server users receive the error "Failed to connect to SharePoint AADSTS700082: The refresh token has expired due to inactivity" every 90 days. My organization is not able to extend the 90 day limit or create non-expiring tokens.

 

If would be great if the SharePoint connectors could automatically refresh the token when it expires so users don't have to open the workflow and do it manually.

There should be an option where an existing SQL query or a complex logic is converted by Alteryx intelligently into an Alteryx high level workflow with tools suggestion which can be modified by the developers. 

 

For e.g. Salesforce Einstein Analytics has an option where an existing dataflow (traditional way of performing data prep.) can be converted to a recipe (premium version of a dataflow with advanced features) using a single click. It gives an option for the user to make additional modifications/enhancements on top of it.  

Introduce CTE Functions and temp tables reading from SQL databases into Alteryx.I have faced use cases where I need to bring in table from multiple source tables based on certain delta condition. However, since the SQL queries turn to be complex in nature; I want to leverage an option to wrap it in a CTE function and then use the CTE function as an input for In-DB processing for Alteryx workflows.  

Top Liked Authors