The Product Idea boards have gotten an update to better integrate them within our Product team's idea cycle! However this update does have a few unique behaviors, if you have any questions about them check out our FAQ.

Alteryx Designer Desktop Ideas

Share your Designer Desktop product ideas - we're listening!
Submitting an Idea?

Be sure to review our Idea Submission Guidelines for more information!

Submission Guidelines

Featured Ideas

I think the Nearest Neighbor Algorithm is one of the least used, and most powerful algorithms I know of.  It allows me to connect data points with other data points that are similar.  When something is unpredictable, or I simply don't have enough data, this allows me to compare one data point with its nearest neighbors.

 

So, last night I was at school, taking a graduate level Econ course.  We were discussing various distance algorithms for a nearest neighbor algorithm.  Our prof discussed one called the Mahalanobis distance.  It uses some fancy matrix algebra.  Essentially it allows it it to filter out the noise, and only match on distance algorithms that are truly significant.  It takes into account the correlation that may exists within variables, and reduces those variables down to only one.  

 

I use Nearest Neighbor when other things aren't working for me.  When my data sets are weak, sparse, or otherwise not predictable.  Sometimes I don't know that particular variables are correlated.  This is a powerful algorithm that could be added into the Nearest Neighbor, to allow for matches that might not otherwise be found.  And allow matches on only the variables that really matter.  

Many legacy applications that use Mainframes have certain data encoded in EBCDIC  in DB2 tables(variable length columns  that can have 0-100s of iterations in a single EBCDIC encoded compressed data). When this data is downloaded to a platform like Hadoop that does'nt understand EBCDIC data , It appears as junk characters . 

 

I solved this issue in my project  using an approach designed/implemented in PySpark script[ separate logic needed for COMP , COMP3, X(alphanumeric)  datatypes] .Having this functionality in a Tool can help many Organizations that  use data from Mainframes applications.

Add a tool to the right-click menu that would place all currently selected tools into a container.

Hi,

 

when I right-click on an Input tool, I can select "Convert To Macro Input" from the context menu. I would like the similar functionality when right-clicking a Browse tool to "Convert To Macro Output".

I always get a chuckle out of seeing PowerPoint presentations which, through numerous tables and illustrations show the "health" of "x" - where "x" is a customer, a market, business division, etc. - where the author spent hours creating graphical indicators such as arrows to render trends.

 

Recalling being provided such a deck from a senior leader who was touting how precisely the deck drew attention to areas of needed action.  The deck, in this case, was a 6 figure investment from a consulting firm. My unwelcomed response was, "Pretty pictures, nice illustrations.  What happens when business shifts tomorrow and we need to see this updated?"  There was no response.

 

In a different past life, I had the opportunity to create a similar deck, but utilized PowerPoint's ability to link to Excel, and VBA that made the tables graphs - and trending arrows all dynamic based on the data.  

 

This feature request, however, is not about me.  The dynamic slide deck solution was a windfall.  Alteryx workflows can certainly push dynamic data into .xlsm files to lighten the burden on the .xslm by eliminating the data pull.

 

BUT - not many organizations care for having production .xlsm files in use, and they're difficult to governance control.

 

Would be GREAT if one of Alteryx's output options was:  A PowerPoint

Give the developer the same ability to render "pages" as one would to populate Worksheets in a Workbook.

 

Have to believe would be a game changer for the product.

Would be nice to make it so if I change from RowCount to index it all is linked together.

 

Something like a $FieldName$ which would automatically get substituted. 

 

Also nice in other formula tools, but most useful in places where name needs to be duplicated over and over

Hi all,

 

Based on the thread here https://community.alteryx.com/t5/Data-Preparation-Blending/Support-for-unsigned-int-database-type/m-... - there would be value in natively supporting the Unsigned Int data type in Alteryx.   

This idea was raised by @jgreene.

 

Note: this does appear to be directly supported in the core OCBC library (as long as it's supported by the ODBC driver for the specific database in question), so hopefully this won't be a huge lift:

https://docs.microsoft.com/en-us/sql/odbc/reference/appendixes/sql-data-types

 

@jgreene- would you mind adding any further information about the DBMS you are using which supports unsigned int, and any other info that may help the team to develop and test this (e.g. any link you can find to an available ODBC driver for this database etc?)

 

Thank you

Sean

Hi there,

 

Similar to @aselameab1 - I was having trouble with using the Linear regression tool because it was giving error messages that were not explanatory or self descriptive.

@chadanaber identified the issue - that a specific field only had one unique value which was causing the regression tool to fail - however the error message provided gives no useful or helpful indication that this is the issue.   You can see that the error message below is pretty tough to understand.

 

Could we add an item to the development backlog to add defensive checks to the predictive analytics tools to check for conditions that will cause them to fail, and rework the error messaging?

 

LinearRegressionError.PNGWorkflow.PNG

 

I've attached the workflow with the sample data that replicates this issue

 

Many thanks

Sean

My employer just started with Alteryx, so I'm sure after a few months, this will be basic simple, but I'm sure I'm not the only person who has made this mistake, so there you go...

 

I was running a larger workflow that ends with outputting a table with about 80 columns.  Everything was working fine, except for one column, which kept coming up as all 0s.  Took me a while to find it, but the error all traced back to an error I made with a "Contains" function in a Formula node.

 

The Alteryx "Contains" function takes two arguments, 'string' and 'target'.  I interpreted this as 'find this string in this target variable', and entered two arguments: the string of characters I was looking for first, and then the name of the column I wanted to search in second. 

 

I think I had this backwards.  You need to enter the column name first, then the substring you want to look for second.  So the column name is the 'string' and the search term is the 'target'.  Me getting these reversed did not produce any errors or warnings, which I understand (e.g. maybe you might want to pairwise search for the text in one column in the text of another).

 

I think the terms 'string' and 'target' are kind of vague labels in this case (in Excel, for example, the analogous "Search" function takes the arguments 'find_text' and 'within_text', which is a lot harder to misinterpret).  So I would suggest changing the argument names in Designer and the online documentation, or at least adding text to the Alteryx Documentation (https://help.alteryx.com/2019.2/Reference/Functions.htm) that clarifies which is which.

 

Normally, I'd say this is a pretty minor thing, but since this is a potential 'silent error', I'd argue it's worthwhile.

 

(Also, the online documentation seems to imply there is a "CaseInsensitive" argument you can pass to "Contains", however I don't think there actually is.  May want to edit that for clarity.)

 

Thanks!

 

1) Contains function.PNG

 

 

I suggest an additional tool that would allow adding columns to the data, if and only if they do not exist already.

 

Currently working with data that has a dynamic set of columns can be a bit tiresome as the Select tool will not allow to select columns that have not been witnessed in the data.

 

Adding a tool that would ensure that certain columns are available downstream can currently be achieved by:

 

  1. 'Append Fields' tool with a 'Text Input' tool which will always append the fields, renaming them on the fly if needed
  2. 'Union' tool with a 'Text Input' tool

Both options do not seem straight forward and I expect have a performance impact.

 

A separate tool to achieve this seems the more user friendly and performance oriented way.

The one single feature I miss the most in Alteryx, is the possibility to restart the workflow from wherever I want by using a built-in cache functionality. I have used the 'Cache Dataset V2' macro, but it really is to inflexible and really doesn't make me a happy Alteryx user. I would like to se a more flexible, quicker way of working with cached data.

 

On a single tool in the workflow I want to be able to set the option to:

 

  • Enable cache
    This would enable me to always use cached data from this node when possible

  • Run to this node
    Run from start OR from node with enabled cache to this node.

 

There should be lots of workflow options regarding the creation/deletion of cached data. Examples:

 

  • Enable data cache on all nodes
    This would enable functionality to always use cached data on all nodes in the workflow

  • Enable data cache on end nodes
    This would enable functionality to cache data on all 'Run to this node'-nodes.

...and so on. These are just a few examples, but there should be lots of options and shortcut keys revolving the cached data functionality in the workflow.

It would be helpful to have the Read Uncommitted listed as a global runtime setting.

Most of the workflows I design need this set, so rather than risk forgetting to click this option on one of my inputs it would be beneficial as a global setting.

For example: the user would be able to set specific inputs according to their need and the check box on the global runtime setting would remain unchecked.

However, if the user checked the box on the global runtime setting for Read Uncommitted then all of the workflow would automatically use an uncommtted read on all of the inputs.

When the user unchecks the global runtime setting for Read Uncommitted, then only the inputs that were set up with this option will remain set up with the read uncommitted.

 

It would be good to be able to fix pie chart colours (either automatically of manually), so that when building a report, categories are given the same colour throughout the report across multiple pie charts.  Currently, if the number of categories being shown varies, there is no way to manually align the colours (In the Bar Chart type there is a Style Mode for the Series where you can use a formula to assign a colour based on a criteria such as the name, but not for pie charts).

 

At present, users can create a new field or update an existing one within the formula tool.

 

When making changes to an additional column, I often have to then add a select tool to rename this (or remove the original column if I make another with the correct name). Therefore, it'd be great if there was an option to rename the output of the formula as part of the configuration. Perhaps a tick box to 'Rename column upon output' along with a text box, where the Data Type selection is.

 

This is a fairly minor QoL thing but it could definitely trim down the number of tools used in some cases. In terms of referencing the field itself, maybe the formula tool could use [This field] or something descriptive to dynamically reference it, rather than the actual name which will be edited.

 

As an aside, I'm not sure if it's a technical limitation, but it'd also be brilliant if the field size could be changed for existing columns (within the limitations of the data type), rather than being static.

It is important to be able to test for heteroscedasticity, so a tool for this test would be much appreciated.

 

In addition, I strongly believe the ability to calculate robust standard errors should be included as an option in existing regression tools, where applicable. This is a standard feature in most statistical analysis software packages.

 

Many thanks!

Currently, in order to change the header and data justification to CENTER, one has to select the "Column Rules" button for each column configuration. In a large report (25+ columns), that means selecting each time. It would be more efficient to have the header justification in the Default Table Settings style editor. There is already a setting for font, font size, bold or itallic, text color and background color. I have never created a report for someone where they did not want the headers centered. The workaround is to only feed one column through, then to change the column rules for that column and the Cynamic or Unknown Fields. This works fine, but when I have a report that creates an Excel workbook with 5+ tabs, it gets annoying. It's even more time consuming when I have a report tab, that I need to create sub headings for, so there are multiple report tools for one Excel tab.

Transfer of records from Python SDK RecordRef seems to be slow sending large amounts of data to the Alteryx Engine (e.g. discussion here). Although unclear of the exact specifics, it seems that there's a copy and convert process in play.

 

Apache Arrow appears to be addressing this issue, and the roadmap and specs are impressive! It seems like (again I have no understanding of the Alteryx Engine specifics) that something like this would be excellent for expanding SDK use cases as well as for other connectors such as the Apache Spark connector.

 

And it looks like it'd be fun to build into Alteryx! 🙂

I would like to be able to pause a module A, run module B then restart A at a later time.   

I have an app that contains 4 check boxes.  Each check box is independent, and when checked, two other prompts open up to the user (in this case, text boxes which are also independent of the other check boxes).

I would like to be able to "Check All", so that with one click, ALL four check boxes are checked, and their prompts all open up.

It would be really helpful if Alteryx server could connect directly to files on cloud file storage such as Dropbox, Box and OneDrive.  For example; a workflow could access specific source files or a folder with multiple files stored on Dropbox and could run the workflow against those files and then write the output to another folder on Dropbox.  We are making less and less use of internal file servers, so accessing files directly from the cloud allow for additional deployment scenerios and flexibility. 

Top Liked Authors