The Product Idea boards have gotten an update to better integrate them within our Product team's idea cycle! However this update does have a few unique behaviors, if you have any questions about them check out our FAQ.

Alteryx Designer Desktop Ideas

Share your Designer Desktop product ideas - we're listening!
Submitting an Idea?

Be sure to review our Idea Submission Guidelines for more information!

Submission Guidelines

Featured Ideas

Even with browse everywhere I see plenty of workflows dotted with browses.  Old habits die hard.  Maybe a limit configuration for browses would be a good thing.  If you've recently browsed a very large set of data you might agree.

 

Cheers,

Mark

There is currently no way to export interactive output from the network graph tool. I would like to be able to export a png of the static network graph image, a pdf of the report, and a complete html of the whole (which means including the JSON and vis.js files necessary for creating the report).

 

 

Would be great if i could just select x number of tools and just say add to a new container vs creating container and then dragging into said container.

The Interactive Chart configure window cannot be resized.  I'm using a single, large monitor, and this window occupies the entire monitor so that I am forced to switch between windows to see other applications.  Most of the space used by the configure window is wasted white space.  Please update this to allow this window to be resized.

 

I'm using a 32:9 monitor running at 5120x1440, and the attached screenshot shows the size of the configure window on this monitor - 4986x1286.

Hi

 

The native email tool is great up to a certain volume, but at high volume it makes more sense to use sendgrid to take advantage of its features like deliverability etc

 

Would be great if there was a ready-made connector tool for SendGrid

 

Thanks

Transfer of records from Python SDK RecordRef seems to be slow sending large amounts of data to the Alteryx Engine (e.g. discussion here). Although unclear of the exact specifics, it seems that there's a copy and convert process in play.

 

Apache Arrow appears to be addressing this issue, and the roadmap and specs are impressive! It seems like (again I have no understanding of the Alteryx Engine specifics) that something like this would be excellent for expanding SDK use cases as well as for other connectors such as the Apache Spark connector.

 

And it looks like it'd be fun to build into Alteryx! 🙂

When we rename a field in canvas it breaks a lot of formulas or config on tools like select tool or unique tool or summarize tool. If designer can automatically update these new field names in all the further tools if will save us a lot of time.

 

Thanks,

Sanju

Dear Alteryx Team,

Dynamic Input Tool is a great tool to import easily multiple files using files paths parameters ... having the same tool for outputs would be great to export many files in pre-established folder.

Many thanks

Arno

Hi all,

 

Based on the thread here https://community.alteryx.com/t5/Data-Preparation-Blending/Support-for-unsigned-int-database-type/m-... - there would be value in natively supporting the Unsigned Int data type in Alteryx.   

This idea was raised by @jgreene.

 

Note: this does appear to be directly supported in the core OCBC library (as long as it's supported by the ODBC driver for the specific database in question), so hopefully this won't be a huge lift:

https://docs.microsoft.com/en-us/sql/odbc/reference/appendixes/sql-data-types

 

@jgreene- would you mind adding any further information about the DBMS you are using which supports unsigned int, and any other info that may help the team to develop and test this (e.g. any link you can find to an available ODBC driver for this database etc?)

 

Thank you

Sean

Hi there,

 

Similar to @aselameab1 - I was having trouble with using the Linear regression tool because it was giving error messages that were not explanatory or self descriptive.

@chadanaber identified the issue - that a specific field only had one unique value which was causing the regression tool to fail - however the error message provided gives no useful or helpful indication that this is the issue.   You can see that the error message below is pretty tough to understand.

 

Could we add an item to the development backlog to add defensive checks to the predictive analytics tools to check for conditions that will cause them to fail, and rework the error messaging?

 

LinearRegressionError.PNGWorkflow.PNG

 

I've attached the workflow with the sample data that replicates this issue

 

Many thanks

Sean

Would be nice to make it so if I change from RowCount to index it all is linked together.

 

Something like a $FieldName$ which would automatically get substituted. 

 

Also nice in other formula tools, but most useful in places where name needs to be duplicated over and over

When I perform an average aggregation the data type changes to increase the precision.  While I could see this is useful sometimes, it is not when I have a currency based field and I want to keep the precision fixed at two decimal places.  The result is I have to either add another select tool to update the data type or find an embedded select in a tool that exists post summarizing.

It is great to see the ability to stage data for bulk loading into Databricks in s3 and ADLS.  Previously this only appeared to allow staging in Databricks DBFS. 

 

However the current connector included in Designer 2022.1 has a key gap in functionality with ADLS Gen 2 

 

The only authentication method provided to the ADLS storage is though a shared key. 

Shared keys provide access that is

  1. Too broad and allows access to the entire storage
  2. Limited or no auditability and traceability to who is using it 

We do not provide users the shared key for the ADLS storage, thus our users cannot take advantage of this new feature. 

 

The preferred method of authentication to ADLS would be

  1. RBAC - Role Based Access Control
  2. ACL - Access Control lists

Either of these options can be provided though a service principal with a tenant id, client id and client secret as inputs to the bulk load tool

 

This request would specifically be to allow the ACL authentication. ACL would help empower our our self service data analysts and data scientists who could have access to a specific container. 

 

For example

 

storageAccount/Container/directory

 

The ACL access in this tool would allow the Alteryx tool to follow the same access patterns where fine grained access was provided at the directory level and not at the storage account or container level.  This would provide self service analysts and data scientists to use Alteryx as they need within their directory without providing higher level access.  

 

Access control model for Azure Data Lake Storage Gen2 | Microsoft Docs

 

 

AStasi_0-1652966213867.png

 

Add a tool to the right-click menu that would place all currently selected tools into a container.

It would be really helpful if Alteryx server could connect directly to files on cloud file storage such as Dropbox, Box and OneDrive.  For example; a workflow could access specific source files or a folder with multiple files stored on Dropbox and could run the workflow against those files and then write the output to another folder on Dropbox.  We are making less and less use of internal file servers, so accessing files directly from the cloud allow for additional deployment scenerios and flexibility. 

The one single feature I miss the most in Alteryx, is the possibility to restart the workflow from wherever I want by using a built-in cache functionality. I have used the 'Cache Dataset V2' macro, but it really is to inflexible and really doesn't make me a happy Alteryx user. I would like to se a more flexible, quicker way of working with cached data.

 

On a single tool in the workflow I want to be able to set the option to:

 

  • Enable cache
    This would enable me to always use cached data from this node when possible

  • Run to this node
    Run from start OR from node with enabled cache to this node.

 

There should be lots of workflow options regarding the creation/deletion of cached data. Examples:

 

  • Enable data cache on all nodes
    This would enable functionality to always use cached data on all nodes in the workflow

  • Enable data cache on end nodes
    This would enable functionality to cache data on all 'Run to this node'-nodes.

...and so on. These are just a few examples, but there should be lots of options and shortcut keys revolving the cached data functionality in the workflow.

It is important to be able to test for heteroscedasticity, so a tool for this test would be much appreciated.

 

In addition, I strongly believe the ability to calculate robust standard errors should be included as an option in existing regression tools, where applicable. This is a standard feature in most statistical analysis software packages.

 

Many thanks!

Often in larger workflows, I will copy data partway down the stream into a new workflow in order to troubleshoot a small section in order to avoid having to run the workflow over and over again which can take a while. I'm aware (and thankful) of cacheing, but sometimes if there are many parallel streams or, I'd rather just copy the data from the data preview built into the tool so I don't have to take the time to run the workflow again. I'm also aware I could output a yxdb file and use that, but again that takes longer than I would like.

 

The issue I run into is if I copy the data and paste in a text input tool, all the field types change to what they would default to. This is fine with new data, but for data that has specific fields throughout the workflow, this can be a hassle. If copying data could also copy the field type and size that would be great.

It would be helpful to have the Read Uncommitted listed as a global runtime setting.

Most of the workflows I design need this set, so rather than risk forgetting to click this option on one of my inputs it would be beneficial as a global setting.

For example: the user would be able to set specific inputs according to their need and the check box on the global runtime setting would remain unchecked.

However, if the user checked the box on the global runtime setting for Read Uncommitted then all of the workflow would automatically use an uncommtted read on all of the inputs.

When the user unchecks the global runtime setting for Read Uncommitted, then only the inputs that were set up with this option will remain set up with the read uncommitted.

 

In a browse tool, sometimes I need to check any metrics haven’t been lost or omitted in different summarise tools (eg the totals still add up with different group by).

I would LOVE the quick win of selecting some cells in the browse tool results pane, and a label somewhere shows me the sum, average, count, distinct count, etc. I hate to use the line but ... “a bit like Excel does”. They even alow a user to choose what metrics they want to show in the label by right clicking. 🙂

This improvement would save me having to export to Excel, check they add up just by selecting, and jumping back, or even worse getting my calculator out and punching the numbers in.

Jay
Top Liked Authors