The Product Idea boards have gotten an update to better integrate them within our Product team's idea cycle! However this update does have a few unique behaviors, if you have any questions about them check out our FAQ.

Alteryx Designer Desktop Ideas

Share your Designer Desktop product ideas - we're listening!
Submitting an Idea?

Be sure to review our Idea Submission Guidelines for more information!

Submission Guidelines

Featured Ideas

The "Manage Data Connections" tool is fantastic to save credentials alongside the connection without having to worry when you save the workflow that you've embedded a password. 

 

Imagine if - there were a similar utility to handle credentials/environment variables. 

 

  • I could create an entry, give it a description, a username, and an encrypted password stored in my options, then refer to that for configurations/values throughout my workflows. 
    • Tableau credentials in the publish to tableau macro
    • Sharepoint Credentials in the sharepoint list connector
  • When my password changes I only have to change it in one place
  • If I handoff the workflow to another user I don't have to worry about scanning the xml to make sure I'm not passing them my password
  • When a user opens my workflow that doesn't have a corresponding entry in their credentials manager they would be prompted using my description to add it.
  • Entries could be exported and shared as well (with passwords scrubbed)

Example Entry Tableau:

Alias Tableau Prod
Description Tableau Production Server
UserID JPhillips
Password *********
+  

Then when configuring a tool you could put in something like [Tableau Prod].[Password] and it would read in the value.

 

Or maybe for Sharepoint:

 

Alias TeamSP
Description Team sharepoint location
UserID JPhillips
Password *********
URL  http://sharepoint.com/myteam
+  

 

Or perhaps for a team file location:

Alias TeamFiles
Description Root directory for team files
Path \\server.net\myteam\filesgohere
+  

 

Any of these values could be referenced in tool configurations, formulas, macro inputs by specifying the Alias and field.

1-3-2019 12-43-52 PM.png

I love this tool,  but think it would be improved by including an option to create a column per delimiting character.  This could be added in the number of columns selector box.  In the case where 1 row has more delimiters than another, null columns can be created.  Without this option you have to Regex count the delimiters, select the max and then embed the Text to columns tools in a macro and then pass the max columns as a param.  Would be nice to resolve all this in the main tool.

 

Thanks, nick

When I use the Comment Tool its difficult to select the tools inside it, but when I use the Container Tool the Container Text doesn't support Font Sizes, and doesn't support multiple lines of text so I end up moving the Comment into the Container, but still have problems selecting a group of tools.

 

So a combined Comment and Container Tools would be wonderful!

 

Bonus: If the Comment Tool could support Multiple Font Sizes.

 

Combined Tools.png

When writing a good amount of code, it is easy to get lost in a sea of parentheses.  Just when you think you're all done, you get an error that can force you to scour through your code to find the missing, extra, or misplaced parenthesis. 

 

A common feature today is to highlight a parenthesis when its partner is clicked on.  This instantly lets you know if you have the wrong number of them and where. 

 

I didn't think this was that important early on in Alteryx, at least for me.  Formulas were meant to be short and easily readable at a glance.  Now as I dig deeper, there's R, Python, SQL and other text-heavy inputs. 

 

I don't need a full-fledged text editor in Alteryx, but I would love some quality of life features like parentheses matching.

I constantly find my using pre and post SQL Commands in the Output tool to run SQL when I don't actually have any data to output.

 

One example is when I load data into S3 and want to load it into Redshift. I have SQL code to run but no data to Output - I end up running a dummy row into a temp table.

 

So can we have an SQL tool that simply acts the same as a Pre-SQL command without the associated data output. Once the command is run we should be able to continue the workflow, so the tool should have an option input and output, like the Run Command tool.

 

 

The following idea might not be as valuable as some of @SeanAdams posts, but it would save this user precious fractions of time.   When I leave the canvas with my mouse (point A) to go up to the pallet I select and drag the tool down to the canvas.  Sometimes I do right-click and go through the menus to add the next tool, but generally I go through that labor only when I'm inserting in-stream the tool.  So here is my idea:

 

Double-Click your NEXT tool and it "Alteryx-ly" appears on your canvas in proximity to the hi-lighted (last) tool.  Better yet, connect it!  Now I can move from the pallet to the configuration panel directly without having to move my mouse down to the canvas and then over to the configuration panel.

 

Hopefully, my friend @Hollingsworth will find this time-saving idea worthy of a star.  Speed demons like @NicoleJohnson and @BenMoss might not need this turbo boost, but at my age it is worth the ask.

 

Cheers,

 

Mark

I love Workflow Meta info, especially the ability to put the Author, the search tags,the version, the description, etc...

workflow meta info.png

But why can't we use it as Engine Constant? It doesn't seem very hard to implement and it would change life for development.

 

engine_constant.png

Tools within a workflow needs to be able to run in parallel whereever applicable.

 

For example: Extracting 10 million rows from one source, 12 million rows from a different source to perform blending.

currently the order of execution is the order in which tools are dragged into the canvas. Hence Source1 first, Source2 second and then the JOIN.

 

Here Source1 & Source2 are completely independent, hence can be run in parallel. Thus saving the workflow execution time.

 

Execution time is quite crucial when you have tight data loading window.

 

Hopefully alteryx considers this in the next release!

  • Feature Request

I like the new cache option in 2018.3, but I would like it to function a little bit different. Let's say you cache at a certain point and then continue to build after that. If I reach another checkpoint and want to cache, it currently re-runs the entire workflow (ie it ignores my cache upstream and just goes back to the beginning of the workflow); instead, I would rather have it utilize the upstream cache. Personally, caching is usually an iterative effort during development where I keep caching along the way. The current functionality of the cache is not conducive to this. Thanks!

  • Feature Request

Given Crew Macro Pack increases Alteryx's capability so much, and is used so pervasively, is there a reason to not include Crew Macro Pack in Alteryx Designer or Alteryx Server by default?

 

Can anyone give a reason why Alteryx wouldn't bundle Crew Macro Pack?

 

If not, can we get Crew Macro Pack bundled into Alteryx and have official support for it?

Most of the finance industry is using HFM or Essbase finance solutions which is flexible in keeping the data and generating reports. Alteryx should create a direct connection with Hyperion or Essbase so that it's easier to connect with the data in those systems and it can directly feed other system reports.

  • Feature Request

Please include IBM DB2 as an in-Database option. Currently, my primary use of Alteryx is for copying DB2 tables into Teradata for use on that server. Copying large tables and particularly joining several tables and copying the results to Teradata is too slow in Alteryx.

  • Feature Request

TIBCO Data Virtualization is a Data Virtualization product focused on creating a virtual data store consolidating data from throughout the enterprise.  It can be accessed via a SQL query engine, and has a variety of supported connectors, including an ODBC driver.

 

This data source can be connected to via ODBC in Alteryx today, but error messaging is unclear/unhelpful, and attempting to use the Visual Query Builder causes Alteryx to crash.

 

Adding TIBCO Data Virtualization as a supported ODBC connection would empower business users to leverage this product and easily utilize this enterprise data store, enhancing the value of the Alteryx platform as a consumer of this data.

As a developer I want to Cache & Run all/selected inputs in a Workflow with one command. Maybe even with a keyboard shortcut?

 

clipboard_image_1.jpeg

Some of the workflows I use have multiple inputs that can take a long time to initially load. The new cache function itself has been amazing, but there is one big drawback for me: I can't cache multiple tools at the same time.  Alteryx will allow me to eventually cache all of the tools I want cached, but it will take multiple times running the file.  This still saves me time in the end, but it feels a bit cumbersome to set up.

Dear Alteryx GUI Gang,

 

I'll create a container and then customize the colours, margins, transparency, border and then want consistency for other containers.  It would be nice to have a format painter function (brush) to apply the format of one container to another.  This of course could be extended to other tools like comments.  There might be a desire to apply this to more tools too, but the comments and containers would be my focus as they are almost always custom configured.

 

Cheers,

 

Mark

In the Gallery, the File Browse tool returns the file location on the server where the file was uploaded. This allows the file to then be read in as input to a workflow.

If you need the file path of the original file location, you have to add a Text Input for the user to manually add it.

 

In my case (#00293302), I used a chained app to populate a list box for the user to select the Sheet Names they would like to process through the application. Unfortunately, since I was not able to capture the original file location the application errored out. This is due to the second app using the file location on the server where the file was uploaded, which is provided by the first workflow. This file location (from the Browse tool) is a temporary file location, where inputs are immediately deleted after processing.

 

Want to test this out? Create an application where you Output the file path from a File Browse tool.

 

i know.....grrr, this doesn't match your original file location!

 

Thank you,

Mark

It would be extremely useful to quickly find which of my many workflows feed other workflows or reports.

 

A quick and easy way to do this would be to export the dependencies of a list of workflows in a spreadsheet format. That way users could create their own mapping by linking outputs of one workflow, to inputs of another.

 

Looking at the simple example below, the Customers workflow would feed the Market workflow.

 

WorkflowDependencyType
CustomersSQL Table 1Input
CustomersSQL Table 2Input
CustomersExcel File 1Input
CustomersExcel File 2Input
CustomersExcel File 3Output
MarketExcel File 3Input
MarketSQL Table 3Output

 

It would be CRAZY AWESOME if we could get a report like this for all scheduled workflows in the scheduler. 

@AdamR_AYX,

 

Limit conversion warning allows for a minimum of 1 message.  Can we set the minimum to 0 to completely ignore the message?

 

Perhaps we can allow warning messages a similar function as ERROR messages and allow the designer to Ignore, Warn or Cancel?

 

ConvError: Imputation (441): Tool #104: No demand: 0.200000000000031 had more precision than a double. Some precision was lost.

ConvError: Summarize (456): Data: 0.360000000004675 had more precision than a double. Some precision was lost.

 

End: Designer x64: Finished running FP Model - Marquee Crew v3.yxmd in 32.3 seconds with 16 field conversion errors and 4 warnings

 

Thanks,

 

Mark

The Source field of the field metadata is very useful, but has some problems.

 

  • It is repetitious. A long connection string repeated for many fields from the same source can bloat the size of the workflow above 10 MB, and when removed is around 0.5 MB.
  • It exposes sensitive information about a company's infrastructure, such as server names, ports, user ids, and proprietary data structures.

I first started paying attention when we found a user's password in the metadata because they had passed it as a string to the Dynamic Input Tool (separate Idea submitted for that - LINK). Then when I had to share an App with the Alteryx Support team for support with an issue, I thought to check the metadata, and I noticed that the file was too big and was exposing information that I would not normally share with another company.

 

I'm not sure how you want to handle this, but here's some thoughts:

 

  • Default the Source field to 'off' and provide users the option to turn it 'on' in the workflow/app settings.
  • Provide a mechanism to strip the 'Source' field at time of saving or exporting the workflow.
  • If nothing else, provide education to users on the implications of including this information in the file.

 

Thanks for listening!

 

Cameron

Top Liked Authors