Advent of Code is back! Unwrap daily challenges to sharpen your Alteryx skills and earn badges along the way! Learn more now.
Free Trial

Alteryx Designer Desktop Ideas

Share your Designer Desktop product ideas - we're listening!
Submitting an Idea?

Be sure to review our Idea Submission Guidelines for more information!

Submission Guidelines

Featured Ideas

Problem: In certain workflows, it becomes necessary to arrange columns in a specific order for the output. While achieving the desired order for a fixed number of columns is feasible using the select tool, difficulties arise when dealing with dynamic outputs that introduce new columns during each workflow run.

 

Example: Consider the following scenario: the INPUT data for the select tool includes a set of Question/Answer columns. However, with every run of the workflow, new columns of this type are introduced. The challenge is to ensure that Question N and Answer N columns are grouped together in the OUTPUT dynamically. Unfortunately, this task is not easily accomplished using the current capabilities of Alteryx.

 

INPUT:

CompanyQuestion 1Question 2Question 3Answer 1Answer 2Answer 3
ContosoBlahBlehBlyNYN

 

DESIRED OUTPUT:

CompanyQuestion 1Answer 1Question 2Answer 2Question 3Answer 3
ContosoBlahNBlehYBlyN

 

 

With Python/Pandas, this problem can be easily resolved by assigning index values to each column and then sorting the columns based on the assigned index:

reorder_py.png

 

 

So, based on the Python solution, if Alteryx could do the same, it would be great. I personally think that if the Dynamic Rename tool could held the Index Info, and the select tool could also held the Sort option, this would work.

Workflow.png

 

Dynamic Rename: Already can hold Description info, could hold Index Info.

dynamic_rename.png

 

Select tool: Could sort by index and hold this info when the workflow is saved.

Index.png

 

Hope this all make sense.

 

Thanks.

Hi is it possible to add sheet names (to spreedsheet files) to the output of a file directory tool

Hey all,

 

At present, if you have an existing canvas and you want to move to a DCM Connection - you are asked something like "this will reset all of your connection details - are you sure".    If you have complex queries; or pre+post SQL - then you first have to copy all of this out into Notepad before you can convert to DCM and then reconfigure it all again.

 

However, if you are not using DCM you can change data sources when you go into Workflow Dependancies without losing your queries etc.

 

 

Capture.PNG

 

 

Could we revisit the user experience of changing to or from a DCM connection to eliminate this "start from scratch" phenomenon - if you are converging from an existing SQL ODBC or ODB or SSVB connection to a SQL connection via DCM then it should allow you to make this conversion without losing your current configuration; and the same for any other database type.

 

cc: @mbarone 

Please consider implementing a consistent case-sensitive option for all tools and functions.

 

To compare string values, including case-sensitivity:   This post had a good description of the challenge, but the post has been archived:

   https://community.alteryx.com/t5/Alteryx-Designer-Discussions/IF-equation-ignoring-case-BUG/m-p/4170...

   For all the time I've used Alteryx, I thought that IF "test" = "TEST" would evaluate to false. Today I realised that isn't the case and I was surprised. I'm very surprised that "equals" performs like it does.

 

A few existing Ideas request case-sensitivity for individual tools:

   Case insensitive option while joining two data sets
   https://community.alteryx.com/t5/Alteryx-Designer-Desktop-Ideas/Case-insensitive-option-while-joinin...

 

   Unique tool enhancement - deal with case sensitive data
   https://community.alteryx.com/t5/Alteryx-Designer-Desktop-Ideas/Unique-tool-enhancement-deal-with-ca...

 

 

This new Idea requests system-wide consideration for case-sensitivity, for all tools and functions.

 

Current state:

 

These tools and functions are case-sensitive:

  • Tool: Join
  • Tool: Tile
  • Function: FindString
  • Functions: MD5_ASCII, MD5_UNICODE, MD5_UTF8

These tools and functions are NOT case-sensitive:

  • Tool: Unique
  • Function: CompareDictionary

These tools and functions can be either case-sensitive or NOT case-sensitive, depending on the options used:

  • Function: Contains
  • Function: EndsWith
  • Function: StartsWith
  • Functions: REGEX_Match, REGEX_Replace, REGEX_CountMatches

Current Challenges:

   How do we easily identify Lower Case, Upper Case, Mixed Case?

   How do we easily compare strings for equality, using case sensitivity?

 

Request:

   Ensure all tools and functions include an option to ignore or consider Case

   Create new functions for IsUpperCase, IsLowerCase, IsMixedCase

   Create a new function for IsEqual, with an option to ignore or consider Case

 

See attached workflow, which

  • uses REGEX_Match to create 3 new fields: IsUpperCase, IsLowerCase, IsMixedCase
  • creates a field [Flag: Original value IsEqual, case-sensitive], to compare strings for equality, using case sensitivity

 

  1. TEXT TO COLUMN TOOL : Check Mark  for “Output/No-Output” next  to  “OUTPUT ROOT NAME”  

 

Most of the time I don't want/need the column that I parsed. Provide a check box for if you want the root column output. 

Hi!

 

Just thought up a simple improvement to the US Geocoder macro that could potentially speed up the results. I'm doing an analysis on some technician data where they visit the same locations over & over again. I'm doing a full year analysis (200k + records) & the geocoder takes a bit to churn thru that much data. In the case of my data though, it's the same addresses over & over again & the geocoder will go thru each one individually.

 

What I did in my process & could be added to the macro is to put a unique tool into the process based off address, city, state, zip, then Geocode the reduced list, then simply join back to the original data stream using a join based off the address, city, state, zip fields (or use record id tool to created a unique process id to join off).

 

In my case, the 200k records were reduced to 25k, which Alteryx completed in under a minute, then joined back so my output was still the 200k records (all geocoded now).

 

Not everyone will have this many duplicates, but I'd bet most data has a few, & every little bit of time savings helps when management is waiting on the results haha!

Please update the Render tool to allow users to name the Excel sheet for the output. Alteryx currently errors when using same naming convention that works in normal Output tool.

When a user wants to use the find nearest to say find the nearest within 200 miles the dropdown stops at 100.

Similar if they want a number in between IE 15 the interface is not intuitive.

While you can just type the number in the interface doesn't look like you are able to.

 

Simply adding a "Custom" selection at the bottom would make this much more intuitive.

Hi,

Would be helpful to have an Input and Output Tool for ProjectOnline like the SharePoint and OneDrive Tools.

This way we can read the projects in a tabular form and automate our project management tasks.

Thank you.

 

 

 

When you start using DCM - you may have existing canvasses which use regular old connection strings which you want to migrate to DCM.

 

Currently (in 2023.1.1.123) - when you select "Use Data Connection Manager" - it shreds the configuration of your input tool which makes it difficult to just convert these from an existing connection to a DCM connection

SeanAdams_0-1685363363519.png

 

 

The only way to then make sure that you don't lose any configuration on the tool then is to use the XML editing functionality of the tools and copy across your old configuration.

 

Could you please add the capability to keep my current tool configuration, but just change from using a regular old connection string to using DCM?

 

Many thanks 

Sean

 

cc: @wesley-siu @_PavelP 

 

I have developed many workflows, macros, and apps, and I have always had to find a workaround for displaying information on the user config page or user interface.

 

For example, I want to input 'Default text' into the Text Box interface tool, but the problem is that it does not accept any external connection.

It would be great if this tool had a Q input anchor that could accept data from a connected tool (in both single or multi-line mode) or from external input (such as a file for DropDown list or List Box tools).

 

TextBox.PNG

TextBox_with Default_text.PNG

 

As an international organization we deal with clients in multiple-countries.

 

Name matches for names including Chinese characters generate a unicode conversation warning and are excluded from the fuzzy match.

 

It would be good if fuzzy match could be enhanced to handle Chinese characters.

After using the PCA can there be a model object to output to be able to "score" new data?

 

Similar to PCA transform here https://stackoverflow.com/questions/26182329/how-do-i-convert-new-data-into-the-pca-components-of-my...

 

As currently there is no way to use this model with new data

It would be great if you could include a new Parse tool to process Data Sets description (Meta data) formatted using the DCAT (W3C) standard in the next version of Alteryx.

DCAT is a standard for the description of data sets. It provides a comprehensive set of metadata that can be used to describe the content, structure, and lineage of a data set.

We believe that supporting DCAT in Alteryx would be a valuable addition to the product. It would allow us to:

  • Improve the interoperability of our data sets with other systems (M2M)
  • Make it easier to share and reuse our data sets
  • Provide a more consistent way to describe our data sets
  • Bring down the costs of describing and developing interfaces with other Government Entities
  • Work on some parts of making our data Findable – Accessible – Interopable - Reusable (FAIR)

We understand that implementing support for this standards requires some development effort (eventually done in stages, building from a minimal viable support to a full-blown support). However, we believe that the benefits to the Alteryx Community worldwide and Alteryx as a top-quality data preparation tool outweigh the cost.

 

I also expect the effort to be manageable (perhaps a macro will do as a start) when you see the standard RDF syntax being used, which is similar to JSON.

 

DCAT, which stands for Data Catalog Vocabulary, is a W3C Recommendation for describing data catalogs in RDF. It provides a set of classes and properties for describing datasets, their distributions, and their relationships to other datasets and data catalogs. This allows data catalogs to be discovered and searched more easily, and it also makes it possible to integrate data catalogs with other Semantic Web applications. 

DCAT is designed to be flexible and extensible, so they can be used to describe a wide variety. They are both also designed to be interoperable, so they can be used together to create rich and interconnected descriptions of data and knowledge.

 

Here are some of the benefits of using DCAT:

  • Improved discoverability: DCAT makes it easier to discover and use KOS, as they provide a standard way of describing their attributes.
  • Increased interoperability: DCAT allows KOS to be integrated with other Semantic Web applications, making it possible to create more powerful and interoperable applications.
  • Enhanced semantic richness: DCAT provides a way to add semantic richness to KOS , making it possible to describe them in a more detailed and nuanced way.

Here are some examples of how DCAT is being used:

  • The DataCite metadata standard uses DCAT to describe data catalogs.
  • The European Data Portal uses DCAT to discover and search for data sets.
  • The Dutch Government made it a mandatory standard for all Dutch Government Agencies.

As the Semantic Web continues to grow, DCAT is likely to become even more widely used.

 

DCAT

 

RDF

 

 

When creating a connection using DCM (example being ODBC for SQL) - the process requires an ODBC Data Source Name (see screenshot 1 below).

However, when you use the alias manager (another way to make database connections) - this does allow for DSN-free connections which are essential for large enterprises (see screenshot 2 below).    

 

NOTE: the connection manager screens do have another option - Quick Connect - which seems to allow for DSN-free connections, but this is non-intuitive; and you're asked to type in the name of the driver yourself which seems to be an obvious failure point (especially since the list of all installed drivers can be read straight from the registry)

 

Please could we change DCM to use the same interfaces / concepts as the alias screens so that all DCM connections can easily be created without requiring an ODBC DSN; and so that DSN-free connections are the default mode of operation?

 

 

 

Screenshot 1: DCM connection:

SeanAdams_0-1685360285460.png

 

screenshot 2

SeanAdams_2-1685360473900.png

 

cc: @wesley-siu  @_PavelP @ToddTarney 

 

 

I would love a tool to be created for looking up a value in a table based on a condition. It could be called "Lookup." One input to the tool would be the lookup list, the other is the main database. Inside the tool you could enter functions that can query the lookup table and return the results either as an overwrite of an existing field in the main DB or as a new field in the main DB, similar to the options in the Multi-Row Formula tool.

 

Here is a link to my post in Community that explains the problem. The solution, in a nutshell, was to create a Join (which resulted in millions of additional rows), run the conditional formula, then filter to get rid of the millions of rows that were created by the Join so only those that met the condition remained (the original database rows).

 

Here is the text of my Community post describing my project (slightly modified for clarity):

 

Table 1:  A list of Pay Dates (the lookup table)

Table 2:  Daily timekeeper data with Week Start and Week End Date fields.

 

The goal:  To find the Pay Date in Table 1 that is greater than the Week Start Date in Table 2 and no more than 13 days after the Week End Date in Table 2.

 

[Table 2: Week Start Date] < [Table 1: Pay Date]

and [Table 2: Week End Date] < [Table 1: Pay Date]

and DateTimeDiff([Table 1: Pay Date], [Table 2: Week End Date], 'Days') <= 13

 

There are many different flows I could use this type of tool for that would save time and simplify the flow.

Thanks!

Sometimes, Control Containers produce error messages even if they are deactivated by feeding an empty table into their input connection.

 

screenshot_error_in_spite_control_container_deactivated.png

(Note that this is a made up example of something which can happen if input tables might be from different sources and have different columns so that they need separated treatment.)

 

According to the product team, this is expected behaviour since a selection does not allow zero columns selected. This might be true (which I doubt a bit), but it is at least counter-intuitive. If this behaviour cannot be avoided in total, I have a proposal which would improve the user experience without changing the entire workflow validation logic.

(The support engineer understands the point and has raised a defect.)

 

Instead of writing messages inside Control Containers directly to the log output (on screen, in logfile) and to mark the workflow as erroneous, I propose to introduce a message (message, warning, error) stack for tools inside Control Containers:

  1. When the configuration validation is executed:
    1. Messages (messages, warnings, errors) produced outside of Control Containers are output to the screen log and to the log files (as today).
    2. Messages (messages, warning, errors) produced inside of Control Containers are not yet output but stored in a message stack.
  2. At the moment when it is decided whether a Control container is activated or deactivated:
    1. If Control Container activated: Write the previously stored message stack for this Control Container to the screen and to the log output, and increase error and warning counts accordingly.
    2. If Control Container deactivated: Delete the message stack for this Control Container (w/o reporting anything to the log and w/o increasing error and warning count).

This would result in a different sequence of messages than today (because everything inside activated Control Containers would be reported later than today). Since there’s no logical order of messages anyways, this would not matter. And it would avoid the apparently illogical case that deactivated Control Containers produce errors.

Our company has a need to link a new data source in Athena.  We have been able to establish a connection using the input functionality however the connection is so slow it is unusable.  We need to have Alteryx build an In Database option for Athena to allow us to link our data lake to Alteryx.  

This is a hybrid idea related to both posts regarding dynamic tool configuration during runtime / without having to run an analytic app.

 

What I would like to propose is a new optional connection type for the interface tools that can be updated with incoming connections (having a Q letter with white background), namely Drop Down, List Box, Tree and Map tools. This could be a simple R letter in a square for example, which would be located to the left of the incoming question anchor.

 

Use Case

 

Imagine an app where there are two control containers and three interface tools (Action tools excluded from the count) outside those containers, one of them is a Text Box connected to a filter tool (via an Action tool) in the first control container with the purpose of limiting the dataset by specifying a city for example, another one is a Numeric Up Down for limiting the dataset by the average transaction amounts that are greater than the specified amount. These two interface tools are contained in a Group Box in the Interface Designer.

 

The third interface tool is a Drop Down tool which obtains the values (which will be Store Name for this example) from the results of the Select tool (in the second control container that is connected to the output anchor of the first control container) that is connected to an incoming filter tool which is modified by the previously mentioned interface tools. Output anchor of this Select tool is connected to the hypothetical R anchor on the top of the Drop Down tool, which is then connected to an outgoing filter tool that is connected to a series of tools which ends with a Browse tool that displays basic KPI information for the store specified from the Drop Down tool.

 

The main difference of the R (Refresh) anchor from the Q anchor is that it will enable the user to dynamically update the incoming values (i.e., choices for a drop down tool) without having to run the workflow. Alteryx Designer will automatically execute only the tools necessary to be able to update the values (up to a certain point of the workflow only, which may also be indicated by the boundaries of the control containers containing the target tool) for the R anchor connected applicable Interface tools specified above. This will be possible by clicking the hypothetical confirm button (same appearance with the Apply Data Manipulations button) which only appears next to the Interface tools (or the Group Boxes containing them instead) that are automatically determined by Alteryx Designer to be providing downstream data to the the tools (T anchor of the Filter tool for example) sending values to the applicable Interface tools having an incoming R anchor connection.

 

I saw that a similar feature recently became available with Alteryx Analytics Cloud Platform with the App Builder product, and I think that Alteryx Designer Desktop could definitely benefit both from this feature and additional App Builder features (that can be adapted to Desktop counterpart) in the upcoming releases.

Apologies if this has been suggested or exists. I often find myself using manual Excel files as a data source. These files frequently use cell formatting elements, such as cell color and text color, to convey important information. However, when these files are imported into Alteryx, this valuable formatting information is unfortunately lost.

 

To address this, a dedicated input tool that can read Excel files with separate fields for these formatting elements would be very helpful. This would be incredibly beneficial, especially when the data lacks other fields that relate to the coloring. Currently, I manage to achieve this using a Python tool, but integrating this as a built-in feature in Alteryx would undoubtedly be more efficient and user-friendly. This enhancement would not only simplify data preparation but also ensure the preservation of the full context of the original Excel file.

Top Liked Authors