Alteryx Designer Ideas

Share your Designer product ideas - we're listening!
Submitting an Idea?

Be sure to review our Idea Submission Guidelines for more information!

Submission Guidelines
Don't forget to submit your entry for the Excellence Awards by October 30! | Need more information about the program? Check out the blog here

Featured Ideas

Improve Help Documentation or in-tool options for handling null values in statistical tools like Weighted Average or Linear Regression. For instance, checkbox to remove null value records, or at least warn users.

 

In the processing of learning to perform linear regression in RStudio and Alteryx, I came across differing outputs depending on how null values were addressed. Take the Weighted Average tool for example.

 

In R, the weighted.mean function treats null values in the variable of interest as if they were not there. If the user does not specify that null values exist, the result is NA. If any null values exist in the weight field, the result is NA.

 

Since I am more familiar with Alteryx, I originally did the data preparation—including calculating the weighted means—in Alteryx. When comparing these weighted means with those generated in R, I found that Alteryx treats the null values as zeros (i.e. includes them in the calculation). The user would have to know this is incorrect and first filter out the null values. See screenshot examples.

 

 

 

This is also the case within the Linear Regression tool. If null values are not omitted prior to regression, the results are wildly different. Perhaps this is known by more experienced users/statisticians, but this incorrect usage would have gone on unbeknownst to be had I not cross-checked with RStudio.

 

Weighted Average in AlteryxWeighted Average in AlteryxWeighted Mean in RWeighted Mean in R

Hi,

 

Recently I was helping a client to design the workflow to do transformation. In the middle of the work, I feel a bit lost on handling so many fields and thinking it would be great if there is a feature that allow me to track the field actions along the workflow. It could be something like a configuration on the canvas, user activate it only when they want to.

And when it is activated, the workflow could become:

23.PNG

So it is easier to find the path of certain field along the whole workflow.

 

Or is there any method to achieve this at the moment?

 

Thanks.

 

Kenneth

Can you add some additional options to the running sum tool?  I know this can be done in the Multi-Row formula, but it just takes too long to program and Alteryx is designed for speed and I know it can be done.  Can you add options to average, standard deviation, etc. also with running sum?  And then add a rolling time window on it.  That would be great!

Thanks.
  • Category Transform

Hi All,

 

It would be a given wherein IT would have invested effort and time building workflows and other components using some of the tools which became deprecated with the latest versions.

 

It is good to have the deprecated versions still available to make the code backward compatible, but at the same time there should be some option where in a deprecated tool can be promoted to the new tool available without impacting the code.

 

Following are the benefits of this approach -

 

1) IT team can leverage the benefits of the new tool over existing and deprecated tools. For e.g. in my case I am using Salesforce connectors extensibly, I believe in contrast to the existing ones the new ones are using Bulk API and hence are relatively much faster.


2) It will save IT from reconfiguring/recoding the existing code and would save them considerable time.

 

3) As the tool keeps forward moving in its journey, it might help and make more sense to actually remove some of the deprecated tool versions (i.e. I believe it would not be the plan to have say 5 working set of Salesforce Input connectors - including deprecated ones). With this approach in place I think IT would be comfortable with removal of deprecated connectors, as they would have the promote option without impacting exsiting code - so it would ideally take minimal change time.

 

In addition, if it is felt that with new tools some configurations has changed (should ideally be minor), those can be published and as part of
promotions IT can be given the option to configure it.

 

Thanks,

Rohit Bajaj

In SQL, obtaining partial sums in a grouped aggregation is as simple as adding "WITH ROLLUP" to the GROUP BY clause.

 

Could we get a "WITH ROLLUP" checkbox in the Summary Tool's confg panel in order to produce partial sums?

 

(This was initiated as a question here.)

 

Here is an example of without rollup vs. with rollup, in SSMS:

Capture.PNG

 

(Replacing "NULL" with the phrase "(any)" or some such, and we have a very useful set of partial sums.

 

  • Category Transform


It would help if there is some option provided wherein one can test the outcome of a formula during build itself rather than creating dummy workflows with dummy data to test same.

 

For instance, there can be a dynamic window, which generates input fields based on those selected as part of actual 'Formula', one can provide test values over there and click some 'Test' kind of button to check the output within the tool itself.

 

This would also be very handy when writing big/complex formulas involving regular expression, so that a user can test her formula without having to
switch screens to third party on the fly testing tools, or running of entire original workflow, or creating test workflows.

Currently if I receive a file where fields are empty rather than Null(), the summarize won't count them with the CountNull option. Fairly easy to put a formula right before when there is just a couple of fields and change with an IF IsEmpty() but with a file with a large number of fields and large file size, a multiple field tool eats up a lot of time converting.  Not sure if it would just be trading processing speed in the summarize to do basically the same thing.
  • Category Transform
Hi ,
Today i stuck in one position where my current module gaves an error because it doesnt found the Fields name. I define the field name in Formula tool for validation and harmonization. So as my Fields changes formula is also changes. But i donot want to make any changes in my Module.

So what i am thinking it will be better that we can define a formula in any file format like (.xlsx or .csv) and take the Input in formula tool.
So we do not have to change the module again and again. We just need to change the mapping file against the lates file coming. So we can check the file and define the formula in mapping file.

Thanks in Advance.
  • Category Transform
0 Likes

When CrossTab is used, string data in fields is converted to field names.  If the data in the data field has a hyphen in it, this is automatically converted to an underscore when it becomes a field name. 

 

Hyphens are legal in field names, so can we make CrossTab tolerate the string as is without changing it?  If that is a breaking change, could a checkbox be added that allows users to get CrossTab to try to use the text as is and exception if the string is illegal as a field name?

 

Hyphens are required in the field name when using the Download tool, as some header names like "Content-Type" have hyphens in them.

0 Likes

The Summarize tool should have an option to ignore warnings like this:

Group Bys on Double or Float are not recommended due to rounding error.

  • Category Transform
0 Likes

Hello Alteryx fans and Devs!

 

For the Summarise tool - would it be useful to have a "Rank" option for the potential Summarise options we can use? Just a thought!

  • Category Transform
0 Likes

It would be great if Alteryx could introduce a tool implementing a Decision Model Notation (DMN)-style Decision Table as an option to remove business logic from a workflow.

 

Arbitrary business rules are frequently implemented against datasets in Alteryx workflows before actual processing occurs. The implementation of complex business logic frequently results in a spider web of join, filter, formula, and union tools. 

0 Likes

amcghee6_0-1580399446000.png

For example, allowing the letters in front of the column titles to be formatted differently to make them stand out on a report. 

0 Likes

Hi 

1:
I'm from Denmark, and like several other european countries we use commas instead of dot as decimal seperator. And we use dot as thousand seperator.

So if im working in a flow with loads of price fields, lets say cost price, amount per unit, amount and amount including vat i need to do a multi field replace. Else I dont get the output i can work with in excel or other programs.

So it would be great beeing able to set seperators on a flow level, like you can in excel when importing.

2.
Beeing able to set a date format on a flow level.
Lets say my input data is 12.12.2019 and i need 2019-12-12 in my output. If i work with several different date fields i need to use several datetime fields.
Alternate could be a multi field datetime ? 

 

3.
Having a search function when using the select ? And maybe a numbers order.

So if i scroll down, i could enter 3 - which means this would now be my 3. shown field? 

0 Likes

Would be nice if in Designer customer's may want to upload and reference  a " DATA DICITIONARY - METADATA REPOSITORY file when working with various input source to transform data . 

Organizations that are mature in their data governance strategy implement special software that extracts, manages and provides access to data dictionary of data assets in multiple databases such as  ERWIN to maintain schema for enterprise. 

Within DESIGNER access to a file METADATA REPOSITORY held in DESIGNER customer may easily select a list of columns  fields or attributes from that file to manipulate data elements using DESIGNER and provide all the relevant information required they wish to massage the data.

Possible Attributes that may be in data dictionary file:

Table name

Column Name

Data Type
Foreign Table

Source

Table Description

Sensitive Data

Required Field

Values

  • Category Transform
0 Likes

Hello Community,

 

 

I was wondering if there is a tool that could de-duplicate records after serializing (or after using Transpose Tool) with a given priority for each field in one of the keys? i.e.

 

ID Origin Field Name Value
1 A NAME JACK
1 B NAME PETER
1 B ZIP CODE 15024
1 C ZIP CODE 15024
1 D TYPE MID
1 H TYPE

PKL

 

Assuming for the field name NAME, the priority should be [ A, B ]

ZIP CODE -> [ C, B ]

TYPE -> [ H, D ]

The expected outcome for Id 1 should be -> JACK, 15024, PKL

Record discarded -> PETER, 15024, MID

In this case I'm using ID and Origin as keys in the Transpose Tool.

 

I just want to make sure there is no other route than the Python Tool.

 

Thank you

 

Luis

  • Category Transform
0 Likes

recently loaded the new V11 and gettting used to it.  one immediate gripe is the new version of the Formula Tool no longer supports multiple field actions.  In the prior version I could change Data Types on many fields at once.  I could move multiple fields in a block at once.  there were a few other things but these are things I am sorely missing on my first use of V11.  I created about 20 fields in quick succession just getting names down and then going back and putting in formula which were variations on a theme.  When done I noticed the default DataType was V_WString and I wanted integer.  In the past it was no big deal because I could select the block or interspersed fields and then right click to change data type for all to the same data type.  it was very handy and now appears to be gone.  please bring these things back.

  • Category Transform
0 Likes

Hi All,

 

I believe the following would help improve the functionality of Select Tool.

 

The idea is to have a defaulting option for each of the field in the Select Tool (which I believe should be a light weight Tool i.e. not adversely impacting performance and gives best exhaustive picture of all columns flowing through a (/particular point in ) pipeline).

 

Following are some of the cases where defaulting might come handy -

 

1) Fields which are supposed to hold monetary data - instead of Null, one can put 0.00 to help roll up summary properly.
2) Fields which are supposed to hold dates (say expiry date) - instead of Null, one can put some enterprise standards like 31-12-2099 to avoid mixing Nulls and 31-12-2099.
3) Fields which are supposed to hold purchase quantity/number of employees/number of merchandise - instead of Null, one can put 0 (and not 0.00) again to help with roll up summary.
4) Fields which are supposed to hold Currency - instead of Null, one can put USD.
5) Fields which are supposed to hold dates (say this time create date) - instead of Null, one can hardcode actual date, or an additional feature to put Now() kind of functions.

 

At present one of the options of achieving same might be to put a Formula Tool and to code whatever is desired inside the Formula Tool.

 

Benefits of having the functionality inside Select Tool would be -

 

1) It would be more user friendly and call for faster build to just write '0.00' or 'USD' or '31-12-2099' as compared to writing IF IsNull()... statements.
2) Inside Formula Tool, user needs to pull desired fields from the drop down and hence exhaustive view of all fields passing through pipeline is not available.


Pain in selection of fields from drop down and writing actual formulas might be aggravated with the number of columns increasing and might be more prone to human omission related errors.

 

Thanks,
Rohit Bajaj

0 Likes

Hey guys!!

 

I was just thinking... they might not need to fully build out a python ide, but could still reach the same objective.

 

You should be able to keep a python file on its own and call it in r.  By doing this, you might be able to have the json/xml handling of python with the visual/stats power of R while it being nicely bundled in your workflow.  This uses base functions in r and does a good job turning a pandas dataset to an r dataframe you can move along your workflow.

 

You could always just use this same idea to write a file somewhere and once it's written, your workflow will continue.  If you do, the code is literally 1 line in r...  Anyway, let me know your thoughts! 🙂

 

Will this work for your organization?

 

https://www.linkedin.com/pulse/using-python-r-windows-7-subhash-jaini?trk=hp-feed-article-title-publ...

0 Likes

The community could benefit from easier integration of splitting and applying functions to grouped data. The summarize tool is great for splitting your data and applying summary statistical functions. It would be super useful to take that block just one step further, and allow users to apply any other (aggregate) function to their grouped data instead of just the built-in functions in the summarize tool. I would envision that aggregate function either being a custom function that is a combination of existing user-specified functions within Alteryx (e.g. in the formula tool) and/or even an interface that allows you to use other Alteryx macros on the grouped data.

 


Apply user-defined functions, or other powerful Alteryx macros to grouped and data is a very common operation in the data analyst's daily workflows and being able to apply them without reverting to batch/iterative macros in a seamless manner would be naturally helpful.

 

https://ramnathv.github.io/pycon2014-r/explore/sac.html

Top Liked Authors