Get Inspire insights from former attendees in our AMA discussion thread on Inspire Buzz. ACEs and other community members are on call all week to answer!
The Product Idea boards have gotten an update to better integrate them within our Product team's idea cycle! However this update does have a few unique behaviors, if you have any questions about them check out our FAQ.

Alteryx Designer Desktop Ideas

Share your Designer Desktop product ideas - we're listening!
Submitting an Idea?

Be sure to review our Idea Submission Guidelines for more information!

Submission Guidelines

Featured Ideas

Please add measures of statistical central tendency to the Summarize tool, such as:

Mid-range

Midhinge

Quartile

Thank you.

I have a process that joins 3 data sets to identify a specific group of data and apply certain ruling. From this created file, I need to extract the data (not the headings) from specific columns and insert into an already existing template. The template has formatting that needs to remain in order for it to function. 

 

Is this possible? 

Hi,

 

Recently I was helping a client to design the workflow to do transformation. In the middle of the work, I feel a bit lost on handling so many fields and thinking it would be great if there is a feature that allow me to track the field actions along the workflow. It could be something like a configuration on the canvas, user activate it only when they want to.

And when it is activated, the workflow could become:

23.PNG

So it is easier to find the path of certain field along the whole workflow.

 

Or is there any method to achieve this at the moment?

 

Thanks.

 

Kenneth

Would be nice if in Designer customer's may want to upload and reference  a " DATA DICITIONARY - METADATA REPOSITORY file when working with various input source to transform data . 

Organizations that are mature in their data governance strategy implement special software that extracts, manages and provides access to data dictionary of data assets in multiple databases such as  ERWIN to maintain schema for enterprise. 

Within DESIGNER access to a file METADATA REPOSITORY held in DESIGNER customer may easily select a list of columns  fields or attributes from that file to manipulate data elements using DESIGNER and provide all the relevant information required they wish to massage the data.

Possible Attributes that may be in data dictionary file:

Table name

Column Name

Data Type
Foreign Table

Source

Table Description

Sensitive Data

Required Field

Values

Hello Community,

 

 

I was wondering if there is a tool that could de-duplicate records after serializing (or after using Transpose Tool) with a given priority for each field in one of the keys? i.e.

 

ID Origin Field Name Value
1 A NAME JACK
1 B NAME PETER
1 B ZIP CODE 15024
1 C ZIP CODE 15024
1 D TYPE MID
1 H TYPE

PKL

 

Assuming for the field name NAME, the priority should be [ A, B ]

ZIP CODE -> [ C, B ]

TYPE -> [ H, D ]

The expected outcome for Id 1 should be -> JACK, 15024, PKL

Record discarded -> PETER, 15024, MID

In this case I'm using ID and Origin as keys in the Transpose Tool.

 

I just want to make sure there is no other route than the Python Tool.

 

Thank you

 

Luis

Was thinking with my peers at work that it might be good to have join module expanded both for desktop and in-database joins.

 

As for desktop join: left and right join shows only these records that are exclusive to that side of operation. Would it be possible to have also addition of data that is in common?

As for in-db join: db join acts like classic join (left with matching, right with matching data). Would it be possible to get as well only-left, only-right join module?

 

 

Can you add some additional options to the running sum tool?  I know this can be done in the Multi-Row formula, but it just takes too long to program and Alteryx is designed for speed and I know it can be done.  Can you add options to average, standard deviation, etc. also with running sum?  And then add a rolling time window on it.  That would be great!

Thanks.

Hi All,

 

It would be a given wherein IT would have invested effort and time building workflows and other components using some of the tools which became deprecated with the latest versions.

 

It is good to have the deprecated versions still available to make the code backward compatible, but at the same time there should be some option where in a deprecated tool can be promoted to the new tool available without impacting the code.

 

Following are the benefits of this approach -

 

1) IT team can leverage the benefits of the new tool over existing and deprecated tools. For e.g. in my case I am using Salesforce connectors extensibly, I believe in contrast to the existing ones the new ones are using Bulk API and hence are relatively much faster.


2) It will save IT from reconfiguring/recoding the existing code and would save them considerable time.

 

3) As the tool keeps forward moving in its journey, it might help and make more sense to actually remove some of the deprecated tool versions (i.e. I believe it would not be the plan to have say 5 working set of Salesforce Input connectors - including deprecated ones). With this approach in place I think IT would be comfortable with removal of deprecated connectors, as they would have the promote option without impacting exsiting code - so it would ideally take minimal change time.

 

In addition, if it is felt that with new tools some configurations has changed (should ideally be minor), those can be published and as part of
promotions IT can be given the option to configure it.

 

Thanks,

Rohit Bajaj

In SQL, obtaining partial sums in a grouped aggregation is as simple as adding "WITH ROLLUP" to the GROUP BY clause.

 

Could we get a "WITH ROLLUP" checkbox in the Summary Tool's confg panel in order to produce partial sums?

 

(This was initiated as a question here.)

 

Here is an example of without rollup vs. with rollup, in SSMS:

Capture.PNG

 

(Replacing "NULL" with the phrase "(any)" or some such, and we have a very useful set of partial sums.

 


It would help if there is some option provided wherein one can test the outcome of a formula during build itself rather than creating dummy workflows with dummy data to test same.

 

For instance, there can be a dynamic window, which generates input fields based on those selected as part of actual 'Formula', one can provide test values over there and click some 'Test' kind of button to check the output within the tool itself.

 

This would also be very handy when writing big/complex formulas involving regular expression, so that a user can test her formula without having to
switch screens to third party on the fly testing tools, or running of entire original workflow, or creating test workflows.

Hi,

 

I think that the sample tool should have a T or F port.

 

Lets say I want to keep first N records but would like to stream the rest of the data (the not sampled one) somewhere else in my workflow, its possible but it would be easier to have that in the sampler. 

 

Simon

Korem

Currently only DateTime based functions are available, Time based functions should be introduced. like TimeAdd(), TimeDiff() etc.

This will help users a lot to calculate different aspects of time based calculations...

 

Ashok Bhatt

Currently if I receive a file where fields are empty rather than Null(), the summarize won't count them with the CountNull option. Fairly easy to put a formula right before when there is just a couple of fields and change with an IF IsEmpty() but with a file with a large number of fields and large file size, a multiple field tool eats up a lot of time converting.  Not sure if it would just be trading processing speed in the summarize to do basically the same thing.
Hi ,
Today i stuck in one position where my current module gaves an error because it doesnt found the Fields name. I define the field name in Formula tool for validation and harmonization. So as my Fields changes formula is also changes. But i donot want to make any changes in my Module.

So what i am thinking it will be better that we can define a formula in any file format like (.xlsx or .csv) and take the Input in formula tool.
So we do not have to change the module again and again. We just need to change the mapping file against the lates file coming. So we can check the file and define the formula in mapping file.

Thanks in Advance.
0 Likes

We had a workflow where we needed to count business days. The standard solve of generating rows for each day between the dates wouldn't work as it would slow down the workflow too much.

 

Something that takes 5 seconds in Excel turned into a tremendous pain.

 

It would be really nice to have a built out tool where you can input the start date / end date (or what field they are tied to).

Which days of the week are considered business days and which days are not.

Which holidays should be excluded and available to add custom holidays.

0 Likes

For the summary tool, allow for the field data definition type of the output.  

0 Likes

Hi 

In all its simplicity, I would like to be able to group by Month based of dates:


Hamder83_0-1668170619078.png


To acieve something like this:

Hamder83_1-1668170691432.png

 

0 Likes

The Summarize tool returns NULL when performing a Mode operation. This doesn't seem to be documented anywhere in Alteryx documentation nor the community. Please fix this behaviour.

0 Likes

Hello

 

I have searched the community but haven't found any obvious solutions to this.

 

When using a cross tab I often find that there shouldn't be any aggregated values and if there is it means there is an issue with my data or workflow.

 

Therefore I think a useful feature would be an option for the cross tab tool to be able to return an error if it trying to aggregate any values.

 

I have a work around by using a summarize tool to count the non unique records and then a test to see if there are any duplicates but I think this could be a useful addition to the tool.

 

Thanks

0 Likes

It would be helpful to be able to toggle the way the Mode calculation handles two or more "ties." Currently if there is a tie between records, the lower is returned. I have a use case where I would rather have the higher value return if there is a tie. I could also see there being a use for an average between the tied records. Ideally I think there would be three options for a tie: use the 1. lowest value 2. highest value 3. average of tied values. I'm not sure if first/last would also be helpful to have as options.

My use case is for product dimensions. We use the mode to normalize the dimensions (height, width, depth) of products. Because we are using the dimensions for space planning, if the lower value is used there may not be enough space for the product on the shelf. We would rather use the higher of the tied values to make sure we aren't creating a plan where the products won't fit.

Top Liked Authors