Free Trial

Alteryx Designer Desktop Ideas

Share your Designer Desktop product ideas - we're listening!
Submitting an Idea?

Be sure to review our Idea Submission Guidelines for more information!

Submission Guidelines

Featured Ideas

0 Likes

When working on the Weekly Challenge #108, I was trying to design a non-macro solution.

 

I ended up settling on the Generate Rows tool and was trying to find a way to generate rows until I had reached or exceeded the maximum density, however, I ran into an issue where I'd always have one too few rows, since the final row I was looking for was the one that broke the condition I specified.

 

In order to get around this, I came up with the following solution:

 

Spoiler

 

Idea_IncludeConditionBreak_GenerateRowsTool.PNG

Idea_IncludeConditionBreak_GenerateRowsTool.PNG

 

Essentially, I just set my condition to twice that of the true threshold I was looking for. This worked because I was always doubling the current value in my Loop Expression, and so anything which broke the 'actual' condition I was looking for ([MaxDensity]), would necessarily also break the second condition if doubled again.

 

However, for many other loop expressions, this sort of solution would not work.

 

My idea is to include a checkbox which, when selected, would also generate the final row which broke the specified condition.

 

By adding such a checkbox, it would allow users to continue using the Generate Rows tool as they already do, but reduce the amount of condition engineering that users are required to do in order to get that one extra row they're looking for, and reduce the number of  potentially unseen errors in their workflows.

0 Likes

 

The CrossValidation tool in Alteryx requires that if a union of models is passed in, then all models to be compared must be induced on the same set of predictors. Why is that necessary -- isn't it only comparing prediction performance for the plots, but doing predictions separately? Tool runs fine when I remove that requirement. Theoretically, model performance can be compared using nested cross-validation to choose a set of predictors in a deeper level, and then to assess the model in an upper level. So I don't immediately see an argument for enforcing this requirement.

 

This is the code in question:

if (!areIdentical(mvars1, mvars2)){
        errorMsg <- paste("Models", modelNames[i] , "and", modelNames[i + 1],
                          "were created using different predictor variables.")
        stopMsg <- "Please ensure all models were created using the same predictors."
      }

As an aside, why does the CV tool still require Logistic Regression v1.0 instead of v1.1?

 

And please please please can we get the Model Comparison tool built in to Alteryx, and upgraded to accept v1.1 logistic regression and other things that don't pass `the.formula`. Essential for teaching predictive analytics using Alteryx.

 

0 Likes

1. The Union tool 

 

When switching to Manual method and then adding fields up stream, the result is a warning "Field was not found". I don't look for warnings. This should create a red error. Having fields fall off the workflow is a pain.

 

2. Unique tool

 

Changing fields upstream causes the tool to error out when the workflow runs. No issues are shown before the run. 

 

3. Having containers all open up when I reopen a workflow is a nightmare when you have 20+ containers all over lapping.

 

0 Likes

Please provide an option to have a the caption of a tool container sent to log/results after all items within it have been processed.

  • Tool Improvement
0 Likes

Can you devise a way to bring out the dynamic network visualisation on to Powerpoint.  Right now, we can only see a static image on a browser

  • Tool Improvement
0 Likes

The "Field Summary" tool and several others have a configuration requirement that provides a list of fields to select or deselect.  The selection action is singular meaning you can only effect the action on one item in the list of many.  As the number of fields we work with grows significantly this becomes a time consuming and tedious task.

 

This should be enhanced to allow highlighting of multiple fields to select or deselect as we can do in tools like the "Select" tool.

  • Tool Improvement
0 Likes

The field summary tool is an excellent resource to get an overview of the data and spot targets for analysis or data cleansing. 

 

Unfortunately it has limitations either in the number of fields included or some combination of the number of fields and one or more of its attributes.   There is nothing in the documentation I found to make a user aware of this. When you exceed fields selected the system just hangs, indicating it is running but there is no connection progress shown and nothing seems to happen, even if you limit input to 1 record.

 

Through trial and error I found an approximate limit in number of fields I can include and still have it work.

 

I request that Alteryx update the tool help info and devise enhancements to dynamically load balance the tool so it can scale to the number of fields requested or at least warn when the limit is reached or approached.  The latter warning could be similar to the red font warnings in the formula tool when you have a malformed expression.  However a load balancing version is most desired.

 

The issue as it stands results in users wasting a lot of time trying to make the tool work as expected, then report it to support as an apparent bug in the tool which can be argued both ways.

 

I realize in the real world there are limits but in this real world we are seeing the number of fields to analyze increase significantly, especially when you have a data license and integrate 3rd party data to you own native data adding a hundred or hundreds more fields.

0 Likes

Hi Alteryx,

 

I tried importing a file into my text input tool and I ran across this error.

 error.png

 

Why is there a limit? Can we get rid of it since we're not living in DOS anymore.

  • Tool Improvement
0 Likes

During execution the user cannot scroll around. Large workflows need to be shrunk to very small icons to be able to follow the progress. Either have an option to automaticaly center on the active icon or allow scrolling during execution.

  • Tool Improvement
0 Likes

It would be helpful to have one of the following methods to disable output modules to prevent overwriting output files each time a workflow is run:

- A global 'Disable All Output Modules' option, which would effectively mute the workflow without removing any connections

- A module-specific 'disable/enable this module' option, to the same end

Top Liked Authors