community
cancel
Showing results for 
Search instead for 
Did you mean: 

Alteryx Designer Ideas

Share your Designer product ideas - we're listening!
Upgrade Alteryx Designer in 10 Steps

Debating whether or not to upgrade to the latest version of Alteryx Designer?

LEARN MORE

1 Review

Our submission guidelines & status definitions before getting started

2 Search

The community for a solution or existing idea before posting

3 Vote

By clicking the star in the top left corner of an idea you support

4 Submit

A new idea to suggest a product enhancement or new feature


Suggest an idea
0 Stars

Hi,

Is there an easy way through Alteryx to rename a file once I have processed it... Would like file name to be- FileName.csv.Date.Time (FileName.txt.20180424.055230)

 

 

Thanks.

 

 

 

 

 

 

0 Stars

 

Why do we need yxmd files? Why shouldn't the default be yxmz? The workflow logic is the same. If you don't add any interface tools it will run, and it you want to have a interface you can. 

 

If you start off with an yxmd and then decide to make it an app you now have two files to worry about.  

 

As a habit I no longer save things as yxmd. As soon as I start a new workflow I save it as an yxmz.

 

Thoughts?

 

0 Stars

The layout of tables in the Visual Query Builder should be persistent.  The placement, made by the user, should be persistent when called-up again.

0 Stars

Hello,

 

It would be nice if Alteryx had a way to create a time lapse video based on a date field.  Create a map showing data over time (day, week, year, etc) that shows it cumulative or individual for each time frame.  I needed to map a set of location points, with a date field, to show where and when they opened throughout time by each year or by each decade, so it starts with only a few locations, then keeps adding the new locations with time to show all locations at the end or present time.

 

Thanks,

Bob

0 Stars

 

It would be a timesaver for workflow documentation if macro metadata could be passed through to the workflow where the macro is deployed.  The image below suggests that the Description encased in the macro could be copied across to the corresponding Annotation panel.  I would also suggest that the Meta Info panel have a checkbox to control this new funtionality.

 

MACRONOTES.png

0 Stars

I have encountered problem with Oracle Direct Connection tool. I have the correct host, port, service name, user name and password (the same configuration works with Toad), but Alteryx still complaints the service name is not found on tnsnames.ora (ORA-12154). Well, the reason to use Direct Connection is that I do not have admin rights on my computer to edit the file, so it seems this kind of problem can only be resolved by reaching out to our IT service to edit tnsnames.ora for me.

 

Alteryx2.png

0 Stars

I would like to able to limit the data being read from the source based on the volume , such as 10GB or 5 GB etc. This will help in case of POC's where we can process portion of the dataset and not the entire dataset. This will have many different used cases as well.

0 Stars

It would be a huge time saver if you had an option to unselect the fields selected and select the fields not selected in the Select tool.

0 Stars

Improve Help Documentation or in-tool options for handling null values in statistical tools like Weighted Average or Linear Regression. For instance, checkbox to remove null value records, or at least warn users.

 

In the processing of learning to perform linear regression in RStudio and Alteryx, I came across differing outputs depending on how null values were addressed. Take the Weighted Average tool for example.

 

In R, the weighted.mean function treats null values in the variable of interest as if they were not there. If the user does not specify that null values exist, the result is NA. If any null values exist in the weight field, the result is NA.

 

Since I am more familiar with Alteryx, I originally did the data preparation—including calculating the weighted means—in Alteryx. When comparing these weighted means with those generated in R, I found that Alteryx treats the null values as zeros (i.e. includes them in the calculation). The user would have to know this is incorrect and first filter out the null values. See screenshot examples.

 

 

 

This is also the case within the Linear Regression tool. If null values are not omitted prior to regression, the results are wildly different. Perhaps this is known by more experienced users/statisticians, but this incorrect usage would have gone on unbeknownst to be had I not cross-checked with RStudio.

 

Alteryx_WeigtedAvg.PNGWeighted Average in AlteryxR_WeightedMean.PNGWeighted Mean in R

0 Stars

@KuoL 

 

Yes, I know, it's weird to have a situation where a decision tree decides that no branches should be created, but it happened, and caused great confusion, panic, and delay among my students.

 

v1.1 of the Decision Tool does a hard-stop and outputs nothing when this happens, not even the succesfully-created model object while v1.0 of the stool still creates the model ("O") and the report ("R") ... just not the "I" (interactive report). Using the v1.0 version of the tool, I traced the problem down to this call:

 

dt = renderTree(the.model, tooltipParams = tooltipParams)

Where `renderTree` is part of the `AlteryxRviz` library.

 

I dug deeper and printed a traceback.

 

9: stop("dim(X) must have a positive length")
8: apply(prob, 1, max) at <tmp>#5
7: getConfidence(frame)
6: eval(expr, envir, enclos)
5: eval(substitute(list(...)), `_data`, parent.frame())
4: transform.data.frame(vertices, predicted = attr(fit, "ylevels")[frame$yval],
       support = frame$yval2[, "nodeprob"], confidence = getConfidence(frame),
       probs = getProb(frame), counts = getCount(frame))
3: transform(vertices, predicted = attr(fit, "ylevels")[frame$yval],
       support = frame$yval2[, "nodeprob"], confidence = getConfidence(frame),
       probs = getProb(frame), counts = getCount(frame))
2: getVertices(fit, colpal)
1: renderTree(the.model)

The problem is that `getConfidence` pulls `prob` from the `frame` given to it, and in the case of a model with no branches, `prob` is a list. And dim(<a list>) return null. Ergo explosion.

 

Toy dataset that triggers the error, sample from the Titanic Kaggle competition (in which my students are competing). Predict "Survived" by "Pclass".

0 Stars

Dear Team

 

If we are having a heavy Workflow in development phase, consider that we are in the last section of development. Every time when we run the workflow it starts running from the Input Tool. Rather we can have a checkpoint tool where in the data flow will be fixed until the check point and running my work flow will start from that specific check point input.

 

This reduces my Development time a lot. Please advice on the same.

 

Thanks in advance.

 

Regards,
Gowtham Raja S

+91 9787585961 

0 Stars

The error message is:

 

Error: Cross Validation (58): Tool #4: Error in tab + laplace : non-numeric argument to binary operator

 

This is odd, because I see that there is special code that handles naive bayes models. Seems that the model$laplace parameter is _not_ null by the time it hits `update`. I'm not sure yet what line is triggering the error.

0 Stars

When output is disabled, Alteryx's output tools are helpfully grayed out and include the message 'output has been disabled by the workflow properties.'

However, if a macro has an output, there is no visual indicator that output is disabled, even though the macro's output will also be suppressed by this workflow configuration.

Obviously, macros can be very complex, and could have both a file and a macro output, or have an optional file output, so these cannot be entirely locked out just because there is an output.

To that end, I suggest some other kind of color-coding/shading be applied visually to these tools, and that a message be added to the interface for these macros that says something like "output has been disabled, this macro may not perform all of its functions".

I just spent about 10 minutes debugging why a macro wasn't working properly in one workflow but was working in another, and it was because I had disabled output, which I wasn't thinking of because this particular macro uses the Render tool to produce a hyperlink.  I wouldn't have spent more than 30 seconds on this if there was some kind of visual indicator showing me what I was doing wrong!

0 Stars

 

The CrossValidation tool in Alteryx requires that if a union of models is passed in, then all models to be compared must be induced on the same set of predictors. Why is that necessary -- isn't it only comparing prediction performance for the plots, but doing predictions separately? Tool runs fine when I remove that requirement. Theoretically, model performance can be compared using nested cross-validation to choose a set of predictors in a deeper level, and then to assess the model in an upper level. So I don't immediately see an argument for enforcing this requirement.

 

This is the code in question:

if (!areIdentical(mvars1, mvars2)){
        errorMsg <- paste("Models", modelNames[i] , "and", modelNames[i + 1],
                          "were created using different predictor variables.")
        stopMsg <- "Please ensure all models were created using the same predictors."
      }

As an aside, why does the CV tool still require Logistic Regression v1.0 instead of v1.1?

 

And please please please can we get the Model Comparison tool built in to Alteryx, and upgraded to accept v1.1 logistic regression and other things that don't pass `the.formula`. Essential for teaching predictive analytics using Alteryx.

 

0 Stars

Hi,

 

i recently started working on Alteryx and would like to ask if its possible to have preview of data when using In database tools working with db tables.

I have to invest more time on seeing the result using browse tools which again slows bit the performance on queries just to see the results.Also each step you have to add browse tool.

 

I have worked heavily on powerbi and see this as prime need as in PBI you can work on the visible data itself and figure it out right away how the transformations are working.if preview feature with cache options are given i guess that would be a big booster for Alteryx users.

 

Thanks,

Brij

0 Stars
In some corporate instances that I have encountered, browsing to the internet is not supported. What I've also seen is private instances of wiki media and yammer. Perhaps we could extend the browse to support these types of instances.

Cheers,
Mark
0 Stars

We are big fans of the In-Database Tools and use them A LOT to speed up workflows that are dealing with large record counts, joins etc.

This is all fine, within the constraints of the database language, but an annoyance is that the workflow is harder to read, and looks messy and complicated.

 

A potential solution would be to have the bottom half of the icon all blue as is, but the top half to show the originaling palette for that tool.

ie  Connect In-DB - Green/Dark Blue

Filter In-DB - Light Blue/Dark Blue

Join In-DB - Purple/Dark Blue

 

etc.

 

in-DB workflows would then look as cool as they are !

 

Thanks

dan

0 Stars

This would allow for a couple of things:

 

Set fiscal year for datasource to a new default.

 

Allow for specific filters on the .tde (We use this for row level security with our datasources).

 

Thanks

0 Stars

Another seemingly minor one that would just make life a little easier and clean up workflows.  I often find myself renaming the Name/Value fields from a Transpose to be more descriptive.  Currently this requires a Select tool after each transpose, and it would be nice to put a couple of text boxes at the bottom of the transpose to rename Name/Value directly in the tool itself.

0 Stars

The Multi-Field Binning tool, when set to equal records, will assign any NULL fields to an 'additional' bin

e.g. if there are 10 tiles set then a bin will be created called 11 for the NULL field

 

However, when this is done it doesn't remove the NULLs from the equal distribution of bins across the remaining items (from 1-10).Assuming the NULLs should be ignored (if rest are numeric) then the binning of remaining items is wrong.

 

Suggestion is to add a tickbox in the tool to say whether or not NULL fields should be binned (current setup) or ignored (removed/ignored completely before binning allocations are made).

Top Starred Authors