community
cancel
Showing results for 
Search instead for 
Did you mean: 

Alteryx Designer Ideas

Share your Designer product ideas - we're listening!

1 Review

Our submission guidelines & status definitions before getting started

2 Search

The community for a solution or existing idea before posting

3 Vote

By clicking the star in the top left corner of an idea you support

4 Submit

A new idea to suggest a product enhancement or new feature


Suggest an idea

While In-db tools are very helpful and cut down the time needed to write complex SQL , there are some steps that are faster by directly writing SQL like window functions- OVER (PARTITION BY .....). In Alteryx, we need to create multiple joins and summaries to perform a window function. It would be immensely helpful if there was a SQL editor tool for in-db workflows where we can edit the SQL code at any point in the workflow, or even better, if they can add an "edit" function to every in-db tool where we can customize the SQL code generated and then send to the next tool.

 

This will cut down the time immensely and streamline the workflow to make Alteryx a true contender for the ETL solution space.

When I use the Comment Tool its difficult to select the tools inside it, but when I use the Container Tool the Container Text doesn't support Font Sizes, and doesn't support multiple lines of text so I end up moving the Comment into the Container, but still have problems selecting a group of tools.

 

So a combined Comment and Container Tools would be wonderful!

 

Bonus: If the Comment Tool could support Multiple Font Sizes.

 

Combined Tools.png

Unsupervised learning method to detect topics in a text document.

 

Helpful for users interested in text mining.

I've seen this question before and have run into it myself.  I'd like to see a new tool that would allow a developer (of a workflow) to choose a path of logic based upon criteria known only during the execution of a module.

 

If LEFT INPUT Count of records < 10,000 THEN Path1 (e.g. use a calgary join)

ELSE Path 2 (e.g. use a standard join)

endif

 

Thanks,

 

Mark

I've seen this question before and have run into it myself.  I'd like to see a new tool that would allow a developer (of a workflow) to choose a path of logic based upon criteria known only during the execution of a module.

 

If LEFT INPUT Count of records < 10,000 THEN Path1 (e.g. use a calgary join)

ELSE Path 2 (e.g. use a standard join)

endif

 

Thanks,

 

Mark

I constantly find my using pre and post SQL Commands in the Output tool to run SQL when I don't actually have any data to output.

 

One example is when I load data into S3 and want to load it into Redshift. I have SQL code to run but no data to Output - I end up running a dummy row into a temp table.

 

So can we have an SQL tool that simply acts the same as a Pre-SQL command without the associated data output. Once the command is run we should be able to continue the workflow, so the tool should have an option input and output, like the Run Command tool.

 

 

I constantly find my using pre and post SQL Commands in the Output tool to run SQL when I don't actually have any data to output.

 

One example is when I load data into S3 and want to load it into Redshift. I have SQL code to run but no data to Output - I end up running a dummy row into a temp table.

 

So can we have an SQL tool that simply acts the same as a Pre-SQL command without the associated data output. Once the command is run we should be able to continue the workflow, so the tool should have an option input and output, like the Run Command tool.

 

 

DELETE from Source_Data Where ID in

SELECT ID from My_Temp_Table where FLAG = 'Y'

 

.... 

 

Essentially, I want to update a DB table with either an update or with the deletion of rows.  I can't delete all of the data.  My work around will be to create/insert into a table the keys that i want to delete and try to use a input/output tool with SQL that performs the delete.  Any other suggestions are welcome, but a tool is best.

 

Thanks,

Mark

DELETE from Source_Data Where ID in

SELECT ID from My_Temp_Table where FLAG = 'Y'

 

.... 

 

Essentially, I want to update a DB table with either an update or with the deletion of rows.  I can't delete all of the data.  My work around will be to create/insert into a table the keys that i want to delete and try to use a input/output tool with SQL that performs the delete.  Any other suggestions are welcome, but a tool is best.

 

Thanks,

Mark

The introduction fo a rank tool would be hugely beneficial. Whilst there are currently means to rank using a combination of other tools formula/running total/multirow etc... a specific "Rank Tool" would be provide a seemless and smoother way to rank your data either for further analysis or purely to output this field.

 

This tool should include a sort by and group by functionaility as well as options for ranking (such as dense ranking or unique ranking) and in addition multi levels of ranking (ie. Rank by "Field A" Then By "Field B" etc...).

The introduction fo a rank tool would be hugely beneficial. Whilst there are currently means to rank using a combination of other tools formula/running total/multirow etc... a specific "Rank Tool" would be provide a seemless and smoother way to rank your data either for further analysis or purely to output this field.

 

This tool should include a sort by and group by functionaility as well as options for ranking (such as dense ranking or unique ranking) and in addition multi levels of ranking (ie. Rank by "Field A" Then By "Field B" etc...).

XGboost regression is now the benchmark for every Kaggle competition and seems to consistently outperform random forest, spline regression, and all of the more basic models. For those of us using predictive modeling on a regular basis in our actual work, this tool would allow for a quick improvement in our model accuracy. And I think, from a marketing standpoint, having a core group of users competing in Kaggle using Alteryx would be a great way to show off Alteryx's power.

 

It is readily available as an R package: https://cran.r-project.org/web/packages/xgboost/index.html

XGboost regression is now the benchmark for every Kaggle competition and seems to consistently outperform random forest, spline regression, and all of the more basic models. For those of us using predictive modeling on a regular basis in our actual work, this tool would allow for a quick improvement in our model accuracy. And I think, from a marketing standpoint, having a core group of users competing in Kaggle using Alteryx would be a great way to show off Alteryx's power.

 

It is readily available as an R package: https://cran.r-project.org/web/packages/xgboost/index.html

I need support for outbound data streams to be gzip compressed.  Ideally, this would be done by a new tool that can be inserted into a workflow (maybe similar to the Base 64 Encoding tool).  Just including it in the Output Tool will not address my needs as I will be sending gzip payloads to a cloud API.  There are two main reasons why this is necessary (and without it, quite possibly a roadblock for our enterprise's use of Alteryx):

 

  1. Some APIs enforce gzip encoding, therefore Alteryx cannot currently be used to interact with such APIs
  2. When transmitting large volumes of data across the Internet, gzip compression will significantly decrease transmission times

It would be handy to have a tool that mirrored another specified tool in the workflow.  I frequently use the same regex or filter configurations in multiple parts of my workflow.  If I could just update one of those tools and have every other one where I want the same logic applied to automatically update, that would make things easier. 

 

Users would just have to enter the tool ID of the tool they want mirrored.  

  • New Tool

Here's a twist on the iterative macro.  Suppose, like a generate rows tool, you could initialize a container to iterate on it's internal processes without having to construct a macro?  The container could include anchors for iterations and for output and allow the user to DoWhile inside of the container.

 

Just a thought....

Hello,
 
We have been storing data on the PostgresDB and would be using variables from these tables for modeling processes in Alteryx. The ODBC connectors for postgres don't seem to be geared towards optimal performance, hence it would be extremely useful if Alteryx came up with an in-db functionality for postgres database in the future release.
 
Thanks
 
  • New Tool

Here's a twist on the iterative macro.  Suppose, like a generate rows tool, you could initialize a container to iterate on it's internal processes without having to construct a macro?  The container could include anchors for iterations and for output and allow the user to DoWhile inside of the container.

 

Just a thought....

Hello,
 
We have been storing data on the PostgresDB and would be using variables from these tables for modeling processes in Alteryx. The ODBC connectors for postgres don't seem to be geared towards optimal performance, hence it would be extremely useful if Alteryx came up with an in-db functionality for postgres database in the future release.
 
Thanks
 
  • New Tool

Visio is our organization's most common method of communicating business processes and workflows. Being able to export an Alteryx workflow to Visio would help us communicate the tool's functionality to process owners.

Top Starred Authors