Alteryx Designer Desktop Ideas

Share your Designer Desktop product ideas - we're listening!
Submitting an Idea?

Be sure to review our Idea Submission Guidelines for more information!

Submission Guidelines

Featured Ideas

Please bring back the following two features/characteristics of the Saleforce input tool:


1. Large text box for SOQL Where statements (previous versions enabled one to space out complex queries in an organized fashion, we typically filter on 3 to 8 variables)

 

2. Remember connection settings on machine (in previous versions, the tool remembered the username/pw/token, all you had to do was select the URL from a drop down - now you have to type in the URL, username, PW, and Token)

 

Thank you much,

Jeremy

It would be nice to have the ability to automatically generate twbx files from a master Tableau workbook so that end users can open the file in Tableau reader.  For example, if I were creating separate CSV files with my data for each state I would similarly want to create them as a ready to consume twbx file with Tableau Reader.

Hi,

 

When you create a data model in excel you can create measures (aka KPI). These is then something you can then use when you pivot the data and measure would dynamically be updated as you segment the data in your pivot table.

 

By example, let's say you have a field with the customer name and a field with the revenue, you could create a measure that will calculate the average revenue per customer (sum of revenue / disctintcount of customer)

Now if you have a 3rd field in your data that inidcates your region, the measure would allow to see the average revenue by customer and by region (but the measure formula would remain the same and wouldn't refer to the region field at all)

 

Excel integrates well with PowerBI and currently these measures flow into PowerBI.

 

While we have a "Publish to PowerBI" in Alteryx I haven't seen any way to create such measures and export them to PowerBI.

 

Hence I still need to load to Excel to create these measures before I can publish to PowerBI, it'd be great to avoid that intermediate tool.

 

Thanks

 

Tibo

 

0 Likes

Hi everybody! As you can read here I had the necessity to insert a macro (publish to tableau server) at the end of a self-made app.

 

I have actually found 2 different ways to solve the problem:

 

1) Turn the macro into an app and use 2 chained apps.

2) Copy and paste tools (normal and interface) from the macro to my app.

 

Both solution work, but both require some (quite a lot to be honest) editing and re-work that actually is already done. It's kind of like to re-invent the wheel!

 

A quick way to merge 2 configuration interfaces would be really usefull.

 

 

I understand the need for "exclusive rights" when using an input tool.  Unfortunately, due to the nature of some corporate data, getting write access to a file is not always possible.  I would like to have the ability to configure an input tool to open a file in "Read Only" mode while producing a warning message that the file was processed in that mode and may not contain the lastest version pf the data.  I envision this as a checkbox option in the tool configuration panel.

Hi,

Recently in Feb 2016, Australia released the geocoded national address file to the public for no extra cost and will continually update this each quarter.

 I think It would be a game changer to build this functionality natively into the alteryx product to enable any alteryx user simple access to it. also I think it would drive a lot of sales for the alteryx product.

 

http://www.data.gov.au/dataset/geocoded-national-address-file-g-naf

 

adrian

I have a problem where bulk upload is failing because the last column of the table that the data is being imported into, is using the DEFAULT data type option. I am not passing through any value to this column as I want the DEFAULT value specified to always be applied.

 

The COPY command fails in this scenario if you don't specify an explicit field list

 

More details of the problem can be seen in this post, along with a workaround:

http://community.alteryx.com/t5/Data-Sources/import-into-redshift-DEFAULT-data-value/m-p/16655#U1665...

 

A tick box option should at least be added to the bulk upload tool to enable explicit field list specification based on column names coming into the bulk upload tool

 

I have been going back and forth between output tool and render tool. The render tool works well when you want formatting. It's also great when you don't want an output created (when rec count=0) while output tool always generates a file regardless if there are records present or not. Output tool works well when you have a lot of fields but then you cannot easily control styling.

My issue is that I have a render tool connected to table tool. In the table tool everything looks neat; there's no wrapped text and no unnecessary white space (auto column width not necessary). However, with the render output, you don't know what to expect. Especially, when you have a lot of fields (30+) data gets truncated and column width is forcibly narrowed due to paper dimensions used in render tool. I skip letter and tabloid formats and now I have to mess with custom paper width (e.g.50) to get my reports looking right and when you have dynamic fields, this is not ideal. 

Would it be possible to make the paper size/width automatic in the render tool just like in the table and layout tool? Then this tool also doesn't negate what layout/table tools do so well.

 

 

I have a module that queries a large amount of data from Redshift (~40 GB). It appears that the results are stored in memory until the query completes; consequently, my machine, which has 30 GB memory, crashes. This is a shame because Alteryx is good with maintaining memory <-> HDD balance.

 

Idea: Create a way to offload the query results onto the HDD as they are received.

In the old charting tool, you could change the order of the series by moving them up or down. This feature has been eliminated from the new charting tool. I want sales percent to be the top bar, but despite which 'series' I put it in, inventory percent is the top bar, as the order appears to be alphabetic. It would be nice to have that option added back in.

 

In the new charting tool, you can adjust the number of decimal places for your series, which is great. Adding '%' as a suffix to the series label would be nice. There is an option to add prefix/suffix to the axis label, but not to the series.

 

Horizontal axis (x-axis) displays at an angle. Having an alignment option (center; vertical; horizontal; etc.) is preferable.

0 Likes

It would be nice if I had a workflow that had multiple paths to be able to run only one line of it.  This would allow me to only run what I needed so instead of having to run the entire workflow, I would be able to run up to a certain point and only run tools that are required to get to that point.  

0 Likes

Other tools that I have used allow you to determine where you are caching from so instead of always having to cache at the input, you could cache after a big join.  This would be great for efficiency as having to run everything through the entire workflow every time is innefficient and I end up spending a lot of time waiting for my workflow to go through the same tools.  

For the purpose of debugging a workflow, I often filter just one customerID or any other ID to analyse the workflow.

 

With the Browse tool (ctrl-shift-B) you can just double click a cell and copy the value of it. This is not possible in the result tool, it would be nice if that would become possible.

 

Thanks,

 

Hans

 

It would be great if you could create default settings for the Tool Containers. As workflows become larger, I use containers a lot. But once I have 10-15 containers, I have to set all of them to have a Transparency of 1 and a margin of None. While the changes don't take long to make, it would be nice if they could be preset.

I'd like to be able to disable a tool container but not minimize it so I can still see what's in there. Maybe disabled containers could be grayed out the way the output tools are when you disable them. We would still need to retain current features in case people like it that way, but it would be nice to choose.

With SSIS, you can invoke user precedence contraint(s) to where you will not run any downstream flows until one or more flows complete.  A simple connector should allow you to do this.  Right now, I have my workflow(s) in containers, and have to disable / enable different workflows, which can be time consuming.  Below is a better definition:

 

Precedence constraints link executables, containers, and tasks in packages in a control flow, and specify conditions that determine whether executables run. An executable can be a For Loop, Foreach Loop, or Sequence container; a task; or an event handler. Event handlers also use precedence constraints to link their executables into a control flow.

0 Likes

Hey guys!!

 

I was just thinking... they might not need to fully build out a python ide, but could still reach the same objective.

 

You should be able to keep a python file on its own and call it in r.  By doing this, you might be able to have the json/xml handling of python with the visual/stats power of R while it being nicely bundled in your workflow.  This uses base functions in r and does a good job turning a pandas dataset to an r dataframe you can move along your workflow.

 

You could always just use this same idea to write a file somewhere and once it's written, your workflow will continue.  If you do, the code is literally 1 line in r...  Anyway, let me know your thoughts! 🙂

 

Will this work for your organization?

 

https://www.linkedin.com/pulse/using-python-r-windows-7-subhash-jaini?trk=hp-feed-article-title-publ...

@AdamR_AYX,

 

Limit conversion warning allows for a minimum of 1 message.  Can we set the minimum to 0 to completely ignore the message?

 

Perhaps we can allow warning messages a similar function as ERROR messages and allow the designer to Ignore, Warn or Cancel?

 

ConvError: Imputation (441): Tool #104: No demand: 0.200000000000031 had more precision than a double. Some precision was lost.

ConvError: Summarize (456): Data&colon; 0.360000000004675 had more precision than a double. Some precision was lost.

 

End: Designer x64: Finished running FP Model - Marquee Crew v3.yxmd in 32.3 seconds with 16 field conversion errors and 4 warnings

 

Thanks,

 

Mark

Field Summary is a great tool, but would be nice to have a count and count not null on it.

Would save an extra select for parsing text files in correct format for dates and times.

 

 

Top Liked Authors