Free Trial

Alteryx Designer Desktop Ideas

Share your Designer Desktop product ideas - we're listening!
Submitting an Idea?

Be sure to review our Idea Submission Guidelines for more information!

Submission Guidelines

Featured Ideas

0 Likes

It would be fantastic if the Detour tool worked outside for a Standard Workflow, and not just for Analytic Apps and Macros. I can think of many instances where this would be very useful.

0 Likes

Please consider making the Count Record tool configurable so that users can receive a visual read of results in canvas. This would quickly assist in ensuring a user's ability to verify a workflow is functioning as expected or not. Currently there is no in canvas visual cue tied to the counter. If a user is expecting a certain count result (e.g. zero), the user has to click on the count records function to see if the result meets expectations or not. Users may spend a lot of time checking each stage of a workflow to ensure everything is flowing appropriately. A visual cue of results would reduce that time to check. Outside of runtime errors, there is currently no visual cue to indicate a possible problem or unexpected exception to a filter or other macro.

 

Two suggestions:

 

1) Allow users to color code count results in specific count results or ranges. (e.g. if a user is expecting a zero count, allow them to change the color of the counter to red for anything not resulting in zero and green if count is zero). Or allow them to set a color range depending on the count total (e.g. a user may need some visual tolerance indicator with a count under 100 as green, 101-200 as yellow, 201-300 as orange and 300+ as red).

 

2) Show the actual count in the Count Record macro icon.

 

Thank you!

0 Likes

Hi Team,

 

I have an idea where we can use Alteryx to build a virtual Assistant. As we are currently using Intelligent suite extract information through pdfs. Now we can connect the VA and Intelligent suite to offer a complete product.

 

Please let us know your views,

0 Likes

Hello,

 

My idea is that the current Download tool does not handle errors and continues its path even if it does not find for example a file in the transmitted URL or if it does not find the hostname it crashes.

 

In the case of a user with several URLs in a row, this is penalizing.


In the case of downloading files with recording, it still writes a file (thus overwriting the existing file) but which is not openable and is not in the correct format. (BLOCKED file!) Which then causes problems in workflows reading these files.

 

The idea would be to put a second output to this tool for all the URLs where there was a problem (non-existent hostname, file not found, HTTP KO) and one where it received the expected elements so as not to prejudice the processing. and allow better management of error cases.

 

Regards,

Bruno

0 Likes

More and more people are making use of Plotly and Plotly Express to develop really great graphics.  However it is extremely difficult to extract either interactive charts (HTML format) or static images (PNG or JPEG).

 

The main issue for Plotly to save static images is the ability to install and use Orca (application) for R and kaleido (library) for Python.  Despite my best efforts I have had no luck getting either approach to work with the respective Alteryx environments

0 Likes

Using the Download Tool, when doing a PUT operation, the tool adds a header "Transfer-Encoding: chunked".  The tool adds this silently in the background.

 

This caused me a huge headaches, as the PUT was a file transfer to Azure Blob Storage, which was not chunked.  At time of writing Azure BS does not support chunked transfer.  Effectively, my file transfer was erroring, but it appeared that I had configured the request correctly.  I only found the problem by downloading Fiddler and sniffing the HTTPS traffic.

 

Azure can use SharedKey authorization.  This is similar to OAuth1, in that the client (Alteryx) has to encrypt the message and the headers sent, so that the Server can perform the same encryption on receipt, and confirm that the message was not tampered with.  Alteryx is effectively "tampering with the message" (benignly) by adding headers.  To my mind, the Download tool should not add any headers unless it is clear it is doing so.

 

If the tool adds any headers automatically, I would suggest that they are declared somewhere.  They could either be included in the headers tab, so that they could be over-written, or they could have an "auto-headers" tab to themselves.  I think showing them in the Headers tab would be preferable, from the users viewpoint, as the user could immediately see it with other headers, and over-ride it by blanking it if they need to.

0 Likes

I still see that companies are expecting BI professionals to write SQL queries manually despite Alteryx giving complete solution. Idea is to spread value add of Alteryx in Indian market and global companies

0 Likes

If we have similar workflows within the gallery saved and we need to make slight changes in the process to all of them, it would be helpful if we could open all workflows from the gallery by either mutliselect or using a checkbox next to those workflows that we wish to open all at once.  

0 Likes

Why don't Alteryx build an "Designer extended edition"?

 

An Alteryx designer download that contains every available 

  • R package and
  • Python package available in the universe to date of release...

all you can eat.png

0 Likes

in a scenario where i pass multiple files to the download tool (to upload to an FTP server), the data in the workflow gets cut off with the message "ended by a downstream tool". This leads to me only uploading one file (out of the two I am trying to upload).

 

based on community posts, this seems to be because i am not outputting anything in the workflow or do not have any browses (browses are automatically turned off because I am using Crew Macro's List Runner)

 

BUG: the download tool should count as an output tool because it is indeed outputting. 

 

BUG 2: if browses are disabled but they exist in the workflow, it should still create all the records.

0 Likes

Yeah, so when you have 15 workflows for some folks and you've actually decided to publish to a test database first, and now you have to publish to a production database it is a *total hassle*, especially if you are using custom field mappings.  Basically you have to go remap N times where N == your number of new outputs.  

 

Maybe there is a safety / sanity check reason for this, but man, it would be so nice to be able to copy an output, change the alias to a new destination, and just have things sing along.  BRB - gotta go change 15 workflow destination mappings. 

0 Likes

The idea is the ability to skip tools in a workflow. Check the attached example:

 

Input (TRX_TABLE): Contains transactional raw data (unstructured & uncleaned data)
Filter (TRX_DATE <= 90 days from TODAY): It only filters the data 90 days old from today. Works as an incremental data loader
If the filter tool is NOT SKIPPED the workflow loads the filtered data (incremental behavior)
If the filter tool is SKIPPED the workflow loads all the data (full behavior)
This way I can quickly switch between both modes for testing & production environments

0 Likes

Create a standardized Mailbox application that could bolt onto Alteryx Server, to handle incoming attachments from sources like a Service Desk (Service Now for example) and other applications.

 

Essentially anything that regularly exports data in the form of an emailed attachments to which Alteryx could, using a series of predefined user rules and a designated email address, put those attachments into various directories ready for processing by automated Alteryx workflows.

 

This would save a huge amount of time as people currently have to manually drag and drop files. At least the on board Alteryx designers here haven't been able to come with a solution. Would also save any messy programming around systems like Outlook and bending any security issues within those systems. Many, many other applications have this simple feature built in to their products, especially service desks. I believe there would be a huge benefit to this very simple bolt on.

0 Likes

I've run into an issue where I'm using an Input (or dynamic input) tool inside a macro (attached) which is being updated via a File Browse tool. Being that I work at a large company with several data sources; so we use a lot of Shared (Gallery) Connections. The issue is that whenever I try to enter any sort of aliased connection (Gallery or otherwise), it reverts to the default connection I have in the Input or Dynamic Input tool. It does not act this way if I use a manually typed connection string.

 

Initially, I thought this was a bug; so I brought it to Support's attention. They told me that this was the default action of the tool. So I'm suggesting that the default action of Input and Dynamic Input tools be changed to allow being overridden by Aliased connections with File Browse and Action tools. The simplest way to implement this would probably be to translate the alias before pushing it to the macro.

0 Likes

This may be too much of an edge case...

 

I would like to be able to feed a dynamic input component with a input file and a format template file, so make the input component completely dynamic.  This is because I have excel spreadsheets that I want to download, read and process hourly throughout the day.  Every 4 hours another 4 columns are added to the spreadsheet, thereby changing the format of the spreadsheet.  This then causes the dynamic input component to error because the input file does not match the static format template.  I would be happy to store the 6 static templates, but feed these into to the dynamic input with the matching input file, thus making the component entirely dynamic.  Does this make sense???

 

BTW, my workaround was to define six dynamic inputs, filter on the file type, then union the results:

 

image.png

 

 

0 Likes

Hi, 

 

I am currently using Oracle Discoverer Tool which has option to drag or place an results column of a table or cross tab on to page items so that I can further filter the results. This helps the user to filter large results set without even downloading the output. I have attached a screenshot of current tool with page items.

 

Regards,

Sunil

0 Likes

Can we have string function that parse the string between 2 indices?

As FindString can find particular string occurrence,we can easily get required part of string easily from that index till required index.

or If we want entire remaining string we can have a function like:

Substring(String,StartIndex,EndIndex) Where endIndex can be : Length(String)-1

 

 

0 Likes

I setup a generic template because the excel files do not have a consistent schema.   This is the template

 

Capture1.JPG 

 

which is interpreted as text

 

  Capture3.JPG

 

When I feed files into the dynamic input tool the output configuration is

 

Capture2.JPG

This is the schema of the first file read by the DI tool.  Why didn't it use the template?

 

The requirement to have all the schemas match is difficult - especially with excel.

 

I would like to suggest the input tool have a checkbox to override the data driver and interpret all columns as text.

 

Read excel as text

https://www.connectionstrings.com/ace-oledb-12-0/treating-data-as-text/

 

0 Likes

It would be great if it were possible in future to adjustment Analytic App buttons.

 

For example Run instead of Finish or so... and so on!?!

 

Best regards

Mathias

0 Likes

Please ignore this post! I just discovered the amazing dynamic input tool can accomplish exactly this 🙂

 

This is no longer a "good future idea" but simply a solution already provided by AlterYX!  Thanks AlterYX!

 

Hello everyone. I think it would be incredibly helpful if all input tools (database / csv files / etc..)  had an input stream for data processing.  

 

This input stream would cause the tool to fire if and only if at least 1 row of "data" was passed into it.  This would work similar to the Email tool, as the input would simply not fire if no records in the workflow were passed into it.

 

By making this change, AlterYX could then control the flow of execution when reading / writing to database tables and files without having to bring in Macros or chained applications.

 

The problem with having to use batch macros and block until done to solve this problem is that it obfuscates your workflow logic, making it much harder for another developer to look at the workflow and understand what is happening.   Below is an extremely simply example of where this is necessary.

 

Example: any time you are doing a standard dimentional warehouse ETL.

 

Step #1: Write data to a table with an auto-incremented Key

Step #2: Read back from that table to get the Key that you just generated

Step #3:  use that key as a foreign key in your primary Fact table.

 

Without this enhancment you need at least 2 different workflows (chained apps or marco's) to accomplish the task above, however with this simple enhancement the above task could be accomplished in 3 tools.

 

This task is the basic building block of almost all ETL's so this enhancement would be very useful in my opinion.

 

 

Top Liked Authors