Alteryx Designer

Share your Designer product ideas - we're listening!

1 Review

Our submission guidelines & status definitions before getting started

2 Search

The community for a solution or existing idea before posting

3 Vote

By clicking the like in the top left corner of an idea you support

4 Submit

A new idea to suggest a product enhancement or new feature


Suggest an idea

Hi!

 

I just installed 2020.2 & I LOVE this update within the Formula tool:

 

  • When you double-click a column name inside of an expression, the brackets around the column name are also selected.

 

Are there any plans to bring this ability to the Filter tool as well, where I can double click on my column name & have it grab the brackets as well?

 

Not going to lie, love that you added the Open/Save/Undo buttons back to the top toolbar too! Latest version looks great, thank you!

The current version of the Publish to Tableau macro retrieves an API key at the start of the workflow run.  Often times the workflow may take several hours to run before it's ready to write to Tableau by which time the API may have expired.  (I think the default tableau server setting times out in 2 hrs)  It's one of those soul crushing "I should've forked the output!"

Sample Log Error - 

  •         Tool #46: TableauServer.UploadChunks (238): Iteration #1: Tool #19: Tool #4: Tableau Server API Request (Upload file) Error Code 401002: Unauthorized Access -- Invalid authentication credentials were provided. 
  •          Tool #46: Tool #252: Tool #4: Tableau Server API Request (Publish file) Error Code 401002: Unauthorized Access -- Invalid authentication credentials were provided.

The idea would be to change when the macro obtains the API from when the workflow is initiated to just before the workflow is ready to write to the Tableau avoiding these timeouts.

 

(If you're having this issue in the meantime you can have your Tableau server admin up the timeout)

7-9-2020 2-53-42 PM.png

Today when we install custom tools that use DLLs, the DLLs must be placed in the Plugins folder inside the Alteryx installation directory.  This requires a second step after the YXI installer runs.  I would like to be able to package the DLL with the YXI installer and Alteryx will search for the DLL inside the tool's directory, just the same as what happens with custom Python tools.  This will allow custom tools that use DLLs to be installed just as easily as the 1-step installation process for Python tools.

 

For example, this today does not work, but I want it to:

Screen Shot 2020-06-05 at 8.35.17 AM.png

Please enhance the dynamic select to allow for dynamic change data type too.  The use case can be by formula or update in an action for a macro.  If you've ever wanted to mass change or take precision action in a macro, you're forced to use a multi-field formula.  It would be rather helpful and appreciated.

 

Cheers,

 

Mark

The Alteryx Python tool currently throws an error if the inbound record set has zero rows (screenshot 1).

In order to manage that - you need to create try-except block around the Alteryx.read that instead creates an empty record set data frame. (screenshot 2).   This is inefficient because every time you change the canvas before the python tool, you need to re-code a static field list into the try-except block (i.e. you can no-longer deal with variable fields)

 

Please could you change the Alteryx.read method to create a zero-record dataframe with the correct column names if the input is zero-length?

 

Thank you

Sean

 

Screenshot 1:

ErrorMessage.png

 

Screenshot 2:

TryExceptBlock.png

Can we have an option to save a workflow in a prior version for backward compatibility? I think Tableau offers this functionality.

Example:

If I have 2019.4.8 and a colleague has 2019.1.x, I cannot share my workflows because my colleague will receive a notice that the workflow was built in a newer version. I want to be able to save my workflow in 2019.1.x and send to my colleague.

 

This is predicated on the workflow not containing any tools/features not present in the older version. In that case, give me a warning about the specific tools/features that are not backward compatible. Thank you.

Environment variables act as a shortcut so that different computers can be configured in different ways, but a particular path will still point to the right place.

 

For example if you open up explorer and go to %TEMP%\  - you will open up whichever folder is set up as Temp on this machine.    This is super useful so that you can use a particular logical folder without knowing the actual placement on every machine (for example the Windows Directory)

 

envVariables.png

 

 

This works partially in the Directory / input - when you put in the environment variable, it is able to search possible subdirectories (screenshot 1) but it does not work once you run the workflow (screenshot 2).

It seems as if the designer hits the Windows API directly, but it does not work within the engine.

 

Please could you alter the engine to be able to make full use of the environment variables on the machine in question in the directory path or input tool path?

 

Screenshot 1 - works in designerScreenshot 1 - works in designer

 

Doesnt work in Engine.png

  • Engine

With an increasing number of different projects, involving different machine learning models, it's becoming difficult to manage different package versions across workflows. Currently, the Python tool has a single virtual environment, so we need to develop models in different projects always using the same Python and package versions as the Python tool venv. While this doesn't bother the code itself too much, it becomes a problem as soon as we store and load pickled models, which are sensitive to even minor changes in packages.

 

This is even more so a problem when we are working on the Alteryx server, where different teams might use different packages. Currently, there is only the server admin who can install packages on the server and there can only be one version per package.

 

So, a more robust venv management in the Python tool would be much appreciated!

I love Workflow Meta info, especially the ability to put the Author, the search tags,the version, the description, etc...

workflow meta info.png

But why can't we use it as Engine Constant? It doesn't seem very hard to implement and it would change life for development.

 

engine_constant.png

Hello,

 

we have several environment in our organization : dev, recept, production.

 

In order to make that change safe we intend to make several connection (standard alias)  like

alias_in_memory_pour_support.PNG

PRODUCTION_HIVE

DEV_HIVE

RECEPT_HIVE

 

In our workflows, we want to use aka:%Question.v_environment%HIVE

 

Sadly, this solution does not work despite the value defaut. 

 aka_et_alias_in_memory.PNG

 

Hello!  I use Alteryx for lots of spatial data blending, and 99% of the time that works perfectly.  However, when I try to analyze data in Puerto Rico or some other USA territories, Alteryx cannot read the data and I have to reproject it in another tool so it can.  Can support for all common projections be added to Alteryx so the data can be read in natively?  The GDAL Python code data file "gcs.csv" contains all of the projections I think would be needed.

It appears that the Workflow Dependencies window does not report dependencies from all tools. In the example image, you can see that the file input from the Amazon S3 Download tool is not listed. Some tools may have dependencies that do not easily fit the current field structure of the window, but maybe the input/download tools could be listed with an asterisk or partial reference.

Missing Amazon S3 DependencyMissing Amazon S3 Dependency

When we industrialize our workflows, we often use a parameter file with a command like :

 

 

AlteryxEngineCmd.exe MyAnalyticApp.yxwz AppValues.xml

 

 

I would like to have the parameter file path with its extension as an engine constant, like we have the workflow name.

Idea:

I know cache-related ideas have already been posted (cache macros; cache tools), but I would like it if cache were simply built into every tool, similar to the way it is on the Input Tool.

 

Reasoning:

During workflow development, I'll run the workflow repeatedly, and especially if there is sizeable data or an R tool involved, it can get really time consuming.

 

Implementation ideas:

I can see where managing cache could be tricky: in a large workflow processing a lot of data, nobody would want to maintain dozens of copies of that data.  But there may be ways of just monitoring changes to the workflow in order to know if something needs to be rebuilt or not: e.g. suppose I cache a Predictive Tool, and then make no changes to any tool preceeding it in the workflow... the next time I run, the engine should be able to look at "cache flags" and/or "modified tool flags" to determine where it should start: basically start at the "furthest along cache" that has no "modified tools" preceeding it.

 

 

Anyway, just a thought.

 

Hi,

     I would like to see Global Variable being made available in Alteryx. I have seen the Global Constant being made available under Workflow "User" configuration. But this is constant and needs to be defined at Design time.

How about a Process Id that needs to be auto genearted and the same needs to be available across the formula tools used with in the workflow.

In order to perform audit-trail logging - it would be valuable to have 2 new capabilities

 

a) environment variables which show the workflow name; filepath; version; run start date and time; etc.   For any worklows we build, we need to have a solid audit trail to be SOX compliant, so having this detail available as a data field to write and manipulate is essential

b) A logging component.   What would be great is a component that you can drop on a workflow, not connected to anything, which is able to trap the start; end; runtime; version; etc of a workflow; and commit this to any output data format (CSV or ODBC etc).   This logging tool would need to be able to capture the full runtime, so it would need to be the last thing that runs (which means it may need to exist in parallel to the main workflow in some way).    This is not currently possible with a complex workflow with outputs, because it's not possible to identify when the entire workflow ended; or the runtime (since output tools don't have an onward connector to pass flow-of-control to catch the final end-time)

 

Again, both of these are necessary to meet audit requirements for workflows and prodcution-quality ETLs for BI data warehouses.

Please offload map rendering, in Browse Tool, to the video card using DirectX or OpenGL, the software rendering currently used is embarrassingly slow and disruptive.  

We often build very large Alteryx projects that breakdown large data processing jobs into multiple self contained workflows.

 

We use CReW Runner tools to automate running the workflows in sequence but it would be nice if Alteryx supported this natively with a new panel for "Projects"

 

Nice features for Projects could be:

  • Set the sequence
  • Conditional sequence
  • Error handling
  • Shared constants
  • Shared aliases
  • Shared dependencies
  • Chained Apps
  • Option to pass data between workflows - Input from yxmd Output - no need to persist intermediary data
  • Input/output folder/project folder setups for local data sources in dependencies window
  • Ability to package like "Export Workflow" for sharing
  • Results log the entire project

A lot of popular machine learning systems use a computer's GPU to speed up some of the math to a huge degree. The header on this article on Medium shows a 15x difference from a high-end CPU vs a high-end GPU. It could also create an improvement in the spatial tools. Perhaps Alteryx should add this functionality in order to speed up these tools, which I can imagine are currently some of the slowest.

Currently we can't use any PaaS MongoDB products (MongoDB Atlas / CosmosDB) as Alteryx Gallery doesn't support SSL for connecting to the MongoDB back end.

 

SSL is good security practice when splitting the MongoDB onto a different machine too.

Top Liked Authors