Bring your best ideas to the AI Use Case Contest! Enter to win 40 hours of expert engineering support and bring your vision to life using the powerful combination of Alteryx + AI. Learn more now, or go straight to the submission form.
Start Free Trial

Alteryx Designer Desktop Ideas

Share your Designer Desktop product ideas - we're listening!
Submitting an Idea?

Be sure to review our Idea Submission Guidelines for more information!

Submission Guidelines

Featured Ideas

0 Likes

I was working on the file and there are multiple sheets need to be pulled from one excel file. I was not sure how we can give one source of information and pull multiple sheet from one source as per the requirement. So wanted to submit this idea to create a toll which can pull any sheet(s) from one input tool as per requirement.

Since we can use Snowflake udfs in Alteryx, when do you think Snowflake stored procedures will be available.

0 Likes

When you import a csv file,  I sometimes use a "TAB" as delimiter. In section 5 Delimiters I want that as an option.

I have learned that it is possible to wright "\t" but a normal choice would bed nice.

Hello all,

As of today, you can populate the Drop Down tool in the interface category with a query launched from a in-memory connection. I would really appreciate the ability to use instead an in-db connection.

simonaubert_bd_0-1675421825343.png

 

Why ?
It means managing two connections instead of one, and finding ways to manage it on server for both of them, etc etc.. Simplicity is key.

Best regards,

Simon

Hello all,

 

As of today, you must set which database (e.g. : Snowflake, Vertica...) you connect to in your in db connection alias. This is fine but I think we should be able to also define the version, the release of the database. There are a lot of new features in database that Alteryx could use, improving User Experience, performance and security. (e.g. : in Hive 3.0, there is a catalog that could be used in Visual Query Builder instead of querying slowly each schema)

I think of a menu with the following choices :
-default (legacy) and precision of the Alteryx default version for the db
-autodetect (with a query launched every time you run the workflow when it's possible). if upper than last supported version, warning message and run with the last supported version settings.
-manual setting a release (to avoid to launch the version query every time). The choices would be every supported alteryx version.

Best regards,

Simon

Hello everybody,

 

I'm seeing an increased usage of geojson and kmz files across many popolar open data.

Would it be possible to add them to the supported Alteryx input formats?

 

Thanks!

0 Likes

Very beneficial will be adding extra row in Input tool (while we importing CSV) where we can define escape character. This functionality will help to avoid errors that currently occur when importing such files.

Hello all,

Big picture : on Hadoop, a table can be

-internal (it's managed by Hive or Impala, and act like any other database)
-external (it's managed by hadoop, can be shared among the different hadoop db such as hive and impala and you can't delete it by default when dropping the table

 

for info, about suppression on external table :

https://docs.cloudera.com/HDPDocuments/HDP3/HDP-3.1.4/using-hiveql/content/hive_drop_external_table_...

Alteryx only creates internal tables while it would be nice to have the ability to create external tables that we can query with several tools (Hive, Impala, etc).

It must be implemented

-by default for connection
-by tool if we want to override the default

Best regards,

Simon

Hello,

As of today, we can't choose exactly the file format for Hadoop when writing/creating a table. There are several file format, each wih its specificity.

Therefore I suggest the ability to choose this file format :

-by default on connection (in-db connection or in-memory alias)

-ability to choose the format for the writing tool itself.

Best regards,

Simon

Connecting to Smartsheets using Alteryx Desktop (and by extension, Alteryx Server) is extremely cumbersome. If a user wants to read data from Smartsheet, they are required to get an API token (preferred) or use a username/password

 

Then do one of the following to read data from Smartsheets:

1. a. Install a ODBC driver

    b. Configure a DSN connection for ODBC
    c. Use the input data using a generic ODBC connection
or
2. Use python

 

To write data to Smartsheets, a user can use Python or upload the data using an API call - both very hard for end users to use especially if they're not Python developers.

 

Regardless, all of these are problematic. On the server I manage, I have over 15 ODBC connections to Smartsheets and it's getting very hard to upgrade the server hardware because of them. Creating a native connector for input/output of data to Smartsheets will eliminate a headache of managing ODBC connections, and make it simple for Alteryx Desktop users to read and write data.

 

 

0 Likes

Right now, you can highlight 1 output tool, right click, convert to input tool.  Im doing that to each of 30 outputs in 1 workflow as part of splitting a workflow that has too much spatial processing. Would be cool to highlight all of the outputs and convert all at once to save the clicker finger.  Keep it rocking Alteryx!

Hello all,

We all love pretty much the in-memory multi-row formula tool. Easy to use, etc. However, the indb counterpart does not exist.

I see that as a wizard that would generate windowing functions like LEAD or LAG
https://mode.com/sql-tutorial/sql-window-functions/

 

Best regards,

Simon

In a similar vein to the forthcoming enhancement of being able to disable a specific output tool, my idea is to have the inverse where you can globally disable all outputs and then enable specific ones only. This should help reduce the number of clicks required/avoid workarounds using containers to obtain this functionality and allow users to be very specific in which outputs run and don't run as required.

Sometimes I need to connect to the data in my Database after doing some filtering and modeling with CTEs. To ensure that the connection runs quicker than by using the regular input tool, I would like to use the in DB tool. But is doesn't working because the in DB input tool doesn't support CTEs. CTEs are helpful for everyday life and it would be terribly tedious to replicate all my SQL logic into Alteryx additionally to what I'm already doing inside the tool. 

I found a lot of people having the same issue, it would be great if we can have that feature added to the tool. 

Using the Output tool to send data to a formatted spreadsheet apparently doesn't preserve formatting if the entire column is formatted.  I'd love this changed to keep the formatting when its applied to an entire column.  See this thread in Designer Discussions. 

0 Likes

Creating tool which summarize all operations performed in workflow might be beneficial for people who are working on others i.e. where SQL code is required for data transformation.

0 Likes

While using Alteryx Desinger with below option to upload files to Azure Conatiner, storage event trigger wont kick off the adf pipeline:

 

Joyeb_0-1664430380461.jpeg


There is issue with the above tool ,They are calling "CreateFile" which simply opens the file for writing, then they are uploading the file content. And finally they should be calling FlushWithClose. However, that last step is not being done.


This below microsoft thread should help Alteryx Developers to resolve the issue, many customers are looking forward to this  issue and resolution.
ADF event based trigger fires when file created manually but not when databricks creates a file (mic...


Please add support for Databricks' Unity Catalog

 

Currently, when selecting a Databricks-connection in the “Connect In-DB”-tool, and opening the “Query Builder”, only tables in the catalog named “hive_metastore” are listed. That is, Alteryx submits the following SQL query to Databricks:

Listing tables 'catalog : hive\_metastore, schemaPattern : %, tableTypes : null, tableName : %'

 

However, with Unity Catalog in Databricks the namespace is three-tier and there may be multiple catalogs (and not just the "hive_metastore" catalog), see https://docs.microsoft.com/en-gb/azure/databricks/lakehouse/data-objects#--what-is-a-catalog

 

I reached out to Alteryx support, which replied that you currently have a feature request for implementing this change (ID TDCB-4056) and they furthermore suggested that I post here.

 

Thanks in advance.

When email body gets imported using latest version of the Outlook 365 tool, this tool removes the new line separators from the message body, which makes it difficult to parse relevant information out of the message body. New line separators are there prior to message being imported into Alteryx as can be verified when importing same message using different tools (for example, Python or Power Automate). Without new line separators it is not possible to accurately parse message body using Alteryx. Please add the enhancement to the Outlook 365 tool so that it doesn't remove new line separators from the message body.

This limitation of the Outlook  365 tool has been discussed in the community 

Outlook 365 Input - Separate Body lines - Alteryx Community

I'm dealing with a database that is not always up when my production workflow runs. When the database is down, the workflow errors out because the input data connection errors. I want the workflow to gracefully execute some other steps if the database is down. I need the input data tool to allow errors to be treated as warnings, using a checkbox, and then I can have Alteryx run different logic when 0 rows are detected.

Top Liked Authors