The Product Idea boards have gotten an update to better integrate them within our Product team's idea cycle! However this update does have a few unique behaviors, if you have any questions about them check out our FAQ.

Alteryx Designer Desktop Ideas

Share your Designer Desktop product ideas - we're listening!
Submitting an Idea?

Be sure to review our Idea Submission Guidelines for more information!

Submission Guidelines

Featured Ideas

Hello all,

 

ADBC is a database connection standard (like ODBC or JDBC) but specifically designed for columnar storage (so database like DuckDB, Clickhouse, MonetDB, Vertica...). This is typically the kind of stuff that can make Alteryx way faster.

 

more info in https://arrow.apache.org/blog/2023/01/05/introducing-arrow-adbc/

 

Here a benchmark made by the guys at DuckDB : 38x improvement

https://duckdb.org/2023/08/04/adbc.html

 

307583921-894b24c0-19ee-4900-b80c-7c7688ea9cd9.png

 

Best regards,

 

Simon

Hello all,

Apache Doris ( https://doris.apache.org/ ) is a modern datawarehouse with a lot of ambitions. It's probably the next big thing.

 

image.png


You can read the full doc here https://doris.apache.org/docs/get-starting/what-is-apache-doris but to sum it up, it aims to be THE reference solution for OLAP by claiming even better performance than Clickhouse, DuckDB or MonetDB. Even benchmarks from the Clickhouse team seem to agree.

Best regards,

Simon

Problem statement - 

Currently we are storing our Alteryx data in .yxdb file format and whenever we want to fetch the data, the whole dataset first load into the memory and then we can able to apply filter tool afterwards to get the required subset of data from .yxdb which is completely waste of time and resources.

 

Solution - 

My idea is to introduce a YXDB SQL statement tool which can directly apply in a workflow to get the required dataset from .YXDB file, I hope this will reduce the overall runtime of workflow and user will get desired data in record time which improves the performance and reduce the memory consumption.

Hello all,


As of today, we use the good old alias in-memory to connect to our datasources in in-memory. We have several environments so we use constants in order to change the name of the in-memory alias during execution.

To illustrate :

simonaubert_bd_0-1701895804110.png

 

 

Depending of the environment, the constant « v_gp_contexte » will take different values :

  • GP_DS08_SIDATA for la dev.
  • GP_EE_SIDATA for prod.

Sounds nice, right? But now, we would like to use DCM and the nightmare begins :

We can't manually change the name and set the question :

simonaubert_bd_2-1701895988262.png

 

if we look at the xml of the workflow, we only find an id so editing it is useless :

simonaubert_bd_3-1701895988262.png

 


(for informationDCM connections are stored in some sqlite db in C:\Users\{yourname}\AppData\Local\Alteryx

So, I would like to use the DCM inside the in-memory alias (the in-memory alias is stored and can be edited), just like for in-db connection alias.

Best regards,

Simon

 

 

When loading multiple sheets from and Excel with either the Input Data tool or the Dynamic Input Tool, I usually want a field to identify which Sheet the data came from. Currently I have to import the Full Path and then remove everything except the SheetName.

 

It would be great if there was an option to output she SheetName as a field.

 

DavidP_0-1657614328613.png

 

For companies that have migrated to OneDrive/Teams for data storage, employees need to be able to dynamically input and output data within their workflows in order to schedule a workflow on Alteryx Server and avoid building batch MACROs.

 

With many organizations migrating to OneDrive, a Dynamic Input/Output tool for OneDrive and SharePoint is needed.

 

  • The existing Directory and Dynamic Input tools only work with UNC path and cannot be leveraged for OneDrive or SharePoint.
  • The existing OneDrive and SharePoint tools do not have a dynamic input or output component to them.
  • Users have to build work arounds and custom MACROS for a common problem/challenge.
  • Users have to map the OneDrive folders to their machine (and server if published to the Gallery)
    • This option generates a lot of maintenance, especially on Server, to free up space consumed by the local version when outputting the data.

The enhancement should have the following components:

 

OneDrive/SharePoint Directory Tool

  • Ability to read either one folder with the option to include/exclude subfolders within OneDrive
  • Ability to retrieve Creation Date
  • Ability to retrieve Last Modified Date
  • Ability to identify file type (e.g. .xlsx) 
  • Ability to read Author
  • Ability to read last modified by
  • Ability to generate the specific web path for the files

OneDrive/SharePoint Dynamic Input Tool

  • Receive the input from the OneDrive/SharePoint Directory Tool and retrieve the data.

Dynamic OneDrive/SharePoint Output Tool

  • Dynamically write the output from the workflow to a specific directory individual files in the same location
  • Dynamically write the output to multiple tabs on the same file within the directory.
  • Dynamically write the output to a new folder within the directory

Hello all,

As of now, you have two very distinct kinds of connection :

-in memory alias

-in database alias

It happens than every single time I use a in-database alias I have to create the same for in memory since some operations cannot be realized in in-database (such as pre-sql or interface tools)

What does that mean for us :

-more complex settings operations/training/tests

-unefficient worflows that have to deal with two kinds of alias.

What I propose :
-a single "connection alias", that can be used either for in-db either for in-memory,

-one place to configure

-the in-db or in-memory being dependant on the tools you use

simonaubert_bd_0-1682053466250.png

 


Best regards,

Simon

Hi everyone,

 

Add two additional features to a directory tool. Something like this:

fmvizcaino_0-1686406240366.png

Use cases:

1. Since it is not possible to use a folder browse on the Gallery, this could help a basic user create a list of possible folders to select from with the help of a drop-down

2. Directory analysis for cleaning purposes - currently, if you want to get a list of the folders with Alteryx, it takes forever for big file servers since Alteryx is mapping all the files

 

Both are achievable today through regex or a bat script.

 

Thank you,

Fernando Vizcaino

Hello all,

Sometimes, when you have too much time to retrieve your tables metadas, you can have this message

 

 

 

Initialization Timed Out: Workflow must be run for field meta info to be accurate.

 

 

image.png

From what I understand, it's Alteryx and the source system that drives the time out value. However, I have some cases where the long time is "normal" and that really hurts the user experience.
So, I would like the ability in settings to change the default value.

Best regards,

Simon

In the Input tool, I rely heavily on the recent connection history list. As soon as a file falls off of this list, it takes me a while to recall where it's saved and navigate to the file I'm wanting to use. It would be great to have a feature that would allow users to set their favorite connections/files so that they remain at the top of the connection history list for easy access.

 

Kenda_0-1677686307640.png

 

hello,

version 2021.4 does not allow workflows to run if any of their input files are open.... would be great to have an option for the input tool that switches on/off the ability to read from open files. Some of my input files have frequent data changes and i tend to keep them open while testing/simulating results

 

Thank you,

abdou 

Please improve the Excel XLSX output options in the Output tool, or create a new Excel Output tool, 
or enhance the Render tool to include an Excel output option, with no focus on margins, paper size, or paper orientation

 

The problem with the current Basic Table and Render tools are they are geared towards reporting, with a focus on page size and margins.
Many of us use Excel as simply a general output method, with no consideration for fitting the output on a printed page.

 

The new tool or Render enhancement would handle different formats/different schemas without the need for a batch macro, and would include the options below.

 

The only current option to export different schemas to different Sheets in one Excel file, without regard to paper formatting, is to use a batch macro and include the CReW macro Wait a Second, to allow Excel to properly shut down before a new Sheet is created, to avoid file-write-contention issues.

 

Including the Wait a Second macro increased the completion time for one of my workflows by 50%, as shown in the screehshots below.

 

I have a Powershell script that includes many of the formatting options below, but it would be a great help if a native Output or Reporting tool included these options:

 

Allow options below for specific selected Sheet names, or for All Sheets


AllColumns_MaxWidth: Maximum width for ALL columns in the spreadsheet. Default value = 50. This value can be changed for specific columns by using option Column_SetWidth.

Column_SetWidth: Set selected columns to an exact width. For the selected columns, this value will override the value in AllColumns_MaxWidth.

 

Column_Centered: Set selected columns to have text centered horizontally.

 

Column_WrapText: Set selected columns to Wrap text.

 

AllCells_WrapText: Checkbox: wrap text in every cell in the entire worksheet. Default value = False.

 

AllRows_AutoFit: Checkbox: to set the height for every row to autofit. Default value False.

 

Header_Format: checkbox for Bold, specify header cells background color, Border size: 1pt, 2pt, 3pt, and border color, Enable_Data_Filter: checkbox

 

Header_freeze_top_row: checkbox, or specify A2:B2 to freeze panes

 

Sheet_overflow: checkbox: if the number of Sheet rows exceeds Excel limit, automatically create the next sheet with "(2)" appended

 

Column_format_Currency: Set selected columns to Currency: currency format, with comma separators, and negative numbers colored red.

 

Column_format_TwoDecimals: Set selected columns to Two decimals: two decimals, with comma separators, and negative numbers colored red.
Note: If the same field name is used in Column_Currency and Column_TwoDecimals, the field will be formatted with two decimals, and not formatted as currency.

Column_format_ShortDate: Set selected columns to Short Date: the Excel default for Short Date is "MM/DD/YYYY".

 

File_suggest_read_only: checkbox: Set flag to display this message when a user opens the Excel file: "The author would like you to open 'Analytic List.xlsx' as read-only unless you need to make changes. Open as read-only?
vb code: xlWB.ReadOnlyRecommended = True

 

File_name_include_date_time: checkboxes to add file name Prefix or Suffix with creation Date and/or Time

 

========

Examples:

 

My only current option: use a batch macro, plus a Wait a Second macro, to write different formats/schemas to multiple Sheets in one Excel file:

Excel output enhancement - must now use a Macro and Wait a Second.jpg

Using the Wait a Second macro, to allow Excel to shut down before writing a new Sheet, to avoid write-contention issues, results in a workflow that runs 50% longer:

Excel output enhancement - Wait a Second increased runtime by 50 percent.jpg

 

In a similar vein to the forthcoming enhancement of being able to disable a specific output tool, my idea is to have the inverse where you can globally disable all outputs and then enable specific ones only. This should help reduce the number of clicks required/avoid workarounds using containers to obtain this functionality and allow users to be very specific in which outputs run and don't run as required.

Referencing the previous idea: Inputs/Output should have the option to read/write a compressed file (ZIP or GZIP)

 

This idea has been implemented for inputting .zip files. However, we still need to use the run command workaround for outputs. It's very common for many users to want to output their .csv, .xlsx, .pdf to a .zip. The functionality would also need to extend to Gallery.

 

See the following links for people that are looking for this type of functionality:

https://community.alteryx.com/t5/Alteryx-Server-Discussions/Download-Multiple-Outputs-from-the-Galle...

https://community.alteryx.com/t5/Alteryx-Designer-Discussions/Output-files-to-ZIP/td-p/163502

https://community.alteryx.com/t5/Alteryx-Designer-Discussions/Zip-files/td-p/151456

 

Feel free to merge this idea with the previous one for continuity.

Hello all,

As specified in the title, this idea is to distinguish between Append Prefix/Suffix to File and to Table on the Output Data Tool.


For most files (csv...), the table name does not really exist. However, for at least Excel files, if you choose this option, the result will be one sheet by suffix and the only option to have one file by suffix will be to change entire file path.

simonaubert_bd_0-1641282924102.png

Best regards,

Simon

 

This is a QoL-request, and I love me some QoL-updates!

 

best idea for 2020.png

 

While I'm developing I often need the output of a workflow as input for the next phase of my development. For example: an API run returns job location, status, and authentication ids. I want to use these in a new workflow to start experimenting what'll work best. Because of the experimenting part, I always do this in a new workflow and not cache and continue in my main flow.

 

Writing a temporary output file always feels like unnescesary steps, and tbh I don't want to write a file for a step that'll be gone before it reaches production. Esp if there is sensitive information in it.

 

Thanks.

As per a recent discussion (https://community.alteryx.com/t5/Alteryx-Designer-Discussions/Geopackages-Can-Alteryx-Open-GeoPackag...), please add the GeoPackage datatype to the Input tool.

 

For reference, the open-source project ogr2ogr has this functionality. (https://gdal.org/programs/ogr2ogr.html)

 

Thanks!

Currently it's possible to use the Output tool to output to either a sheet, a place in a sheet or a named range in Excel, but it is not possible to output to a preformatted excel table - it would be really good if the output tool had an option to output to [Table1] in an Excel workbook for example. This enhancement would be incredibly helpful for reporting purposes.

Hello all,

As you may know, Alteryx use the Active Query Builder component. However this component itself evolves with cool new features :
https://www.activequerybuilder.com/blog/2018-04-28-much-faster-visual-sql-query-building-in-the-new-...

 

You can also try the online demo

https://www.activequerybuilder.com/

simonaubert_bd_0-1628595636399.png

 

Best regards,

Simon

Hi all,

 

At present, Alteryx does not support DSN-free connections to Snowflake using the Bulk Connector.    This is a critical functionality for any large company that uses Alteryx - and so I'm hoping that this can be changed in the product in an upcoming release.    As a corollary - every DB connection type has to be able to work without DSNs for any medium or large size server instance - so it's worth extending this to check every DB connection type available in Alteryx.

 

Here are the details:

 

What is DSN-Free?

In order to be able to run our Alteryx canvasses on a multi-node server - we have to avoid using DSNs - so we generally expand connection strings that look like this:

odbc:DSN=DSNSnowFlakeTest;UID=Username;PWD=__EncPwd1__|||NEWTESTDB.PUBLIC.MYTESTTABLE

 

to instead have the fully described connection string like this:
odbc:DRIVER={SnowflakeDSIIDriver};UID=Username;pwd=__EncPwd1__;authenticator=Snowflake;WAREHOUSE=compute_wh;SERVER=xnb27844.us-east-1.snowflakecomputing.com;SCHEMA=PUBLIC;DATABASE=NewTestDB;Staging=local;Method=user

 

For Snowflake BL:

Now - for the Snowflake Bulk Loader the same process does not work and Alteryx gives the classic error below

 

With DSN:

snowbl:DSN=DSNSnowFlakeTest;UID=Username;pwd=__EncPwd1__;Staging=local;Method=user|||NEWTESTDB.PUBLIC.MYTESTTABLE

 

Without DSN:

snowbl:driver=SnowflakeDSIIDriver;UID=SeanBAdamsJPMC;pwd=__EncPwd1__;SERVER=xnb27844.us-east-1.snowflakecomputing.com;WAREHOUSE=compute_wh;SCHEMA=PUBLIC;DATABASE=NewTestDB;Staging=local;Method=user|||NEWTESTDB.PUBLIC.MYTESTTABLE

 

Output Data (6) Error SQLDriverConnect: [Microsoft][ODBC Driver Manager] Data source name not found and no default driver specified" which indicates that the driver details may be wrong

 

 

Many thanks

Sean

Top Liked Authors