Bring your best ideas to the AI Use Case Contest! Enter to win 40 hours of expert engineering support and bring your vision to life using the powerful combination of Alteryx + AI. Learn more now, or go straight to the submission form.
Start Free Trial

Alteryx Analytics Cloud Product Ideas

Share your Alteryx Analytics Cloud product ideas, including Designer Cloud, Intelligence Suite and more - we're listening!

Currently you have one top level folder. In order to keep flows and other objects organized it would be nice to create folders within folders. For example create a top folder called "Sales Data" and a subfolder called "Sales Pipeline Data".

Please allow connections to be created from Trifacta to SharePoint online using SSO authentication, just like for Azure SQL/DWH.

Allow for more then 1 job to be deleted at a time.

The ability to apply various interpolation methods (cspline, linear, etc.) between sorted columns of integers.

My use case is that by looking at the target table data, i need to have a column which will indicate which flow has loaded data into that target table. this will be useful for bug fix and tracing back data issues to a flow.

Rgt now we are hardcoding this value as a new column in the recipe step, but if some developer changes the flow name he/she has to manually change the recipe step to reflect the flow name, instead if we can have a dynamic flow name like we have $Filepath for filepath on similar fashion it will be useful.

We need the full steps on GCP Dataprep and GCP to allow us to run scheduled jobs as a true service account (not a user account) and not require authentication of the owning user account (which is timing out in the night due to 16 hour policy for users we have)

So when we schedule a job we should be able to choose a true technical account to "run the job as".

We have an issue as our AD users are synchronised from on premise and a 16 hour timeout policy is applied to each user so any job scheduled with a user will fail after 16 hours and job will be disabled . There is no way for us at our company to sync ad users to GCP IAM without this policy from on premise so we need to be able to run with Service Account.

A client is organizationally structured where GCP project administration is separated from tool (Dataprep) administration. Currently, usage charts are only visible in the admin console which is only accessible by project admins. We would like an additional IAM role or an ability to see usage charts without the ability to make project changes.

Have an option when scheduling jobs and if they fail to restart after X minutes. Most of the time when I have a job failure and rerun, it completes fine.

We would like to connect Trifacta to a SharePoint file in a folder.

When I export a flow that contains a reference dataset, the name of the JSON file downloaded doesn't match the name of the flow that was exported. Instead it matches the name of the reference dataset inside the flow. I would like to change this so that a flow always keeps its original name when exported from Trifacta.

Allow functionality in app for customizing support page, users to be able to contact our team when there is an issue with the application, page to show our email address, not Trifacta support email address

Users to be able to create multiple connections to data lake. Currently user needs to add new data lake path and browse it in order to import data

We need a custom viewer role so that user is able only to use connections shared to him, but not re-share those connections to others. In our case, admin will set up the connections for users and they will just use them. Users should not be able to create or share connections. This will improve the connection security and access to data.

In our organization we would like to export path to data lake to be enforced and thus user not able to export to any location but a location desired by the application admins.

Allow a connection to a geo coding system, like USPS or Google, that allow you to join and run a demographics dataset through to have longitude and latitude added the output for mapping. I can see a lot of uses for this and especially in the Marketing and Advertising sector.

When output a Trifacta formated date, like a derived year (year(mydate) ) or formatted to mmddyyyy that when it is published to a SQL table that it maintains that format or transformation. Currently putting out a formatted or transformed date is put out to the SQL Table in the input format and all transformation and formatting is lost.

When creating a destination to and SQL table on an SQL server, allow you to set the field length and not use the Trifacta output standard of varchar (1024). This would help keep the maintenance on the table easier.

I'm looking for a way to discover which datasets, recipes, or outputs are taking up the most time and resources.

it would also be nice if we were able to view this over time as well.


an example would be sometime like the Unity3d profiler.

https://docs.unity3d.com/uploads/Main/profiler-window-layout.png

this is for a video game engine, but i hope the system can be similar.

in this profiler you can see what resource (ram,cpu, gpu) is being used and by what character/object in your video game.

similarly it would be nice to see what database is being used by what flow in trifacta.

Currently is scheduling allowed only on Instance level.

The request is to be able to allow scheduling for particular user, instead for all instance users.

When migrating flows from one environment to another, currently one must click on each flow to export it, and then import it into another environment. It would be great to do so in bulk by ticking a bunch of flows to export. Or do it by exporting an entire folder. Or do it by exporting a plan.