Advent of Code is back! Unwrap daily challenges to sharpen your Alteryx skills and earn badges along the way! Learn more now.
Free Trial

Alteryx Analytics Cloud Product Ideas

Share your Alteryx Analytics Cloud product ideas, including Designer Cloud, Intelligence Suite and more - we're listening!

Hello,

Correct me if I'm wrong but we can't export or import a workflow. That sounds a very easy and very useful function, especially when you develop something and want to share with others users.

 

Screen Shot 2024-03-21 at 10.04 AM.PNG

Best regards,

Simon

Use a linked datasets created by GCP Analytic Hub as data source in DataPrep. Detailed informations in link below:

Can I use linked dataset (created by Analytic Hub in GCP) to build flows in DataPrep? (trifacta.com)

As of now the once a user deletes the flow, the flow will not be visible to anyone, except in the database. But the flow is soft deleted in the database. So can enable the option for admins to see all the deleted flows and recover those flows if required, so that in case some one deletes the flow by mistake then admins can retrieve it by recover option. This has to be an option by check box, where they can recover those flows all at once if it is a folder. This option can also be given to folder recovery where they can recover all the flows in the folder.

I want to create a dynamic output file name in both the Trifacta UI and New Designer Cloud UI. I know that I can create a timestamp or a parameter, but it seems like the parameterization in Trifacta is static. Instead of having to manually override the parameter every run, I am wanting to grab the desired filename from a field in the dataset. This is a functionality commonly used in Desktop Designer, so I'm hoping for it to be added to the Cloud soon!

Currently there is support for parameterizing variables in custom SQL dataset in Dataprep. However it requires that the tables using this feature have the same table structure. This request is to allow this same functionality but with tables that have different table structures.

Example:

Table A

dev.animals.dogs

name | height | weight


Table B

dev.animals.cats

name| isFriendly


Would like to use a query where we have 1 custom SQL dataset where we just say

SELECT * FROM dev.animals.[typeOfAnimal]

typeOfAnimal being the parameterized variable with a default of dogs.

Hi team,

We would need a page where a user can handle all the email notifications they are receiving from all the flows (success and failure).

Thank you

We would strongly like the ability to be able to edit datasets, created with custom SQL that have been shared with us. We think of Trifacta in part as a shared development space so if 1 users needs to make an update to a dataset but wasn't originally the owner - this slows down our workflow considerably.

Create a connector to Mavenlink.

If a flow is shared between multiple editors and someone make changes in it, there should be a way we can see all the changes made to that flow by different users, like creating a trigger that will notify the users about the changes made in the flow by someone as soon as the recipe changes or if we can extract the information about the flow or the job. I have attached the snippet of data that can be useful to us.

Right now there is no place where team members can collectively create flows and share at one place. If given the option to share the Folders among different members just like we have for flows it will be lot easier. For Example: If there is a folder with 4 different flows, and I share the folder with my team mates they can edit and created new flows over there and can see all the 4 different flows already present. But if out of 4 flows if I share 2 flows with someone, they see the folder but they don't see the flows not shared with them.

In order to monitor the status of the plan that has been running several different flows inside, in my case it is around 300, I send the HTTP request to Datadog to display the result of failed and success on a dashboard. The problem is, DATADOG understands only epoch timestamp and not the datetime value. Right now we cannot convert the timestamp into epoch. I was thinking of approaching this problem in the following ways:

1) Having a pre-request script

2) Creating dynamic parameters in Dataprep instead of using a fixed value, that can be used further in the HTTP request body

3) This is just the turnaround - Creating a table that stores the flow name and timestamp in it, and we are supposed to use this table in a plan every time we are running a flow. But this is not the right way. It will work but it is waste of time as we will end up creating separate tables like this one for each flow.

We at Grupo Boticário, who currently have 13k Dataprep licenses and close to the official launch internally, have noticed a recurring request for a translation of the tool. Bearing in mind that it will be an enabler for more users to use in their day-to-day work, I would like to formalize and reinforce the importance of our request for translation into Brazilian Portuguese as well as a forecast of this improvement.

We often use hashing functions like fingerprint in SQL (Big Query) to mark or identify rows that match for specific attributes or to generate UUIDs. I know it's possible to do so by adding UDFs, but it would be more convenient to have a native function.

We can migrate flows from one environment to other environment using Trifacta APIs.

Export and Import the flow from source to target.

Rename the flow.

Share flow with appropriate user according to environment.

Change the input and output of the flow.

Currently, when a recipe is copied, any data qualities within the original recipe are not duplicated in the copied version. In order to implement a systematic data quality program, the rules must be manually created for every single recipe, which obviously takes a lot of time. It would be great if the data quality rules could persist when the recipe is copied.

Being able to Publish outputs directly to Google Sheets would be a major benefit for Sheets users.

We need the ability the create folders underneath the plans. We can create folders underneath flows, but not underneath plans. Additionally, having the ability to create sub folders inside of these parent flow and plan folders is needed. Hard to organize flows and plans without the ability to put them in categories (folders) and subcategories (sub folders) when you approach hundreds of plans and flows.

I often receive data sets which have rows above the column headers that I don't need. When importing the data set, there is a dropdown on the edit menu to "make the first row a column header". However, I would like for this dropdown to include an option to for example, "make row 20 the column header and delete all preceding rows". This would allow me to import the data already with column headers. When dealing with one dataset, I can always choose any row to make it the column headers, but when you have to join 20 similar datasets, it is not possible to do the same. Not sure if my idea is clear (lol), but it seems like it's something that could be easily incorporated into the tool. Thanks!

Hello,

I need to store many variable RegEx in column to use it in MATCHES function (for example).

But Dataprep doesn't currently support a column as an input in a way that the pattern inside it is read as an actual regular expression.

I think this feature could be a great feature !

Thanks.

more informations about this cese : https://community.trifacta.com/s/question/0D53j00007kB5UmCAK/matches-function-using-pattern-regex-st...


Please redirect the user back to the page where the session has expired, instead of redirecting to home once the user re-authenticates.

Current scenario we are seeing that the user will be redirected to home page instead of the page he was in when the session has expired after set time in the config, in my case after 30 mins. ( this is because sometime user goes to a meeting forgets about the page he was working and he has to re-open everything from home page after re-auth )