The Product Idea boards have gotten an update to better integrate them within our Product team's idea cycle! However this update does have a few unique behaviors, if you have any questions about them check out our FAQ.

AACP Product Ideas

Share your AACP product ideas, including Designer Cloud, Intelligence Suite and more - we're listening!

When i use Dataprep full day, i would prefer to have a dark theme to preserve my eyes ! If we could swith between a dark or a light theme, it would be the best solution !

:-)

PS : Sorry for my bad English, I'm a French user.

Please add features to your current "Folder" feature by allowing us to share them with other users, move a flow that has been "shared" with us from the root folder to a sub-folder, etc.

Someone has already submitted an idea for multiple levels of sub-folders which was another request that we had. Thanks.

Use a linked datasets created by GCP Analytic Hub as data source in DataPrep. Detailed informations in link below:

Can I use linked dataset (created by Analytic Hub in GCP) to build flows in DataPrep? (trifacta.com)

Parquet has been more performant than publishing to csv and love to have this feature implemented

We use heavily Tibco Data Virtualization server views and web services in our organization.

We need to have official connector supported by Trifacta to connect and fetch data from those.

Currently, when pivoting a specific field into multiple columns, all other fields you want present in the resulting table must be individually added to "row labels".

First off - It is very time consuming when you have a lot of columns to add.

Secondly - When new columns are added in the source data, these new fields are not automatically included. When this happens we need to:

  1. re-sample the data,
  2. make sure the new column is present,
  3. manually add it to the list in the row labels.


When using an automation tool such as Trifacta I would expect that my flow can deal with new columns being added without having to go and fix my flow every time. Adding an option to add "All other fields" or being able to select the Fields to exclude would make this process much smoother and it would ensure that our flow is future proof.

We often use hashing functions like fingerprint in SQL (Big Query) to mark or identify rows that match for specific attributes or to generate UUIDs. I know it's possible to do so by adding UDFs, but it would be more convenient to have a native function.

When I'm on the Flow workspace, if I do a click on a recipe, the steps displays on the float right box.

Unfortunately, i'can't select and copy steps :-/

I'm forced to load the recipe to copy steps before past it in another recipe. I think that everyone would gain some time to copy steps directly from the Flow view !

:-)

PS : Sorry for my bad English, I'm a French user.

Currently is scheduling allowed only on Instance level.

The request is to be able to allow scheduling for particular user, instead for all instance users.

There is no explicit sharing functionality for user macros between users or groups of users. Currently, the closest thing is users gain access to an owner’s macros within a flow once the flow is shared with them. It would be good to be able to share macros in the same way that users can share flows.

A client is organizationally structured where GCP project administration is separated from tool (Dataprep) administration. Currently, usage charts are only visible in the admin console which is only accessible by project admins. We would like an additional IAM role or an ability to see usage charts without the ability to make project changes.

Please redirect the user back to the page where the session has expired, instead of redirecting to home once the user re-authenticates.

Current scenario we are seeing that the user will be redirected to home page instead of the page he was in when the session has expired after set time in the config, in my case after 30 mins. ( this is because sometime user goes to a meeting forgets about the page he was working and he has to re-open everything from home page after re-auth )

We would like to connect Trifacta to a SharePoint file in a folder.

We would strongly like the ability to be able to edit datasets, created with custom SQL that have been shared with us. We think of Trifacta in part as a shared development space so if 1 users needs to make an update to a dataset but wasn't originally the owner - this slows down our workflow considerably.

In order to monitor the status of the plan that has been running several different flows inside, in my case it is around 300, I send the HTTP request to Datadog to display the result of failed and success on a dashboard. The problem is, DATADOG understands only epoch timestamp and not the datetime value. Right now we cannot convert the timestamp into epoch. I was thinking of approaching this problem in the following ways:

1) Having a pre-request script

2) Creating dynamic parameters in Dataprep instead of using a fixed value, that can be used further in the HTTP request body

3) This is just the turnaround - Creating a table that stores the flow name and timestamp in it, and we are supposed to use this table in a plan every time we are running a flow. But this is not the right way. It will work but it is waste of time as we will end up creating separate tables like this one for each flow.

Hi,

I use the import by folder for GCS files to import many files in one time (present and futurs files droped in the same folder). Somtimes, the dataschema of files is not exactly the same for all the files but the columns names are always the same ! I'would like to use the "union by name" for the first union of the many files included in the folder that i've imported. With this function, if the dataschema change in the futur, my importation will be ok whatever !

We could have a screen like "recipe union screen" for the "import with union" (for the inports by folder) to select the columns to import and the type of matching for exemple... 

This is a real issue for me because when the datascema of one file has changed, the scheduled RUNs are KO...

Sorry for my bad English, I'm French :-)

Thanks !

As far as I know, the current error logs for a failed Trifacta job do not tell the user which recipe, which recipe step, and on what data the error was thrown.

This lack of basic information on the Trifacta level makes it hard for a normal user to debug Trifacta jobs. Typically, I will have to work backwards in the flow, attaching and running an output for each recipe until I can find the culprit recipe causing the issues. Then, I will have to disable steps one by one until I find the step that causes the recipe to fail. This is time and resource consuming.

As for the offending data triggering the problem, I still don't know how to get that, and that's actually crucially important for an ongoing issue we're having with Spark execution.

Therefore, I suggest that improved and simplified error logging would be very helpful in fixing problems in the future. Thank you for your consideration.

Support for reading XML files natively with Trifacta.

My use case is that by looking at the target table data, i need to have a column which will indicate which flow has loaded data into that target table. this will be useful for bug fix and tracing back data issues to a flow.

Rgt now we are hardcoding this value as a new column in the recipe step, but if some developer changes the flow name he/she has to manually change the recipe step to reflect the flow name, instead if we can have a dynamic flow name like we have $Filepath for filepath on similar fashion it will be useful.

We need the full steps on GCP Dataprep and GCP to allow us to run scheduled jobs as a true service account (not a user account) and not require authentication of the owning user account (which is timing out in the night due to 16 hour policy for users we have)

So when we schedule a job we should be able to choose a true technical account to "run the job as".

We have an issue as our AD users are synchronised from on premise and a 16 hour timeout policy is applied to each user so any job scheduled with a user will fail after 16 hours and job will be disabled . There is no way for us at our company to sync ad users to GCP IAM without this policy from on premise so we need to be able to run with Service Account.